You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
> -[Azure Create a resource](https://learn.microsoft.com/en-us/azure/cognitive-services/openai/how-to/create-resource?pivots=web-portal#create-a-resource)
140
+
> -[Azure-OAI Rest API versoning](https://learn.microsoft.com/en-us/azure/ai-services/openai/reference#rest-api-versioning)
141
+
136
142
## Embedding
137
143
<table>
138
144
<tr>
@@ -146,16 +152,24 @@ chatCompletion()
146
152
importOpenAIfrom'openai';
147
153
148
154
const openai =newOpenAI({
149
-
apiKey: '', // defaults to process.env["OPENAI_API_KEY"]
Copy file name to clipboardExpand all lines: docs/docs/new/faq.md
+17-5Lines changed: 17 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,18 +3,30 @@ title: FAQs
3
3
slug: /faq
4
4
---
5
5
6
-
### 1. Is Nitro the same as Llama.cpp with an API server?
6
+
<details>
7
+
<summary>1. Is Nitro the same as Llama.cpp with an API server?</summary>
7
8
8
9
Yes, that's correct. However, Nitro isn't limited to just Llama.cpp; it will soon integrate multiple other models like Whisper, Bark, and Stable Diffusion, all in a single binary. This eliminates the need for you to develop a separate API server on top of AI models. Nitro is a comprehensive solution, designed for ease of use and efficiency.
9
10
10
-
### 2. Is Nitro simply Llama-cpp-python?
11
+
</details>
12
+
13
+
<details>
14
+
<summary>2. Is Nitro simply Llama-cpp-python?</summary>
11
15
12
16
Indeed, Nitro isn't bound to Python, which allows you to leverage high-performance software that fully utilizes your system's capabilities. With Nitro, learning how to deploy a Python web server or use FastAPI isn't necessary. The Nitro web server is already fully optimized.
13
17
14
-
### 3. Why should I switch to Nitro over Ollama?
18
+
</details>
19
+
20
+
<details>
21
+
<summary>3. Why should I switch to Nitro over Ollama?</summary>
15
22
16
23
While Ollama does provide similar functionalities, its design serves a different purpose. Ollama has a larger size (around 200MB) compared to Nitro's 3MB distribution. Nitro's compact size allows for easy embedding into subprocesses, ensuring minimal concerns about package size for your application. This makes Nitro a more suitable choice for applications where efficiency and minimal resource usage are key.
17
24
18
-
### 4. Why is the model named "chat-gpt-3.5"?
25
+
</details>
26
+
27
+
<details>
28
+
<summary>4. Why is the model named "chat-gpt-3.5"?</summary>
29
+
30
+
Many applications implement the OpenAI ChatGPT API, and we want Nitro to be versatile for any AI client. While you can use any model name, we've ensured that if you're already using the chatgpt API, switching to Nitro is seamless. Just replace api.openai.com with localhost:3928 in your client settings (like Chatbox, Sillytavern, Oobaboga, etc.), and it will work smoothly with Nitro.
19
31
20
-
Many applications implement the OpenAI ChatGPT API, and we want Nitro to be versatile for any AI client. While you can use any model name, we've ensured that if you're already using the chatgpt API, switching to Nitro is seamless. Just replace api.openai.com with localhost:3928 in your client settings (like Chatbox, Sillytavern, Oobaboga, etc.), and it will work smoothly with Nitro.
0 commit comments