This is an example project for Float16
We have included a few examples to help you get started with Float16.
Model Name | Model Size | Description |
---|---|---|
OpenThaiGPT-70b | 70b | A Thai language model |
OpenThaiGPT-13b | 13b | A Thai language model |
OpenThaiGPT-7b | 7b | A Thai language model |
SQLCoder-7b-2 | 7b | A model for converting natural language to SQL queries |
SeaLLM-7b-v2 | 7b | A South East Asia (SEA) languages model |
SeaLLM-7b-v2.5 | 7b | A South East Asia (SEA) languages model |
Pricing for each model is available on the Float16.Cloud website.
-
[2024-04-15] Added brand new SeaLLM-7b-v2.5 model.
-
[2024-04-08] Added an example for Bare-metal request. (OpenThaiGPT-70b)
-
[2024-03-31] Added an example for Bare-metal request. (SQLCoder-7b-2)
-
[2024-03-05] Added an basic tutorial for LlamaIndex.
-
[2024-02-04] Added an example for LlamaIndex Streaming. (SeaLLM-7b-v2)
-
[2024-02-02] Added an example for Bare-metal request. (SeaLLM-7b-v2)
-
[2024-01-10] Added an example for Bare-metal request. (SeaLLM-7b-v2)
-
[2024-01-05] Added an example for LlamaIndex. (SeaLLM-7b-v2)
-
[2023-12-18] Added an example for Langchain. (SeaLLM-7b)
You can visit other directories to see how to use Float16 with various models.
Float16.Cloud is a platform providing open-source models and APIs for developers. It is similar to the OpenAI API.
There are, however, some differences between Float16 and the OpenAI API. Open-source models come in various sizes. For instance, the Llama2 model is available in 3 different sizes: 7b, 13b, and 70b. The larger the model, the better its performance. However, a larger model is generally more expensive to run. Hence, we provide a range of model sizes for developers to choose based on their needs.
To achieve better performance from smaller model sizes, we encourage using techniques such as prompt engineering. We plan to provide more examples of prompt engineering in the future.
Prompt engineering allows one to extract better performance from smaller model sizes. For instance, it's possible to achieve better performance with the 7b model than with the 13b or 70b models.
-
Which prompt engineer techniques are used in this project ? We have used the following prompt engineering techniques in this project:
- Few-shot learning
- Chain of Thought (CoT)
- Tree of Thought (ToT)
- Reason + Act (ReACT)
-
How do I use Float16? You can use Float16 by visiting Float16.Cloud. Register and get your API key. You can then use the API key to access the models.