Use the CLI to upload, manage, and query documents based on fine-tuned LLM models. It uses the smartloop API to manage projects and documents and gives you an easy way to quickly process contents and reason based on it.
- Python 3.11
Install the CLI with the following command:
pip install -U smartloop
Once installed, check that everything is setup correctly:
First you will need to create a free account, verify and configure your account. Once verified, copy your developer token to the clipboard. You will need a invitation code as of writing this document, please reach out to us at [email protected]
and we should be able to get you started.
Once you have your token, run the following command in your terminal:
smartloop login
This command will prompt you for your token, copy and pase the token that you have received in your email. Next step it to create a project, you can do so with the following command:
smartloop project create --name microsoft
Use the following command to interactively select a project:
smartloop project select
Once the project is selected , upload documents from your folder or a specific file, in this case I am uploading the a document describing Microsoft online services form my local machine:
smartloop upload --path=~/document1.pdf
Next step, once the project is selected, document(s) you have uploaded are processed, execute the following to start prompting:
smartloop run
This will bring up the prompt to query your information from your uploaded document(s)
Current project: Microsoft(microsoft-24-07-2024)
Enter message (Ctrl-C to exit): what the SLA for azure open ai
⠋
The SLA (Service Level Agreement) for Azure OpenAI is not explicitly mentioned in the provided text. However, it's possible that the SLA for Azure OpenAI might be similar to the one mentioned below:
"Uptime Percentage"
* Service Credit:
+ < 99.9%: 10%
+ < 99%: 25%
+ < 95%: 100%
Please note that this is not a direct quote from the provided text, but rather an inference based on the format and structure of the SLA mentioned for other Azure services (e.g., SAP HANA on Azure High Availability Pair). To confirm the actual SLA for Azure OpenAI, you should check the official Microsoft documentation or contact their support team.
Enter message (Ctrl-C to exit):
In order to set temperature
of your conversation, which ranges from 0.0 to 1.0, use the following command:
smartloop project set --id=project_id --temp=0.3
To enable memory to retain context in the conversation, use the following command:
smartloop project set --id=project_id --memory
To disable, use the following command:
smartloop project set --id=project_id --no-memory
LLM temperature is a parameter that influences the language model's output, determining whether the output is more random and creative or more predictable.
The higher value tends towards more creative answer
- DOCX
- TXT
- CSV
Contributions are welcome! Please create a pull request with your changes.
If you have any questions or suggestions, please feel free to reach out to [email protected]
This project is licensed under the terms of the MIT license.