LLM plugin providing access to Groqcloud models.
Install this plugin in the same environment as LLM:
llm install llm-groqFirst, obtain an API key for Groqcloud.
Configure the key using the llm keys set groq command:
llm keys set groq<paste key here>
You can now access the three Mistral hosted models: groq-llama2 and groq-mixtral.
To run a prompt through groq-mixtral:
llm -m groq-mixtral 'A sassy name for a pet sasquatch'To start an interactive chat session with groq-mixtral:
llm chat -m groq-mixtralllm chat -m groq-mixtral
Chatting with groq-mixtral
Type 'exit' or 'quit' to exit
Type '!multi' to enter multiple lines, then '!end' to finish
> three proud names for a pet walrus
Here are three whimsical and proud-sounding names for a pet walrus:
1. Regalus Maximus
2. Glacierus Royalty
3. Arctican Aristocat
These names evoke a sense of majesty and grandeur, fitting for a noble and intelligent creature like a walrus. I hope you find these names fitting and amusing! If you have any other requests or need assistance with something else, please don't hesitate to ask.
To use a system prompt with groq-mixtral to explain some code:
cat example.py | llm -m groq-mixtral -s 'explain this code'TBD
To set up this plugin locally, first checkout the code. Then create a new virtual environment:
cd llm-groq
python3 -m venv venv
source venv/bin/activate