-
Notifications
You must be signed in to change notification settings - Fork 153
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Compatibility with Llama #9
Comments
I tested it with 405b and works fine. Nvidia gives 1000 free credits. You can see here how to adjust the OpenAI client. https://build.nvidia.com/explore/discover#llama-3_1-405b-instruct |
@2187Nick do you give a example with using llama 3.1? |
@dangyuuki123 Here is an example using Gemini. I will work on llama 3.1 version next. |
That'd be lovely and would help receive further contributions from our side. As a side note, recommending using |
@dangyuuki123 @vmayoral Llama 405b is not producing consistent JSON. You can use my new example that uses free Nvidia hosted models which offers Llama 405b. If you are able to modify the model settings or prompt to get consistent JSON output, please let me know. Here is a working example that uses Nvidia hosted Mistral Large 2: I found Codestal-22b will also work. |
Can be quick and dirty swapped if using local ollama, search and replace
change to
and change the parser.add_argument models from default to what you want, i.e.
llama3.1 8B has inconsistencies that make the output iffy at best, gest caught on json format errors every single iteration. You can run the optional models with for example:
|
Is it possible to add out of the box functionality with OS models like llama 405b?
The text was updated successfully, but these errors were encountered: