Skip to content

haalven/LLM_terminal

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

49 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ask LLMs in the Terminal

ask_openai.py – call the OpenAI API, fill in your API_KEY

ask_ollama.py – call the local Ollama API (localhost:11434)

required:

OpenAI API

ask_openai.py uses the openai Python library to call the OpenAI API. An internet connection is required. You have to insert your API_KEY by editing the my_api_key = 'sk-***' line before running the script.

Choose a specific OpenAI model by editing the openai_model = models['o1pre'] line. See also: platform.openai.com/docs/models.

Ollama API

ask_ollama.py uses the requests Python library to call the local Ollama API (based on llama.cpp). An internet connection is not required. Install Ollama first and start the local server (localhost:11434). To download LLMs use the ollama run <model> command.

The best general LLMs for computers with 8…16 GB of RAM in Oct 2024 are gemma2 by Google (9B, 5.4GB), llama3.1 by Meta (8B, 4.7GB), phi3.5 by Microsoft (4B, 2.2GB), and mistral by MistralAI (7B, 4.1GB). See also: ollama.com/models.

Example:

example screenshot