The purpose of this reading is to give a guidance on setting up and running personal co-pilot similar to one that you can buy from Github.
- There is a great company anyscale that develops ray framework and those folks are offering serving of open source LLMs. All you need is to REGISTER HERE
- Once you register - generate and API key
- Models that are avialable at anyscale:
- Open-Orca/Mistral-7B-OpenOrca
- mistralai/Mixtral-8x7B-Instruct-v0.1
- HuggingFaceH4/zephyr-7b-beta
- codellama/CodeLlama-34b-Instruct-hf
- Meta-Llama/Llama-Guard-7b
- BAAI/bge-large-en-v1.5
- thenlper/gte-large
- mistralai/Mistral-7B-Instruct-v0.1
- meta-llama/Llama-2-70b-chat-hf
- meta-llama/Llama-2-13b-chat-hf
- meta-llama/Llama-2-7b-chat-hf
- You can query the api that is compatible with OpenAI, for instance to get a list of models:
curl --location 'https://api.endpoints.anyscale.com/v1/models' \
--header 'accept: application/json' \
--header 'Authorization: Bearer esecret_key-goes-here'
- The nice stuff is that once registerd you get a 10USD of credits for generation requests.
- Now install continue plugin for vscode - its on of opensource analogs of github co-pilot.
- Add CodeLlama model as an engine for co-pilot in
config.json
configuration file of continue :{ "title": "CodeLlama-34b", "model": "codellama/CodeLlama-34b-Instruct-hf", "contextLength": 16384, "apiBase": "https://api.endpoints.anyscale.com", "completionOptions": {}, "apiKey": "esecret_your-key-goes-here", "provider": "openai" }
- You are set and can query the model to assist you with your coding.