Skip to content

ganga-devs/ganga-ai

Repository files navigation

Installation of dependencies

Ollama

Ollama is used for downloading and serving a large number of llm packages. To install it follow the instructions given here https://ollama.com/download. If you are on linux and your Ollama is available in your distro's repository then you can install it through distro's package manager and set up a systemd service. For Arch linux the steps would be

$ sudo pacman -S ollama

Then create the file /home/usr/.config/systemd/user/ollama.service with the content

[Unit]
Description=Ollama Serve
After=network-online.target

[Service]
ExecStart=/usr/bin/ollama serve
Restart=always
RestartSec=3

[Install]
WantedBy=default.target

Then run

$ systemctl --user enable ollama.service 
$ systemctl --user start ollama.service 

Python packages

To create the python environment do

pyenv virtualenv 3.8 ganga-ai8
pyenv activate ganga-ai8
pip install -r requirements.txt

Running the code

To run the code do

$ ipython
$ %load_ext ganga_ai
$ %%assist <your query>

To enable the rag do

$ ipython
$ %load_ext ganga_ai
$ %%enable_rag <path-to-your-local-ganga-repository>

Then use it normally.

Docker

To run the code in docker do

docker build --tag ganga-ai . # ganga-ai can be named anything
docker run -it --name test ganga-ai # test can be named anything

To remove the current image do

docker rmi test

About

LLM for your Ganga tasks

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published