Skip to content

Latest commit

 

History

History
32 lines (20 loc) · 991 Bytes

llm-linux-cpu.md

File metadata and controls

32 lines (20 loc) · 991 Bytes

Setup llm-inference using CPU on windows

This is a short guide to setup llm-inference project to run on your Linux machine using CPU.

NOTE: Python 3.12 breaks torch instllation. Please use Python 3.10

  • Create Python Virtual Environment:

    • python -m venv venv
  • Activate the virtual environment:

    • source venv/bin/activate
  • Install pytorch with cpu support:

  • Install required packages:

  • Install Bitsandbytes

    • pip3 install bitsandbytes
  • Create .env file based on .env.example or env-samples/env.cpu.example

    • Change the Model path and config then Run the server:
      • python3 main.py --multiprocess

Back to main doc