Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: support quantization for local LLMs #189

Open
kdziedzic68 opened this issue Nov 17, 2024 · 0 comments
Open

feat: support quantization for local LLMs #189

kdziedzic68 opened this issue Nov 17, 2024 · 0 comments
Labels
feature New feature or request
Milestone

Comments

@kdziedzic68
Copy link
Collaborator

Feature description

the workflow should be planned mean-less with respect to the blogpost: https://huggingface.co/blog/4bit-transformers-bitsandbytes

Motivation

Huggingface has an easy to use interface for configurable model quantization - we should have support for it in order to run local LLMs with lower GPU capacity available

Additional context

No response

@kdziedzic68 kdziedzic68 added the feature New feature or request label Nov 17, 2024
@mhordynski mhordynski moved this to Backlog in ragbits Nov 18, 2024
@kdziedzic68 kdziedzic68 moved this from Backlog to Ready in ragbits Nov 18, 2024
@mhordynski mhordynski added this to the Ragbits 0.6 milestone Nov 18, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature New feature or request
Projects
Status: Ready
Development

No branches or pull requests

2 participants