Skip to content

Releases: 0cc4m/GPTQ-for-LLaMa

GPTQ-KoboldAI 0.0.6

12 Jun 06:23
Compare
Choose a tag to compare
Add v2 with bias support (e.g. for Tulu-30b)

GPTQ-KoboldAI 0.0.5

23 May 04:57
81469a4
Compare
Choose a tag to compare
Merge pull request #14 from TehVenomm/latestmerge

Incorrect bit shift applied to 8bit models

GPTQ-KoboldAI 0.0.4

19 May 19:50
Compare
Choose a tag to compare
2023-05-19-2

Bump version

GPTQ-KoboldAI 0.0.3

18 May 19:47
Compare
Choose a tag to compare
2023-05-18-2

Fix setup.py

GPTQ-KoboldAI 0.0.2

09 May 20:15
Compare
Choose a tag to compare
Add support for upstream gptq cuda version

Co-authored-by: qwopqwop200 <[email protected]>

GPTQ Python module

06 May 18:46
Compare
Choose a tag to compare
2023-05-06-2

Add MPT support

quant_cuda for CUDA 11.8/ROCm 5.4.2

02 May 20:06
Compare
Choose a tag to compare
Revert "Add wheel links file for pip"

This reverts commit 539af97b13e2b8ab0da6b26733159c22d3fe0963.

quant_cuda for CUDA 11.7

27 Apr 14:02
Compare
Choose a tag to compare
Revert "Add wheel links file for pip"

This reverts commit 539af97b13e2b8ab0da6b26733159c22d3fe0963.

2023-04-10

10 Apr 20:18
Compare
Choose a tag to compare

First wheel release