Skip to content

Releases: YellowRoseCx/koboldcpp-rocm

KoboldCPP-v1.79.1.yr1-ROCm

05 Dec 03:43
d6949d6
Compare
Choose a tag to compare
attempt 6700xt fix for cmake-rocm-windows.yml

KoboldCPP-v1.79.1.yr0-ROCm

03 Dec 11:02
d6949d6
Compare
Choose a tag to compare
attempt 6700xt fix for cmake-rocm-windows.yml

KoboldCPP-v1.78.yr0-ROCm

19 Nov 06:36
Compare
Choose a tag to compare

koboldcpp-rocm-1.78

image

  • NEW: Added support for Flux and Stable Diffusion 3.5 models: Image generation has been updated with new arch support (thanks to stable-diffusion.cpp) with additional enhancements. You can use either fp16 or fp8 safetensor models, or the GGUF models. Supports all-in-one models (bundled T5XXL, Clip-L/G, VAE) or loading them individually.
  • Debug mode prints penalties for XTC
  • Added a new flag --nofastforward, this forces full prompt reprocessing on every request. It can potentially give more repeatable/reliable/consistent results in some cases.
  • CLBlast support is still retained, but has been further downgraded to "compatibility mode" and is no longer recommended (use Vulkan instead). CLBlast GPU offload must now maintain duplicate a copy of the layers in RAM as well, as it now piggybacks off the CPU backend.
  • Added common identity provider /.well-known/serviceinfo Haidra-Org/AI-Horde#466 PygmalionAI/aphrodite-engine#807 theroyallab/tabbyAPI#232
  • Reverted some changes that reduced speed in HIPBLAS.
  • Fixed a bug where bad logprobs JSON was output when logits were -Infinity
  • Updated Kobold Lite, multiple fixes and improvements
    • Added support for custom CSS styles
    • Added support for generating larger images (select BigSquare in image gen settings)
    • Fixed some streaming issues when connecting to Tabby backend
    • Better world info length limiting (capped at 50% of max context before appending to memory)
    • Added support for Clip Skip for local image generation.
  • Merged fixes and improvements from upstream

To use, download and run the koboldcpp_rocm.exe, which is a one-file pyinstaller.
If you're using Linux, clone the repo and build in terminal with make LLAMA_HIPBLAS=1 -j

Run it from the command line with the desired launch parameters (see --help), or manually select the model in the GUI.
and then once loaded, you can connect like this (or use the full koboldai client):
http://localhost:5001

For more information, be sure to run the program from command line with the --help flag.
Release notes from: https://github.com/LostRuins/koboldcpp/releases/tag/v1.78

KoboldCPP-v1.77.yr1-ROCm

06 Nov 18:57
Compare
Choose a tag to compare
  • Bring Speed Back

upstream llama.cpp introduced a change to calculate certain values in full 32 bit precision by default which introduced a major slow down for some users with AMD GPUs, this reverts that change until improvements are made

KoboldCPP-v1.77.yr0-ROCm

03 Nov 09:42
8d3449d
Compare
Choose a tag to compare
Update dependencies in cmake-rocm-windows.yml

KoboldCPP-v1.76.yr1-ROCm

14 Oct 21:41
Compare
Choose a tag to compare

KoboldCPP-v1.76.yr0-ROCm

13 Oct 17:44
f27e9a9
Compare
Choose a tag to compare

Upstream changes & rocBlas GPU file reconstruction to attempt fixing the issue some RX 7000 GPU users were experiencing

Oct/14/2024 2:36PM CST - This build may be broke for some users.
try https://github.com/YellowRoseCx/koboldcpp-rocm/releases/tag/v1.76.yr1-ROCm
I'm sorry for the inconvenience, please work with me as I try to solve any errors🙏

KoboldCPP-v1.75.2.yr1-ROCm

27 Sep 17:44
Compare
Choose a tag to compare
  • Recompiled with gfx906 support
  • disable mmq by default
  • update make_pyinstaller.sh (used to create a single linux executable)

KoboldCPP-v1.75.2.yr0-ROCm

23 Sep 18:56
42edf71
Compare
Choose a tag to compare
Update cmake-rocm-windows.yml remove openblas

KoboldCPP-v1.74.yr0-ROCm

01 Sep 23:26
Compare
Choose a tag to compare
Merge remote-tracking branch 'upstream/concedo'