Skip to content

Releases: helixml/helix

1.4.12 - apps drag'n'drop filestore, initial MCP support

06 Jan 16:16
78bba78
Compare
Choose a tag to compare

What's Changed

security/refactoring/quality changes:

Full Changelog: 1.4.11...1.4.12

1.4.11 - out of the box local models support for gptscript; fix RAG over files on k8s

18 Dec 15:38
Compare
Choose a tag to compare

What's Changed

Full Changelog: 1.4.10...1.4.11

1.4.10 - rag improvements

15 Dec 23:44
8b7fcb6
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: 1.4.9...1.4.10

1.4.9 - support function calling in helix runners, add gptscript runner

11 Dec 19:06
9d6a4c7
Compare
Choose a tag to compare

What's Changed

Full Changelog: 1.4.8...1.4.9

1.4.8 - Runner slimming, add llama3.3:70b

08 Dec 09:09
ffd2818
Compare
Choose a tag to compare

What's Changed

Runner slimming, add llama3.3:70b by @lukemarsden in #634:

  • switch default model back to llama3.1:8b-instruct-q8_0, it was set to llama3.1:8b-instruct-q4_K_M by mistake
  • remove rarely used models to reduce size of -large model, remove image model from -small, prune default startup models
  • switch to symlinks instead of copying in the case that the model cache isn't bind-mounted (will dramatically reduce ephemeral storage usage in production)
  • add new llama3.3:70b model which is apparently as good as llama3.1:405b
  • upgrade to ollama v0.5.1
  • add new OLLAMA_KV_CACHE_TYPE=q8_0 option to quantize kv store, should reduce long context memory usage (we can monitor memory usage in the dashboard and adjust the predicted memory values accordingly)

Full Changelog: 1.4.7...1.4.8

1.4.7 - [Breaking] Replace some Ollama Models to aid cache stability

06 Dec 12:10
Compare
Choose a tag to compare

Ollama regularly updates their "default" models to use new weights. For example llama3.1:7b gets updated roughly once every two weeks. Each time this happens we have to rebuild our base images from scratch.

To help avoid this, we've switched to use the same model, but with a more specific tag.

New sessions will use this new models by default. API users should change their model request to use the newer models.

Here is a list of changed models:

llama3.1:8b-instruct-q8_0
llama3.1:70b
aya:35b

Switching to:

llama3.1:8b-instruct-q4_K_M
llama3.1:70b-instruct-q4_K_M
aya:35b-23-q4_0

What's Changed

  • refactor(runner): swap some ollama models to make them more stable and bust model cache to update by @philwinder in #631

Full Changelog: 1.4.6...1.4.7

1.4.6 - fix regression of frontend using chat models with togetherai

05 Dec 15:59
Compare
Choose a tag to compare

What's Changed

Full Changelog: 1.4.5...1.4.6

1.4.5 - Bugfix for openshift

03 Dec 16:35
Compare
Choose a tag to compare

What's Changed

  • fix(diffusers): add --no-dev because it shouldn't install dev dependencies by @philwinder in #622

Full Changelog: 1.4.4...1.4.5

1.4.4 - New Image Generation Codebase and Capabilities

03 Dec 09:42
Compare
Choose a tag to compare

This release contains a new image generation codebase that allows Helix much more flexibility over what image generation models are added.

To celebrate this release we've baked a selection of the latest image generation models from Stability.AI and black-forest-labs into the -large runner images only.

In the future we will tinker with the baked model selection to reduce the size of the base images and add the code to allow fine-tuning.

What's Changed

New Contributors

Full Changelog: 1.4.3...1.4.4

1.4.4 RC1 Diffusers Testing

03 Dec 09:14
Compare
Choose a tag to compare
Pre-release
1.4.4-rc1

chore(ci): bump runner base-images