Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: stream tokens usage #4415

Merged
merged 6 commits into from
Dec 18, 2024
Merged

Conversation

mintyleaf
Copy link
Contributor

@mintyleaf mintyleaf commented Dec 16, 2024

Description

This PR fixes #4334

Notes for Reviewers

Got some time to take a look at log of issue with hanging webui response - found that llama sends the empty message from the start for some reason

There was an unnecessary usage of such case to break the channel loop, which was overlooked
Still can't reproduce that by myself, since i can't really run more advanced models, than phi2 on my 8/256 machine
Yet, tokens count still works like before and code from http/openai/chat remains untouched

@mudler can you confirm that everything is working now?

Copy link

netlify bot commented Dec 16, 2024

Deploy Preview for localai ready!

Name Link
🔨 Latest commit 4f7b373
🔍 Latest deploy log https://app.netlify.com/sites/localai/deploys/676260f7c3dde00008fea92f
😎 Deploy Preview https://deploy-preview-4415--localai.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

@mudler mudler changed the title Fix/stream tokens usage feat: stream tokens usage Dec 18, 2024
@mudler
Copy link
Owner

mudler commented Dec 18, 2024

Just did tested this and seems to work, thanks!

@mudler mudler merged commit 2bc4b56 into mudler:master Dec 18, 2024
28 of 30 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

After updating to 2.24 LLM hangs after first response
2 participants