-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix rule and instructions for TGI #18
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
consider also changing the docker image name on README.md
docs/source/howto/serving.mdx
Outdated
Text-Generation-Inference (TGI) is a highly optimized serving engine enabling serving Large Language Models (LLMs) in a way | ||
that better leverages the underlying hardware, Cloud TPU in this case. | ||
|
||
TGI comes as ready to use Docker containers |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
add full stop at the end of the line. Also, should we mention the name?
docs/source/howto/serving.mdx
Outdated
-e HF_TOKEN=${HF_TOKEN} \ | ||
-e HF_BATCH_SIZE=1 \ | ||
-e HF_SEQUENCE_LENGTH=1024 \ | ||
huggingface/tpu-tgi \ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
is this name correct?
docs/source/howto/serving.mdx
Outdated
--max-batch-total-tokens 1024 | ||
``` | ||
|
||
### Executing requests against |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I guess you forgot this sentence
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
No description provided.