The pychatml
package allows you to convert chat interfaces from and to the ChatML format.
You can install the pychatml
package using pip:
pip install pychatml
Makes it easy to integrate between different chat formats and models.
from pychatml.llama2_converter import Llama2
PROMPT = """[INST] Hi, how are you? [/INST] Good thanks!
[INST] Can you help me with this math program? [/INST]"""
converter = Llama2()
converter.to_chatml(PROMPT)
[
{"role": "user", "content": "Hi, how are you?"},
{"role": "assistant", "content": "Good thanks!"},
{"role": "user", "content": "Can you help me with this math program?"},
]
- Llama 2
- Anthropic
- Alpaca
- Vicuna/ShareGPT (from https://github.com/lm-sys/FastChat/blob/e67b21dcbde91a5ad9740b081c59c433093f54da/fastchat/conversation.py#L394C1-L404C2)
Create an issue or discussion in this repository.
Or, reach out to our team! @jakob_frick, @__anjor, @maxnajork on X or [email protected].
A new release will automatically be published to pypi. For this to happen you need to
- update the version specificed in
setup.py
- create a
git tag
with the version you want to release and push it viagit push --tags
- create a new release in Github
A Github action will automatically publish the new version to pypi.
Thank you for your interest in contributing to our project! Before you begin writing code, it would be helpful if you read these contributing guidelines. Following them will make the contribution process easier and more efficient for everyone involved.
Please note that the project is released with an MIT License.