-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reproduce the result on llama3. #11
Comments
Forget to mention, I am using llama-3-8b-instruct as the model. I wonder whether this is because the model is too weak to follow the instruction. Please let me know. |
Hi! Sorry for the inconvenience.
|
OK! Thanks for pointing that out! Unfortunately, there are some bad news. (I do add The result is that, the five blanks were never replaced, which actually is the result described in issue #9. As he also found there seems no difference for replacing I'm not sure if the problem occurs because of llama3 architecture or something else. I use llama2-style token and the problem seems to be the same.
|
input_ids
being on a device type different than your model's device
That's weird... Using llama2-7b-chat-hf seems not replicate the results either. Maybe that's issue related to dependency? I established the environment using miniconda, and here is the result by running
For further reference:
(I should also mention that running |
Hi! Here's the env I have
|
Can you also share the full code you are running? |
sorry for not being active these days, I'm using the updated code with the only modification on |
Sorry for another silly question. I am aware that the was an warning on your original
demo.ipynb
saying:on cell
In[4]
. I wonder how that will affect the results.So similar to issue #9. I couldn't reproduce a reasonable result. But I didn't change the
InterpretationPrompt
as issue #9 did. I changed thetokens_to_interpret
. But the result seems very weird. Would you mind checking my modification and pointing a way for me?I truly appreciate your exciting work, and am looking forward to your generous help.
Raw Code Blocks
The text was updated successfully, but these errors were encountered: