You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
vllm under any version mismatch with current env and if you separate eval and train. I still need a version of vllm
in separate env, using the command below, I try different engine (4o and 4-turbo) and get some numbers dose not make sense. Have you ever try different annotators when use 4o, it give me a result where DPO>SimPO, while 4-turbo gives the opposite
Had a similar issue with ckpt [princeton-nlp/Llama-3-Instruct-8B-SimPO-v0.2](https://huggingface.co/princeton-nlp/Llama-3-Instruct-8B-SimPO-v0.2)
3 AE runs gives consistent LC/WR (48/44) where the number reported is 53.7/47.5.
One potential issue could be #75 and there is no vllm version specified in https://github.com/princeton-nlp/SimPO/blob/main/environment.yml
The text was updated successfully, but these errors were encountered: