You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Have you ever experimented with LLama2-7B as the model to do C-RLFT? How about the performance? Because OpenChat-3.5-0106 is based on Mistral, performance is really high, I have tried using LLama2-7B, the performance is not satisfied.
Another 2 questions: can chat model be used as the model to do C-RLFT? I think, some code needs to be done, e.g., chat template, etc.
How about LLama3-8B-instuct, how to easy train, any performance data?
Thanks in advance.
The text was updated successfully, but these errors were encountered:
Thank you for your wonderful work!
Have you ever experimented with LLama2-7B as the model to do C-RLFT? How about the performance? Because OpenChat-3.5-0106 is based on Mistral, performance is really high, I have tried using LLama2-7B, the performance is not satisfied.
Another 2 questions: can chat model be used as the model to do C-RLFT? I think, some code needs to be done, e.g., chat template, etc.
How about LLama3-8B-instuct, how to easy train, any performance data?
Thanks in advance.
The text was updated successfully, but these errors were encountered: