Replies: 5 comments
-
Any model that is based on HF-Transformers CaualLM should be usable without problems. If you have compute you could try fine-tuning a GPT JT model on the OA dataset. |
Beta Was this translation helpful? Give feedback.
-
Would GPT JT be considered as an option as a pretrained model to fine-tune into the final model? |
Beta Was this translation helpful? Give feedback.
-
Yes, that's definitely something we should try. Also probably the larger togethercomputer/GPT-NeoXT-Chat-Base-20B. |
Beta Was this translation helpful? Give feedback.
-
Yeah, that sounds good. Maybe an open assistant model trained from GPT JT can be Open Assistant Large and one trained from GPT NeoXT chat base can be Open Assistant Extra Large |
Beta Was this translation helpful? Give feedback.
-
@sanagno what are your thoughts on this? |
Beta Was this translation helpful? Give feedback.
-
According to Yannic Kilcher’s video about Open Assistant, the MVP is a model like InstructGPT. However, a model similar to InstructGPT already exists called GPT JT. Can’t we skip training an InstructGPT like model and use GPT JT?
Beta Was this translation helpful? Give feedback.
All reactions