Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Would you consider adding Pythia12B, LLaMa and Alpaca since that's what you're directly compared against/based on?

GPT3.5/GPT4 is what everyone would also love to see but I understand you're performance is inline with GPT-neoX.

Vicuna/GPT4all would be intersting but IMO are less important.

RWKV would be interesting because it's a completely different model from the transformers.

EDIT: Also thanks for the opensource contributions! Highly appreciated!



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: