Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The pretraining (so 99% of training) is fully global, in over 1000 languages without special weighting. The posttraining (See section 4 of the paper) had also as many languages as we could get, and did upweight some languages. The posttraining can easily be customized to any other target languages


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: