Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Pre-training refers to unsupervised training that's done before a model is fine-tuned. The model still starts out random before it's pre-trained.

Here's where the Othello paper's weights are (randomly) initialized:

https://github.com/likenneth/othello_world/blob/master/mingp...



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: