You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, respected developers of Open Assistant. @andreaskoepf While studying your reward model training code, I noticed that besides the ranking loss, there is an additional L2 regularization term. What is the purpose of this regularization term? Are there any papers that mention it?
My apologies I don’t know if there is any research that covers it as I am
new to the AI system. However I can explain that I am interested in
attempting to make a conversational AI companion that responds with miare
personality so it’s less boring, I understand if you would like me to cease
some of my testing on your online version.--
The one, the only, Kaegan R. Bruce
Hello, respected developers of Open Assistant. @andreaskoepf While studying your reward model training code, I noticed that besides the ranking loss, there is an additional L2 regularization term. What is the purpose of this regularization term? Are there any papers that mention it?
Open-Assistant/model/model_training/utils/losses.py
Line 76 in 7e40ee3
The text was updated successfully, but these errors were encountered: