Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
labmlai
GitHub Repository: labmlai/annotated_deep_learning_paper_implementations
Path: blob/master/translate_cache/neox/utils/finetune.si.json
4922 views
1
{
2
"<p> </p>\n": "<p> </p>\n",
3
"<p>No need to train the mlp bias because we are adding it with attention output </p>\n": "<p>mlp\u0db1\u0dd0\u0db9\u0dd4\u0dbb\u0dd4\u0dc0 \u0db4\u0dd4\u0dc4\u0dd4\u0dab\u0dd4 \u0d9a\u0dd2\u0dbb\u0dd3\u0db8 \u0d85\u0dc0\u0dc1\u0dca\u0dba \u0db1\u0ddc\u0dc0\u0dda, \u0db8\u0db1\u0dca\u0daf \u0d85\u0db4\u0dd2 \u0d91\u0dba \u0d85\u0dc0\u0db0\u0dcf\u0db1\u0dba \u0dba\u0ddc\u0db8\u0dd4 \u0d9a\u0dd2\u0dbb\u0dd3\u0db8\u0dda \u0db4\u0dca\u0dbb\u0dad\u0dd2\u0daf\u0dcf\u0db1\u0dba \u0dc3\u0db8\u0d9f \u0d91\u0d9a\u0dad\u0dd4 \u0d9a\u0dbb\u0db1 \u0db6\u0dd0\u0dc0\u0dd2\u0db1\u0dd2 </p>\n",
4
"<p>Set <span translate=no>_^_0_^_</span> to <span translate=no>_^_1_^_</span> for the entire layer. </p>\n": "<p>\u0dc3\u0db8\u0dca\u0db4\u0dd6\u0dbb\u0dca\u0dab\u0dc3\u0dca\u0dad\u0dbb\u0dba <span translate=no>_^_0_^_</span> <span translate=no>_^_1_^_</span> \u0dc3\u0db3\u0dc4\u0dcf \u0dc3\u0d9a\u0dc3\u0db1\u0dca\u0db1. </p>\n",
5
"finetune.py": "finetune.py"
6
}
7