Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
labmlai
GitHub Repository: labmlai/annotated_deep_learning_paper_implementations
Path: blob/master/translate_cache/RWKV/configs.zh.json
4922 views
1
{
2
"<h2>Transformer Configurations</h2>\n<p>This defines configurations for a transformer. The configurations are calculate using option functions. These are lazy loaded and therefore only the necessary modules are calculated.</p>\n": "<h2>Transformer Configurations</h2>\n<p>This defines configurations for a transformer. The configurations are calculate using option functions. These are lazy loaded and therefore only the necessary modules are calculated.</p>\n",
3
"<p>Dropout probability </p>\n": "<p>Dropout probability </p>\n",
4
"<p>Number of attention heads </p>\n": "<p>Number of attention heads </p>\n",
5
"<p>Number of layers </p>\n": "<p>Number of layers </p>\n",
6
"<p>Number of tokens in the source vocabulary (for token embeddings) </p>\n": "<p>Number of tokens in the source vocabulary (for token embeddings) </p>\n",
7
"<p>Number of tokens in the target vocabulary (to generate logits for prediction) </p>\n": "<p>Number of tokens in the target vocabulary (to generate logits for prediction) </p>\n",
8
"<p>Transformer embedding size </p>\n": "<p>Transformer embedding size </p>\n",
9
"configs.py": "configs.py"
10
}
11