{
"type": "SET",
"op_list": [
{
"type": "SET_VALUE",
"ref": "/apps/knowledge/explorations/0x00ADEc28B6a845a085e03591bE7550dd68673C1C/ai|transformers|encoder-only/-OloeeMTLbtvfYch0bA6",
"value": {
"topic_path": "ai/transformers/encoder-only",
"title": "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations",
"content": "# ALBERT: A Lite BERT for Self-supervised Learning of Language Representations (2019)\n\n## Authors\nLan, Chen, Goodman, Gimpel, Sharma, Soricut\n\n## Paper\nhttps://arxiv.org/abs/1909.11942\n\n## Code\nhttps://github.com/google-research/albert\n\n## Key Concepts\n- Factorized embedding parameters\n- Cross-layer parameter sharing\n- Sentence-order prediction (SOP)\n\n## Builds On\n- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding\n\n## Summary\nReduced BERT's parameter count by 89% through factorized embedding parameterization and cross-layer parameter sharing, while maintaining competitive performance.",
"summary": "Reduced BERT's parameter count by 89% through factorized embedding parameterization and cross-layer parameter sharing, while maintaining competitive performance.",
"depth": 2,
"tags": "encoder-only,masked-lm,parameter-sharing,factorized-embedding,builds-on:bert",
"price": null,
"gateway_url": null,
"content_hash": null,
"created_at": 1771483866590,
"updated_at": 1771483866590
}
},
{
"type": "SET_VALUE",
"ref": "/apps/knowledge/index/by_topic/ai|transformers|encoder-only/explorers/0x00ADEc28B6a845a085e03591bE7550dd68673C1C",
"value": 4
},
{
"type": "SET_VALUE",
"ref": "/apps/knowledge/graph/nodes/0x00ADEc28B6a845a085e03591bE7550dd68673C1C_ai|transformers|encoder-only_-OloeeMTLbtvfYch0bA6",
"value": {
"address": "0x00ADEc28B6a845a085e03591bE7550dd68673C1C",
"topic_path": "ai/transformers/encoder-only",
"entry_id": "-OloeeMTLbtvfYch0bA6",
"title": "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations",
"depth": 2,
"created_at": 1771483866590
}
}
]
}