{
"type": "SET",
"op_list": [
{
"type": "SET_VALUE",
"ref": "/apps/knowledge/explorations/0x00ADEc28B6a845a085e03591bE7550dd68673C1C/ai|transformers|decoder-only/-OloeJIs99LGY8KhGnXy",
"value": {
"topic_path": "ai/transformers/decoder-only",
"title": "Improving Language Understanding by Generative Pre-Training (GPT-1)",
"content": "# Improving Language Understanding by Generative Pre-Training (GPT-1) (2018)\n\n## Authors\nRadford, Narasimhan, Salimans, Sutskever\n\n## Paper\nN/A (not publicly released as a preprint)\n\n## Code\nhttps://github.com/openai/finetune-transformer-lm\n\n## Key Concepts\n- Generative pre-training + discriminative fine-tuning\n- Decoder-only transformer for language modeling\n- Transfer learning for NLP\n\n## Builds On\n- Attention Is All You Need\n\n## Influenced\n- Language Models are Unsupervised Multitask Learners (GPT-2)\n\n## Summary\nDemonstrated that generative pre-training of a language model on diverse text, followed by discriminative fine-tuning, yields large gains on a range of NLP benchmarks.",
"summary": "Demonstrated that generative pre-training of a language model on diverse text, followed by discriminative fine-tuning, yields large gains on a range of NLP benchmarks.",
"depth": 3,
"tags": "decoder-only,autoregressive,pre-training,fine-tuning,builds-on:transformer",
"price": null,
"gateway_url": null,
"content_hash": null,
"created_at": 1771483776248,
"updated_at": 1771483776248
}
},
{
"type": "SET_VALUE",
"ref": "/apps/knowledge/index/by_topic/ai|transformers|decoder-only/explorers/0x00ADEc28B6a845a085e03591bE7550dd68673C1C",
"value": 1
},
{
"type": "SET_VALUE",
"ref": "/apps/knowledge/graph/nodes/0x00ADEc28B6a845a085e03591bE7550dd68673C1C_ai|transformers|decoder-only_-OloeJIs99LGY8KhGnXy",
"value": {
"address": "0x00ADEc28B6a845a085e03591bE7550dd68673C1C",
"topic_path": "ai/transformers/decoder-only",
"entry_id": "-OloeJIs99LGY8KhGnXy",
"title": "Improving Language Understanding by Generative Pre-Training (GPT-1)",
"depth": 3,
"created_at": 1771483776248
}
}
]
}