{
    "author": null,
    "date_published": "2018-05-01T01:24:00.000Z",
    "dek": null,
    "direction": "ltr",
    "domain": "arxiv.org",
    "excerpt": "We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers. Unlike recent language representation models, BERT is designed&hellip;",
    "lead_image_url": "https://static.arxiv.org/icons/twitter/arxiv-logo-twitter-square.png",
    "next_page_url": null,
    "rendered_pages": 1,
    "title": "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding",
    "total_pages": 1,
    "url": "https://arxiv.org/abs/1810.04805v2",
    "word_count": 167
}