{
    "author": null,
    "date_published": null,
    "dek": null,
    "direction": "ltr",
    "domain": "kylrth.com",
    "excerpt": "It turns out that neural networks can reach training loss of 0 even on randomly labeled data, even when the data itself is random. It was previously thought that some implicit bias in the model&hellip;",
    "lead_image_url": null,
    "next_page_url": null,
    "rendered_pages": 1,
    "title": "Understanding deep learning requires rethinking generalization",
    "total_pages": 1,
    "url": "https://kylrth.com/paper/understanding-requires-rethinking-generalization/",
    "word_count": 1
}