学术报告

您所在的位置:首页  学术交流  学术报告

Be aware of model capacity when talking about generalization in machine learning

发布时间:2025-04-07阅读次数:10

In this talk, I will discuss how the test error will behave if a more suitable metric than model size for model capacity is used. To be specific, I will present a unified perspective on generalization by analyzing how norm-based model capacity control reshapes our understanding of these foundational concepts: there is no bias-variance trade-offs; phase transition exists from under-parameterized regimes to over-parameterized regimes while double descent doesn't exist; scaling law is formulated as a multiplication style under norm-based capacity. Additionally, I will briefly discuss which norm is suitable for neural networks and what are the fundamental limits of learning efficiency imposed by such norm-based capacity from the perspective of function space. Talk based on https://arxiv.org/abs/2502.01585, https://arxiv.org/abs/2404.18769

学术海报.pdf