查看原文
其他

【直播】【青年科学半月谈】On the Statistical analysis of Scientific……

KouShare 蔻享学术 2023-07-09



活动名称:

On the Statistical analysis of Scientific Machine learning

活动时间

2023年4月13日(周四)10:00

报告嘉宾:

Yiping Lu

主办单位:

蔻享学术

直播通道

蔻享学术直播间

识别二维码,即可观看直播。


报告人介绍


Yiping Lu


Yiping Lu is a 4th year Ph.D. student at Stanford University working with Jose Blanchet and Lexing Ying. Before joining Stanford, he received his B.Sc. degree in Math from Peking University. Yiping is broadly interested in non-parametric statistics, applied probability, and stochastic simulation. He is particularly interested in applying machine learning and experimental design techniques to learn and solve models from applied probability and industrial engineering models.


报告简介


Massive data collection and computational capabilities have enabled data-driven scientific discoveries and control of engineering systems. However, there are still several questions that should be answered to understand the fundamental limits of just how much can be discovered with data and what is the value of additional information. For example, 1) How can we learn a physics law or economic principle purely from data? 2) How hard is this task, both computationally and statistically? 3) What’s the impact on hardness when we add further information (e.g., adding data, model information)? I’ll answer these three questions in this talk in two learning tasks. A key insight in both two cases is that using direct plug-in estimators can result in statistically suboptimal inference.For the first learning task, we focus on variational formulations for differential equation models. We discuss a prototypical Poisson equation. We provide a minimax lower bound for this problem. Based on the lower bounds, we discover that the variance in the direct plug-in estimator makes sample complexity suboptimal. We also consider the optimization dynamic for different variational forms. Finally, based on our theory, we explain an implicit acceleration of using a Sobolev norm as the objective function for training.The second learning task I’ll discuss is (linear) operator learning, which has wide applications in causal inference, time series modeling, and conditional probability learning. We build the first min-max lower bound for this problem. The min-max rate has a particular structure where the more challenging parts of the input and output spaces determine the hardness of learning a linear operator. Our analysis also shows that an intuitive discretization of the infinite-dimensional operator could lead to a sub-optimal statistical learning rate. Then, I’ll discuss how, by suitably trading-off bias and variance, we can construct an estimator with an optimal learning rate for learning a linear operator between infinite dimension spaces. We also illustrate how this theory can inspire a multilevel machine-learning algorithm of potential practical use.




推荐阅读

【青年科学半月谈】Uni-Mol: A Universal 3D Molecular...>>

【青年科学半月谈】机器学习贝叶斯力场及量化不确定性的分子动力学模拟>>

【青年科学半月谈】快速物理模拟中的降维方法>>

【青年科学半月谈】德国马普光科学研究所顾雪梅 学术报告>>

【青年科学半月谈】Simultaneous State Estimation and Dynamics……>>

【青年科学半月谈】Interaction-Driven Metal-Insulator Transition……>>

【青年科学半月谈】人工智能的产业化——人工智能科研的互补视角>>

【青年科学半月谈 】 平均场博弈论>>

【青年科学半月谈】制备多体张量网络态>>

【青年科学半月谈】可原位调控的集成光学克尔非线性>>

编辑:吴良秀

蔻享学术 平台


蔻享学术平台,国内领先的一站式科学资源共享平台,依托国内外一流科研院所、高等院校和企业的科研力量,聚焦前沿科学,以优化科研创新环境、传播和服务科学、促进学科交叉融合为宗旨,打造优质学术资源的共享数据平台。

识别二维码,

下载 蔻享APP  查看最新资源数据。


点击阅读原文,查看更多精彩报告!

您可能也对以下帖子感兴趣

文章有问题?点此查看未经处理的缓存