查看原文
其他

一周活动预告(6.26-7.2)

请查收。


目录:
  1. 变分方法及其应用的若干问题 (邹文明)

    https://mathcenter.ctgu.edu.cn/info/1028/2031.htm

  2. Learning functionals using deep ReLU networks(樊军)

    https://math.fudan.edu.cn/bc/44/c30472a506948/page.htm 

  3. Thinking Fast with Transformers: Algorithmic Reasoning via Shortcuts (Bingbin Liu) https://www.mis.mpg.de/calendar/lectures/2023/abstract-35914.html


1. 变分方法及其应用的若干问题

  • 报告人: 邹文明(清华大学)

  • 报告时间: 2023-06-26  15:00

  • 报告地点: 腾讯会议ID: 498-372-609

  • 信息来源: 

    https://mathcenter.ctgu.edu.cn/info/1028/2031.htm

  • 报告摘要:

介绍变分方法及其应用方面若干进展和存在的问题,包括临界椭圆方程、Lane-Emden方程孤立奇点分析、Dancer 猜想、约束变分问题等等。

(灰色区域内上下滑动阅读全部内容)


2. Learning functionals using deep ReLU networks

  • 报告人: 樊军(香港浸会大学)

  • 报告时间: 2023-06-27  10:30-12:00

  • 报告地点: 腾讯会议ID: 273-665-687 密码: 200433

  • 信息来源: https://math.fudan.edu.cn/bc/44/c30472a506948/page.htm 

  • 报告摘要:

Functional neural networks have been proposed and studied as a means of approximating nonlinear continuous functionals defined on Lp spaces. However, their theoretical properties are largely unknown beyond the universality of approximation or the existing analysis does not apply to the rectified linear unit (ReLU) activation function. In this talk, I will present functional deep ReLU networks and explore their convergence rates of approximation and generalization errors under different regularity conditions. Additionally, I will discuss their applications to functional data analysis. This talk is based on joint works with Dirong Chen, Ying Liu, Linhao Song, and Dingxuan Zhou.


(灰色区域内上下滑动阅读全部内容)


3. Thinking Fast with Transformers: Algorithmic Reasoning via Shortcuts 


  • Speaker: Bingbin Liu (Carnegie Mellon University)

  • Time: 2023-6-29, 17:00, Berlin Time, 23:00 Beijing Time‍

  • Registration and Source Link: 

    https://www.mis.mpg.de/calendar/lectures/2023/abstract-35914.html

  • Abstract:

Algorithmic reasoning requires capabilities which are most naturally understood through recurrent models of computation, like the Turing machine. However, Transformer models, while lacking recurrence, are able to perform such reasoning using far fewer layers than the number of reasoning steps. This raises the question: what solutions are these shallow and non-recurrent models finding? In this talk, we will formalize reasoning in the setting of automata, and show that the computation of an automaton on an input sequence of length T can be replicated exactly by Transformers with o(T) layers, which we call "shortcuts". We provide two constructions, with O(log T) layers for all automata and O(1) layers for solvable automata. Empirically, our results from synthetic experiments show that shallow solutions can also be found in practice.

(灰色区域内上下滑动阅读全部内容)




我们是信息搬运工,

若活动临时有变化,

请以网站信息为准。



长按上方二维码,关注我们吧!
联系我们:sharecomptmath@foxmail.com
期待您和我们一起分享信息


欢迎转发,点赞,在看。👇


您可能也对以下帖子感兴趣

文章有问题?点此查看未经处理的缓存