From Theoretical Understanding of Neural Networks to Practical Applications

Thumbnail

Event details

Date 10.07.2023
Hour 10:3012:30
Speaker Yongtao Wu
Location
Category Conferences - Seminars
DIC candidacy exam
Exam president: Prof. Nicolas Flammarion
Thesis advisor: Prof. Volkan Cevher
Co-examiner: Prof. Martin Jaggi

Abstract
Deep learning has demonstrated unprecedented success in influential applications ranging from vision tasks to language modeling. The design of network architecture plays a pivotal role in its performance, as evident from the development of ResNet, EfficientNet, and Transformer. These achievements have ignited a profound interest in theoretically understanding neural networks across various topics, such as convergence, generalization, and learnability, which can also significantly contribute to practical applications. In this write-up, we first delve into the convergence of feedforward neural networks. Subsequently, we will examine a study on Transformer from the perspective of generalization. Lastly, we will introduce a theoretical work on in-context learning within the Transformer model.

Background papers

Practical information

  • General public
  • Free

Tags

EDIC candidacy exam

Share