Multiplication Free Neural Network Architectures

Thumbnail

Event details

Date 31.08.2023
Hour 12:0014:00
Speaker Bettina Messmer
Location
Category Conferences - Seminars
EDIC candidacy exam
Exam president: Prof. François Fleuret
Thesis advisor: Prof. Martin Jaggi
Co-examiner: Prof. Mathieu Salzmann

Abstract
The remarkable progress of deep learning models
has come with increased energy consumption, impacting the
environment, and limiting deployment on resource constrained
devices, such as mobile devices. Our research goal is to explore
resource-efficient alternatives to state-of-art network architectures,
specifically, we focus on designs that substantially reduce
multiplications. We discuss three significant works: a baseline
architecture notably reducing the multiplicative parameters, a
theoretical model reducing the number of multiplications to a
minimum while retaining universal approximation capabilities,
and a more general exploration of resource efficiency in the
natural language processing (NLP) domain.

Background papers
  • AdderNet: Do We Really Need Multiplications in Deep Learning? (here)
  • Min-Max-Plus Neural Networks (here)
  • pNLP-Mixer: an Efficient all-MLP Architecture for Language (here)

Practical information

  • General public
  • Free

Tags

EDIC candidacy exam

Share