Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Towards Sequence-to-Sequence Models Without Activation Functions

Simons Institute via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
This talk by Grigorios Chrysos from the University of Wisconsin-Madison explores the fundamental question of whether neural networks truly need activation functions. Examine how high-order interactions among input elements might provide sufficient expressivity for complex tasks without traditional activation functions. Learn about the challenges activation functions pose for deep learning theory, network dynamics analysis, interpretability, and privacy. Discover research findings on networks that achieve strong performance in demanding static tasks like ImageNet recognition and sequence-to-sequence tasks such as arithmetic operations and language modeling without conventional activation functions. Part of the "The Future of Language Models and Transformers" series at the Simons Institute, this 56-minute presentation challenges core assumptions about neural network architecture design.

Syllabus

Towards sequence-to-sequence models without activation functions

Taught by

Simons Institute

Reviews

Start your review of Towards Sequence-to-Sequence Models Without Activation Functions

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.