# Optimization Seminar: Saeed Damadi

### UMBC

Location

Mathematics/Psychology : 401

Date & Time

March 7, 2024, 10:30 am – 12:00 pm

Description

**Title:**

*Sparsification of Neural Networks*

**Speaker:**Saeed Damadi

**Abstract:**

Neural networks are structured as vector-valued functions, which are not only easy to manage but can also be systematically expanded. When equipped with an appropriate loss function and minimization algorithm, they have the capability to learn from almost any dataset presented to them. This is not surprising to mathematicians due to the possibility of defining a function for a set comprising pairs of x (input) and y (output). The widespread adoption of neural networks is largely due to our ability to define these networks and approximate the mapping between a given domain and codomain. With this problem addressed, the subsequent question arises: Is this approximated function the most efficient in terms of the number of parameters used? In simpler terms, is it possible to eliminate some parameters without compromising the efficiency of the mapping? This leads to the concept of sparse optimization, where the goal is to solve it directly, without approximation. By doing so, we aim to find an approximate function that is less complex while maintaining its performance. In the context of computer science, solving sparse optimization allows us to provide a sparse neural network that is as effective as a dense neural network. This presentation will progressively elucidate this problem, requiring no prior knowledge of neural networks or optimization.

**Tags:**