EE Systems Seminar

Thursday June 26, 2014 4:00 PM

Inexactness, geometry, and optimization: recurrent themes in modern data analysis

Speaker: Suvrit Sra, Intelligent Systems, Max Planck Institut
Location: Moore B270

<!-- /* Font Definitions */ @font-face {font-family:Cambria; panose-1:2 4 5 3 5 4 6 3 2 4; mso-font-charset:0; mso-generic-font-family:auto; mso-font-pitch:variable; mso-font-signature:3 0 0 0 1 0;} /* Style Definitions */ p.MsoNormal, li.MsoNormal, div.MsoNormal {mso-style-parent:""; margin-top:0in; margin-right:0in; margin-bottom:10.0pt; margin-left:0in; mso-pagination:widow-orphan; font-size:12.0pt; font-family:"Times New Roman"; mso-ascii-font-family:Cambria; mso-ascii-theme-font:minor-latin; mso-fareast-font-family:Cambria; mso-fareast-theme-font:minor-latin; mso-hansi-font-family:Cambria; mso-hansi-theme-font:minor-latin; mso-bidi-font-family:"Times New Roman"; mso-bidi-theme-font:minor-bidi;} @page Section1 {size:8.5in 11.0in; margin:1.0in 1.25in 1.0in 1.25in; mso-header-margin:.5in; mso-footer-margin:.5in; mso-paper-source:0;} div.Section1 {page:Section1;} -->

The current data-age is witnessing an unprecedented confluence of

disciplines, blurring traditional domain boundaries. But what

aspects of data are driving this rich interaction? We can single out

at least two: size and structure.

 

Today, I will talk about both size and structure of data. In

particular, I will demonstrate how a large number of machine

learning problems (for instance regularized risk minimization,

dictionary learning and matrix factorization) fall into a generic

framework for scalable nonconvex optimization. I will highlight a

few applications that have benefited from this framework, while

commenting on ongoing and future work that strives for even greater

scalability.

 

Beyond size, I shall talk about "structure", specifically geometric

structure of data. My motivation lies in a number of applications of

machine learning and statistics to data that are not just vectors,

but richer objects such as matrices, strings, functions, graphs,

trees, etc. Processing such data in their "intrinsic representation"

can be of great value. Notably, we'll see examples where exploiting

the data geometry allows us to efficiently minimize a class of

nonconvex cost functions, not to local, but to global optimality.

 

Time permitting, I will also mention some surprising connections of

our work to areas beyond machine learning and data analysis.

 

Bio

Suvrit Sra is a Sr. Research Scientist at the Max Planck Institute

for Intelligent Systems, in Tübingen, Germany. He obtained his

Ph.D. in Computer Science from the University of Texas at Austin. He

has held visiting faculty positions at Carnegie Mellon University

(2013-14) in the Machine Learning Department and at UC Berkeley

(2013) in EECS.

 

His research involves several data-driven real-world applications,

which rely on a variety of tools from different mathematical areas

such as geometry, analysis, statistics, and optimization.

 

His work has won several awards; notably, the "SIAM 2011 Outstanding

Paper Prize". He regularly organizes a workshops on "Optimization

for Machine Learning" at the well-known Neural Information

Processing Systems (NIPS) conference, and has recently (co)-edited a

book with the same title.

Series Electrical Engineering Special Seminar

Contact: Shirley Slattery at x4715 shirley@systems.caltech.edu