BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//University of California\, Berkeley//UCB Events Calendar//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
BEGIN:VTIMEZONE
TZID:America/Los_Angeles
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
DTSTART:19701029T020000
RRULE:FREQ=YEARLY;BYMONTH=11;BYDAY=1SU
END:STANDARD
BEGIN:DAYLIGHT
DTSTART:19700402T020000
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
RRULE:FREQ=YEARLY;BYMONTH=3;BYDAY=2SU
END:DAYLIGHT
END:VTIMEZONE
BEGIN:VEVENT
DTSTAMP:20171003T203145Z
DTSTART;TZID=America/Los_Angeles:20171009T153000
DTEND;TZID=America/Los_Angeles:20171009T170000
TRANSP:OPAQUE
SUMMARY:Peter Bartlett - Representation\, optimization and generalization in deep learning
UID:112266-ucb-events-calendar@berkeley.edu
ORGANIZER;CN="UC Berkeley Calendar Network":
LOCATION:3108 Etcheverry Hall
DESCRIPTION:Peter Bartlett\, UC Berkeley Departments of Statistics and EECS\n\nDeep neural networks have improved state-of-the-art performance for prediction problems across an impressive range of application areas\, and they have become a central ingredient in AI systems. This talk considers factors that affect their performance\, describing some recent results in two directions. First\, we investigate the impact of depth on representation and optimization properties of these networks. We focus on deep residual networks\, which have been widely adopted for computer vision applications because they exhibit fast training\, even for very deep networks. We show that as the depth of these networks increases\, they are able to represent a smooth invertible map with a simpler representation at each layer\, and that this implies a desirable property of the functional optimization landscape that arises from regression with deep function compositions: stationary points are global optima. Second\, we consider the generalization behavior of deep networks\, that is\, how their performance on training data compares to predictive accuracy. In particular\, we aim to understand how to measure the complexity of functions computed by these networks. For multiclass classification problems\, we present a margin-based generalization bound that scales with a certain margin-normalized "spectral complexity\," involving the product of the spectral norms of the weight matrices in the network. We show how the bound gives insight into the observed performance of these networks in practical problems.\n\nJoint work with Steve Evans and Phil Long\, and with Matus Telgarsky and Dylan Foster.
URL:http://events.berkeley.edu/index.php/calendar/sn/pubaff.html?event_ID=112266&view=preview
SEQUENCE:0
CLASS:PUBLIC
CREATED:20171003T203145Z
LAST-MODIFIED:20171003T203145Z
X-MICROSOFT-CDO-BUSYSTATUS:BUSY
X-MICROSOFT-CDO-INSTTYPE:0
X-MICROSOFT-CDO-IMPORTANCE:1
X-MICROSOFT-CDO-OWNERAPPTID:-1
END:VEVENT
END:VCALENDAR