Colloquium | November 15 | 4-5 p.m. | Soda Hall, 306 (HP Auditorium)
About seventy years ago analog computing was regarded as having equal prospects as digital computing. Operational amplifiers could provide analog differentiation and integration functions. Nonetheless analog computing disappeared, being unable to provide the precision and dynamic range required for solving real problems.
The emergence of Deep Learning has been accompanied by the realization that only modest precision is required. This has taken us from regular Floating Point, to half-precision (16 bits), to quarter-precision, and with some difficulty even single-bit precision. At 8 bits and below, analog can do that, suggesting that analog matrix multiplication could provide more efficient Deep Learning accelerators.
In this brain-storming Colloquium, we will examine three different potential forms of analog computing.
(a) analog matrix multipliers for Deep Learning.
(b) literal annealing, not simulated annealing.
(c) adiabatic computing, (classical not quantum).
Let us examine whether any of these technologies could become the fore-runner of a new computing paradigm.
Eli Yablonovitch is Professor in Berkeley EECS, and Co-Chairman of the EECS Colloquium.