• Skip to content
  • Skip to primary sidebar

Dharmendra S. Modha

My Work and Thoughts.

  • Brain-inspired Computing
    • Collaborations
    • Videos
  • Life & Universe
    • Creativity
    • Leadership
    • Interesting People
  • Accomplishments
    • Prizes
    • Papers
    • Positions
    • Presentations
    • Press
    • Profiles
  • About Me

Archives for September 2016

Breaking News: Energy-Efficient Neuromorphic Classifiers

September 23, 2016 By dmodha

A paper entitled “Energy-Efficient Neuromorphic Classifiers” was published this week in Neural Computation by Daniel Martí, Mattia Rigotti, Mingoo Seok, and Stefano Fusi.

Here is the abstract:

Neuromorphic engineering combines the architectural and computational principles of systems neuroscience with semiconductor electronics, with the aim of building efficient and compact devices that mimic the synaptic and neural machinery of the brain. The energy consumptions promised by neuromorphic engineering are extremely low, comparable to those of the nervous system. Until now, however, the neuromorphic approach has been restricted to relatively simple circuits and specialized functions, thereby obfuscating a direct comparison of their energy consumption to that used by conventional von Neumann digital machines solving real-world tasks. Here we show that a recent technology developed by IBM can be leveraged to realize neuromorphic circuits that operate as classifiers of complex real-world stimuli. Specifically, we provide a set of general prescriptions to enable the practical implementation of neural architectures that compete with state-of-the-art classifiers. We also show that the energy consumption of these architectures, realized on the IBM chip, is typically two or more orders of magnitude lower than that of conventional digital machines implementing classifiers with comparable performance. Moreover, the spike-based dynamics display a trade-off between integration time and accuracy, which naturally translates into algorithms that can be flexibly deployed for either fast and approximate classifications, or more accurate classifications at the mere expense of longer running times and higher energy costs. This work finally proves that the neuromorphic approach can be efficiently used in real-world applications and has significant advantages over conventional digital devices when energy consumption is considered.

Filed Under: Brain-inspired Computing, Collaborations

Deep learning inference possible in embedded systems thanks to TrueNorth

September 21, 2016 By dmodha

See the beautiful blog by Caroline Vespi on our PNAS paper “Convolutional networks for fast, energy-efficient neuromorphic computing“.

Filed Under: Accomplishments, Brain-inspired Computing, Press

Breaking News: Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing

September 20, 2016 By dmodha

Guest Blog by Steven K. Esser

We are excited to announce that our work demonstrating convolutional networks for energy-efficient neuromorphic hardware just appeared in the peer-reviewed Proceedings of the National Academy of Sciences (PNAS) of the United States of America. Here is an open-access link. In this work, we show near state-of-the-art accuracy on 8 datasets, while running at between 1200 and 2600 frames per second and using between 25 and 275 mW running on the TrueNorth chip. Our approach adapts deep convolutional neural networks, today’s state-of-the-art approach for machine perception, to perform classification tasks on neuromorphic hardware, such as TrueNorth. This work overcomes the challenge of learning a network that uses low precision synapses, spiking neurons and limited-fan in, constraints that allow for energy-efficient neuromorphic hardware design (including TrueNorth) but that are not imposed in conventional deep learning.

Low precision synapses make it possible to co-locate processing and memory in hardware, dramatically reducing the energy needed for data movement. Employing spiking neurons (which have only a binary state) further reduces data traffic and computation, leading to additional energy savings. However, contemporary convolutional networks use high precision (at least 32-bit) neurons and synapses. This high precision representation allows for very small changes to the neural network during learning, facilitating incremental improvements of the response to a given input during training. In this work, we introduce two constraints to the learning rule to bridge deep learning to energy-efficient neuromorphic hardware. First, we maintain a high precision “shadow” synaptic weight during training, but map this weight to a trinary value (-1, 0 or 1) implementable in hardware for operation as well as, critically, to determine how the network should change during the training itself. Second, we employ spiking neurons during both training and operation, but use the behavior of a high precision neuron to predict how the spiking neuron should change as network weights change.

Limiting neuron fan-in (the number of inputs each neuron can receive) allows for a crossbar design in hardware that reduces spike traffic, again leading to lower energy consumption. However, in deep learning no explicit restriction is placed on inputs per neuron, which allows networks to scale while still remaining integrated. In this work, we enforce this connectivity constraint by partitioning convolutional filters into multiple groups. We maintain network integration by interspersing layers with large topographic coverage but small feature coverage with those with small topographic coverage but large feature coverage. Essentially, this allows the network to learn complicated spatial features, and then to learn useful mixtures of those features.

Our results show that the operational and structural differences between neuromorphic computing and deep learning are not fundamental and indeed, that near state-of-the-art performance can be achieved in a low-power system by merging the two. In a previous post, we mentioned one example of an application running convolutional networks on TrueNorth, trained with our method, to perform real-time gesture recognition.

The algorithm is now in the hands of over 130 users at over 40 institutions, and we are excited to discover what else might now be possible!

Filed Under: Accomplishments, Brain-inspired Computing, Papers

Career Opportunity: Brain-inspired Computing

September 1, 2016 By dmodha

Apply here.

Filed Under: Brain-inspired Computing

Primary Sidebar

Follow Me

  • Linkedin
  • Twitter

Recent Tweets

  • "The Air Force has embedded a member of its team, for instance, with IBM researchers working on chips for a neuromo… https://t.co/uUMXDtd00m49 days ago
  • RT @cgvprasad: Fitting end to the #IBMTLE NorCal with another eye opening session by @DharmendraModha !! Overall some interesting topics to…50 days ago
  • An open secret of project management is 'closing open loops'. Loops are of 4 types: tasks that you have assigned s… https://t.co/t9npXYPN6l77 days ago
  • AFRL Anticipates Arrival of Neuromorphic Supercomputer https://t.co/uo34hy7Lbz @IBMResearch80 days ago
  • CALL FOR SUBMISSIONS: The Third Misha Mahowald Prize for Neuromorphic Engineering https://t.co/2cVLV1fq3N121 days ago
  • What’s Next With Computing? https://t.co/GUQFX37ffJ122 days ago
  • The head of IBM's San Jose research center is at home in both the kitchen and the lab https://t.co/7pNMeeLbbQ via… https://t.co/tmdAB2rkXS133 days ago
  • Brain-inspired Computing Positions https://t.co/6clTNBogrM @IBMResearch133 days ago
  • Research Summer Internship in Brain Inspired Computing @IBMResearch https://t.co/IrCSD49t1t134 days ago
  • "The US Air Force Research Lab is testing a 64-chip array of our IBM TrueNorth Neurosynaptic System designed for de… https://t.co/Lp5uhr0Qxa134 days ago

Recent Posts

  • Spiking Optical Flow for Event-based Sensors Using IBM’s TrueNorth Neurosynaptic System
  • Video of Keynote (Commencement Speech) at 2017 UCSD School of Engineering Ring Ceremony
  • Brain-inspired Computing Video Gallery
  • IBM demos event-based gesture recognition using a brain-inspired chip at CVPR 2017
  • Keynote Speech at the Ring Ceremony at UCSD’s Jacobs School of Engineering

Archives by Month

  • 2017: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2016: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2015: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2014: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2013: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2012: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2011: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2010: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2009: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2008: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2007: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
  • 2006: Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec

Evolution: Brain-inspired Computing

Copyright © 2018