Focus & Approach
Designing algorithmic methods to optimize the compute complexity of large-scale deep learning/spiking networks based on principled and interpretable metrics for robust, explainable and energy-efficient AI.
Exploring novel non-Von Neumann architecture solutions (such as analog dot product engines) with both standard CMOS and emerging technology for improved and energy-efficient AI hardware.
Designing novel hardware solutions (such as, discretization etc.) to address the algorithm vulnerabilities (such as, adversarial attacks, non-interpretability) of today’s AI systems and vice-versa.
Exploring bio-plausible algorithms and hardware guided by natural intelligence (how the brain learns, the internal fabric of the brain etc.) to define the next generation of robust and efficient AI systems for beyond-vision static recognition tasks with the ability to perceive, reason, decide autonomously in real-time.
Building adaptive learning solutions that are secure and reliable for enabling ubiquitous intelligence centered around IoT devices.
Our goal is to co-design solutions that can optimize 3-dimensional tradeoff between energy-accuracy/capability-robustness with advancements in algorithms-architectures-theory.
Research directions - Algorithm:
Learning algorithms for Accurate, Robust and Interpretable Deep Spiking Networks:
Spiking neural networks (SNNs) have emerged as next-generation deep learning models due to their huge energy efficiency benefits and biological plausibility. However, the development of energy-efficient training and interpretability still is immature compared to conventional deep learning, which limits its utility in real-world applications. To fill the gap, we designed new algorithms and architectures that improve the accuracy, robustness, and Interpretability of deep SNNs. We demonstrated that SNNs have a huge potential to synchronize their performance with conventional deep learning frameworks. We also proposed visualization tool (Spike Activation Map - SAM) for SNNs and observed SNNs provide reliable accuracy and explain (i.e., heatmap) regarding the adversarial attacks.
Figure. (a) Illustration of SNN with our proposed BNTT. (b) The average value of at each layer over all time-steps. (c) Early exit time can be calculated as since values at every layer are lower than threshold after time-step 20.
Beyond Vision Applications with Spiking Neural Networks:
Spiking Neural Networks (SNNs) are one of the emerging technologies providing an energy efficient alternative to traditional neural networks. As these are often targeted towards low power edge devices, these models need to be regularly updated as new data is generated, without leaking sensitive data. Federated Learning offers a practical solution to train neural networks on edge devices while preserving data privacy. We explore state of the art training methods and analyze the advantages of SNNs in a federated setting and study the advantages of SNNs over ANNs. Further, we apply SNNs to event-camera datasets and explore the advantage of sparse SNNs for time-series data.
Figure. Schematic diagram of Federated Learning.
Research directions - Hardware:
Using algorithm-hardware codesign approaches for improving adversarial robustness and energy efficient solutions for adversarial input detection. Specific layers in digital CMOS based accelerators can be designed using hybrid 8T-6T SRAM memories . Likewise, using algorithmic metrics like adversarial perturbations in intermediate layers can lead to lightweight and more structured adversarial input detection .
Hardware evaluation tool for large scale Spiking Neural Networks: