Please use this identifier to cite or link to this item:
https://dspace.iiti.ac.in/handle/123456789/18153
Full metadata record
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Lokhande, Mukul | en_US |
| dc.contributor.author | Sankhe, Akash | en_US |
| dc.contributor.author | Jaya Chand, S.V. | en_US |
| dc.contributor.author | Mishra, Shivangi | en_US |
| dc.contributor.author | Vishvakarma, Santosh Kumar | en_US |
| dc.date.accessioned | 2026-05-14T12:28:14Z | - |
| dc.date.available | 2026-05-14T12:28:14Z | - |
| dc.date.issued | 2025 | - |
| dc.identifier.citation | Lokhande, M., Sankhe, A., Jaya Chand, Mishra, S., & Vishvakarma, S. K. (2025). FERMI-ML: A Flexible and Resource-Efficient Memory-In-Situ SRAM Macro for TinyML acceleration. Proceedings of the International Conference on Microelectronics, ICM. https://doi.org/10.1109/ICM66518.2025.11321332 | en_US |
| dc.identifier.isbn | 979-833159370-4 | - |
| dc.identifier.issn | 2332-7014 | - |
| dc.identifier.other | EID(2-s2.0-105032437259) | - |
| dc.identifier.uri | https://dx.doi.org/10.1109/ICM66518.2025.11321332 | - |
| dc.identifier.uri | https://dspace.iiti.ac.in:8080/jspui/handle/123456789/18153 | - |
| dc.description.abstract | The growing demand for low-power and area-efficient TinyML inference on AIoT devices necessitates memory architectures that minimise data movement while sustaining high computational efficiency. This paper presents FERMI-ML, a Flexible and Resource-Efficient Memory-In-Situ (MIS) SRAM macro designed for TinyML acceleration. The proposed 9T XNOR-based RX9T bit-cell integrates a 5T storage cell with a 4T XNOR compute unit, enabling variable-precision MAC and CAM operations within the same array. A 22-transistor (C22T) compressor-tree-based accumulator facilitates logarithmic 1-64-bit MAC computation with reduced delay and power compared to conventional adder trees. The 4 KB macro achieves dual functionality for in-situ computation and CAM-based lookup operations, supporting Posit-4/FP-4 precision. Post-layout results at 65 nm show operation at 350 MHz with 0.9 V, delivering a throughput of 1.93 TOPS and an energy efficiency of 364 TOPS/W, while maintaining a Quality-of-Result (QoR) above 97.5% with Inception-V4 and ResNet-18. FERMI-ML thus demonstrates a compact, reconfigurable, and energy-aware digital Memory-In-Situ macro capable of supporting mixed-precision TinyML workloads. © 2025 IEEE. | en_US |
| dc.language.iso | en | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | en_US |
| dc.source | Proceedings of the International Conference on Microelectronics, ICM | en_US |
| dc.title | FERMI-ML: A Flexible and Resource-Efficient Memory-In-Situ SRAM Macro for TinyML acceleration | en_US |
| dc.type | Conference Paper | en_US |
| dc.rights.license | All Open Access | - |
| dc.rights.license | Green Open Access | - |
| Appears in Collections: | Department of Electrical Engineering | |
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.
Altmetric Badge: