In‑memory Euclidean distance enables energy‑efficient unsupervised learning, cutting data movement and hardware complexity for edge AI and next‑generation neuromorphic systems.
Neuromorphic computing aims to collapse the traditional memory‑processing divide that plagues von Neumann systems, especially as AI workloads grow in size and complexity. While memristor cross‑bars have already demonstrated in‑situ vector‑matrix multiplication for supervised networks, unsupervised algorithms such as self‑organizing maps require rapid Euclidean distance calculations—a subtraction‑heavy operation that earlier hardware could only approximate with peripheral digital units. The inability to perform subtraction natively has limited the energy‑saving promise of memristive platforms for clustering and pattern discovery tasks.
The breakthrough from Hanyang University introduces a vertically stacked 2 × 32 × 32 memristor array where the lower layer stores weight values and the upper layer stores their squares. A shared middle electrode simultaneously carries opposite‑polarity currents from both layers, naturally summing them to produce a net current proportional to the squared Euclidean distance. This elegant current‑subtraction pathway delivers a correlation coefficient of 0.96 with ideal calculations and a standard deviation of roughly 50 nA, despite device variations of about 10 %. The array supports 32 distinct conductance levels with ±37.5 nS tolerance, endures more than 10⁴ switching cycles, and retains data for years at room temperature. Real‑world tests—optimizing a ten‑city traveling‑salesman route, clustering the Iris dataset, and performing image‑color quantization—showed hardware outputs closely mirroring software baselines, confirming the architecture’s practical viability.
Beyond the laboratory, this in‑memory distance engine opens a path toward ultra‑low‑power edge AI modules that can pre‑process and cluster sensor streams before handing compact representations to downstream processors. By integrating computation and storage, the stacked design reduces data shuttling, cuts energy per operation, and occupies roughly half the silicon area of comparable planar solutions. Scaling to larger arrays will require careful management of voltage drops and parasitic coupling, and future work must integrate on‑chip learning updates to achieve fully autonomous operation. Nonetheless, the demonstrated approach aligns with industry roadmaps that prioritize compact, energy‑frugal neuromorphic accelerators for Internet‑of‑Things devices, autonomous vehicles, and next‑generation data centers.
Comments
Want to join the conversation?
Loading comments...