0% found this document useful (0 votes)
70 views12 pages

The Development of General-Purpose Brain-Inspired Computing

The document discusses the development of general-purpose brain-inspired computing (GPBIC), which aims to create efficient computing systems by mimicking biological mechanisms. It explores hardware and software advancements, emphasizing the integration of neuroscience and computer science to enhance computing capabilities across various applications. The authors highlight the need for strategic initiatives in application-level, hardware-level, and software-level generalization to advance the field of GPBIC.

Uploaded by

PhilFoden
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
70 views12 pages

The Development of General-Purpose Brain-Inspired Computing

The document discusses the development of general-purpose brain-inspired computing (GPBIC), which aims to create efficient computing systems by mimicking biological mechanisms. It explores hardware and software advancements, emphasizing the integration of neuroscience and computer science to enhance computing capabilities across various applications. The authors highlight the need for strategic initiatives in application-level, hardware-level, and software-level generalization to advance the field of GPBIC.

Uploaded by

PhilFoden
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 12

nature electronics

Perspective https://doi.org/10.1038/s41928-024-01277-y

The development of general-purpose


brain-inspired computing

Received: 5 May 2023 Weihao Zhang 1,2,3,4,5,7, Songchen Ma , Xinglong Ji1,2,3,4,5, Xue Liu1,2,3,4,5,
1,2,3,4,5,7

Yuqing Cong1,6 & Luping Shi 1,2,3,4,5


Accepted: 2 October 2024

Published online: 7 November 2024


Brain-inspired computing uses insights from neuroscience to develop more
Check for updates efficient computing systems. The approach is of use in a broad range of
applications—from neural simulation to intelligent computing—and could
potentially be used to create a general-purpose computing infrastructure.
Here we explore the development of general-purpose brain-inspired
computing. We examine the hardware and software that have so far been
used to create brain-inspired computing systems. We then consider the
potential of combining approaches from neuro­science and computer
science to build general-purpose brain-inspired computing systems,
highlighting three areas: temporal, spatial and spatiotemporal capabilities;
approximate computing and precise computing; and control flow and
data flow. Finally, we discuss initiatives that will be needed to develop
general-purpose brain-inspired computing, highlighting three potential
strategies: application-level pattern generalization, hardware-level
structural generalization and software-level systematic generalization.

Brain-inspired computing1, which is also known as neuromorphic diversity in workloads, to accommodate different resource demands
computing, aims to mimic biological mechanisms to provide new in terms of computation, storage and other computing resources;
approaches to studying the brain and developing artificial intelligent and diversity in performance requirements, to encompass specific
systems. Initially, neuromorphic research primarily sought to replicate requirements for latency, throughput and power consumption.
or validate biological findings on hardware platforms, with insights A practical general-purpose system should also offer programmability
from neuroscience guiding the development of such systems2,3. The that allows convenient deployment of applications and flexible control
recent explosion of big data, along with new technologies such as over computing resources.
artificial intelligence and the Internet of Things, has led to a surge in Modern computing models tend to solve problems through a
demand for high-performance computing, and thus researchers have control-flow mechanism with fast instruction and state switching,
also turned to brain-inspired computing as a potential approach to which leverages temporal processing capabilities. Parallelism of such
deliver more efficient computing systems. Brain-inspired computing a computing model is typically achieved by replicating control-flow
could be of use in a broad range of applications4 and could provide a processes or directing multiple uniform data operations with one
basis for future computing infrastructures5. But what would it take to control-flow process. In contrast, the human brain relies on mas-
develop general-purpose brain-inspired computing (GPBIC)? sively interconnected neurons that transmit information via sparse
A general-purpose system needs to address the diversity of prac- spikes. Dynamic topology and neural dynamics play a crucial role in
tical applications. This includes diversity in functionality, to support intelligence6,7, and the temporal correlation of spikes establishes the
a range of functions and algorithms with acceptable performance; basis for learning. In essence, the brain tends to solve problems by

1
Center for Brain-inspired Computing Research (CBICR), Tsinghua University, Beijing, China. 2Department of Precision Instruments, Tsinghua University,
Beijing, China. 3Optical Memory National Engineering Research Center, Tsinghua University, Beijing, China. 4Tsinghua University- China Electronics
Technology HIK Group Co. Joint Research Center for Brian-inspired Computing, Tsinghua University, Beijing, China. 5IDG/McGovern Institute for Brain
Research at Tsinghua University, Beijing, China. 6Department of Computer Science and Technology, Tsinghua University, Beijing, China. 7These authors
contributed equally: Weihao Zhang, Songchen Ma. e-mail: [email protected]

Nature Electronics | Volume 7 | November 2024 | 954–965 954


Perspective https://doi.org/10.1038/s41928-024-01277-y

Application scope synaptic dynamics, types and numbers of ion channels and compart-
Neuroscience Intelligence General ments), the number of modelling neurons, the density and complexity
of connections, and timing on multiple scales (such as long- and short-
term synaptic plasticity). These factors affect the circuit designs and
Brain Classical
paradigm neuromorphic final performance.
engineering When designing chips, researchers typically address design
Brain-
inspired trade-offs from two perspectives: research goals and biological
computing references (such as spike firing rates, the locality of neuronal con-
Dual-driven
Dual-driven nections and cortex orientation references). This leads to each chip
paradigm
brain-inspired computing
being designed with common metrics, while also having its specific
focus (Fig. 2).
Computer Initially, neuromorphic engineering focused on developing neuro-
paradigm Dedicated Classical
super- morphic sensors such as artificial retinas or cochlea based on relatively
computing computing
established biological principles. Among various neuromorphic sen-
sors, visual sensors have received the most attention (such as DVS10,
Paradigm ATIS11, CeleX12, Vidar13 and Tianmouc14) and have then been designed by
Fig. 1 | Guiding paradigms and applications of brain-inspired computing. drawing from the peripheral perception of motion and the fine texture
Neuromorphic engineering has historically been guided by brain paradigms, and pattern recognition function of the central fovea. These sensors
replicating features or functions of biological nervous systems for exploring serve as effective interfaces between the environment and cognitive
neuroscience and accelerating intelligent tasks. Beyond these, the fusion of systems with sensitivity to dynamic signals and substantially reduce
computational paradigms inherent to both the brain and classical computers signal processing complexity and communication costs.
could lead to the efficient and flexible management of a broader range of general- When modelling a small-scale network of silicon neurons, a hard-
purpose tasks. wired approach usually suffices15. However, a programmable approach
is necessary to model different types of neuron, ion channel and
dynamic connection with growing complexity. To meet this demand,
adopting spatiotemporal processing capabilities, integrating memory different technical approaches are adopted. Analogue circuits are effi-
and computation efficiently in an event-driven data-flow manner. cient at implementing dynamic and differential processes with various
Thus, a promising approach to developing an advanced computing time constants, making them suitable for building spiking neurons with
system—and GPBIC—is to combine features of modern computers and rich spatiotemporal complexity (such as Neurogrid16, BrainScaleS17
the human brain8 (Fig. 1). (formerly known as FACETS18,19) and DYNAPs20). To model more neurons
In this Perspective, we examine the development of brain-inspired within a limited area, various circuit-sharing and parameter-sharing
computing and consider the hardware and software that could pro- techniques have been developed16. Furthermore, many-core architec-
vide a foundation for GPBIC. We explore the potential of combining tures with a hierarchical communication infrastructure21 have been
approaches from neuroscience and computer science to build GPBIC, employed for higher scalability, while digital circuits provide more
highlighting three areas: temporal, spatial and spatiotemporal capabili- reliable and reconfigurable connections.
ties; approximate computing and precise computing; and control flow The inter-chip communication bandwidth is an important factor of
and data flow. Finally, we consider the short-term and long-term initia- neuromorphic chips. Inspired by supercomputers, SpiNNaker22 organ-
tives that will be required to develop GPBIC, highlighting three poten- izes large-scale digital-logic neurons instead of implementing specific
tial strategies: application-level pattern generalization, hardware-level neuron circuits. It consists of optimized advanced RISC machine (ARM)
structural generalization and software-level systematic generalization. processors interconnected through multiple levels of communication
infrastructure. Owing to resource limits, the design of neuromorphic
The development of brain-inspired computing chips with large-scale neurons and synapses involves a careful bal-
The field of brain-inspired computing has undergone considerable ance of various trade-offs, such as circuit complexity, communication
developments along four main routes (Fig. 2). Initially, it centred around bandwidth and energy efficiency, to achieve optimal performance for
neuromorphic sensing engineering, then advanced to brain simula- a specific application.
tion and intelligent computing, and more recently, general-purpose In light of the advancements in big data and artificial intelli-
computing. We highlight here the key hardware and software advances gence, coupled with the end of Moore’s law, the field of intelligent
along each of these routes, as well as the construction of large-scale computing23–25 has garnered considerable attention within the
computing systems. (Note, algorithms are not covered in this article, brain-inspired computing community. Along this route, there has been
and have been documented extensively elsewhere4,9.) a noticeable shift in the emphasis on hardware design objectives. Ini-
tially centred on replicating biological behaviour with complex neuro-
Hardware scientific modelling, the current emphasis is on practical applications
The precise mechanism underlying the brain’s computational power with simplified neuron models26,27. This has led to the development of
remains unclear, but it is believed to arise from the coordinated activity several chips using fully digital circuits to build logic neurons (such as
of numerous neurons and their connections, which undergo dynamic TrueNorth28,29, IMPULSE30, Tianjic31, Darwin32,33, Novena34 and µBrain35).
changes in connection strength and availability to support learning TrueNorth features a hybrid synchro–asynchronous architecture for
and adaptability. increased neuron capacity and multi-chip scalability. Loihi-136 offers
To emulate this process in hardware, researchers aim to replicate a specialized instruction set, facilitating the programming of various
the brain’s structural features to create a computational substrate local learning rules. ODIN37 supports long-term potentiation and 20
as well as to explore neural mechanisms. The primary challenge is to different neuronal models with Izhikevich response behaviours.
construct a functional processor that can organize a large number of Recent advances in network-level hardware and increasing appli-
fine-grained parallel computing units with intrinsic temporal dynam- cation demands have led to the development of hardware designs for
ics. Constraints imposed by manufacturing and limited resources general-purpose computing. To improve the practicality of neuro-
require careful consideration of trade-offs across multiple aspects, morphic computing, heterogeneous hardware has been developed8
such as the granularity of neuron modelling (such as neuron and by acting as a co-processor and accelerating specific workloads under

Nature Electronics | Volume 7 | November 2024 | 954–965 955


Perspective https://doi.org/10.1038/s41928-024-01277-y

central processing unit (CPU) or field-programmable gate array (FPGA) programming frameworks and algorithm abstractions for convenient
control8,38. For example, BrainScaleS utilizes a reduced instruction modelling and hardware-agnostic development of brain-inspired algo-
set computer (RISC) CPU for reward-based learning39, while Loihi-140 rithms; compilers for efficiently deploying algorithms on specific or
directly integrates three embedded x86 processor cores on the chip various hardware with compatibility; runtimes and operating systems
to facilitate communication and encodings between conventional for flexible and dynamic task commitment and resource management;
computing and spike-based neuromorphic computing. SpiNNaker hardware abstractions for decoupling hardware details from other
is a specialized supercomputer based on ARM processors, balancing software and providing low-level control interfaces; and hardware
between general-purpose computing and dedicated parallel design. simulators and design tools for facilitating the engineering of systems.
Fusion with modern computing has also become a widely adopted We examine here the first four types of system software along each
strategy to extend applications of brain-inspired computing hardware. route (Fig. 2).
Tianjic41,42, for instance, achieves efficient support for artificial neural Route I emphasizes the use of hard-wiring neural models, mini-
networks (ANNs), spiking neural networks (SNNs) and hybrid models, mizing the reliance on software, whereas route II prioritizes the pro-
providing a unified computing platform catering to both computer grammability and configurability of deploying diverse neural models.
science and neuroscience research. It establishes a unified algorithmic To enable the rapid construction and analysis of these models, vari-
framework, operator representation and resource-shared circuit archi- ous simulation frameworks, including GENESIS59, NEURON60, NEST61,
tecture that leverages the strengths of both computing paradigms. PyNN62 and Brian63,64, have been developed. Modelling theories, such as
Some chips also incorporate a similar hybrid architecture, includ- NEF50, have also been established for dynamic modelling, leading to the
ing Shenjing43, NEUBULA44, SpiNNaker-245, Loihi-246, BrainScaleS-247 creation of practical modelling frameworks such as Nengo65. In route II,
and TCAS-II202148. Approximating general functions with neural net- toolchains are used to identify appropriate hardware configurations
works is another approach to realizing general computing. One such that match hardware behaviours and simulation expectations66,67. Spe-
example is Braindrop49, a chip designed specifically for the Neural cifically, some research in this area focuses on parameter calibration of
Engineering Framework50 (NEF) to enable neural networks to generi- analogue circuits or analogue–digital mixed circuits49,68, and runtime
cally approximate functions through system dynamics with minimal systems have been developed to accommodate various drivers and
computation cost. These efforts collectively highlight a growing effort interacting modes of the hardware67,69–71.
towards advancing general computing and expanding the applications Route III places greater emphasis on abstraction for improving
of brain-inspired hardware. portability, programmability and performance. Algorithm abstrac-
The deceleration of Moore’s law has made the continued tions enhance the modularity and composability of constructing SNNs.
miniaturization of transistors both physically and economically Hybrid support for ANNs and SNNs is one of the important research
unsustainable. Consequently, various post-complementary metal– topics across all system levels72. For instance, frameworks such as
oxide–semiconductor (CMOS) technologies have emerged51–53, such as Spyketorch73 and SpkingJelly74 integrate SNN modelling and training
redox-based resistive random-access memory, phase-change memo- with established ANN frameworks. Hardware abstractions, ranging
ries, spin-transfer torque magneto resistive random-access memory, from inference to learning instruction set75, play a key role in improving
and memristors. These technologies offer potential solutions to the system functionality. Some abstractions focus on the structural and
demands of brain-inspired computing for high computational density behavioural features of the hardware, such as Corelet25, and establish a
and low power consumption. Optical memristors, alongside their comprehensive low-level programming/execution model. The execu-
electronic counterparts, offer significant benefits, including high tion model provides concrete and operable interfaces that enable the
bandwidth, low crosstalk and fast speeds, paving the way for all-optical runtime and operating system to achieve flexible task scheduling and
brain-inspired computing, such as photonic tensor cores54. Recently, resource allocation42,76. Finally, compilers link high-level representa-
fully parallel on-chip neural networks have been implemented by mem- tions and low-level abstractions to map neural tasks on massive parallel
ristors, making substantial progress in the feasibility and diversity of infrastructures, satisfying diverse resource constraints and achieving
memristor systems55,56. Compared with CMOS technology, memristive load-balancing for high performance29,75,77–79.
circuits show superior power and area efficiency, particularly for matrix Route IV introduces novel abstractions that facilitate general-
multiplication and accumulation tasks. They can serve as independent purpose computing using brain-inspired patterns. Two primary
processors or co-processors to accelerate neural networks, boosting approaches are used: embedding general computer languages into
computing efficiency in in-memory or near-memory computing57. the neural network structure80,81 and utilizing brain-inspired repre-
However, achieving versatility for general-purpose computations sentations to approximate general functions82,83. While compilers
still requires additional CMOS circuitry, which increases power con- and some NEF implementations leverage the general approximation
sumption and area overhead, which, to a certain extent, potentially ability of neuron systems to deploy non-neural network applications84,
diminishes the efficiency advantages of memristors. the main challenge of the approximation approach is to ensure the
quality of computing with limited resources and improve program-
Software mability to support complex control flow. Portability is another key
Effective computing relies heavily on system software. For brain- objective. LAVA81,85 pioneers cross-platform deployment by defining
inspired computing, the system software bridges neural models and a series of protocols, and some studies directly use general-processors
hardware with specific constraints, adapting to changes in algorithms ARM infrastructure80, or extend SNN functionalities from the RISC-V
and hardware, while translating abstract neural models into concrete instruction set architecture (ISA)86. However, more dedicated designs
hardware operations58. System software falls into five main categories: for GPBIC hardware abstractions are still needed.

Fig. 2 | Hardware and software advancements in brain-inspired computing. applications, including ANNs, SNNs and hybrid neural networks (HNNs); and
Top: key contributions to software (SW) development. The location of entries the pathway towards general-purpose applications. The vertical positioning
indicates it belongs to a specific route or crosses multiple routes. Bottom: of each entry indicates the specific route to which it belongs. Due to extensive
trends in hardware (HW) development. Circle sizes represent citation counts, research in brain-inspired computing software and hardware, certain critical
normalized logarithmically. There are four primary routes for brain-inspired contributions may be inadvertently omitted, and some works may span
computing: neuromorphic engineering for intelligent sensory processing, which multiple target routes. SIMD stands for single instruction multiple data parallel
involves minimal software; brain simulation for investigating brain functions mechanism. References 1–118 in the figure, along with their citation counts, are
and mechanisms; intelligent computing to accelerate neural network (NN) listed in Supplementary Information.

Nature Electronics | Volume 7 | November 2024 | 954–965 956


Perspective https://doi.org/10.1038/s41928-024-01277-y

Software stack NEF81 Brian112 SpykeTorch116 Random walk representation89


Nengo 115
Darwin framework 98
Fugu80
GENESIS118 Brian2113
SpikingJelly117 Graph of communicating entities92
Graph of communicating entities94
Programming framework 110
NEST83 99
and algorithm abstraction PyNN HNN programming Communicating sequential process107
Corelet96 NxTF106
NEURON111 NeuCube114 Tree-organized compartments105
Neuromorphic completeness representation101
82 86
Dynamical system guided mapping Cross-SIMD compilation Dynamics approximation38
NN approximates NN108
Neural accelerator for general programs90,91
SNN partition and mapping88
Analogue circuit calibration38,87 Coarse-fine-grained hybrid mapping98
Compiler Minimum communication mapping97 Machine graph generation92
95
Splitting, grouping, mapping Magma107
Homogeneity-aware mapping84 Minimum core mapping105 Generic graph reforming, tuning103
Semi-folded mapping100 Simulation-interactive compilation hierarchy104
Event-driven runtime library94,95 Event-based operating93
85
Batch and hybrid operation mode Time-annotated control flow86
TrueNorth runtime
78 Mutual scheduling69
Runtime and OS 85
Availability and calibration database Neural task operation and resource management98
Neural state machine scheduling102 Magma107
Loihi runtime105
Dimensionless model for mixed analogue–digital34 Extended HW coordinate8
Corelet96 Machine graph9
Hardware Hierarchical coordinate85 Microcode for learning105
abstraction
Darwin modelling98 Spatiotemporal execution model69
Interrupt handler API94 RISC-V SNN ISA109

1988 2
1 Hardware
Brains in silicon Tianjic
3
FACETS\BrainScaleS Darwin
4 SpiNNaker Intel-loihi
IBM ODIN/MorphIC
1995 UZH\ETH Zurich Others
5 16
6

17
2000
7
19
18
8 20
21
9
Year

22
2005 10 24 23
25 26
11
27 39
29 28

2010 40 41
12 72
13 30 31
42 73 74
43 75
14 32
46 45 44
33 76
34 48
2015 49 47 78
15 51 77
35 53 50
52 79
37 54 55 38
36
59 58
38 57 56 60
62
2020 65 64
63 61 67 68
68
66
69 67 66
70 71 69

Route I: Route II: Route III: Route IV:


Neuromorphic engineering Brain simulation Intelligent computing GPBIC
Neuron Diverse Brain Bio-inspired Paradigm
structure Intelligent Bio-inspired Intelligent principles fusing
neurons and mechanisms General
and sensory principles applications +
organizations exploration purpose
properties Neuron Neuron Algorithm–HW advanced SW–HW
modelling programming co-design computing decoupling
• Compact/circuit area • Programmability • AI model support • Versatility
• Power consumption • Scalability • Energy efficiency • Elasticity
• Biologically realistic • Simulation latency/power • Latency • Interactivity
• Biologically plausibility • Programmability

Nature Electronics | Volume 7 | November 2024 | 954–965 957


Perspective https://doi.org/10.1038/s41928-024-01277-y

For systems based on post-CMOS technologies, studies on soft- Recently, neuromorphic completeness has been proposed as
ware mainly include two key aspects: hardware simulators and algo- a more adaptive and broader definition of completeness for brain-
rithm deployment software. When it comes to system software stacks, inspired computing83. It relaxes the completeness requirement for
such as compilers and runtimes, the progress still lags behind tradi- brain-inspired computing hardware, enlarging the design space
tional CMOS-based counterparts, despite some initial efforts87. This lag by introducing a new dimension, the approximation granularity.
is largely attributed to the fact that the hardware has not yet matured Guided by this theory, a system hierarchy can be developed includ-
enough for large-scale systems. As a result, manual compilation and ing Turing-complete software abstractions and versatile hardware.
resource management are still viable for small-scale systems. Achieving Moreover, distinct roles and optimization objectives for different
compatibility with the existing brain-inspired computing ecosystem levels of hardware and software components can be defined with
and developing end-to-end compilation tools are still challenges. greater clarity with enhanced compatibility among different hardware
In the development of system software, three emerging trends and/or software designs.
are noticeable. First, there is an increasing emphasis on decoupling To some extent, the completeness theory addresses fundamental
software stacks through different levels of abstraction. While co-design questions about brain-inspired computing’s capabilities, discussing
or coupling between software and hardware can enhance performance, the problem its systems can solve and delineating their functional
decoupling improves usability and iterative ability, and standardizes boundaries. The performance advantage it offers is rooted in how it
interdisciplinary cooperation. Co-evolution will promote long-term represents and processes information. Modern computers excel in
development. Second, there is an increasing focus on extending the temporal processing capabilities driven by rapid instruction and state
performance benefits to various areas, from domain-specific to gen- switching. The brain, in contrast, is characterized by highly intercon-
eral parallel computing, and the responsibility of general purpose nected connections, dynamics and adaptivity, enabling simultaneous
has shifted partly from hardware to system software. Third, the com- processing of spatial and spatiotemporal capabilities. Furthermore,
munity of software is becoming diverse and richer, from deployment modern computers are engineered for precise processing, whereas
toolchains to middleware for large-scale systems. the brain possesses the attribute of approximate processing. In addi-
tion, modern computers with their von Neumann architecture are
Large-scale computing systems and applications recognized as control-flow architectures with separated storage and
To meet the growing demands for high-performance computing, computation, whereas the brain is typically characterized as a data-flow
large-scale brain-inspired systems have been developed to support event-driven architecture with co-located storage and computation.
expansive neural networks or simulations. Typically, these systems are To advance the computing model of GPBIC, it is important to
constructed hierarchically, starting with smaller components (such as fuse the brain’s spatiotemporal processing capability, approximation
board, reticle) that utilize brain-inspired chips and scaling up to larger computing and data flow with the computer’s temporal processing
systems (Fig. 3) with various methods. The integration and intercon- capability, precise computing and control flow. We explore here the
nection are crucial to ensure high bandwidth and effective cross-mode fusion of these three aspects—temporal, spatial and spatiotemporal
communication. However, current large-scale systems primarily use capabilities; approximate computing and precise computing; and
custom-developed chips and system software. control flow and data flow—providing a model that can be conceptual-
Current large-scale systems primarily use custom-developed ized as brain-inspired automata (Fig. 4). Our primary goal in providing
chips and system software and show heterogeneity of the von this theoretical computing model is to bring together researchers
Neumann-based host and brain-inspired computing slave modules. from diverse disciplines and inspire further research within a cohesive
FPGAs are frequently used for data conversion, interface translation theoretical framework. While these aspects are pivotal, they are not
and control information transmission. Most systems use an ethernet intended to be exhaustive or fixed, and further developments could
link and peripheral component interconnect express (PCIe) as the phys- be made by examining additional features.
ical interface and adopt a hierarchical connection scheme to improve
scalability, including routing topology, communication protocols and Temporal, spatial and spatiotemporal capabilities
schemes. During runtime, general-purpose processors are responsible The formalized descriptions of the concepts of ‘temporal’ and ‘spatial’
for running the operating system and managing data exchange. underpin this computing model, although there are diverse interpre-
Brain-inspired computing chips or systems have been used in tations of these terms at different levels within the research domain.
various applications, including artificial perception, signal processing, To provide a more coherent framework and enhance communication
robotics and neural simulation (Fig. 3f). Applications beyond neural among researchers, we consider a collection of parallel processing units
network or brain simulation88 have also been explored, such as commu- that have topological connections in space (analogous to synapses)
nity detection89, scientific computing90 and agent-based modelling91. and show internal state transitions (comparable to spiking neurons),
By demonstrating their capability to solve non-deterministic polyno- which is referred to as brain-inspired automata. This model is inspired
mial complete problems, these systems have highlighted their poten- by the cellular automaton92 and each unit features finite state machines
tial for complex computations. As system software continues to evolve, and input–output connection patterns. Its effectiveness relies on both
these systems are expected to provide greater high-performance temporal and spatial features in processing and information repre-
computing power, versatility and efficiency. sentation. In subsequent discussions, ‘automata’ specifically denotes
‘brain-inspired automata’.
Combining neuroscience and computer science
The integration of approaches from neuroscience and computer sci- Temporal and spatial processing. The computing capabilities of
ence is key to developing GPBIC. To do this, it is necessary to establish brain-inspired automata rely significantly on some intrinsic attributes,
a theoretical foundation that defines the scientific scope and provides notably temporal richness and spatial richness. Temporal richness
essential guidance, including complete and formalized computing reflects neuron dynamics and measures the processing capabilities
models. Turing completeness lays the theoretical foundation for com- of each processing unit, such as the number of states and the flex-
puters, and the hierarchical system structure decouples modern com- ibility of state transitions of state machines. Spatial richness origi-
puters’ software and hardware to ensure their independent progress as nates from the neuron interconnections and reflects the potential
well as compatibility. However, the lack of guidance in brain-inspired for information exchange, evaluated by the connectivity and het-
systems leads to more tight coupling development of dedicated soft- erogeneity among processing units. While computers prioritize
ware and hardware, hindering the establishment of GPBIC. temporal richness, the brain leverages massive neurons with spatial

Nature Electronics | Volume 7 | November 2024 | 954–965 958


Perspective https://doi.org/10.1038/s41928-024-01277-y

460 billion

100 billion
a Chips f
Neurons

4 billion

4 billion
460 million
Synapses Bio-data (signal)
Cognitive

1 billion
1 billion

100 million
control processing
Perception,
45 million

auditory 9%
5%

16 million
Perception,
4 million 3%

4 million
olfactory
1 million

2%
200,000

Brain simulation

27%
Robotics

7%
29,000
7,000

%
768
384

21
9%
Scientific

100
computing Neural
64

17%
simulation
16

Perception,
vision

FACETS Neurogrid BrainScaleS SpiNNaker TrueNorth Loihi Tianjic

b c e
USB
CPLD Neurocore

N N N FX2 N N N

Brain-inspired computing chip


CPLD
N N N N
4 wafer modules per rack


N N N N N N
FPGA
d 24 boards per card frame, 5 card frames 384
C C … C
HICANN B B B
per rack Rack 0–4
chips B … B … Chips C C … C

HICANN
board
S S … S B B B
SpiNNaker

… … … C
48 chips S S … S Rack 0–9
board(0) … … S C C … C
S S … S …
3 Link
Inner-FPGA Link

FPGA
FPGA Link Ethernet board
FPGA

chips Ethernet PCIe/ethernet


FPGA

BMP BM P Switch Switch


Switch
Switch CPU

Host PCIe/ethernet
Switch Switch

CPU
system
Ctrl
CPU

Host system Data IO


Host system

Fig. 3 | Large-scale brain-inspired computing systems. a, The scale of six e, The system architecture of other fully digital projects, including Tianjic,
representative large-scale systems. b, The system architecture of Neurogrid16. Loihi and TrueNorth 38,40,98. C indicates chips. Ctrl, controller: details may be
The input/output (IO) data of the system is sent through the universal serial bus different for these systems. f, Publication statistics on applications deployed on
(USB) and converted by complex programmable logic device (CPLD). FX2 is a brain-inspired computing chips or systems. Here we specifically focused on the
USB controller. c, The system architecture of BrainScaleS-NM-PM-117. B indicates articles about applications deployed on brain-inspired computing chips over the
BrainScaleS dies. HICANN, high input count analogue neural network. d, The past decade. Following this, we categorized the themes and contents of the 157
system architecture of SpiNNaker22. S indicates SpiNNaker chips. Board(0) is the articles obtained.
first board. Each board is controlled by a board management processor (BMP).

parallelism and their connections to process complex information, input processing units and the sequence symbols. Both languages
demonstrating significant spatial richness. Despite advancements can be complete, but the spatial-temporal language is more efficient
such as graphics-processing-unit and neural-network accelerators for handling various information distributions and data structures
that enable high parallelism, these technologies lack the abundant because of its additional dimension. It also offers greater flexibility
spatial richness observed in the human brain. Therefore, a critical in capturing underlying dependencies within algorithms, making it
research topic is exploring effective designs that fuse both temporal promising for leveraging both temporal and spatial richness to achieve
and spatial richness. cost-efficient execution.
The spatial-temporal processing and coding underpin the versatil-
Temporal and spatial coding. In terms of coding, the Turing machine- ity of the automata and determine the adaptivity for diverse resource
based computing model encodes information as a symbol sequence and performance requirements. This adaptivity can be assessed through
(temporal language). In contrast, the automata take a spatial-temporal complexity (cost) metrics of an automaton, including time complexity,
language, where information is represented through the locations of space complexity and transmission complexity, which corresponds to

Nature Electronics | Volume 7 | November 2024 | 954–965 959


Perspective https://doi.org/10.1038/s41928-024-01277-y

Paradigm fusion for GPBIC

Brain paradigm Computer paradigm

• Topology and • Logic based


dynamic based • Precise
• Approximate • Control flow
• Data flow

A theoretical perspective

Features formed
Brain-inspired Design considerations
through fusion
automata model
Time and space TST elasticity

complexity
Time
Processing
unit
Processing
stage Connection Temporal richness
pattern
Spatial richness
Space complexity
Precision and
approximation

pl gy
ity
m er
ex
co En
10
Output Versatility
GPBIC
0 1
Performance

Performance
Scope Scope

0 1 Control flow and Interactivity and programmability


Algorithm: data flow
1 1 Data flow Control flow
Timing
1 0
Processing units

Coding
Schedule

stage 0 0
1 1
Input:
0 1 Functionality
Time

Fig. 4 | Fusing the key features of brains and computers. Hybrid features generates output symbols. During execution, in-processing units interpret the
and design considerations for GPBIC are explored with the brain-inspired corresponding symbol sequences they receive. Processing units that receive
automata formed by a set of processing units. Each processing unit has a finite valid symbols generate corresponding symbols to the output connections and
state machine and is interconnected via specific patterns. These connections undergo state transitions according to predefined rules. Finally, out-processing
include input connections for symbol reception and output connections for units generate output sequences, which are decoded to the final result. The
symbol transmission to other units. A subset of processing units known as in- brain-inspired automata integrate the features of temporal, precise and control-
processing units receive input symbols encoding algorithms and specific input flow processing and representations with spatial, approximate and data-flow
data. Similarly, another subset of processing units, termed out-processing units, patterns to achieve TST elasticity, versatility, interactivity and programmability.

the time required to solve specific problems, the number of process- Temporal richness and spatial richness can be interchangeable.
ing units involved and the volume of data exchange among processing For instance, the computation capability of a biologically detailed
units. These complexity metrics are similar to those in modern com- neuron model is functionally similar to that of a multilayer ANN93. This
puters with certain redefinitions. For automata, a single task can be insight is valuable for designing neuromorphic devices and GPBIC
deployed through different strategies with distinct processing units systems. Ideally, GPBIC hardware should balance between ‘simple’ and
and process orders, resulting in varying time, space and transmission ‘complex’ processing units, coupled with dynamic spatial topological
complexity. Some strategies may use more processing units to reduce connections to handle diverse workloads adaptively and elastically.
time complexity, while others may extend processing time to mini- Consequently, an ‘adaptive’ neuromorphic device should possess
mize transmission complexity. The range of valid strategies defines the capability to switch between ‘high temporal richness with more
a scope of time–space–transmission (TST) elasticity (Fig. 4), allowing available internal states’ and ‘high spatial richness with more exter-
tailored strategies for specific problem requirements. The elasticity is nal connections’. Currently, many neuromorphic devices show the
influenced by spatial-temporal richness and coding, which helps the capacity to mimic complex neuronal or synaptic behaviours, but
computing architecture adapt to different scenarios. Hence, a critical achieving flexible spatial connections remains a significant chal-
challenge of the GPBIC is to leverage spatial and temporal richness and lenge. Inspiration can be drawn from recent innovations, such as IBM
coding to achieve appropriate TST elasticity for target applications. NorthPole94, which adopts four kinds of networks-on-chip to enhance

Nature Electronics | Volume 7 | November 2024 | 954–965 960


Perspective https://doi.org/10.1038/s41928-024-01277-y

Intelligent applications and Applications with more brain-


brain simulation inspired patterns
Versatile
Application applications
ANN, SNN, HNN CG, HPC,
multi-agent simulation, …
Inference, learning

Brain-inspired pattern-inclined programming


Abstraction NN programming framework and representation
language

System
software Compiler Domain-specific compilation stack Domain-agnostic compilation stack

Brain-inspired co-operating Brain-inspired self-operating


Operating system Brain-inspired operating system
system and runtime system

ISA

General-purpose Brain-inspired
processor central processor
Hardware General-purpose
brain-inspired
processor
Brain-inspired Control
co-processor co-processor

Fig. 5 | The development of GPBICs. A full-stack computer system comprises level converts applications with various patterns on brain-inspired hardware
three levels: application, system software and hardware. Efforts aimed at and develops self-organized systems. Structural generalization at the hardware
advancing general-purpose computers can be directed across all levels. level rearranges the architecture and circuit components and eventually gets a
Specifically, the pattern generalization at the application level defines the general-purpose brain-inspired processor. CG, computational graphics; HPC,
scope of applications that can be efficiently supported and enlarges the scope high-performance computing.
by relaxing the supporting patterns. Systematic generalization at the software

spatial computing and improve ANN inference throughput. Future for approximation coding can show increased resilience to minor
developments may involve advanced technologies such as chiplet data variations, leading to fault tolerance during execution95,96. At
and three-dimensional stacking, necessitating close collaboration the processing stage, approximation can be implemented either at
between academia and industry. the hardware level or within operation systems. This differs from the
coding stage, where approximation is achieved through compilers or
Approximate computing and precise computing runtime encoders.
Modern computers typically adhere to precise logic defined by pro- For specific tasks, ongoing efforts are fusing the high precision of
grams to ensure reliable programmability, enabling the construc- digital computing and energy-efficient and/or area-efficient in-memory
tion of complex systems with multilevel abstractions. In contrast, computing. For example, research97 introduced mixed-precision
brain-inspired automata fuse approximate computing, which aligns in-memory computing, which combines the accuracy of the von Neu-
with the inherent nature of approximation in existing brain-inspired mann machine with a low-precision yet computationally powerful
computing and also offers distinct advantages. This can be explained phase-change memory unit, to solve matrix equations by conjugate
from two aspects: the processing stage and the coding stage. gradient method. The phase-change memory unit takes the lead in
At the processing stage, approximation involves executing an handling the bulk of the computational workload, while the von Neu-
algorithm that may not precisely adhere to its exact program logic. mann cores use a backwards method to iteratively enhance accuracy.
In the context of brain-inspired automata, processing can be defined This hybrid system achieves 4× to 20× speed-ups and energy-saving
as y = f(x), where x and y represent input and output. The introduc- compared with conventional approaches.
tion of approximation in processing entails substituting the original From our perspective, the fusion of approximate and precise
algorithm f(·) with an approximate algorithm f′(·). To ensure satis- computing expands the design space between performance and fidelity
factory results, the automata output should be constrained within a for developing GPBIC. It involves fine-grained balancing in the selec-
specific margin of error ε, where |f(x) − f′(x)| ≤ ε. By loosening ε, the tion of precise and approximate components, as well as ε, tailored to
TST complexity would be reduced, providing a design space for trad- the specific requirements of a given program. Various approaches
ing off computing resources and service quality. Allowing a tolerable exist for implementing approximation, and we introduce two exam-
level of error ε can also expand the range of efficiently supported ples for GPBIC. One potential solution is to utilize precise computing
applications, leading to enhanced versatility. Noise during execu- at the critical control locus, constraining the approximation pro-
tion is an inherent approximate feature of some technologies, and cess within a stable trajectory. Second, due to the decentralization,
certain studies utilize this feature to develop unique computing self-organized scheduling strategies can be applied during runtime to
functionalities49. achieve dynamic adaptivity. The scheduling process can be approxi-
Approximation at the coding stage involves employing a less pre- mate as sub-optimal solutions may suffice. Consequently, it becomes
cise data representation for the automata. Specifically, an approximate viable to tailor a specialized scheduling algorithm for approximate
input x′ may be encoded with the constraint |f(x) − f(x′)| ≤ ε. Beyond hardware components, while employing high-precision components
the potential performance enhancements, processing units designed for task-related computation.

Nature Electronics | Volume 7 | November 2024 | 954–965 961


Perspective https://doi.org/10.1038/s41928-024-01277-y

Control flow and data flow control-flow programs as control co-processors. The long-term
In the field of computation, control-flow and data-flow architectures goal is to align control logic into a data-flow manner for a complete
are two frequently adopted schemas. The former relies on sequential brain-inspired hardware solution.
execution of instructions to determine the processing unit’s function, Innovations such as in/near-memory computing and in-sensor
while the latter usually takes an event-driven approach to achieve low computing using technologies such as memristors and non-volatile
latency and low power consumption. Current brain-inspired comput- memories have improved power efficiency and area utilization in some
ing architecture primarily adopts the data-flow architecture16,29,36,41 domain-specific hardware. However, challenges such as immature
due to its high performance and bio-plausibility. Combining data fabrication processes, lack of established design methodologies and
flow with control flow has the potential to increase controllability underdeveloped architectures have limited their use to co-processing
and interactivity while retaining the performance advantages of units at the macro-core level. Future use of these technologies could
event-driven. Control flow can be incorporated in three key aspects, improve memory hierarchy flexibility, promoting the computing
namely, functionality, timing and scheduling. Some architectures ability and power efficiency of GPBIC systems76. Moreover, flexible
establish synchronous groups or barriers to improve controllability memory protocols with reconfigurable storage and computation
in the original data-flow architecture, which typically adopts an asyn- resource ratios are expected to better support various computation
chronous pattern. Moreover, distributed runtime scheduling could operators.
be adopted in the future to enable efficient allocation of computa- Systematic generalization emphasizes creating a complete system
tional resources while retaining flexibility. Decentralized co-located software that works seamlessly with domain/pattern-specific hardware
resources, such as those utilized in in-memory processing, enhance to improve the generality of the entire computing system, wherein
locality and alleviate the data movement bottleneck. However, this general purpose is defined systematically and collectively by hardware
approach can restrict adaptivity when dealing with diverse workloads and software. Current brain-inspired computing software stacks are
with varying proportions of computation, storage and communica- often tailored to specific hardware, including abstractions for decou-
tion. In-memory computing also poses challenges for flexible data pling and modelling, compilers for deploying neural networks or brain
access due to the close coupling of storage and processing. To address simulation tasks, and co-operating systems with drivers and runtimes
these complexities, a hybrid solution can be devised, combining the for hardware interaction, resource allocation and task scheduling. The
strengths of in-memory processing with near-memory processing term co-operating system refers to a program that still executes under
with established memory control protocols. the control of a general operating system, such as Linus and Windows.
As applications diversify, there is potential for domain-agnostic but
Outlook pattern-specific intermediate programming languages and compilers
Unlike domain-specific architectures, GPBIC systems offer versatil- for fusion pattern programs. In addition, a brain-inspired operating
ity across a range of applications, flexibility for various deployment system could directly run on the brain-inspired central processor,
strategies, interactivity for environmental changes and program- leveraging their scheduling ability.
mability for various usage. Successfully developing GPBIC systems Advances such as universal approximation expand software
will though require creating a computing ecosystem that encourages responsibilities. For example, the fusion of approximation and preci-
exploration and innovation at every level. This can be achieved through sion offers a vast exploration space for the compilers to search for
three potential strategies: application-level pattern generalization, appropriate deployment strategies based on performance require-
hardware-level structural generalization and software-level systematic ments, acceptable task error margins and available hardware
generalization (Fig. 5). resources83. Following this direction, a philosophy of ‘efficient spe-
Pattern generalization involves expanding the range of appli- cialized hardware, flexible general-purpose system’ can eventually
cations that a system can support, progressing from domain spe- lead to a general-purpose computer solely based on brain-inspired
cific to pattern specific, and ultimately general purpose. Initially, computing hardware, providing a self-operating system that can adap-
brain-inspired computers were designed to efficiently run specific tively manage the entire computing environment. Such a system could
algorithms, like those used in spike-based intelligence and brain optimize resource allocation and task scheduling, and also dynamically
simulations. Contemporary systems are increasingly optimized to adjust to changing workloads and application demands. By decoupling
support ANNs alongside SNNs41,45–47, which can be achieved with mini- applications and algorithms from hardware through system software,
mal additional costs due to the similarities in the underlying patterns. brain-inspired computing could develop into a viable and practical
Key characteristics of brain-inspired patterns include large-scale solution, potentially serving as the next-generation general-purpose
parallelism, sparse or dense spatial-temporal representation and computing infrastructure.
processing, decentralized resource utilization, and low-precision
tolerance. With appropriate generalization, systems can become pat- References
tern specific, enabling efficient support for a range of applications 1. Mead, C. Analog VLSI and Neural Systems (Addison-Wesley, 1989).
that align with these patterns, including computational graphics, 2. Maher, M. A. C., Deweerth, S. P., Mahowald, M. A. & Mead, C. A.
high-performance computing, graph processing and multi-agent Implementing neural architectures using analog VLSI circuits.
simulation. Incorporating temporal processing capabilities and IEEE Trans. Circuits Syst. 36, 643–652 (1989).
control-flow flexibility allows these systems to support increasingly 3. Mead, C. Neuromorphic electronic systems. Proc. IEEE 78,
complex control logic and non-native functionalities, enhancing 1629–1636 (1990).
their versatility. 4. Schuman, C. D., Kulkarni, S. R., Parsa, M., Mitchell, J. P. & Kay,
Structural generalization at the hardware level focuses on reor- B. Opportunities for neuromorphic computing algorithms and
ganizing various hardware components and rearranging circuits to applications. Nat. Comput. Sci. 2, 10–19 (2022).
establish a complete hardware infrastructure that supports the fusion 5. Waldrop, M. M. The chips are down for Moore’s law. Nature 530,
theoretical computing model. Recently, combining brain-inspired 144–148 (2016).
computing chips with traditional general-purpose processors (such as 6. Zhou, Q., Du, C. & He, H. Exploring the brain-like properties
CPUs or FPGAs) has become popular for increasing usability. As chips of deep neural networks: a neural encoding perspective.
continue to evolve, and their applications leverage brain-inspired pat- Mach. Intell. Res. 19, 439–455 (2022).
terns, brain-inspired chips might supplant traditional CPUs as central 7. Indiveri, G. & Liu, S.-C. Memory and information processing in
processors and CPUs could be relegated primarily to handling complex neuromorphic systems. Proc. IEEE 103, 1379–1397 (2015).

Nature Electronics | Volume 7 | November 2024 | 954–965 962


Perspective https://doi.org/10.1038/s41928-024-01277-y

8. Zhang, B., Shi, L. & Song, S. in Brain-inspired Intelligent Robotics: 28. Merolla, P. A. et al. A million spiking-neuron integrated circuit with
The Intersection of Robotics and Neuroscience (eds Sanders, S. & a scalable communication network and interface. Science 345,
Oberst, J.) 4–9 (Science/AAAS, 2016). 668–673 (2014).
9. Roy, K., Jaiswal, A. & Panda, P. Towards spike-based machine 29. Akopyan, F. et al. TrueNorth: design and tool flow of a 65 mW 1
intelligence with neuromorphic computing. Nature 575, million neuron programmable neurosynaptic chip. IEEE Trans.
607–617 (2019). Comput. Aided Des. Integr. Circuits Syst. 34, 1537–1557 (2015).
10. Lichtsteiner, P., Posch, C. & Delbruck, T. A 128 x 128 120db 30. Agrawal, A. et al. IMPULSE: a 65-nm digital compute-in-memory
30mw asynchronous vision sensor that responds to relative macro with fused weights and membrane potential for
intensity change. In 2006 IEEE International Solid State Circuits spike-based sequential learning tasks. IEEE Solid State Circuits
Conference-Digest of Technical Papers 2060–2069 (IEEE, 2006). Lett. 4, 137–140 (2021).
11. Posch, C. et al. Live demonstration: asynchronous time-based 31. Shi, L. et al. Development of a neuromorphic computing system.
image sensor (ATIS) camera with full-custom AE processor. In In 2015 IEEE International Electron Devices Meeting 4.3.1–4.3.4
Proc. 2010 IEEE International Symposium on Circuits and Systems (IEEE, 2015).
1392–1392 (IEEE, 2010). 32. Shen, J. et al. Darwin: a neuromorphic hardware co-processor
12. Guo, M., Ding, R. & Chen, S. Live demonstration: a dynamic based on spiking neural networks. Sci. China Inf. Sci. 59, 1–5
vision sensor with direct logarithmic output and full-frame (2016).
picture-on-demand. In 2016 IEEE International Symposium on 33. Ma, D. et al. Darwin3: a large-scale neuromorphic chip with a
Circuits and Systems (ISCAS) 456–456 (IEEE, 2016). novel ISA and on-chip learning. Natl Sci. Rev. 11, nwae102 (2024).
13. Dong, S., Huang, T. & Tian, Y. Spike camera and its coding 34. Nambiar, V. P. et al. 0.5 V 4.8 pJ/SOP 0.93 μW leakage/
methods. In 2017 Data Compression Conference 437 (IEEE, 2021). core neuromorphic processor with asynchronous NoC and
14. Yang, Z. et al. A vision chip with complementary pathways for reconfigurable LIF neuron. In 2020 IEEE Asian Solid-State Circuits
open-world sensing. Nature 629, 1027–1033 (2024). Conference 1–4 (IEEE, 2020).
15. Silver, R., Boahen, K., Grillner, S., Kopell, N. & Olsen, K. L. 35. Stuijt, J., Sifalakis, M., Yousefzadeh, A. & Corradi, F. μBrain: an
Neurotech for neuroscience: unifying concepts, organizing event-driven and fully synthesizable architecture for spiking
principles, and emerging tools. J. Neurosci. 27, 11807–11819 neural networks. Front. Neurosci. 15, 538 (2021).
(2007). 36. Davies, M. et al. Loihi: a neuromorphic manycore processor with
16. Benjamin, B. V. et al. Neurogrid: a mixed-analog–digital multichip on-chip learning. IEEE Micro 38, 82–99 (2018).
system for large-scale neural simulations. Proc. IEEE 102, 699–716 37. Frenkel, C., Lefebvre, M., Legat, J.-D. & Bol, D. A 0.086-mm2
(2014). 12.7-pJ/SOP 64k-synapse 256-neuron online-learning digital
17. Schemmel, J. et al. A wafer-scale neuromorphic hardware spiking neuromorphic processor in 28-nm CMOS. IEEE Trans.
system for large-scale neural modeling. In 2010 IEEE International Biomed. Circuits Syst. 13, 145–158 (2018).
Symposium on Circuits and Systems 1947–1950 (IEEE, 2010). 38. DeBole, M. V. et al. TrueNorth: accelerating from zero to 64 million
18. Schemmel, J., Meier, K. & Mueller, E. A new VLSI model of neural neurons in 10 years. Computer 52, 20–29 (2019).
microcircuits including spike time dependent plasticity. In 2004 39. Friedmann, S., Frémaux, N., Schemmel, J., Gerstner, W. & Meier,
IEEE International Joint Conference on Neural Networks Vol. 3, K. Reward-based learning under hardware constraints—using a
1711–1716 (IEEE, 2004). RISC processor embedded in a neuromorphic substrate. Front.
19. Schemmel, J., Fieres, J. & Meier, K. Wafer-scale integration Neurosci. 7, 160 (2013).
of analog neural networks. In 2008 IEEE International Joint 40. Davies, M. et al. Advancing neuromorphic computing with Loihi:
Conference on Neural Networks (IEEE World Congress on a survey of results and outlook. Proc. IEEE 109, 911–934 (2021).
Computational Intelligence) 431–438 (IEEE, 2008). 41. Pei, J. et al. Towards artificial general intelligence with hybrid
20. Moradi, S., Qiao, N., Stefanini, F. & Indiveri, G. A scalable multicore Tianjic chip architecture. Nature 572, 106–111 (2019).
architecture with heterogeneous memory structures for dynamic 42. Ma, S. et al. Neuromorphic computing chip with spatiotemporal
neuromorphic asynchronous processors (DYNAPs). IEEE Trans. elasticity for multi-intelligent-tasking robots. Sci. Robot. 7,
Biomed. Circuits Syst. 12, 106–122 (2017). eabk2948 (2022).
21. Merolla, P., Arthur, J., Alvarez, R., Bussat, J.-M. & Boahen, K. A 43. Wang, B., Zhou, J., Wong, W.-F. & Peh, L.-S. Shenjing: a low power
multicast tree router for multichip neuromorphic systems. reconfigurable neuromorphic accelerator with partial-sum and
IEEE Trans. Circuits Syst. I 61, 820–833 (2013). spike networks-on-chip. In 2020 Design, Automation & Test in
22. Painkras, E. et al. SpiNNaker: a 1-W 18-core system-on-chip for Europe Conference & Exhibition 240–245 (IEEE, 2020).
massively-parallel neural network simulation. IEEE J. Solid State 44. Singh, S. et al. Nebula: a neuromorphic spin-based ultra-low
Circuits 48, 1943–1953 (2013). power architecture for SNNs and ANNs. In 2020 ACM/IEEE 47th
23. Indiveri, G., Chicca, E. & Douglas, R. J. Artificial cognitive Annual International Symposium on Computer Architecture
systems: from VLSI networks of spiking neurons to neuromorphic 363–376 (IEEE, 2020).
cognition. Cogn. Comput. 1, 119–127 (2009). 45. Höppner, S. et al. The SpiNNaker 2 processing element
24. Chicca, E., Stefanini, F., Bartolozzi, C. & Indiveri, G. Neuromorphic architecture for hybrid digital neuromorphic computing. Preprint
electronic circuits for building autonomous cognitive systems. at https://arxiv.org/abs/2103.08392 (2021).
Proc. IEEE 102, 1367–1388 (2014). 46. Taking Neuromorphic Computing to the Next Level with Loihi
25. Amir, A. et al. Cognitive computing programming paradigm: 2 Technology Brief (Intel Labs, 2021); https://www.intel.com/
a corelet language for composing networks of neurosynaptic content/www/us/en/research/neuromorphic-computing-loihi-
cores. In 2013 International Joint Conference on Neural Networks 2-technology-brief.html
1–10 (IEEE, 2013). 47. Pehle, C. et al. The BrainScaleS-2 accelerated neuromorphic
26. Izhikevich, E. M. Simple model of spiking neurons. IEEE Trans. system with hybrid plasticity. Front. Neurosci. 16, 795876 (2022).
Neural Netw. 14, 1569–1572 (2003). 48. Kuang, Y. et al. A 64K-neuron 64M-1b-synapse 2.64 pJ/SOP
27. Brunel, N. & Latham, P. E. Firing rate of the noisy quadratic neuromorphic chip with all memory on chip for spike-based
integrate-and-fire neuron. Neural Comput. 15, 2281–2306 models in 65nm CMOS. IEEE Trans. Circuits Syst. II 68, 2655–2659
(2003). (2021).

Nature Electronics | Volume 7 | November 2024 | 954–965 963


Perspective https://doi.org/10.1038/s41928-024-01277-y

49. Neckar, A. et al. Braindrop: a mixed-signal neuromorphic 74. Fang, W. et al. SpikingJelly (Multimedia Learning Group, Institute
architecture with a dynamical systems-based programming of Digital Media, 2020).
model. Proc. IEEE 107, 144–164 (2018). 75. Lin, C.-K. et al. Programming spiking neural networks on Intel’s
50. Eliasmith, C. & Anderson, C. H. Neural Engineering: Computation, Loihi. Computer 51, 52–61 (2018).
Representation, and Dynamics in Neurobiological Systems 76. Lin, J. et al. SongC: a compiler for hybrid near-memory and
(MIT Press, 2003). in-memory many-core architecture. IEEE Trans. Comput. 73,
51. Mehonic, A. et al. Memristors—from in‐memory computing, deep 2420–2433 (2024).
learning acceleration, and spiking neural networks to the future 77. Müller, E. et al. A scalable approach to modeling on accelerated
of neuromorphic and bio‐inspired computing. Adv. Intell. Syst. 2, neuromorphic hardware. Front. Neurosci. 16, 884128 (2022).
2000085 (2020). 78. Deng, S. et al. Darwin-S: a reference software architecture for
52. Song, M.-K. et al. Recent advances and future prospects for brain-inspired computers. Computer 55, 51–63 (2022).
memristive materials, devices, and systems. ACS Nano 17, 79. Deng, L. et al. SemiMap: a semi-folded convolution mapping for
11994–12039 (2023). speed-overhead balance on crossbars. IEEE Trans. Comput. Aided
53. Liu, L. et al. Domain wall magnetic tunnel junction-based artificial Des. Integr. Circuits Syst. 39, 117–130 (2018).
synapses and neurons for all-spin neuromorphic hardware. 80. Rowley, A. G. et al. SpiNNTools: the execution engine for the
Nat. Commun. 15, 4534 (2024). SpiNNaker platform. Front. Neurosci. 13, 231 (2019).
54. Feldmann, J. et al. Parallel convolutional processing using an 81. LAVA (Intel Labs, 2021); https://lava-nc.org/
integrated photonic tensor core. Nature 589, 52–58 (2021). 82. Smith, J. D. et al. Neuromorphic scaling advantages for energy-
55. Yao, P. et al. Fully hardware-implemented memristor efficient random walk computations. Nat. Electron. 5, 102–112
convolutional neural network. Nature 577, 641–646 (2020). (2022).
56. Zhang, W. et al. Edge learning using a fully integrated 83. Zhang, Y. et al. A system hierarchy for brain-inspired computing.
neuro-inspired memristor chip. Science 381, 1205–1211 (2023). Nature 586, 378–384 (2020).
57. Mannocci, P. et al. In-memory computing with emerging memory 84. Neckar, A., Stewart, T. C., Benjamin, B. V. & Boahen, K. Optimizing
devices: status and outlook. APL Mach. Learn. 1, 010902 (2023). an analog neuron circuit design for nonlinear function
58. Qu, P., Yang, L., Zheng, W. & Zhang, Y. A review of basic software approximation. In 2018 IEEE International Symposium on Circuits
for brain-inspired computing. CCF Trans. High Perform. Comput. and Systems 1–5 (IEEE, 2018).
4, 1–9 (2022). 85. Rueckauer, B. et al. NxTF: an API and compiler for deep spiking
59. Wilson, M., Bhalla, U., Uhley, J. & Bower, J. GENESIS: a system neural networks on Intel Loihi. ACM J. Emerg. Technol. Comput.
for simulating neural networks. Adv. Neural Inf. Process. Syst. 1, Syst. 18, 1–22 (2022).
485–492 (1988). 86. Wang, J. et al. RISC-V toolchain and agile development based
60. Hines, M., Davison, A. P. & Muller, E. NEURON and Python. open-source neuromorphic processor. Preprint at https://arxiv.
Front. Neuroinf. 3, 1 (2009). org/abs/2210.00562 (2022).
61. Gewaltig, M.-O. & Diesmann, M. NEST (neural simulation tool). 87. Khan, A. A., Farzaneh, H., Friebel, K. F., Chelini, L. & Castrillon,
Scholarpedia 2, 1430 (2007). J. CINM (Cinnamon): a compilation infrastructure for
62. Davison, A. P. et al. PyNN: a common interface for neuronal heterogeneous compute in-memory and compute near-memory
network simulators. Front. Neuroinf. 2, 11 (2009). paradigms. Preprint at https://arxiv.org/abs/2301.07486 (2022).
63. Goodman, D. F. & Brette, R. Brian: a simulator for spiking neural 88. Aimone, J. B. et al. Non-Neural Network Applications for Spiking
networks in Python. Front. Neuroinf. 2, 5 (2008). Neuromorphic Hardware (Sandia National Lab, 2018).
64. Stimberg, M., Brette, R. & Goodman, D. F. Brian 2, an intuitive and 89. Hamilton, K. E., Imam, N. & Humble, T. S. Community detection
efficient neural simulator. eLife 8, e47314 (2019). with spiking neural networks for neuromorphic hardware. In Proc.
65. Bekolay, T. et al. Nengo: a Python tool for building large-scale Neuromorphic Computing Symposium 1–8 (ACM, 2017).
functional brain models. Front. Neuroinf. 7, 48 (2014). 90. Severa, W., Parekh, O., Carlson, K. D., James, C. D. & Aimone, J. B.
66. Gao, P., Benjamin, B. V. & Boahen, K. Dynamical system guided Spiking network algorithms for scientific computing. In 2016
mapping of quantitative neuronal models onto neuromorphic IEEE International Conference on Rebooting Computing 1–8
hardware. IEEE Trans. Circuits Syst. I 59, 2383–2394 (2012). (IEEE, 2016).
67. Brüderle, D. et al. A comprehensive workflow for general-purpose 91. Araújo, R., Waniek, N. & Conradt, J. Development of a dynamically
neural modeling with highly configurable neuromorphic extendable spinnaker chip computing module. In Artificial Neural
hardware systems. Biol. Cybern. 104, 263–296 (2011). Networks and Machine Learning—ICANN 2014: 24th International
68. Kauderer-Abrams, E. & Boahen, K. Calibrating silicon-synapse Conference on Artificial Neural Networks 821–828 (Springer, 2014).
dynamics using time-encoding and decoding machines. In 2017 92. Toffoli, T. & Margolus, N. Cellular Automata Machines: A New
IEEE International Symposium on Circuits and Systems 1–4 Environment for Modeling (MIT Press, 1987).
(IEEE, 2017). 93. Beniaguev, D., Segev, I. & London, M. Single cortical neurons as
69. Müller, E. et al. The operating system of the neuromorphic deep artificial neural networks. Neuron 109, 2727–2739.e3 (2021).
BrainScaleS-1 system. Neurocomputing 501, 790–810 (2022). 94. Modha, D. S. et al. Neural inference at the frontier of energy,
70. Brown, A. D. et al. SpiNNaker—programming model. IEEE Trans. space, and time. Science 382, 329–335 (2023).
Comput. 64, 1769–1782 (2014). 95. Qian, W., Li, X., Riedel, M. D., Bazargan, K. & Lilja, D. An
71. Galluppi, F. et al. A hierarchical configuration system for a architecture for fault-tolerant computation with stochastic logic.
massively parallel neural hardware platform. In Proc. 9th IEEE Trans. Comput. 60, 93–105 (2010).
Conference on Computing Frontiers 183–192 (ACM, 2012). 96. Brown, B. D. & Card, H. C. Stochastic neural computation. I.
72. Liu, F. et al. Advancing brain-inspired computing with hybrid Computational elements. IEEE Trans. Comput. 50, 891–905 (2001).
neural networks. Natl Sci. Rev. 11, nwae066 (2024). 97. Le Gallo, M. et al. Mixed-precision in-memory computing.
73. Mozafari, M., Ganjtabesh, M., Nowzari-Dalini, A. & Masquelier, T. Nat. Electron. 1, 246–253 (2018).
Spyketorch: efficient simulation of convolutional spiking neural 98. Pei, J., Deng, L., Ma, C., Liu, X. & Shi, L. Multi-grained system
networks with at most one spike per neuron. Front. Neurosci. 13, integration for hybrid-paradigm brain-inspired computing.
625 (2019). Sci. China Inf. Sci. 66, 1–14 (2023).

Nature Electronics | Volume 7 | November 2024 | 954–965 964


Perspective https://doi.org/10.1038/s41928-024-01277-y

Acknowledgements Correspondence should be addressed to Luping Shi.


This work was supported by the National Key Research and
Development Program of China (number 2021ZD0200300) and the Peer review information Nature Electronics thanks the anonymous
National Nature Science Foundation of China (numbers 62088102 and reviewers for their contribution to the peer review of this work.
61836004).
Reprints and permissions information is available at
Author contributions www.nature.com/reprints.
W.Z., S.M., X.J., X.L. and Y.C contributed to the data collection.
W.Z., S.M. and X.J. contributed to writing the paper. All the authors Publisher’s note Springer Nature remains neutral with regard to
contributed to discussion and data analysis. L.S. contributed to work jurisdictional claims in published maps and institutional affiliations.
planning and development.
Springer Nature or its licensor (e.g. a society or other partner) holds
Competing interests exclusive rights to this article under a publishing agreement with
The authors declare no competing interests. the author(s) or other rightsholder(s); author self-archiving of the
accepted manuscript version of this article is solely governed by the
Additional information terms of such publishing agreement and applicable law.
Supplementary information The online version contains supplementary
material available at https://doi.org/10.1038/s41928-024-01277-y. © Springer Nature Limited 2024

Nature Electronics | Volume 7 | November 2024 | 954–965 965

You might also like