02 Whole
02 Whole
by
Doctor of Philosophy
in
Electrical and Electronic Engineering,
Faculty of Engineering, Computer and Mathematical Sciences
The University of Adelaide, Australia
2014
Supervisors:
Dr Said Al-Sarawi, School of Electrical & Electronic Engineering
Dr Nicolangelo Iannella, School of Electrical & Electronic Engineering
Prof Derek Abbott, School of Electrical & Electronic Engineering
© 2014
S. Mostafa Rahimi Azghadi
All Rights Reserved
To my dearest wife, Maryam
and to my Mum and Dad,
with all my love.
Page iv
Contents
Contents v
Abstract xi
Acknowledgments xv
Publications xxiii
Chapter 1. Introduction 1
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.1 Neural Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.2 Spiking Neural Networks . . . . . . . . . . . . . . . . . . . . . . . 3
1.1.3 Neuromorphic Engineering . . . . . . . . . . . . . . . . . . . . . . 5
1.2 Research Gaps and Objectives of the Thesis . . . . . . . . . . . . . . . . . 6
1.3 Summary of Original Contributions . . . . . . . . . . . . . . . . . . . . . 8
1.4 Thesis Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
Page v
Contents
Page vi
Contents
Chapter 6. First VLSI Designs for Triplet-based Spike Timing Dependent Plas-
ticity 131
6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
6.2 VLSI Implementation of Pair-based STDP . . . . . . . . . . . . . . . . . . 133
6.2.1 Pair-based STDP Model . . . . . . . . . . . . . . . . . . . . . . . . 133
6.2.2 Indiveri’s PSTDP Circuit Model . . . . . . . . . . . . . . . . . . . 134
6.2.3 Bofill and Murray’s PSTDP Circuit Model . . . . . . . . . . . . . . 135
6.3 VLSI Implementation of Triplet-based STDP . . . . . . . . . . . . . . . . 137
Page vii
Contents
Page viii
Contents
Chapter 8. Compact Low Energy Neuromorphic Circuit for Triplet STDP 191
8.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192
8.2 Minimal Representation of Triplet STDP Model . . . . . . . . . . . . . . . 193
8.3 Proposed Low Energy and Compact STDP Circuit . . . . . . . . . . . . . 195
8.4 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 198
8.4.1 Experimental Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . 198
8.4.2 Synaptic Plasticity Experiments with the Proposed TSTDP Mini-
mal Circuits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203
8.5 Synaptic Plasticity Circuit Comparison . . . . . . . . . . . . . . . . . . . . 212
8.5.1 Synaptic Plasticity Ability for Reproducing Experimental Data . 213
8.5.2 Area and Power Consumption . . . . . . . . . . . . . . . . . . . . 215
8.5.3 Process Variation and Transistor Mismatch . . . . . . . . . . . . . 217
8.6 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 221
8.7 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 224
Bibliography 245
Page ix
Contents
Index 263
Biography 265
Page x
Abstract
This thesis presents a versatile study on the design and Very Large Scale Integration
(VLSI) implementation of various synaptic plasticity rules ranging from phenomeno-
logical rules, to biophysically realistic ones. In particular, the thesis aims at developing
novel spike timing-based learning circuits that advance the current neuromorphic sys-
tems, in terms of power consumption, compactness and synaptic modification (learn-
ing) abilities. Furthermore, the thesis investigates the usefulness of the developed de-
signs and algorithms in specific engineering tasks such as pattern classification. To fol-
low the mentioned goals, this thesis makes several original contributions to the field of
neuromorphic engineering, which are briefed in the following.
In the next stage, VLSI designs and implementations of a variety of synaptic plasticity
rules are studied and weaknesses and strengths of these implementations are high-
lighted. In addition, the applications of these VLSI learning networks, which build
upon various synaptic plasticity rules are discussed. Furthermore, challenges in the
way of implementing these rules are investigated and effective ways to address those
challenges are proposed and reviewed. This review provides us with deep insight into
the design and application of synaptic plasticity rules in VLSI.
Next, the first VLSI designs for the triplet STDP learning rule are developed, which
significantly outperform all their pair-based STDP counterparts, in terms of learning
capabilities. It is shown that a rate-based learning feature is also an emergent property
Page xi
Abstract
of the new proposed designs. These primary designs are further developed to gener-
ate two different VLSI circuits with various design goals. One of these circuits that has
been fabricated in VLSI as a proof of principle chip, aimed at maximising the learning
performance—but this results in high power consumption and silicon real estate. The
second design, however, slightly sacrifices the learning performance, while remark-
ably improves the silicon area, as well as the power consumption of the design, in
comparison to all previous triplet STDP circuits, as well as many pair-based STDP cir-
cuits. Besides, it significantly outperforms other neuromorphic learning circuits with
various biophysical as well as phenomenological plasticity rules, not only in learning
but also in area and power consumption. Hence, the proposed designs in this thesis
can play significant roles in future VLSI implementations of both spike timing and rate
based neuromorphic learning systems with increased learning abilities. These systems
offer promising solutions for a wide set of tasks, ranging from autonomous robotics to
brain machine interfaces.
Page xii
Statement of Originality
I certify that this work contains no material, which has been accepted for the award of
any other degree or diploma in my name, in any university or other tertiary institution
and, to the best of my knowledge and belief, contains no material previously published
or written by another person, except where due reference has been made in the text. In
addition, I certify that no part of this work will, in the future, be used in a submission in
my name, for any other degree or diploma in any university or other tertiary institution
without the prior approval of the University of Adelaide and where applicable, any
partner institution responsible for the joint-award of this degree.
I give consent to this copy of my thesis when deposited in the University Library, being
made available for loan and photocopying, subject to the provisions of the Copyright
Act 1968.
The author acknowledges that copyright of published works contained within this the-
sis resides with the copyright holder(s) of those works.
I also give permission for the digital version of my thesis to be made available on
the web, via the University’s digital research repository, the Library Search and also
through web search engines, unless permission has been granted by the University to
restrict access for a period of time.
15/03/2014
Signed Date
Page xiii
Page xiv
Acknowledgments
First and foremost, I would like to convey my deepest gratitude to my supervisors Dr
Said Al-Sarawi, Dr Nicolangelo Iannella and Prof. Derek Abbott for their guidance
and support throughout my candidature. My principal supervisor, Dr Al-Sarawi ad-
vised me with his open view and broad knowledge in the field of electronic engineer-
ing and circuit design. His critical, and thoughtful comments were always constructive
and fruitful to improve the quality of my research. In addition, my co-supervisor Dr
Iannella also provided me with his solid knowledge in the field of computational neu-
roscience and synaptic plasticity in Spiking Neural Networks. He has significantly
helped me to embark in the field of neuromorphic engineering and find my direction
in implementing new circuits for unexplored synaptic plasticity rules. Further, I would
like to express my gratitude to my other co-supervisor, Prof. Abbott, who has been of
great help, support and advice, when it counted most. With his enthusiastic supervi-
sion, he always encouraged me to conduct high quality research and publications.
Page xv
Acknowledgments
I am also thankful to David Lawrence, Kathrin Aguilar, and Simone Schumacher for
their administrative assistance and support in the INI.
Zurich is really memorable for me because of my other great friends and colleagues
in the INI. My great friend in the INI, Saurabh Bhargava, was always there for me to
discuss both scientific and non-scientific life issues together and laugh at our prob-
lems and difficulties as PhD students. I am also thankful to other marvellous PhD
and master students in the INI, Gabriela Michel, Hongjie Liu, Mitra Javadzadeh, Suraj
Honnuraiah, David Bontrager, Nawal El Boghdady, Dennis Goldschmidt, Jonas Klein,
Gina Paolini, Petar Ivanov, Karlis Kanders, Atanas Stankov, Asim Iqbal, Ivan Voitov,
and Sofia Jativa.
I would like to express my deep gratitude to Dr Tara Hamilton from the University of
New South Wales. I had the chance to first meet Dr Hamilton and discuss with her
my neuromorphic research plan and ideas, in a special neuromorphic session at the
IEEE ISSNIP conference in 2011, in Adelaide. Tara supported me and I was invited
to the leading Telluride Neuromorphic Engineering Workshop that is yearly organised by
the Institute of Neuromorphic Engineering (INE), in Telluride, Colorado, USA, where
attendance to this workshop is by invitation only. In this three-week workshop I had
great opportunity to work with other recognised researchers from around the world to
discuss and verify some of my research approaches and methodology. Here, I would
like to thank Prof. Ralph Etienne-Cummings from Johns Hopkins University for invit-
ing me to this great scientific event. Also I would like to thank other workshop organ-
isers and supporters for providing all the needed support in terms of accommodation
and logistics. During the workshop I spent some time implementing STDP and TSTDP
synaptic plasticity rules on the Spinnaker, a neuromorphic architecture developed at
the University of Manchester. I learned about Spinnaker from two talented and en-
thusiastic researchers, Dr Sergio Davies, and Dr Francesco Galluppi. Besides, while in
Telluride, I met Prof. Gert Cauwenberghs, one of the pioneers in the field of neuromor-
phic engineering, and his group members Dr Theodore Yu, and Mr Jongkil Park, from
University of California, San Diego. They kindly introduced me to their developed
large-scale neuromorphic chip, named HiAER IFAT. It was great to learn about lead-
ing neuromorphic projects in the world and at the same time, discuss my own research
ideas with these renowned researchers. In addition, I had great pleasure to discuss my
research ideas and plans with other experts in the field such as Prof. Tobi Delbruck and
Dr Michael Pfeiffer from the INI, as well as Prof. Jonathan Tapson from University of
Page xvi
Acknowledgments
Western Sydney. I would also like to include my gratitude to Prof. Andre van Schaik
from University of Western Sydney for his constructive comments and suggestions on
one of my TSTDP circuit designs presented in The 2012 International Joint Conference on
Neural Networks (IJCNN), in Brisbane and also during the Telluride Workshop.
Back to Australia, there are many people who helped me throughout my PhD candi-
dature. Within the school of Electrical & Electronic Engineering, I am indebted to Dr
Yingbo Zhu, who provided me with valuable help in Cadence, whenever I asked. I
would also like to thank Dr Braden Philips, Dr Brian Ng and Associate Prof. Michael
Liebelt for inviting me to attend and present my research at their weekly group jour-
nal club. My first days in the school was great and happy because of the support and
friendliness of Dr Jega Balakrishnan, Dr Benjamin Ung and Dr Hungyen Lin. Indeed,
one of the most respected colleagues I had in the school, has been Dr Muammar Kabir,
who was always there for me and I could count on his help and advice. Thank you
very much Muammar. I also like to thank the office & support staff of the school
including Danny Di Giacomo for the logistical supply of required tools and items,
IT officers, David Bowler, Mark Innes, and Ryan King, and the administrative staff,
Stephen Guest, Greg Pullman, Ivana Rebellato, Rose-Marie Descalzi, Deborah Koch,
Lenka Hill, and Jodie Schluter for their kindness and assistance. I would also like to
thank the head of school, Associate Prof. Cheng-Chew Lim, for his support regarding
my research travels and software required for my PhD thesis. I am also thankful to my
other friends and colleagues within the school. First of all, I express my deep gratitude
to my dear friend, Dr Ali Karami Horestani for his endless help and invaluable advice
in all aspects of my life in Adelaide. In addition, I sincerely thank my other friends
and colleagues Dr Pouria Yaghmaee, Amir Ebrahimi, Sam Darvishi, Mehdi Kasaee,
Arash Mehdizadeh, Zahra Shaterian, Neda Shabi, Muhammad Asraful Hasan, Sarah
Immanuel, Robert Moric, Tran Nguyen, and Yik Ling Lim, for making such a friendly
research environment. Also my wife and I appreciate help and support of our family
friends in Adelaide, Reza Hassanli, Zahra Ranjbari, Javad Farrokhi, Sahar Daghagh,
Amir Mellati, Hosna Borhani, Sara Chek, Hedy Minoofar, and Mehregan Ebrahimi for
their help to make our life in Adelaide so good. In addition, I am thankful to my dear
Australian friend, Ben Chladek and his great family for their kindness and support.
Page xvii
Acknowledgments
I recognise that this research would not have been possible without the financial assis-
tance of Australian Government via a generous International Postgraduate Research
Scholarship (IPRS) and Adelaide University Scholarship (AUS). During my candida-
ture, I was awarded several other travel grants, scholarships and awards by several
other organisations. Here, I would like to deeply appreciate these organisations sup-
port including annual travel grants from the School of Electrical & Electronic Engineer-
ing (2011-2013), OIST Japanese Nerual Network Society Travel grant (2011), the Ade-
laide University D. R. Stranks Postgraduate Fellowship (2012), the Adelaide Univer-
sity Research Abroad Scholarship (2012), the IEEE Computational Intelligence Society
travel grant (2012), Brain Corporation Travel Fellowships for Spiking Neural Networks
(2012), AFUW-SA Doreen McCarthy Research Bursary (2013), IEEE SA Section travel
award (2013), and Australia’s Defence Science and Technology Organisation (DSTO)
Simon Rockliff Supplementary Scholarship (2013).
Back to my home country, my endless gratitude goes to my father and mother who
always endow me with infinite support, wishes, continuous love, encouragement, and
patience. I also thank my best and kindest sisters and brothers for their love and sup-
port. I also wish to express my warm and sincere thanks to my father- and mother-in-
law for their kindness, guidance, and earnest wishes. Also, I would like to thank my
sister-in-law for her support, inspiration, and kindness.
Last but not least, my most heartfelt thanks are due to my dearest stunning wife,
Maryam. Words are unable to express my deepest appreciation and love to her. She
stood by me in all ups and downs of my PhD and always endowed me with her end-
less love, and support. Darling, I love you from the bottom of my heart.
Page xviii
Thesis Conventions
The following conventions have been adopted in this Thesis:
Typesetting
This document was compiled using LATEX2e. Texmaker and TeXstudio were used as
text editor interfaced to LATEX2e. Inkscape and Xcircuit were used to produce schematic
diagrams and other drawings.
Referencing
The Harvard style has been adopted for referencing.
System of units
The units comply with the international system of units recommended in an Aus-
tralian Standard: AS ISO 1000–1998 (Standards Australia Committee ME/71, Quan-
tities, Units and Conversions 1998).
Spelling
Australian English spelling conventions have been used, as defined in the Macquarie
English Dictionary (A. Delbridge (Ed.), Macquarie Library, North Ryde, NSW, Aus-
tralia, 2001).
Page xix
Page xx
Awards and Scholarships
2013
• Simon Rockliff Scholarship, DSTO
2012
• Brain Corporation Fellowships for Spiking Neural Networks, IEEE WCCI2012
2011
• Japanese Neural Network Society Travel Award, Okinawa Institute of Science
and Technology
2010
• International Postgraduate Research Scholarships, The Australian Government
Page xxi
Page xxii
Publications
Journal Articles
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., I NDIVERI -G., A BBOTT-D. (2014b).
Spike-based synaptic plasticity in silicon: Design, implementation, application,
and challenges, Proceedings of the IEEE, 102(5), pp. 717–737. *
A ZGHADI -M. R., M ORADI -S., FASTNACHT-D., O ZDAS -M. S., I NDIVERI -G. (2014c).
Programmable spike-timing dependent plasticity learning circuits in neuromor-
phic VLSI architectures, ACM Journal on Emerging Technologies in Computing
Systems, Submitted on 15 Dec. 2013. *
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2014a). Tunable
low energy, compact and high performance neuromorphic circuit for spike-based
synaptic plasticity, PLoS ONE, 9(2), art. no. e88326. *
A ZGHADI -M. R., A L -S ARAWI -S., A BBOTT-D., AND I ANNELLA -N. (2013a). A neuro-
morphic VLSI design for spike timing and rate based synaptic plasticity, Neural
Networks, 45, pp. 70–82. *
H ASHEMI -S., A ZGHADI -M. R., Z AKEROLHOSSEINI -A., AND N AVI -K. (2014). A novel
FPGA programmable switch matrix interconnection element in quantum-dot cel-
lular automata, International Journal of Electronics, DOI: 10.1080/00207217.2014.
936526.
A ZGHADI -M. R., K AVEHEI -O., AND N AVI -K. (2007b). A novel design for quantum-
dot cellular automata cells and full adders, Journal of Applied Sciences, 7(22),
pp. 3460–3468.
N AVI -K., F OROUTAN -V., A ZGHADI -M. R., M AEEN -M., E BRAHIMPOUR -M., K AVEH -
M., AND K AVEHEI -O. (2009). A novel low-power full-adder cell with new tech-
nique in designing logical gates based on static CMOS inverter, Microelectronics
Journal, 40(10), pp. 1441–1448.
Page xxiii
Publications
N AVI -K., S AYEDSALEHI -S., FARAZKISH -R., AND A ZGHADI -M. R. (2010b). Five-input
majority gate, a new device for quantum-dot cellular automata, Journal of Com-
putational and Theoretical Nanoscience, 7(8), pp. 1546–1553.
N AVI -K., FARAZKISH -R., S AYEDSALEHI -S., AND A ZGHADI -M. R. (2010a). A new
quantum-dot cellular automata full-adder, Microelectronics Journal, 41(12), pp.
820–826.
FARAZKISH -R., A ZGHADI -M. R., N AVI -K., AND H AGHPARAST-M. (2008). New
method for decreasing the number of quantum dot cells in QCA circuits, World
Applied Sciences Journal, 4(6), pp. 793–802.
Book Chapter
B ONYADI -M., A ZGHADI -M. R., AND S HAH - HOSSEINI -H. (2008). Population-based
optimization algorithms for solving the travelling salesman problem, Traveling
Salesman Problem, Federico Greco (Ed.), ISBN: 978-953-7619-10-7, InTech, DOI:
10.5772/5586, pp. 1–34.
Page xxiv
Publications
Conference Articles
A ZGHADI -M. R., K AVEHEI -O., A L -S ARAWI -S., I ANNELLA -N., A BBOTT-D. (2011c).
Novel VLSI implementation for triplet-based spike-timing dependent plasticity,
Proceedings of the 7th International Conference on Intelligent Sensors, Sensor
Networks and Information Processing, Adelaide, Australia, pp. 158–162. *
A ZGHADI -M. R., K AVEHEI -O., A L -S ARAWI -S., I ANNELLA -N., A BBOTT-D. (2011d).
Triplet-based spike-timing dependent plasticity in silicon, The 21st Annual Con-
ference of the Japanese Neural Network Society, Okinawa, Japan, art. no. P3–
35. *
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2011b). Physical
implementation of pair-based spike-timing-dependent plasticity, 2011 Australian
Biomedical Engineering Conference, Darwin, Australia, Vol. 34, art. no. 141. *
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2011a). Emergent
BCM via neuromorphic VLSI synapses with STDP, 5th Australian Workshop on
Computational Neuroscience, Sydney, Australia, p. 31. *
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2012b). Efficient
design of triplet based spike-timing dependent plasticity, Proc. IEEE 2012 Interna-
tional Joint Conference on Neural Networks (IJCNN), Brisbane, Australia, DOI:
10.1109/IJCNN.2012.6252820. *
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2012a). Design
and implementation of BCM rule based on spike-timing dependent plasticity,
Proc. IEEE 2012 International Joint Conference on Neural Networks (IJCNN),
Brisbane, Australia, DOI: 10.1109/IJCNN.2012.6252778. *
A ZGHADI -M. R., M ORADI -S., AND I NDIVERI -G. (2013b). Programmable neuro-
morphic circuits for spike-based neural dynamics, 11th IEEE International New
Circuit and Systems (NEWCAS) Conference, Paris, France, DOI: 10.1109/NEW-
CAS.2013.6573600. *
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2013c). A new
compact analog VLSI model for spike timing dependent plasticity, 2013 IFIP/IEEE
21st International Conference on Very Large Scale Integration (VLSI-SoC), Istan-
bul, Turkey, pp. 7–12. *
Page xxv
Publications
A ZGHADI -M. R., A L -S ARAWI -S., A BBOTT-D., AND I ANNELLA -N. (2013b). Pairing
frequency experiments in visual cortex reproduced in a neuromorphic STDP cir-
cuit, 2013 20th IEEE International Conference on Electronics, Circuits, and Sys-
tems, Abu-Dhabi, UAE, pp. 229–232. *
A ZGHADI -M. R., B ONYADI -M. R., AND S HAHHOSSEINI -H. (2007). Gender classifi-
cation based on feedforward backpropagation neural network, Artificial Intelli-
gence and Innovations 2007: from Theory to Applications, Springer US, Athens,
Greece, pp. 299–304.
A ZGHADI -M. R., B ONYADI -M. R., H ASHEMI -S., AND M OGHADAM -M. E. (2008). A
hybrid multiprocessor task scheduling method based on immune genetic algo-
rithm, Proceedings of the 5th International ICST Conference on Heterogeneous
Networking for Quality, Reliability, Security and Robustness, Hong Kong, pp. 561–
564.
K AVEHEI -O., A ZGHADI -M. R., N AVI -K., AND M IRBAHA -A.-P. (2008). Design of
robust and high-performance 1-bit CMOS full adder for nanometer design, IEEE
Computer Society Annual Symposium on VLSI, ISVLSI’08, Paris, France, pp. 10–
15.
B ONYADI -M. R., A ZGHADI -M. R., AND H OSSEINI -H. S. (2007b). Solving traveling
salesman problem using combinational evolutionary algorithm, Artificial Intelli-
gence and Innovations 2007: from Theory to Applications, Springer US, Athens,
Greece, pp. 37–44.
B ONYADI -M., A ZGHADI -M. R., R AD -N., N AVI -K., AND A FJEI -E. (2007a). Logic op-
timization for majority gate-based nanoelectronic circuits based on genetic algo-
rithm, IEEE International Conference on Electrical Engineering, Lahore, Pakistan,
pp. 1–5.
H ASHEMI -S., A ZGHADI -M. R., Z AKEROLHOSSEINI -A. (2008). A novel QCA multi-
plexer design, IEEE International Symposium on Telecommunications, Tehran,
Iran, pp. 692–696.
K AZEMI - FARD -N., E BRAHIMPOUR -M., A ZGHADI -M. R., T EHRANI -M., AND N AVI -
K. (2008). Performance evaluation of in-circuit testing on qca based circuits,
IEEE East-West Design & Test Symposium (EWDTS), Lviv, Ukraine, pp. 375–378.
Page xxvi
Publications
A DINEH -VAND -A., L ATIF -S HABGAHI -G., AND A ZGHADI -M. R. (2008). Increasing
testability in QCA circuits using a new test method, IEEE International Design
and Test Workshop, Monastir, Tunisia, pp. 40–44.
A DINEH -VAND -A., PARANDIN -F., A ZGHADI -M. R., AND K HALILZADEH -A. (2009).
Solving multi-processor task scheduling problem using a combinatorial evolu-
tionary algorithm, The 9th Workshop on Models and Algorithms for Planning
and Scheduling Problems (MAPSP09), Kerkrade, the Netherlands, pp. 91–93.
Note: Articles with an asterisk (*) are directly relevant to this thesis.
Page xxvii
Page xxviii
List of Figures
1.1 Thesis outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
Page xxix
List of Figures
Page xxx
List of Figures
7.1 Proposed circuit for the full triplet-based STDP rule . . . . . . . . . . . . 162
7.2 Proposed minimal triplet-based STDP circuit . . . . . . . . . . . . . . . . 164
7.3 Exponential learning window produced by the proposed minimal TSTDP
circuit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
7.4 Weight changes produced by the proposed minimal TSTDP circuit un-
der a frequency-dependent pairing protocol . . . . . . . . . . . . . . . . . 168
7.5 Weight changes produced by the proposed minimal TSTDP circuit un-
der triplet protocol for two different spike triplet combinations . . . . . . 169
7.6 Weight changes produced by the proposed minimal TSTDP circuit un-
der quadruplet protocol . . . . . . . . . . . . . . . . . . . . . . . . . . . . 170
7.7 Synaptic weight changes in result of the extra triplet protocol and using
the proposed minimal TSTDP circuit . . . . . . . . . . . . . . . . . . . . . 173
7.8 The proposed TSTDP circuit can generate BCM-like behaviour . . . . . . 175
7.9 The proposed TSTDP circuit can generate pre-synaptically driven BCM-
like weight changes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 178
Page xxxi
List of Figures
7.10 STDP learning windows produced in 1000 MC runs using the optimised
bias parameters for the hippocampal data set . . . . . . . . . . . . . . . . 180
7.11 NMSEs obtained to reproduce the visual cortex data set in 1000 MC
runs, using the optimised bias parameters for this data set . . . . . . . . 181
7.12 NMSEs obtained to reproduce the hippocampal data set in 1000 MC
runs, using the optimised bias parameters for this data set . . . . . . . . 182
7.13 The MN256R1 multi-neuron chip under the microscope . . . . . . . . . . 184
7.14 The layout of the TSTDP circuit implemented on the MN256R1 chip . . . 185
7.15 Measurement results of the fabricated TSTDP circuit . . . . . . . . . . . . 186
Page xxxii
List of Figures
Page xxxiii
Page xxxiv
List of Tables
4.1 STDP parameters for producing STDP learning window on the IFMEM
chip . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
4.2 STDP parameters for producing competitive Hebbian learning behaviour
on the IFMEM chip . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
4.3 Optimised TSTDP model parameters for generating BCM-like behaviour
on the IFMEM chip . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
6.1 Indiveri’s PSTDP circuit bias voltages for mimicking two different data
sets and their resulting NMSEs . . . . . . . . . . . . . . . . . . . . . . . . 146
6.2 Bofill and Murray’s PSTDP circuit bias currents for mimicking two dif-
ferent data sets and their resulting NMSEs . . . . . . . . . . . . . . . . . . 146
6.3 Proposed voltage-mode TSTDP circuit bias voltages for mimicking two
different data sets and their resulting NMSEs . . . . . . . . . . . . . . . . 148
6.4 First TSTDP circuit bias currents and its resulting NMSE . . . . . . . . . 152
6.5 Second TSTDP circuit bias currents and its resulting NMSE . . . . . . . . 153
7.1 Minimal TSTDP circuit bias currents and the resulted NMSEs for the
two data sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 171
7.2 Retuned TSTDP circuit bias currents and the resulted NMSEs in the
presence of the worst case variation in 1000 MC runs . . . . . . . . . . . 183
8.1 Optimised biases for the minimal TSTDP circuits and two data sets . . . 206
8.2 Comparison of various synaptic plasticity VLSI circuits . . . . . . . . . . 214
8.3 Area and power comparison for various synaptic plasticity circuits . . . 216
Page xxxv
Page xxxvi
Chapter 1
Introduction
T
HIS chapter provides the reader with introductory background
on Spiking Neural Network (SNN) and discusses why neuromor-
phic engineering is important. Identified research gaps and the
motivations behind the current study are also outlined in this chapter. Fur-
thermore, the objectives of the thesis and the research questions and goals
are discussed. Besides, the original contributions made in this thesis to
reach the mentioned goals are highlighted. Finally, the structure and out-
line of this thesis are described.
Page 1
1.1 Introduction
1.1 Introduction
The human being is one of the most mysterious and complicated creatures to our
knowledge. They can understand, learn, deduce, recognise, and judge. Many re-
searchers are inspired by various aspects of human genetic and neural mechanisms.
Scientists have proposed a variety of human- and nature-inspired problem solving
techniques such as the Genetic Algorithm (GA) (Goldberg 1989, Azghadi et al. 2008,
Bonyadi et al. 2007) and Artificial Neural Network (ANN) (Haykin 1994, Azghadi et al.
2007). Artificial neural networks have attracted much attention, during the last few
decades. This has resulted in three various generations of these networks (Vreeken
2003), which are based on three different types of neurons.
The second generation of artificial neural networks that is more realistic and closer
to real neurons introduces sigmoid (i.e. logistic) neurons, which present a continuous
output function rather than just a binary or quantised output (DasGupta and Schnitger
1994). In the first two generations of neural networks, the timing of input-output is not
important and just their rate can carry information, but it is known that in real neurons,
which can carry out many complicated learning and recognition tasks, there is a more
intricate input-output relation (Dayan and Abbott 2001). This relation brings timing
into action and makes a spatial-temporal information coding and transition model of
neurons, so-called spiking neurons.
Spiking neurons are basic building blocks of the third generation of artificial neural
networks, so-called SNN. Since synaptic plasticity in SNNs is the focus of this thesis,
in the following, a brief introduction on this type of neural networks is presented.
Page 2
Chapter 1 Introduction
A SNN is composed of spiking neurons and synapses. As its name infers, a spik-
ing neuron sends information in the form of electrical pulses, so-called spike (i.e. ac-
tion potentials), to other neurons and synapses in the network. This is through the
propagation of these spikes that information is transferred in the SNN (Gerstner and
Kistler 2002). There is general agreement on the information coding in the form of
the rate, timing, and spatial/temporal correlations of the spikes communicated among
neurons. Therefore, a neuron depending on its type, its information type, its place in
the network, and other factors, is able to fire spikes in various ways. For this reason,
various neuron models were proposed by computational and theoretical neuroscien-
tists to mimic the operation of real neurons. This thesis briefly discusses neurons and
addresses their various models and behaviours, in the next chapter.
Besides spiking neurons, synapses are the other basic building blocks in a SNN. A
synapse is the connection point of one neuron to its neighbouring neurons. It is widely
believed that learning, computation and memory processes take place in synapses
(Sjöström et al. 2008). Since the learning and computation in a SNN is a dynamic pro-
cess, so the synapses that are the main blocks for learning and computation should
also be dynamic and modifiable. However, the open question is how the modifica-
tion takes place in the synapses within the brain, in a way that many complicated
and real-time tasks such as learning, computation, and cognition are performed so
smoothly and accurately. Although there is no general agreement as to the answer
to this question, there are several hypotheses stating that these modifications take
place in relation to the activity of pre- and post-synaptic neurons connected to the
synapse (Sjöström et al. 2008). These activities, which are the basis of this thesis are
elaborated in the following chapters.
After investigating the SNNs and their structure, scientists are able to propose models
and techniques to implement these spiking networks of neurons and synapses. The
implementation can be performed either in software or hardware. However, there are
fundamental differences in software and hardware implementations of SNNs.
Since neuron and synapse models and behaviours are described by algorithms and
mathematical terms, they can be easily implemented as software programs and run
on ordinary computers. These implementations are also easily modifiable and can be
altered according to the needs, in every step of the implementation process. These
programs that represent neurons and synapses can then be integrated and connected
Page 3
1.1 Introduction
Table 1.1. Neuron and synapse quantity. Number of neurons and synapses in humans as com-
pared to various animals (Ananthanarayanan et al. 2009).
to each other, in order to from a SNN. This approach is usually noise and error-free
and does not require a special technology to realise them. Despite all these benefits
though, it requires a supercomputer to simulate a mouse or cat cortex (Frye et al. 2007, Anan-
thanarayanan et al. 2009).
This is because of the fact that the Von Neumann architecture, which is the founda-
tion of all today’s computers, runs programs sequentially. Therefore a large amount
of processing, which in turn needs a large amount of memory, must be performed,
in order to realise a portion of cortex containing billions of neurons and trillions of
synapses (Frye et al. 2007, Ananthanarayanan et al. 2009). Table 1.1 shows how very
large the neuron and synapse numbers are in the mouse, rat, cat, and human cortex.
These numbers demonstrate why simulating a neural network requires supercomput-
ers.
Researchers in IBM have utilised and set up a very large supercomputer with 147,456
CPUs and 144 TB of main memory, in order to perform a cat-scale cortical simula-
tion, which is demonstrated in Ananthanarayanan et al. (2009). To this complexity
and very large resource requirements, one may add the complexity and resources for
implementing special purpose programs for neurons, synapses, and their interconnec-
tions. In addition, in terms of power consumption and area, a supercomputer, by no
means is close to a biological cortex. Furthermore, a simulated cortex on the men-
tioned supercomputers, is an order of magnitude slower than biological neural sys-
tems (Frye et al. 2007, Ananthanarayanan et al. 2009). Therefore, this raises the question
”how can researchers implement a reasonably large scale SNN, which consumes rea-
sonable power, takes moderate area, and process in biologically plausible time scales
or even faster time scales?”
Carver Mead as one of the founding fathers of silicon chips and modern electronic
engineering, in his seminal neuromorphic engineering paper (Mead 1990) states,
Page 4
Chapter 1 Introduction
This provides us a rather clear answer to the above mentioned questions and opens
horizons to a new field of engineering, as first posed by Carver Mead, Neuromorphic
Engineering.
Page 5
1.2 Research Gaps and Objectives of the Thesis
Software neural systems though, have some advantages such as reconfigurability, ease
of implementation, and noise-tolerance, over the neuromorphic VLSI systems. How-
ever, when considering a large scale neural network, the advantages of VLSI neuro-
morphic systems are significant. That is the main reason why neuromorphic VLSI
engineering has been growing and attracting attention since the pioneering ideas of
Carver Mead in the late 80’s (Mead 1989). Since then, neuromorphic engineers have
been developing various analog, digital, and mixed-signal VLSI circuits of neurons
and synapse models that have been proposed by computational and theoretical neuro-
scientists (Indiveri 2003, Indiveri et al. 2006, Hamilton et al. 2008, Indiveri et al. 2009, In-
diveri et al. 2011, Hamilton and van Schaik 2011, Azghadi et al. 2013a).
Page 6
Chapter 1 Introduction
causes of these plastic behaviours? and ii) what are the connections of these behaviours
to learning?
The variety of viewpoints in answering the above mentioned questions can be due to
the intrinsic complexity of learning and memory phenomena. This fact causes var-
ious classes of synaptic plasticity rules, with different mechanisms of induction and
expression, have been being proposed (Abbott and Nelson 2000, Mayr et al. 2010,
Shouval 2011, Graupner and Brunel 2012). As these various rules are proposed, neu-
romorphic engineers implement silicon models of them. Probably the most recognised
example of synaptic plasticity rules among neuromorphic engineers is Spike Timing
Dependent Synaptic Plasticity (STDP) (Markram et al. 1997, Bi and Poo 1998). This
rule has been implemented in different studies and by various groups (Bofill-I-Petit
and Murray 2004, Cameron et al. 2005, Indiveri et al. 2006, Tanaka et al. 2009, Bam-
ford et al. 2012b).
However, the research gap here is that, although the traditional form of STDP, Pair-
based Spike Timing Dependent Plasticity (PSTDP), has shown success in solving some
computational and learning problems both in computational neuroscience (Song et al.
2000, Lisman and Spruston 2010, Masquelier and Thorpe 2007, Masquelier and Thorpe
2010, Davison and Frégnac 2006) and in neuromorphic engineering (Bofill-I-Petit and
Murray 2004, Cameron et al. 2005, Indiveri et al. 2006, Koickal et al. 2007, Tanaka et al.
2009, Arena et al. 2009, Seo et al. 2011), recent studies show that the simple form of
PSTDP that changes the synaptic weights according to a linear summation of weight
changes, is not able to account for a variety of biological experiments (Froemke and
Dan 2002, Pfister and Gerstner 2006, Gjorgjieva et al. 2011) and hence may lack learn-
ing and computational capabilities compared to more elaborate new synaptic plasticity
models. Although a variety of other synaptic plasticity models, rather than just sim-
ple PSTDP exist, the implementation of these detailed synaptic plasticity rules in VLSI
and their use in various engineering applications is a rather unexplored research area,
and gives rise to a gap in neuromorphic engineering. Therefore, there is a need to ex-
plore these new synaptic plasticity rules in VLSI and test them in terms of performance
for generating the outcome of various biological experiments (and therefore mimick-
ing real synapses), as well as their use in real-world applications such as in pattern
classification tasks.
The main objectives of this thesis are to investigate new methods for design, analy-
sis, and implementation of a number of novel and unexplored synaptic plasticity rules
Page 7
1.3 Summary of Original Contributions
in VLSI. The main specific rule targeted in this thesis is a derivation of the classical
STDP rule that is named Triplet-based Spike Timing Dependent Plasticity (TSTDP).
This rule is shown to be able to account for a variety of biological experiments (Pfister
and Gerstner 2006). In addition, TSTDP exploited for pattern selection and classifi-
cation (Gjorgjieva et al. 2011). This thesis verifies the performance of a number of
proposed VLSI designs in terms of replication of the outcomes of a variety of bio-
logical experiments. It also investigates the power consumption, and silicon area of
the novel proposed designs and compares them to previous VLSI designs of a vari-
ety of synaptic plasticity rules, to justify if the proposed circuits are able to help reach
the long-lasting goal of having a large-scale SNN with features close to that of the
brain (Poon and Zhou 2011). Furthermore, the thesis also investigates the useful-
ness of the TSTDP rule in pattern classification tasks and reproducing the outcomes
of a number of other synaptic plasticity rules, such as rate-based Bienenstock Cooper
Munro (BCM) (Bienenstock et al. 1982, Cooper et al. 2004) or timing-based suppressive
model of STDP (Froemke and Dan 2002).
The newly introduced synaptic plasticity VLSI circuits tend to be used in various large-
scale SNNs with increased ability of learning, and improved synaptic plasticity ability.
This will lead to neuromorphic cognitive systems with higher degree of applicability
in real-time cognition tasks such as pattern recognition and classification. The next
Section represents a summary of original contributions made to reach the above men-
tioned objectives and to fill the mentioned research gaps.
Page 8
Chapter 1 Introduction
when higher processing speed is required. The results of this study is published
in the IEEE International New Circuit and System Conference (Azghadi et al. 2013d).
• For the first time, the PSTDP, TSTDP and BCM rules are successfully imple-
mented on the IFMEM neuromorphic chip. Furthermore, a pattern classification
neural network is also implemented on this multi-neuron chip. In more detail:
(i) It is successfully shown that both PSTDP and TSTDP rules implemented us-
ing silicon neurons and programmable synapses, can demonstrate the expected
behaviours, similar to those seen in biological experiments.
(ii) It is also shown how the STDP window can be generated using the silicon
neurons and synapses available on the system.
(iii) The PSTDP rule is used and tested for generating a competitive Hebbian
learning behaviour observed in computational STDP experiments.
(iv) For the first time the TSTDP learning algorithm is implemented on the IFMEM
neuromorphic hardware.
(v) In order to test the TSTDP implementation, the rate-based BCM learning
behaviour is reproduced by this implementation. This experiment shows the
usefulness of this timing-based learning algorithm for generating the rate-based
BCM learning behaviour.
(vi) Finally, the implemented TSTDP learning mechanism is utilised to train a
simple feedforward spiking neural network to classify some complex rate-based
patterns.
Obtained results show the high performance of the TSTDP rule in the targeted
classification task. In addition, the preformed research in this part provides good
view of the TSTDP rule and its properties and features, which are essential when
designing VLSI TSTDP synapses in the next parts of this research. The results of
the above mentioned study are presented in The ACM Journal on Emerging Tech-
nologies in Computing Systems (Azghadi et al. 2014c).
Page 9
1.3 Summary of Original Contributions
• The first VLSI design for the TSTDP rule is presented in this thesis. It is shown
that this voltage-based TSTDP circuit is able to mimic the outcomes of a wide
range of synaptic plasticity experiments including timing-based, hybrid rate/-
timing based, and rate-based synaptic plasticity experiments, under various pro-
tocols and conditions. The circuit and these primary results are presented in The
21st Japanese Neural Network Society Annual Conference (Azghadi et al. 2011d). In
addition, a previous voltage-based PSTDP VLSI design proposed by Indiveri et al.
(2006) is investigated, optimised and simulated to show the various synaptic
plasticity experiments. The results show that this PSTDP circuit, similar to our
previous modified PSTDP VLSI circuit (Azghadi et al. 2011b), fails to account
for many experiments. The comparison between the performance of this PSTDP
circuit and the first proposed TSTDP circuit is presented in The IEEE Interna-
tional Conference on Intelligent Sensors, Sensor Networks and Information Processing
(Azghadi et al. 2011c).
• The first voltage-based TSTDP design is not able to account for the exponential
behaviour of the STDP learning rule. In addition, it is not able to reproduce
the exponential learning window generated by the computational model of the
STDP rule presented in Song et al. (2000). Therefore, in order to remove this
deficiency, the voltage-based circuit was modified and a new synaptic plasticity
circuit is proposed, which uses the current-mode design strategy based on the
design proposed by Bofill-I-Petit and Murray (2004). This circuit is verified to
generate the exponential learning window as well as the outcomes of all required
complicated experiments. The results are presented in The 2012 IEEE International
Joint Conference on Neural Networks (Azghadi et al. 2012b).
Page 10
Chapter 1 Introduction
• Further investigation reveals that although the first two TSTDP designs can gen-
erate the outcomes of many experiments, none of them are able to correctly ac-
count for other set of experiments performed based on other STDP-based synap-
tic plasticity rules such as the suppressive STDP rule proposed by Froemke and
Dan (2002). In addition, these circuits are not able to correctly regenerate the
outcomes of BCM experiments under a specific condition, where the synapse
is driven pre-synaptically as in the original experiments (Kirkwood et al. 1996).
Therefore, a new high-performance VLSI design for the TSTDP rule is proposed
that outperforms the other TSTDP VLSI designs in several respects. It is shown
that the new proposed design has significantly lower synaptic plasticity predic-
tion error, in comparison with previous designs for TSTDP and PSTDP rules.
In addition, it is also shown that this new design can successfully account for a
number of new experiments, including experiments involved with various spike
triplet combinations, as well as pre-synaptic and post-synaptic driven rate-based
BCM-like experiments, where the previous TSTDP and PSTDP designs do not
show suitable performance and cannot mimic the experiments effectively. This
new design is also tested against process variation and device mismatch. It is
shown that, although the circuit is susceptible to process variation, it is possi-
ble to mitigate the effect of variations and fine-tune the circuit to its desired be-
haviour. In addition, the power consumption and area of the proposed design are
also investigated and discussed. Obtained results are presented mainly in Neural
Networks (Azghadi et al. 2013a).
• The proposed circuit has been recently fabricated as a proof of principle and the
measurement results testify to the correct functionality of the fabricated circuit in
performing triplet-based synaptic weight modification.
• Although the previous TSTDP design has a very high performance in reproduc-
ing the outcomes of all required experiments, compared to its PSTDP and TSTDP
counterparts, it consumes significantly high power to process each spike. Fur-
thermore, considering the number of transistors and the use of five capacitors,
from which one is a very large capacitor of the size of 10 pF, this high perfor-
mance TSTDP circuit occupies a large silicon area. However, in order to follow
the long-lasting goal of integrating synaptic plasticity circuits in a large-scale neu-
romorphic system, which includes millions of these circuits, they should be of a
practicable size, consume very low energy, have an acceptable performance and
Page 11
1.3 Summary of Original Contributions
tolerate the process variation to some extent. A new compact, ultra low energy,
high performance, and variation tolerant design is proposed to satisfy the needs
for such large-scale systems. This design utilises a 50 fF capacitor instead of the
very large capacitors used in the previous TSTDP circuit, while retaining its abil-
ity to reproduce the STDP learning window, and the triplet and quadruplet ex-
perimental data. This design and its experimental results are presented in the
2013 IFIP/IEEE 21st International Conference on Very Large Scale Integration (VLSI-
SoC) (Azghadi et al. 2013c).
• Due to its small capacitor, the new TSTDP design presented in Azghadi et al.
(2013c), cannot account for the frequency-dependent pairing experiments and is
suitable only for experiments with high spike frequencies. Therefore, the weight
capacitor in the design was increased, so that it can account for all types of ex-
periments and is useful for processing inputs with either high or low spike rates.
The presented results in the 2013 IEEE International Conference on Electronics, Cir-
cuits, and Systems (Azghadi et al. 2013b), show that the new design with larger
capacitor can very closely mimic the frequency-dependent pairing experiments,
and is suitable for processing various input spike rates.
• Further investigations on the performance of the new design show that this de-
sign, which its synaptic plasticity performance is slightly compromised com-
pared to the previous large high-performance circuit, significantly outperforms
previous TSTDP and PSTDP designs, in all important aspects in neuromorphic
engineering including power consumption, required silicon real estate, and tol-
erance to device mismatch. The results are presented in PLoS ONE (Azghadi et al.
2014a).
All these features make the new proposed design an ideal device for use in large
scale SNNs, which aim at implementing neuromorphic systems with an inherent
capability that can adapt to a continuously changing environment, thus leading
to systems with significant learning and computational abilities. This system
then can be used in real-world tasks such as pattern classification.
• In addition to all these contributions, a general overview on the design and VLSI
implementations of various synaptic plasticity rules, ranging from phenomeno-
logical ones (i.e. timing-based, rate-based, or hybrid rules) to biophysically re-
alistic ones (e.g. based on calcium dependent models) is provided in this thesis.
Page 12
Chapter 1 Introduction
The thesis also discusses the application domains, weaknesses and strengths of
the various representative approaches proposed in the literature and provides
deeper insight into the challenges that engineers face when designing and im-
plementing synaptic plasticity rules in order to utilise them in real-world appli-
cations. Furthermore, counter approaches to tackle the challenges in designing
neuromorphic engineering circuits are discussed and proposed. Besides, with a
focus more on the system aspects of neuromorphic engineering, the use of vari-
ous synaptic plasticity rules and circuits in real neuromorphic learning systems
is discussed and these systems are analysed in terms of power consumption and
silicon real estate. Also an example of an effective neuromorphic system is men-
tioned and it is described in detail how it learns to perform an engineering task.
The results, review and discussion mentioned above are mainly presented in The
Proceedings of the IEEE (Azghadi et al. 2014b).
Chapter 2 provides a brief background on neuron and synapse models and their VLSI
implementations. This will provide the required background material to build on in
the thesis. It then focuses on the main topic of the thesis, synaptic plasticity rules.
It reviews various plasticity rules ranging from simple phenomenological, to hybrid,
to biophysically grounded rules, and compare them in terms of biological plausibility.
The chapter also provides results of several Matlab simulations that were performed in
the beginning of the present study to gain knowledge and better understanding of the
targeted rules. These simulation were utilised as benchmark to verify the performance
of the rules, when implemented in VLSI. The chapter also reviews some important
Page 13
1.4 Thesis Outline
Figure 1.1. Thesis outline. The thesis is composed of 9 chapters including background and con-
clusion. The original contributions are distributed in chapters 3 to 8. The thesis can be
divided into two parts. The first part that includes chapters 3, 4 and parts of chapter
5 is mainly concerned with the design of a neuromoprhic system. However, the second
part, which includes some sections of chapter 5, and chapters 6 to 8, is dedicated to
the circuit designs for synaptic plasticity rules. All chapters are virtually self-contained.
Page 14
Chapter 1 Introduction
synaptic plasticity protocols that have been utilised in synaptic plasticity experiments,
as well as in many experiments performed in the present thesis. In short, this chapter
gives the reader an insight about the rest of the experiments and studies performed
within the thesis.
Page 15
1.4 Thesis Outline
Chapter 6 demonstrates that the VLSI implementations of the classical model of STDP
is incapable of reproducing synaptic weight changes similar to those seen in biological
experiments, which investigate the effect of either higher order spike trains (e.g. triplet
and quadruplet of spikes), or simultaneous effect of the rate and timing of spike pairs
on synaptic plasticity. This chapter shows that, a previously described spike triplet-
based STDP rule succeeds in reproducing all of these synaptic plasticity experiments.
In this chapter, synaptic weight changes using a number of widely used PSTDP circuits
are investigated and it is shown how the class of PSTDP circuits fails to reproduce the
mentioned complex biological experiments. In addition, a number of new STDP VLSI
circuits, which act based on the timing among triplets of spikes and are able to re-
produce all the mentioned experimental results, are presented. The presented circuits
in this chapter are the first VLSI implementations of TSTDP rules that are capable of
mimicking a wide range of synaptic plasticity experiments.
Page 16
Chapter 1 Introduction
real estate and process variations. In order to test the design against variation that
leads to device mismatch, a 1000-run Monte Carlo (MC) analysis is conducted on the
proposed circuit. The presented MC simulation analysis and the simulation result from
fine-tuned circuits show that, it is possible to mitigate the effect of process variations
in the proof of concept circuit. In addition to the performed simulations, the proposed
circuit has been fabricated in VLSI as a proof of principle. The shown chip measure-
ment results testify to the correct functionality of the fabricated circuit in performing
triplet-based synaptic weight modification.
Chapter 8 introduces a new accelerated-time circuit that has several advantages over
its previous neuromorphic counterparts, which were discussed in previous chapters,
in terms of compactness, power consumption, and capability to mimic the outcomes
of biological experiments. The proposed circuit is investigated and compared to other
designs in terms of tolerance to mismatch and process variation. Monte Carlo (MC)
simulation results show that the proposed design is much more stable than its previ-
ous counterparts in terms of vulnerability to transistor mismatch, which is a significant
challenge in analog neuromorphic design. All these features make the proposed design
an ideal circuit for use in large scale SNNs, which aim at implementing neuromorphic
systems with an inherent capability that can adapt to a continuously changing environ-
ment, thus leading to systems with significant learning and computational abilities.
Chapter 9 provides concluding remarks of this thesis. It also discusses future research
directions that can be followed based on the study carried out and presented in the
current thesis. In addition, an outlook to the neuromorphic engineering for synaptic
plasticity rules is also written in this chapter.
In a nutshell, considering the background material provided in this thesis and with re-
gards to the versatile study performed on the design, implementation and application
of spike timing-based synaptic plasticity rules, the current thesis can be of great help
for readers with various backgrounds. It is useful for engineering students who want
to grasp an idea around the field of neuromorphic engineering, as well as the more
experienced neuromorphic engineers who need to review and learn about the VLSI
implementation of synaptic plasticity rules and their applications. In addition, compu-
tational and experimental neuroscientists who would like to be familiar with the field
of neuromorphic engineering and its relation with their fields of research, will find this
thesis useful.
Page 17
Page 18
Chapter 2
T
HIS chapter starts by providing a brief background on neuron
and synapse models and their VLSI implementations, that is re-
quired to elucidate the material within this thesis. It then focuses
on reviewing various synaptic plasticity rules ranging from simple phe-
nomenological, to hybrid, to biophysically grounded rules, and compares
them in terms of biological capabilities. The chapter provides several Mat-
lab simulation results that were performed in the beginning of this study
to gain the knowledge and better understanding of the targeted rules. This
will provide the needed understanding required for implementing these
rules in VLSI in the next parts of the study. The chapter also reviews some
important synaptic plasticity protocols that have been utilised in synap-
tic plasticity experiments, as well as in many experiments performed in
the present thesis. In short, this chapter gives the reader an insight of the
neuron and synapse structures and provides the required information and
terms that will be used in the experiments performed within this thesis.
Page 19
2.1 Introduction
2.1 Introduction
The basic building blocks of a SNN are neurons and synapses (Gerstner and Kistler
2002). There are various types of neurons that can be classified based on their shapes
and biophysics. However, all types of cortical neurons produce electric signals—so
called spikes or action potentials—and generally have a shape as demonstrated in
Fig. 2.1.
In addition to the neuron, the synapse is another crucial building block of a SNN. Simi-
lar to neurons, synapses also have complex structures and behaviours. They are widely
thought to be the essential components responsible for learning, memory and compu-
tational ability in the neural networks (Sjöström et al. 2008). A synapse, as shown in
Fig. 2.1, is the contact apparatus between a pre-synaptic neuron’s axon and a post-
synaptic neuron soma or dendrite (see Fig. 2.1). As the figure shows, the synapse is the
site for transmitting various neurotransmitter molecules to the post-synaptic neuron,
through different receptors on the post-synaptic side. The underlying mechanisms of
this transmission and the interactions happening in the synapse are termed synaptic
plasticity rules, which are the focus of the present chapter.
This chapter is organised as follows. Section 2.2 briefly discusses the structure and
behaviour of a typical neuron. Section 2.3 provides information on synapse and its
structure, and also discusses various VLSI implementations of a synapse circuit. Af-
ter these two background sections, Section 2.4 explains the synaptic plasticity phe-
nomenon. Section 2.5 shows various synaptic plasticity experiments that were con-
ducted in biology. These experiments were performed to gain an understanding of the
synaptic plasticity mechanisms, in order to propose plasticity models. These synaptic
plasticity models and rules are reviewed in Section 2.6. The chapter ends in Section 2.7
that includes concluding remarks on synaptic plasticity rules and sheds light on the
future topics discussed in the thesis.
Page 20
Chapter 2 Neurons, Synapses and Synaptic Plasticity
out of the neuron body (soma) and is called an axon. The axon is connected to other
neurons dendrites or soma. The synapse is where the axon of one neuron is connected
to the dendrite or soma of another neuron—moreover, it possesses the apparatus for
learning.
Besides Izhikevich’s neuron model, there are many other neuron models that describe
the behaviour of a biological neuron by means of some equations and formulations.
The most common models that can be found in the literature are HH (Hodgkin and
Huxley 1952), Leaky Integrate and Fire (LIF) (Smith 2006), IF with Adaptation, Integrate-
and-Fire-or-Burst (IFB) (Smith et al. 2000), Resonate-and-Fire model (Izhikevich 2001),
Quadratic Integrate and Fire (QIF) or theta-neuron (Ermentrout 1996), Izhikevich (2003),
FitzHugh (1961), Morris and Lecar (1981), Rose and Hindmarsh (1989), and Wilson
(1999).
Page 21
2.2 Spiking Neurons
Figure 2.1. A spiking neuron and synapse construction. The figure shows various parts of a
neuron and also magnifies the synapse structure, which in the case of this figure, is the
connection point of pre-synaptic neuron’s axon to the soma of the post-synaptic neuron.
Source: US National Institutes of Health.
Since spiking neuron models can be modelled in terms of mathematical formulas, they
can be fairly easily implemented in software. However, because neural systems oper-
ate in parallel, and ordinary computer systems are sequential, implementing a neuron
model in software is slow. Present-day computers are very high performance, allow-
ing straightforward simulation of a neuron in real-time. However, for a real-world
application such as pattern classification, a significantly large number of these spiking
neurons is needed. In addition, the software simulation will be very time consuming
for a large-scale network of spiking neurons. Furthermore, utilising an expensive par-
allel computer to implement the required neural network system, is not economical
since it requires specialised parallel software programs that are costly. Therefore, it is
preferable to implement neurons in hardware, rather than via simulations. Generally,
a hardware neuron, is much more practical than software neurons when considering a
large-scale network of neurons for real-world applications (Smith 2006, Hamilton and
van Schaik 2011, Indiveri et al. 2011). For a review of VLSI implementations of various
neuron models, the reader is directed to Indiveri et al. (2011).
Page 22
Chapter 2 Neurons, Synapses and Synaptic Plasticity
(A) tonic spiking (B) phasic spiking (C) tonic bursting (D) phasic bursting
input dc-current
20 ms
(E) mixed mode (F) spike frequency (G) Class 1 excitable (H) Class 2 excitable
adaptation
(M) rebound spike (N) rebound burst (O) threshold (P) bistability
variability
DAP
Page 23
2.3 Synapses
2.3 Synapses
Similar to neurons, synapses also have complex structures and behaviours. As already
mentioned, they are widely thought to be the essential components responsible for
learning, memory and computational ability in neural networks (Sjöström et al. 2008).
It is believed that through some activity-dependent rules, which control some com-
plex chemical reactions, synapses alter their efficacies (Abbott and Nelson 2000). This
efficacy determines the strength and the depressing/potentiating effect, a synapse has
on the spiking activity of its afferent neurons. Since neurons are associated to each
other through synapses, and because they transfer data and information in the form of
spikes, therefore it is absolutely essential to control the way, through which a synapse
manages spiking behaviour of its post neurons.
There are different VLSI implementations for synapse. A synapse can be implemented
as a simple multiplier circuit (Satyanarayana et al. 1992). Alternatively, a synapse can
be a current source that conducts current to the post-synaptic neuron only in the du-
ration of pre-synaptic spikes (Chicca et al. 2003). Here the amount of the current con-
veyed by the current source represents the synaptic weight. In addition to these simple
implementations that do not consider the detail dynamics and behaviours of the Ex-
citatory Post-Synaptic Current (EPSC) and Inhibitory Post-Synaptic Current (IPSC),
some other VLSI implementations of synapse, take into account more detailed synap-
tic behaviours. A well-known example of these types of synaptic circuits is the Dif-
ferential Pair Integrator (DPI) circuit proposed by Bartolozzi and Indiveri (2007). This
implementation is able to account for short-term dynamics of synapse as well as repro-
ducing the EPSC effects observed in biological synapses (Bartolozzi and Indiveri 2007).
Page 24
Chapter 2 Neurons, Synapses and Synaptic Plasticity
plasticity circuit functions according to a specific synaptic plasticity rule, and therefore
can vary in different synaptic circuit implementations and applications. The focus of
this thesis is on the design and implementation of these synaptic plasticity circuits.
However, before discussing the circuit implementations, the synaptic plasticity rules
are discussed and reviewed to help understanding the design and implementation of
the targeted synaptic plasticity circuits.
As already mentioned, there exists a significant number of hypotheses that try to ap-
proximate synaptic efficacy alterations (Mayr and Partzsch 2010). These hypothe-
ses that govern the synaptic weight changes, are so-called synaptic plasticity models
(rules). Identical to neuron models, there are a variety of synaptic plasticity models,
some of which are closer to biology and have meaningful relationships to biological
synapses, therefore, they are complex. On the other hand, some other models only ap-
proximate a number of biological experiments via mathematical modelling, and hence
they are simpler than the former group. Generally, the main purpose of the second
group of synaptic plasticity rules is to propose effective and simple rules, which are
able to produce the outcomes of as many synaptic plasticity experiments as possible.
In this chapter, a number of important synaptic plasticity rules are reviewed and high-
lighted and their abilities in reproducing the result of various biological experiments
are compared while discussing their complexities and structures. In order to have a
fair comparison among various synaptic plasticity rules, these rules are compared from
two aspects. The first aspect is their strength in reproducing various synaptic plasticity
experiments, while the second aspect is their simplicity and suitability to be employed
in large-scale neural simulations, and/or large-scale hardware realisations. Therefore,
prior to investigating and reviewing various synaptic plasticity rules, a variety of bi-
ological experimental protocols that were used in the experiments performed in the
neocortex have been reviewed, in order to provide the reader with an understanding
of under which protocols and conditions various synaptic plasticity rules are simu-
lated and compared. In the following sections, first some important synaptic plasticity
protocols are reviewed and their structures are described. And second, some signifi-
cant synaptic plasticity models are reviewed and their structures and various synaptic
plasticity abilities are highlighted.
Page 25
2.5 Synaptic Plasticity Experiments
where ∆wiexp , ∆wimodel and σi are the mean weight change obtained from biological ex-
periments, the weight change obtained from the model or circuit under consideration,
and the standard error mean of ∆wiexp for a given data point i, respectively. Here, p
represents the number of data points in a the data set under consideration. In order to
minimise the resulting NMSEs for the model/circuit and fit their output to the experi-
mental data, there is a need to adjust the model parameters or circuit bias parameters
and time constants. This is an optimisation process of the model parameters or circuit
biases to reach a minimum NMSE value and so the closest possible fit to the experi-
mental data. A powerful synaptic plasticity model/circuit, therefore, closely mimics
the outcomes of a variety of biological experiments, and reaches a minimal NMSE.
Hence, the number of various synaptic plasticity experiments a single model can ac-
count for is a good measure and an indication of the model/circuit’s ability to mimic
biology. In the following, some of these synaptic plasticity experiments are reviewed.
Page 26
Chapter 2 Neurons, Synapses and Synaptic Plasticity
These experiments have been utilised throughout this thesis, to first investigate a num-
ber of recognised biological models, and then verify the functionality and strength of
the VLSI implementations of various synaptic plasticity rules.
The pair-based STDP protocol has been extensively used in electrophysiological exper-
iments and simulation studies (Bi and Poo 1998, Iannella et al. 2010). In this protocol,
60 pairs of pre- and post-synaptic spikes with a delay of ∆t = tpost − tpre , as shown
in Fig. 2.3, are conducted with a repetition frequency of ρ Hz (in many experiments
ρ = 1 Hz). This protocol has been utilised in experiments reported in Bi and Poo
(1998), Froemke and Dan (2002), and Wang et al. (2005) and also have been employed
in simulations and circuit designs for synaptic plasticity such as Bofill-I-Petit and Mur-
ray (2004), Indiveri et al. (2006), and Azghadi et al. (2011c).
Figure 2.3. Spike pairing protocol. The figure shows how pairs of pre- and post-synaptic spikes
in a pairing protocol are timed for reproducing an STDP window.
In the simple pairing protocol, the repetition frequency of spike pairs kept constant and
it is usually 1 Hz. However, it has been illustrated in Sjöström et al. (2001) that altering
the pairing repetition frequency, ρ, affects the total change in weight of the synapse.
The spike pairing under this protocol is shown in Fig. 2.4. It shows that in higher
pairing frequencies, the order of pre-post or post-pre spike pairs does not matter and
both cases will lead to LTP. However, in lower pairing frequencies, pre-post results in
LTP and post-pre combination results in LTD (Sjöström et al. 2001, Sjöström et al. 2008).
There are two types of triplet patterns that are used in the hippocampal experiments
performed in Wang et al. (2005). These triplet patterns are adopted in this thesis to
Page 27
2.5 Synaptic Plasticity Experiments
Figure 2.4. Frequency-dependent pairing protocol. The figure shows how pairs of pre- and
post-synaptic spikes in a frequency-dependent pairing protocol are conducted. Here, ρ
determines the repetition frequency, at which the pre-post (∆t > 0) or post-pre (∆t < 0)
spike pair arrives.
compute the synaptic weight prediction error as described in Pfister and Gerstner
(2006). Both of these patterns consist of 60 triplets of spikes that are repeated at a
given frequency of ρ = 1 Hz. These triplet patterns are shown in Fig. 2.5. The first
pattern is composed of two pre-synaptic spikes and one post-synaptic spike in a pre-
post-pre configuration. As a result, there are two delays between the first pre and
the middle post, ∆t1 = tpost − tpre1 , and between the second pre and the middle post
∆t2 = tpost − tpre2 . The second triplet pattern is analogous to the first but with two
post-synaptic spikes, one before and the other one after a pre-synaptic spike (post-pre-
post). Here, timing differences are defined as ∆t1 = tpost1 − tpre and ∆t2 = tpost2 − tpre .
Figure 2.5. Triplet protocol. The figure shows how triplets of pre- and post-synaptic spikes in a
triplet protocol are conducted.
Page 28
Chapter 2 Neurons, Synapses and Synaptic Plasticity
is added either pre- or post-synaptically to the pre-post spike pairs, to form a triplet.
Then this triplet is repeated 60 times at 0.2 Hz to induce synaptic weight changes. In
this protocol, there are two timing differences shown as ∆t1 = tpost − tpre , which is
the timing difference between the two most left pre-post or post-pre spike pairs, and
∆t2 = tpost − tpre , which is the timing difference between the two most right pre-post
or post-pre spike pairs. Fig. 2.6 demonstrates different combinations of these spike
triplets.
Figure 2.6. Triplet protocol for extra triplet patterns. The figure shows how extra triplets of
pre- and post-synaptic spikes in the triplet protocol are timed.
In order to test the ability of the targeted timing-based plasticity rules and timing-based
synaptic plasticity circuits in generating a rate-based learning rule, which mimics the
Page 29
2.6 Synaptic Plasticity Rules
Figure 2.7. Qudruplet protocol. The figure shows how quadruplet of pre- and post-synaptic spikes
in the qudruplet protocol are timed.
Abstract models, themselves can be classified into three various groups including: i)
spike timing-based models such as STDP (Song et al. 2000), ii) spike rate-based models
Page 30
Chapter 2 Neurons, Synapses and Synaptic Plasticity
Although there are a variety of synaptic plasticity rules and experiments, this chapter
will only investigate some of the important biophysical and phenomenological rules,
which either have been already designed in VLSI, or are more suited for implementa-
tion in VLSI.
Pair-based STDP
The pair-based rule is the classical description of STDP, which has been widely used
in various computational studies (Song et al. 2000, Iannella and Tanaka 2006, Ian-
nella et al. 2010) as well as several VLSI implementations (Bofill-I-Petit and Murray
2004, Cameron et al. 2005, Indiveri et al. 2006, Tanaka et al. 2009, Mayr et al. 2010,
Meng et al. 2011, Bamford et al. 2012b). The original rule expressed by Eq. 2.2 is a
mathematical representation of the pair-based STDP (PSTDP) rule (Song et al. 2000)
∆w+ = A+ e( −τ+∆t ) if ∆t > 0
∆w = (2.2)
∆w− = − A− e( τ∆t− ) if ∆t ≤ 0,
where ∆t = tpost − tpre is the timing difference between a single pair of pre- and post-
synaptic spikes. According to this model, the synaptic weight will be potentiated if
a pre-synaptic spike arrives in a specified time window (τ+ ) before the occurrence of
a post-synaptic spike. Analogously, depression will occur if a pre-synaptic spike oc-
curs within a time window (τ− ) after the post-synaptic spike. The amount of poten-
tiation/depression will be determined as a function of the timing difference between
pre- and post-synaptic spikes, their temporal order, and their relevant amplitude pa-
rameters (A+ and A− ). The conventional form of STDP learning window, which is
generated using Eq. 2.2 is shown in Fig. 2.8.
Page 31
2.6 Synaptic Plasticity Rules
Figure 2.8. STDP learning window. Simulation results are produced under pairing protocol, and
using pair-based STDP model. The synaptic parameters for generating the three shown
curves (windows) using numerical simulations are as follows: τ− = 16.8 ms, and τ+ =
33.7 ms kept fixed for all three simulations, while the amplitude parameters, A− , A+
were altered for each simulation as A− = 0.5, and A+ = 1 for the first graph with
the maximum ∆w = 1 and the minimum ∆w = −0.5; A− = 0.36, and A+ = 0.54
for the middle graph; and A− = 0.18, and A+ = 0.276 for the third graph with the
maximum ∆w = 0.276 and the minimum ∆w = −0.18. The first experimental data
shown in black are two data points with their standard deviations that are extracted
from Pfister and Gerstner (2006), and the second experimental data are the normalised
data extracted from Bi and Poo (1998).
Triplet-based STDP
In this model of synaptic plasticity, changes to synaptic weight are based on the timing
differences among a triplet combination of spikes (Pfister and Gerstner 2006). There-
fore, compared to the pair-based rule, this rule uses higher order temporal patterns of
spikes to modify the weights of synapses. Triplet STDP (TSTDP) is described by
− ∆t − ∆t − ∆t
∆w+ = A+ e( τ+ 1 ) + A+ e( τy 2 ) e( τ+ 1 )
2 3
∆w = ∆t1 − ∆t3 ∆t (2.3)
( ) ( 1)
∆w− = − A2− e τ− − A3− e( τx ) e τ− ,
where the synaptic weight is potentiated at times when a post-synaptic spike occurs
and is depressed at the time when a pre-synaptic spike occurs. The potentiation and
depression amplitude parameters are A2+ , A2− , A3+ and A3− , while, ∆t1 = tpost(n) −
Page 32
Chapter 2 Neurons, Synapses and Synaptic Plasticity
tpre(n) , ∆t2 = tpost(n) − tpost(n−1) − ǫ and ∆t3 = tpre(n) − tpre(n−1) − ǫ, are the time dif-
ferences between combinations of pre- and post-synaptic spikes. Here, ǫ is a small
positive constant, which ensures that the weight update uses the correct values occur-
ring just before the pre- or post-synaptic spike of interest, and finally τ− , τ+ , τx and τy
are time constants (Pfister and Gerstner 2006).
According to the numerical simulation results presented in Pfister and Gerstner (2006),
beside the full TSTDP rule shown in Eq. 2.3, a minimised version of TSTDP rule that
excludes the depression contribution of triplet of spikes, is also capable of reproducing
the outcomes of several synaptic plasticity experiments. This rule that is called the first
minimal TSTDP rule, is presented as
− ∆t − ∆t − ∆t
∆w+ = A+ e( τ+ 1 ) + A+ e( τy 2 ) e( τ+ 1 )
2 3
∆w = ∆t1 (2.4)
− ( )
∆w− = − A2 e τ− .
Page 33
2.6 Synaptic Plasticity Rules
Figure 2.9 demonstrates the results for quadruplet experimental protocol and shows
how the first minimal TSTDP learning rule shown in Eq. 2.4 can generate an approx-
imation of the quadruplet experimental results. In order to reach a good approxi-
mation of the quadruplet experiments, there is a need to optimise the six synaptic
parameters shown in Eq. 2.4. These six parameters are optimised in a way that the
minimal NMSE is reached for a number of experiments including pairing (window),
triplet and quadruplet experiments, when compared to biological experiments. The
optimised synaptic parameters that are utilised to approximate these experiments are
those shown for minimal TSTDP rule in Table 2.1. These parameters are optimised
in a way that approximate 13 specific synaptic weight change values shown in black
data points and standard error mean bars, with minimum possible NMSE. The 13 data
points include (i) three points on Fig. 2.9, (ii) eight data points in Fig. 2.10 (a) and (b),
and (iii) two data points in Fig. 2.11.
0.4
∆w
0.2
−0.2
−100 −50 0 50 100
T [ms]
Figure 2.9. Quadruplet experiment in the hippocampus can be approximated using the first
minimal TSTDP model. Simulation results are produced under quadruplet proto-
col. The minimal TSTDP model parameters for generating the shown synaptic weight
changes are listed in Table 2.1. The experimental data shown in black are extracted
from Pfister and Gerstner (2006).
Figure 2.10 shows how the first minimal TSTDP rule using the optimised synaptic
parameters can approximate triplet STDP experiments. Besides, using the same opti-
mised parameters, as those utilised for quadruplet and triplet experiments, the STDP
Page 34
Chapter 2 Neurons, Synapses and Synaptic Plasticity
Table 2.1. Optimised minimal TSTDP model parameters. These parameters have been ex-
tracted from Pfister and Gerstner (2006) and used in our numerical simulations to gen-
erate the results shown in Fig. 2.9 to Fig. 2.13. Note that these parameters have been
used for a minimal nearest-neighbour TSTDP model. In this table, those values that are
shown as ’x’, do not have any affect in the results.
Parameter name A2+ A2− A3+ A3− τ+ (ms) τ− (ms) τy (ms) τx (ms) NMSE
First Minimal TSTDP 4.6 × 10−3 3 × 10−3 9.1 × 10−3 0 16.8 33.7 48 x 2.9
2nd Minimal TSTDP 0 8 × 10−3 5 × 10−2 0 16.8 33.7 40 x 0.34
learning window, which is shown in Fig. 2.11 can be approximated using the first min-
imal TSTDP rule. The minimum NMSE that was achieved using the optimised param-
eters for these three experiments was equal to 2.9 as shown in Table 2.1.
which is simpler and takes lower number of synaptic parameters, and therefore needs a
new set of parameters, in comparison with the previous minimal model for hippocam-
pal experiments. The optimised parameters for generating a close approximation of
the frequency-dependent pairing experiments are shown in the third row of Table 2.1.
The experimental data and the approximated weight changes, which are computed by
the second TSTDP minimal model, are shown in Fig. 2.12.
Page 35
2.6 Synaptic Plasticity Rules
Figure 2.10. Triplet experiments in the hippocampus can be approximated using the first
minimal TSTDP model. Simulation results are produced under triplet protocol
presented in Pfister and Gerstner (2006). The minimal TSTDP model parameters for
generating the shown synaptic weight changes are listed in Table 2.1. The experimental
data, shown in black and their standard deviations are extracted from Pfister and
Gerstner (2006).
of the BCM learning rule has been demonstrated in different ways, but a general, yet
simple form of this model is given as,
∆w
= φ(ρpost , θ ) · ρpre , (2.6)
∆t
Page 36
Chapter 2 Neurons, Synapses and Synaptic Plasticity
Figure 2.11. STDP learning window experiment in the hippocampus can be approximated
using the first minimal TSTDP model. Simulation results are produced under
pairing protocol. The minimal TSTDP model parameters for generating the shown
synaptic weight changes are listed in Table 2.1. The first experimental data shown in
black are two data points with their standard deviations that are extracted from Pfister
and Gerstner (2006), and the second experimental data are the normalised experimental
data that are extracted from Bi and Poo (1998).
where ρpre and ρpost represent the pre-synaptic and post-synaptic neurons spiking ac-
tivities and θ is a constant that represents some threshold (Pfister and Gerstner 2006).
In addition, when φ(ρpost < θ, θ ) < 0 synaptic weight will be decreased (depression),
and when φ(ρpost > θ, θ ) > 0, it will be increased (potentiation) and if φ(0, θ ) = 0,
there will be no change in synaptic weight (Pfister and Gerstner 2006).
According to the literature, the BCM rule can emerge from pair-based and triplet-based
STDP rules. In 2003, Izhikevich and Desai (2003) demonstrated that, the nearest-spike
interaction1 version of PSTDP can replicate BCM-like behaviour. Furthermore, Pfister
and Gerstner (2006) have reported, a triplet-based model of STDP that can also produce
BCM behaviour, when long-time spike statistics are taken into account. According to
1 Nearest-spike model considers the interaction of a spike only with its two immediate succeeding
and immediate preceding nearest neighbours.
Page 37
2.6 Synaptic Plasticity Rules
0.5
∆w
−0.5
0 20 40
ρ [Hz]
Figure 2.12. Pairing frequency experiments in the visual cortex can be approximated us-
ing the second minimal TSTDP model. Simulation results are produced under
frequency-dependent pairing protocol and using the second TSTDP minimal model.
The synaptic parameters for generating the shown weight changes are listed in the last
row of Table 2.1. The experimental data, shown in black and their standard deviations
are extracted from Pfister and Gerstner (2006).
Izhikevich and Desai (2003), under the assumption of Poissonian distribution of spike
times for pre-synaptic and post-synaptic spike trains, nearest-spike pair-based STDP
can give rise to the BCM rule; i.e. BCM emerges from nearest neighbour pair-based
STDP; while all-to-all2 spike interaction cannot. Furthermore, based on Pfister and
Gerstner (2006), if the pre-synaptic and post-synaptic spike trains in a triplet-based
STDP model are Poissonian spike trains, then BCM learning is an emergent property
of the model. A mapping between BCM learning rule and the TSTDP learning rule is
shown in Pfister and Gerstner (2006).
To analyse how BCM-like behaviour emerges from TSTDP, we need to go through the
same analysis used by Pfister and Gerstner (2006). In this circumstance, the triplet
2 All-to-all model considers the interaction of every single spike with all other spikes, not only with
its nearest neighbours.
Page 38
Chapter 2 Neurons, Synapses and Synaptic Plasticity
learning rule can be recast into a simpler form by considering the statistical properties
of TSTDP weight changes which leads to the following time averaged equation,
dw
= − A2− τ− ρpre ρpost + A2+ τ+ ρpre ρpost
dt
− A3− τ− τx ρ2pre ρpost + A3+ τ+ τy ρ2post ρpre ,
(2.7)
where ρpre and ρpost are the pre- and post-synaptic mean firing rates, respectively. The
other parameters in the above equation τ− , and τ+ , are time constants for the pair-
based contribution and τx , and τy are the corresponding time constants for the triplet-
based contribution of the original triplet learning rule shown in Eq. 2.3.
Based on the rate-based BCM rule, synaptic weight change is linearly dependent on
ρpre and non-linearly depends on ρpost (see Eq. 2.6). In order to satisfy this condition
in Eq. 2.7, A3− must be equal to zero and also ρpre ≪ τ+−1 . This is a minimal case of
the triplet-based STDP model—please refer to Pfister and Gerstner (2006). Also, based
on the BCM learning rule definition, the synaptic
D weight
E modification threshold is a
p
function of post-synaptic activity, i.e. θ = α ρpost where p > 1. For triplet-based
STDP, consider the case where all-to-all interactions between triplets of pre- and post-
synaptic spikes; it is possible to redefine A2− , A2+ and A3+ in a way that the threshold be
p
dependent on the post-synaptic firing rate, ρpost . However, in the nearest-spike model
D E
p
it is not possible to change these parameters in a way to satisfy θ = α ρpost . Although
the triplet-based nearest-spike STDP model cannot fully satisfy the second condition of
a BCM learning rule (the dependency of threshold on ρpost ), it can elicit the properties
of BCM for a limited range of frequencies. Numerical simulation results (Fig. 2.13)
show how the threshold is modulated by controllable amplitude parameters (A2− , A2+
and A3+ ) for nearest spike interaction. For further details on the relation between the
TSTDP and the BCM rules, refer to the text and also supplementary materials of Pfister
and Gerstner (2006). Fig. 2.13 demonstrates how the second minimal TSTDP rule, with
the parameters shown in Table 2.1 for visual cortex experiments, produces a BCM-
fashion behaviour.
Page 39
2.6 Synaptic Plasticity Rules
−5
x 10
10 λ=0.35
8
λ=0.64
dw/dt [ms ]
−1
λ=1
6
4
2
0
0 10 20 30 40 50
ρpost [Hz]
Figure 2.13. The BCM learning rule can be mapped to the TSTDP learning model. The
results shown in this figure are produced under Poissonian protocol and using the
second minimal TSTDP model, as the one used for generating the results in Fig. 2.12.
The synaptic parameters for generating the weight changes shown in this figure are
those for visual cortex and are shown in the third row (second minimal TSTDP) of
Table 2.1. The three curves showcase the sliding threshold feature of the BCM rule,
p p
that in TSTDP model can be controlled by changing λ = ρpost /ρ0 (see text for
details).
Prior to the experiments performed by Froemke and Dan in 2002, STDP experiments
were mainly directed at varying the time intervals between a pair of pre-post or post-
pre of spikes. In these conventional STDP experiments, synaptic weight modification
was only based on the timing difference between pairs of spikes. However, based on a
new model and experiments reported in Froemke and Dan (2002), synaptic plasticity
not only depends on the timing differences between pre- and post-synaptic spikes, but
also depends on the spiking pattern and the inter-spike intervals of each neuron.
In light of their experimental data from slices of cat visual cortex, Froemke and Dan,
proposed a synaptic plasticity rule, the so-called suppression model. According to this
Page 40
Chapter 2 Neurons, Synapses and Synaptic Plasticity
rule, spike pairs are not independent and there is a suppressive interaction among con-
secutive spikes, i.e. a pre-synaptic spike before a pre-synaptic spike, or a post-synaptic
before a post-synaptic spike causes suppression of the synaptic efficacy of the second
spike in interaction with its neighbour spikes from the other type. This suppression
subsides in time and therefore, the longer time passed from arrival of a pre/post spike,
the more synaptic efficacy the incoming pre/post spike will cause. The suppression
model is described by the following equation:
pre post
∆wij = ǫi ǫ j F(∆tij ), (2.8)
where ∆wij is the synaptic weight change due to the ith pre-synaptic spike and the
jth post-synaptic spike, ǫi = 1 − e−(ti −t(i−1))/τs is the efficacy of ith spike and τs is the
suppression constant. In addition F(∆tij ) is defined in a similar way as defined by
Eq. 2.2.
Demonstrated results in Froemke and Dan (2002) show that the suppression model is
able to account for the experiments using higher order spike trains such as triplet and
quadruplet of spikes, and the LTP and LTD observed in the STDP learning window.
However, this rule cannot account for frequency-dependent pairing experiments re-
ported in Sjöström et al. (2001). In order to remove the deficiency of the suppression
model in replicating this frequency-based experiment, Froemke et al. proposed a re-
vised version of their suppression model in 2006 (Froemke et al. 2006). Using this new
model, the frequency dependent experiments also can be generated. However, this
feature makes the new suppression model to require more components, which in turn
makes the new model to become more complex compared to the first proposed rule.
In addition to the timing-based synaptic plasticity rules, mentioned above, there is an-
other synaptic plasticity rule, which acts based on the recent activities and the cur-
rent state of the post-synaptic neuron, and not only according to its spike timing.
This rule has been developed to resolve a shortcoming of PSTDP rule, in being un-
able to learn patterns of mean firing rates (Abbott and Nelson 2000). This weakness is
due to the high sensitivity of the PSTDP rule to the spike timings and temporal pat-
terns (Mitra et al. 2009).
Page 41
2.6 Synaptic Plasticity Rules
If the required conditions are not satisfied, there will be no potentiation or depression.
When there is no spike coming and therefore there is no synaptic weight change, the
synaptic weight, W, will drift toward either high or low synaptic weight asymptotes.
The direction of the drift will depend on the values of the weights at that specific time,
which can be above/below a certain threshold, θW (Brader et al. 2007, Mitra et al. 2009)
dW (t)
dt = α; if W (t) > θW
dW (t) (2.10)
dt = − β; if W (t) ≤ θW .
The internal state, C(t), which represents the calcium concentration, depends on the
neuron’s spiking activity and changes by the following equation
dC(t) C (t)
=− + JC ∑ δ(t − ti ), (2.11)
dt τC i
where JC determines the amount of calcium contributed by a single spike (Brader et al.
2007, Fusi et al. 2000, Sheik et al. 2012b).
The main difference between the SDSP rule and the STDP-type rules is that, in the
SDSP rule, the timing of the spikes is replaced by the membrane potential of the post-
synaptic neuron, which has a close relation with the timing of the post-synaptic neuron.
One could simply assume that a neuron with the membrane potential above a certain
level, a threshold, is most likely to fire a spike, and therefore a post-synaptic spike
will be fired at that time. Therefore, there is a close analogy between SDSP and STDP
Page 42
Chapter 2 Neurons, Synapses and Synaptic Plasticity
rules. In addition to SDSP, some other rules have also utilised this specific feature of the
membrane potential available in the post-synaptic neuron site, and combined it with
either pre-synaptic spike timings or with its rate, and proposed new and more pow-
erful rules, compared to the conventional timing-based-only rule, i.e. STDP. A brief
review of two membrane potential (voltage)-based rules is presented in the following
sections.
Voltage-based STDP
The voltage-based STDP model proposed in Clopath et al. (2010) and Clopath and Ger-
stner (2010), is a modification of TSTDP rule proposed by Pfister and Gerstner (2006).
In this rule, the combination of the post-synaptic membrane potential and the pre-
synaptic spike arrival time, govern the plasticity mechanism. At the arrival of a pre-
synaptic spike, the synaptic weight will be depressed, if the post-synaptic neuron has
been depolarised for some time, since it shows that a post-synaptic spike happened
recently. The post-synaptic spike history window, depends on the time constant, by
which the post-synaptic membrane is filtered. The following equation shows the mem-
brane potential, u, which is low-pass-filtered with the time constant τ− ,
d
τ− u − ( t ) = − u − ( t ) + u ( t ). (2.12)
dt
Depression will occur when a pre-synaptic spike arrives, if at that moment, the low-
pass filtered membrane potential, u− (t), is above a certain threshold, θ− . Therefore,
depression can be mathematically modelled as
d −
w = − ALTD X (t)[u − (t) − θ− ]+ if w > 0, (2.13)
dt
where w shows the synaptic weight, ALTD is depression amplitude, and X (t) = ∑i δ(t −
ti ), with ti as the spike times, represents a pre-synaptic spike train. In Eq. 2.13, [ x ]+ is
equal to x, if x is positive and is 0 otherwise.
Page 43
2.6 Synaptic Plasticity Rules
where x is a low-pass-filter of the pre-synaptic spike train with time constant τx , and
can be modelled as
d
τx x ( t ) = − x ( t ) + X ( t ). (2.15)
dt
According to the mentioned equations for depression (Eq. 2.13) and potentiation (Eq.
2.14), the overall synaptic weight change can be calculated by
d
w = − ALTD X (t)[u − (t) − θ− ]+ + ALTP x (t)[u(t) − θ+ ]+ [u+ (t) − θ− ]+ . (2.16)
dt
Clopath et al. have shown that their voltage-based STDP rule is capable of repro-
ducing the outcomes of a variety of STDP experiments (Clopath and Gerstner 2010,
Clopath et al. 2010). Their rule that is a modification of the TSTDP rule, possesses
higher synaptic modification capabilities compared to the TSTDP rule. However, one
should keep in mind that the complexity of this rule is higher than the TSTDP rule as
well. This higher complexity along with the dependence of the rule on the membrane
potential of the post-synaptic neuron, increases the complexity of VLSI implementa-
tion of this rule. This is the case for other rules with higher capabilities, and therefore
higher complexity.
Voltage-based BCM
Another rule that has been recently implemented in VLSI and has a phenomenological
and computational background is the Local Correlation Plasticity (LCP) rule reported
in Mayr and Partzsch (2010). This rule modifies the synaptic plasticity in a BCM-like
fashion. In this rule, the weight changes take place, in a relation with the current state
of the post-synaptic membrane voltage, and the recent dynamic of the pre-synaptic
spikes. The synaptic alteration rule that has been implemented in VLSI for this BCM-
based learning circuit employs the following rule to modify the synaptic weight ac-
cording to Eq. 2.17
where w(t) is the synaptic weight, u(t) is the neuron’s membrane potential, φu is a
threshold between potentiation and depression, g(t) is a conductance variable that
Page 44
Chapter 2 Neurons, Synapses and Synaptic Plasticity
represents the post-synaptic current, Ipsc , and therefore has its maximum value at the
time of a pre-synaptic arrival and decays afterwards. The main difference between this
rule and the BCM learning model, is replacing the non-linear function, φ in BCM model
(Eq. 2.6), with a constant multiplier, B, as shown in Eq. 2.17. This results in a linear
dependence of the plasticity to the membrane potential, however as shown in Mayr
and Partzsch (2010), in this model, this linear dependence is translated to a non-linear
dependence between plasticity and the post-synaptic rate, identical to BCM.
This rule has been shown to reproduce the outcomes of many experimental protocols
including triplet, quadruplet (Wang et al. 2005), and pairing frequency experiments
performed in the visual cortex (Sjöström et al. 2001). Although this rule is able to
replicate many plasticity outcomes, it is prone to large errors when parameters are
fitted to closely replicate experimental results. These errors are rather high compared
to the TSTDP rule which is simpler and possesses fewer state variables (Mayr and
Partzsch 2010, Pfister and Gerstner 2006). In addition, as already mentioned, the com-
plexity of a synaptic plasticity rule is an essential issue when designing a rule in VLSI,
and the LCP rule has higher complexity than the TSTDP rule. Besides, the LCP rule’s
dynamic is dependent to the neuron model, as well as synaptic conductance, compar-
ing to the TSTDP rule that only depends on the timing of the spikes, that are easily
available at the synapse pre-synaptic and post-synaptic sites. Generally, when design-
ing a synaptic plasticity rule, there is a need to consider the complexity versus strength
of the rule and set a trade-off between them. This issue will be discussed later in this
thesis.
In addition to the aforementioned abstract synaptic plasticity models, there are sev-
eral other rules, using which a variety of the biological experiments performed in the
neocortex can be regenerated. These rules range from simple timing-based models to
other calcium-based models. The dynamical two-component long-term synaptic plas-
ticity rule that was proposed in Abarbanel et al. (2002), has the ability to produce some
LTP and LTD experiments, and also can be mapped to the BCM rule. In addition, the
STDP rule that has been combined with the BCM sliding threshold feature is another
synaptic plasticity mechanism that is able to account for a range of experiments in the
hippocampal (Benuskova and Abraham 2007).
Page 45
2.6 Synaptic Plasticity Rules
For a review of phenomenological synaptic plasticity models, see Morrison et al. (2008).
Further, for a review and discussion of biophysical and phenomenological rules refer
to Mayr and Partzsch (2010).
None of the rules mentioned so far truly map to the biophysics of the synapse and bio-
chemical reactions that take place in the synapse to induce synaptic weight changes.
Instead, they are all models that curve fit the outcomes of as many biological experi-
ments as possible under a unified mathematical expression. Advances in experimen-
tal techniques, including optogenetic and molecular methods, will permit researchers
to investigate intricate aspects of the biochemical network, including protein-protein
interactions, which result in plastic changes at the level of synapses. This now per-
mits the development of complicated biophysical models that take into account the
observed molecular processes underlying changes in synaptic strength. Such models
are expected to naturally reproduce the correct synaptic alteration for all experimental
protocols. Due to the close analogy of these models with the dynamics of the synapse,
these rules are usually called biophysical rules. In the following, we describe a few of
these rules, in particular those, for which a VLSI implementation is also available.
This rule not only considers calcium and its level for inducing synaptic weight changes,
but also introduces the effect of other ion channels and receptors as the pathways for
calcium to change in the post-synaptic neuron and therefore causes either potentiation
or depression. The synaptic weight change mechanism is as follows: pre-synaptic ac-
tion potentials release glutamate neurotransmitters that binds to N-methyl-D-aspartate
(NMDA) receptors, and when post-synaptic activities that provide large membrane
depolarisations are simultaneously present, it leads to an increase in the level of cal-
cium (Meng et al. 2011). This rule is capable of reproducing both BCM (rate-based)
and PSTDP (timing-based) mechanisms using a unified model. However, this model
is complex and requires a large number of state variables (Meng et al. 2011).
Page 46
Chapter 2 Neurons, Synapses and Synaptic Plasticity
on the dynamics of the ions and channels within the synapse. The rule was originally
proposed by Shouval et al. (2002) and Shouval et al. (2010) and modified to be imple-
mented in VLSI. The weight changes for the VLSI circuit are given by
Similar to the previous ion channel-based plasticity, this rule is shown to be capable
of reproducing BCM and spike pairing synaptic plasticity experiments. However, the
model is very complicated and needs several state variables to induce synaptic weight
changes in a biophysical form. Another limitation is that its ability to reproduce the
behaviour observed in triplet, quadruplet, and frequency-dependent pairing experi-
ments, has not been reported.
Since the calcium ion and its dynamics in the synaptic cleft appear to play an essen-
tial role in the synaptic plasticity, in addition to the aforementioned calcium-mediated
plasticity rules, several other rules, which build upon the calcium dynamics, have been
also proposed and investigated. In 2006, Shah et al. (2006) proposed a modified version
of the Shouval’s calcium-medicated rule (Shouval et al. 2002). This rule has been mod-
ified in order to account for the non-linear contributions of spike pairs to the synaptic
plasticity when considering a natural spike trains as reported in Froemke and Dan
(2002).
Besides these calcium-based models, there are some other new rules, which also utilise
calcium dynamics to induce synaptic weight changes. One of these rules, that can
be also counted as a phenomenological rule, is a simplified version of the Shouval’s
calcium-based rule. This rule utilised calcium dynamics to account for various bio-
logical experiments, and try to investigate the effects of a biophysical parameter, i.e.
calcium, on the synaptic plasticity (Graupner and Brunel 2012). The dynamic of the
synaptic plasticity rule is as follows:
dρ
τ = −ρ(1 − ρ)(ρ∗ − ρ) + γ p (1 − ρ)Θ[c(t) − θ p ] − γd ρΘ[c(t) − θd ] + Noise(t), (2.19)
dt
Page 47
2.6 Synaptic Plasticity Rules
where τ is the time constant of synaptic efficacy changes happening on the order of
seconds to minutes. Synaptic efficacy is shown with ρ and ρ∗ , which are the boundary
of the basins of attraction of two stable states, one at ρ = 0, a DOWN state correspond-
ing to low efficacy, and one at ρ = 1, an UP state corresponding to high efficacy. Here,
c(t) determines the momentous calcium concentration. There will be a potentiation if
c(t) is above a potentiation threshold, θd . Similarly, there will be a depression if c(t)
is above a depression threshold, θd . γ p and γd are potentiation and depression rates,
respectively, that will be in effect when the potentiation or depression threshold are
exceeded. In addition, Θ is a Heaviside function, in which Θ[c − θ ] = 0 for c < θ,
otherwise Θ[c − θ ] = 1 and Noise(t) is an activity-dependent noise term. For further
details refer to Graupner and Brunel (2012).
In addition, Badoual et al. (2006) proposed two synaptic plasticity mechanisms, one
with a biophysical background, and the other as a phenomenological model. Both
rules are based on multiple spike interactions and utilise the feature of conventional
STDP rule to reproduce the outcomes of some biological experiments including the
triplet experiments using the suppression STDP model presented in Froemke and Dan
(2002) and the STDP learning window (Song et al. 2000). Although these rules are able
to account for these experiments, reproducing other important experiments including
the pairing frequency experiments, quadruplet and BCM, are not reported using these
models.
Furthermore, another calcium-based STDP rule was recently proposed (Uramoto and
Torikai 2013), which utilises three state variables in an Ordinary Differential Equa-
tion (ODE) form. This model is simpler than previous calcium-based and voltage-
based STDP models as it uses lower number of state variables and does not use extra
parameters such as delay in the model as used by Graupner and Brunel (2012). Al-
though this model is simpler than some voltage- and calcium-based models and also
has higher synaptic plasticity abilities than the TSTDP rule, it still is more complicated
than TSTDP rule, which is a merely timing-based rule.
All the phenomenological rules mentioned above have stronger links to the biophysics
of synapses than the simpler phenomenological ones. However, these rules are also
more complex in structure, and therefore need more resources to be implemented in
VLSI. In fact, there exist only a few implementations of biophysical rules in the liter-
ature and those implementations are against the main needs of neuromorphic engi-
neers, i.e. low power consumption and compactness (Rachmuth et al. 2011, Meng et al.
Page 48
Chapter 2 Neurons, Synapses and Synaptic Plasticity
2011). Therefore, implementing a simpler, low power, and smaller circuit with ade-
quate synaptic capabilities, which accounts for a number of essential synaptic experi-
ments is absolutely promising for the realisation of a large-scale biophysically plausible
neuromorphic system.
In the next two chapters, three of the reviewed plasticity rules, i.e. PSTDP, TSTDP
and BCM that are the rules of main interest in this thesis are implemented and tested
in a programmable multi-neuron hardware neuromorphic system. Using this system
the hardware implementation of these synaptic plasticity rules are verified. Then the
implemented rules are utilised for classification of complex rate-based patterns.
Page 49
Page 50
Chapter 3
Programmable
Neuromorphic Circuits for
Spike-based Neural
Dynamics
T
HIS chapter describes the architecture and structure of a pro-
grammable hybrid analog/digital neuromorphic circuit, called
IFMEM, that can be used to build compact low-power neural
processing systems. Here, first the architecture of the IFMEM neuromor-
phic system is described and then it is explained how this hybrid analog-
digital CMOS circuit operates correctly over a wide range of input frequen-
cies; a feature that is essential for many engineering applications. The chap-
ter shows measurement results from available silicon neurons, and neuron-
synapse combinations and demonstrates how specific neural behaviours
can be generated by programming the chip and calibrating the silicon neu-
rons and synapses parameters. The provided information in this chapter
elucidates the presented results in the following chapter.
Page 51
3.1 Introduction
3.1 Introduction
As already mentioned, the two main components when implementing a neural sys-
tem are neurons and synapses. Synapses are essential components of spiking neu-
ral networks that represent the site of memory (as they store the network’s synap-
tic weight values), and play a fundamental role in computation (as they implement
crucial temporal and non-linear dynamics). In spiking neural networks, the synap-
tic weight is directly associated with the activity of pre-synaptic and post-synaptic
neurons (Kempter et al. 1999). Different types of learning algorithms have been pro-
posed, to update the synaptic weight as functions of both pre- and post-synaptic ac-
tivity (Brader et al. 2007, Morrison et al. 2008). The different learning strategies have
a profound effect on the post-synaptic neuron functionality and on the spiking neu-
ral network behaviour (Laughlin and Sejnowski 2003). Implementing such types of
synapses and learning mechanisms in compact electronic systems is crucial, for devel-
oping efficient large-scale spiking neural networks, which learn, and for brain-inspired
computing technologies that can adapt. However, as the implementation of the learn-
ing algorithm often depends on the specific application domain and on the nature of
Page 52
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
In this chapter a programmable neuromorphic circuit that has been fabricated using
standard CMOS VLSI process (Moradi and Indiveri 2011) and that can support any
weight-update mechanism of interest and learning strategies, is presented. Specifi-
cally, a set of experimental results measured from the fabricated neuron and synapse
circuits demonstrating how they can be calibrated to a specific targeted behaviour, is
shown. Using this circuit in the following chapter, it is shown how STDP and TSTDP
learning algorithms can be implemented and reproduce a number of synaptic plasticity
experimental outcomes.
The results in this chapter have been mainly presented in The 2013 New Circuits and
Systems conference, in Paris, France (Azghadi et al. 2013d), as well as in The ACM Journal
on Emerging Technologies in Computing Systems (Azghadi et al. 2014c). Note that the
IFMEM neuromorphic architecture is designed and fabricated in the Neuromorphic
Cognitive Systems (NCS) group, Institute of Neuroinformatics (INI), University and
ETH Zurich, Switzerland. The architecture and main characteristics of the IFMEM
neuromorphic device are presented mainly in Moradi and Indiveri (2011) and Moradi
and Indiveri (2014).
The multi-neuron chip used in this chapter is characterised by the fact that it comprises
circuits that implement models of IF neurons, and a programmable memory for storing
the synaptic weights. Therefore, it is referred to this device as the “IFMEM” (Integrate
and Fire neurons with synaptic Memory) chip.
The photo micro-graph of the IFMEM chip comprising the programmable synapse cir-
cuits is shown in Fig. 3.1. This chip was fabricated using a standard 0.35 µm CMOS
technology and is fully characterised in Moradi and Indiveri (2014). The micro-graph
depicts various parts of the chip, including an on-chip 32-bit programmable bias gener-
ator (Delbrück et al. 2010), SRAM cells, an arbitration part, an asynchronous controller,
and the “Neural Core”.
Page 53
3.2 The IFMEM Chip
Figure 3.1. The IFMEM neuromorphic device chip micro-graph. The multi-neuron IFMEM
chip was fabricated using a standard 0.35 µm CMOS technology and occupies an area
of 2.1×2.5 mm2 . The programmable synapses are integrated inside the neural core
block (Moradi and Indiveri 2014, Azghadi et al. 2013d).
All circuits on the chip that implement the neural and synapse dynamics are in the
neural core block. The neuron circuits are implemented using an “adaptive exponential
integrate and fire” model (Brette and Gerstner 2005, Indiveri et al. 2010), while the part
of the synapse circuits responsible for integrating input spikes and producing temporal
response properties that have biologically plausible time constants are implemented
using a DPI circuit (Bartolozzi and Indiveri 2007).
The block diagram of the chip architecture is shown in Fig. 3.2(a), which demonstrates
the working scheme of this programmable neural system. Depending on the input
address-event, different types of synapse dynamics can be triggered: excitatory with
slow time constants (e.g., to emulate NMDA-type synapses), excitatory synapses with
faster time constants (e.g., to emulate AMPA-synapses), or inhibitory synapses (e.g., to
emulate GABA-type synapses). Since the DPI can be used as a linear low-pass filter, it
is possible to make use of a single integrator circuit for any of the synapse dynamics
Page 54
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
REQ
ACK
A
E
input [4:0]
DE-MULTIPLEXER
Asynchronous Controller
R
AER-IN [17:0]
DECODER-Y
O
18 SRAM U
T
32 X32 W O R D S P
E A C H W O R D = 5 b its U
T
NEURAL CORE
DECODER-X DECODER-S
AER I/O
Bundled-data input [9:5] input [12:10]
Dual-rail Synapse Neuron
(a)
(b)
Figure 3.2. The IFMEM chip block diagram. (a) The device comprises a neural-core module
with an array of synapses and integrate-and-fire neurons, an asynchronous SRAM mod-
ule to store the synaptic weight values, a bias generator to set the parameters in the
analog circuits, and asynchronous control and interfacing circuits to manage the AER
communication. (b) Layout picture comprising the SRAM, neural core and AER out-
put blocks. In particular, the layout of the SRAM block measures 524×930 µm2 ; the
synapse array measures 309 µm in length, the synapse de-multiplexer measures 132 µm,
the neuron array 60 µm, and the output AER arbiter 105 µm in length, while the width
of all of them is equal to 930µm (Moradi and Indiveri 2014, Azghadi et al. 2014c).
considered (e.g., NMDA, AMPA, or GABA), and multiplex it in time to integrate the
contributions from multiple spiking inputs (e.g., via multiple SRAM cells).
The analog neural components available on the chip have programmable bias param-
eters that can be set with an on-chip 32-bit temperature compensated programmable
bias generator (Delbrück et al. 2010). The synaptic weights of the synapses are stored in
Page 55
3.2 The IFMEM Chip
a 32×32 5-bit digital SRAM block, designed with asynchronous circuits for interfacing
to the AER components. The digital weight values are converted into currents with
an on-chip Digital to Analog Converter (DAC), so that the addressed synapse circuits
produce EPSCs with amplitudes proportional to their weights.
The on-chip synapse circuits integrate incoming spikes and produce EPSCs with am-
plitudes proportional to their corresponding stored weights. The temporal response
properties of the circuit exhibit dynamics that are biophysically realistic and have
biologically plausible time constants (Bartolozzi and Indiveri 2007). The part of the
synapse circuit that produces the slow temporal dynamics is the log-domain DPI fil-
ter (Bartolozzi et al. 2006, Mitra et al. 2010), shown in Fig. 3.3. By using the DPI in its
linear regime, it is possible to time-multiplex the contributions from multiple spiking
inputs (e.g., via multiple SRAM cells), thus requiring one single integrating element
and saving precious silicon real-estate. This time multiplexing scheme, and circuits
implemented on the chip of Fig. 3.1 have been fully characterised in Moradi and In-
diveri (2011), while the description of the DPI synapse dynamics has been presented
in Bartolozzi et al. (2006).
Using the synapse time-multiplexing scheme, the total number of synapses that a neu-
ron sees is equivalent to the total number of SRAM cells present in each row. The
SRAM cells can work in “feed-through” mode or in storage mode. In feed-through
mode, input events contain both the address of the destination SRAM cell and the
synaptic weight bits, and the synapses generate EPSCs on-line, as the data is received.
In storage mode, the input events contain only the address of the destination SRAM
cell, and the weight bits used by the synapses are the ones stored in the addressed
SRAM cell (Moradi and Indiveri 2011). Therefore it is possible to interface the device
to a workstation and use it in “feed-through” mode to train the spiking neural net-
work on-line, with all of the hardware components in the loop, eventually storing the
final synaptic weight matrix in the SRAM block at the end of the training phase. Once
the training has completed, it is possible to use the device in stand-alone mode, with-
out requiring a PC in the loop, and use the stored weights to carry out the learned
task (Azghadi et al. 2013d, Azghadi et al. 2014c).
Figure 3.2(b) shows a section of the layout of the IFMEM chip comprising the main
blocks mentioned above. As shown, each block is extremely compact, so it is possible
in principle to scale up the network to very large sizes (e.g., a chip fabricated using an
Page 56
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
Figure 3.3. Schematic diagram of the programmable synapse circuit. The top part of the
diagram represents a DPI circuit that implements the temporal dynamics. The bottom
part of the diagram represents the DAC that converts the SRAM 5-bit weight into a
corresponding synaptic current (Moradi and Indiveri 2014, Azghadi et al. 2013d).
The experimental setup, shown in Figures 3.4, consists of the three main components:
a Linux PC, a generic AER interface board and, directly attached to it, a daughter-
board containing the IFMEM chip. The PC is used to control and interact with the
neuromorphic system. It generates spike trains that are sent to the IFMEM chip via an
AER interface. The PC also monitors, records and analyses the AER output of the chip.
Via a separate channel, the PC also sends bias values to IFMEM chip, which control its
various circuit parameters (Azghadi et al. 2014c).
The AEX board, shown in Fig. 3.5, is a generic AER communication platform derived
from the board first presented in Fasnacht et al. (2008). It consists of a high-speed
Page 57
3.3 Experimental Setup
Figure 3.4. Experimental setup of the hardware-software neuromorphic system. Dashed lines
represent the control path for setting analog parameters and configuring the IFMEM
chip, solid lines represent the path for the address-events data flow; from and to the
IFMEM chip (Azghadi et al. 2014c).
(480 MHz) USB2.0 interface and an FPGA device. The USB interface enables the FPGA
to communicate bi-directionally with the PC attached. The FPGA receives spike trains
from the PC via USB and then generates them accordingly on its Parallel AER output
interface to stimulate the IFMEM chip. Vice versa, the FPGA monitors the AER out-
put of the IFMEM chip: each address-event received by the FPGA is sent to the PC,
together with a 128 ns resolution time stamp of when exactly the spike was received at
the Parallel AER input of the FPGA. The AEX board also contains a high-speed Serial
AER interface to communicate with other AEX boards. Since only one such board is
required in the single-chip experimental setup described, the Serial AER interface was
not used (Azghadi et al. 2014c).
The IFMEM system utilises the AER protocol to transfer the required events to the chip
and at the same time, record the spikes being generated by the neurons on the chip.
Page 58
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
Figure 3.5. Printed circuit board of the experimental setup of the system. The AEX printed
circuit board with the attached daughterboard carrying the IFMEM chip (Azghadi et al.
2014c):
A: High-speed USB interface for AER communication,
B: USB interface chip,
C: FPGA for AER monitoring and sequencing,
D: Parallel AER interface chip to FPGA,
E: Parallel AER interface FPGA to chip,
F: Serial AER section (unused),
G: Full-speed USB interface for IFMEM bias control,
H: Microcontroller for bias control,
I: The IFMEM chip,
K: An analog voltage output connection.
Each pre-synaptic AER address contains four slots of information including 18 bits as
shown in Fig. 3.2(a). These bits describe different specifications including:
Page 59
3.4 Silicon Neuron and Programmable Synapse Response Properties
• the address of the SRAM block containing the required synaptic weight (5 bits),
• the type (either inhibitory or excitatory) and the address of the desired synapse
(3 bits), and
• the desired digital value for the synaptic weight that will be written to the ad-
dressed SRAM block (5 bits).
The weight across these virtual synapses can be modified using AER protocol and
according to various learning and synaptic plasticity rules including timing-based ones
e.g. STDP and triplet STDP (Azghadi et al. 2012a), as well as rate-based rules e.g.
Bienenstock-Cooper-Munro (BCM) rule (Azghadi et al. 2012b). Implementing these
rules will be discussed in the following chapter. Prior to implementing any synaptic
plasticity algorithm on the IFMEM neuromorphic system, the response properties of
the neuron and synapse circuits should be characterised.
In the following Section, first the silicon neuron response properties are shown, and
then the response properties of the synapse circuits as a function of input spike fre-
quency and of programmable weight values are characterised to evaluate their linear
characteristics and dynamic range properties.
Page 60
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
Figure 3.6. Silicon neuron response properties. Silicon neuron membrane potential (Vmem ) in
response to constant current injection is shown. Note that silicon neuorn biases are
set specifically to show this biologically plausible regular spiking neural behaviour with
biologically realistic time constants (Azghadi et al. 2014c).
Apart from the spiking behaviour of the neuron, setting a meaningful relationship
between the input current injected to the neuron and its mean spiking frequency is
another important characteristic of the neuron that was tested. Figure 3.7 shows an al-
most linear relationship between the amount of the injected current to the neuron and
its output mean firing rate. Note that the spiking dynamic of the neuron in this case,
that is shown in the inset of Fig. 3.7, is different from that shown in Fig. 3.6 (Azghadi et al.
2013d, Azghadi et al. 2014c).
In addition to the characteristics of the neurons, the properties of the available synapses
on the chip should also be correctly characterised. Therefore, another set of measure-
ments are carried out to calibrate the response properties of the combined synapse-
neuron. For this experiment, one post-synaptic neuron and one single input synapse
are employed. The response properties of the combined synapse-neuron circuit is char-
acterised by sending input spikes to the synapse, and measuring output spikes from
the neuron.
Note that the synapse circuit integrates input spikes to produce an output current (the
synapse EPSC) that has a mean steady-state amplitude, which depends on both the
input spike train frequency and its synaptic weight. The synapse, and if required the
neuron parameters can be calibrated to achieve the desired spiking activity in relation
Page 61
3.4 Silicon Neuron and Programmable Synapse Response Properties
Figure 3.7. Input current versus frequency charachtristics of silicon neurons. Mean firing
rates of all neurons on the chip, as a function of input current. The figure inset shows
the membrane potential of a single neuron (Azghadi et al. 2013d).
to the synaptic weight and the input spike frequency. This spiking activity depends
directly on the application, for which the Spiking Neural Network is going to be used.
In order to optimise the use of the neuron and synapse circuits in various applications
like computation, brain machine interface, pattern recognition etc., it is necessary to
tune their analog bias parameters to specific values to result in the required features
and expected behaviour of the neuromorphic chip for those applications (Azghadi et al.
2013a, Azghadi et al. 2014c).
Figure 3.8 shows how it is possible to optimise the circuit biases for a specific range of
low pre-synaptic frequencies, so that the combined synapse-neuron circuits respond
almost linearly to their afferent synaptic inputs, for all possible weight values that can
be programmed. Under this condition, the neuron is able to show identical (gain = 1)
or linearly higher (gain > 1) or lower (gain < 1) post-synaptic output frequency,
compared to afferent pre-synaptic input frequency. Here, gain is defined as the frac-
tion of post-synaptic to pre-synaptic firing rates in the highest synaptic weight setting
(w[4...0] = (11111)2 = 31). When there are synaptic inputs with various firing rates,
Page 62
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
Figure 3.8. Synapse-neuron output response properties for low input frequencies. (a) The
synapse-neuron bias parameters are optimised only for 31 Hz. (b) Regular spike trains
with pre-synaptic input frequencies of 31, 62, and 93 Hz are applied to the synapse-
neuron circuit, for different synaptic weight settings. The synapse-neuron biases are
optimised to have an almost linear relationship for these three different input frequen-
cies (Azghadi et al. 2013d).
the neuron and synapses should be tuned to act linearly for the whole possible in-
put firing range of frequencies. In the calibration of the bias values in our circuit, the
main sets of parameters that were tuned are those related to the five synaptic currents
Iw0 − Iw4 depicted in Fig. 3.3. Those parameters were tuned in a way to establish an
Page 63
3.4 Silicon Neuron and Programmable Synapse Response Properties
almost linear relationship between the 32-state digital synaptic weight values and the
neuron post-synaptic firing rates.
Note that, figure 3.8 demonstrates the neuron-synapse response properties using a set
of bias parameters optimised, in both neuron and synapse, for biologically plausible
firing rates. Figure 3.8(a) shows the neuron-synapse response properties that are opti-
mised only for 31 Hz. However, Fig. 3.8(b) demonstrates the response properties when
the biases are optimised for three various pre-synaptic input frequencies, simultane-
ously. While Fig. 3.8(a) demonstrates a close match to the expected linear behaviour,
the second figure loses some linearity.
Figure 3.9 shows similar results to Fig. 3.8, but with parameters optimised for high
firing rates (e.g., for applications that need to process incoming data quickly, and for
neuromorphic systems that do not need to interact with the environment). This fig-
ure shows that the biases can be optimised to achieve a very good linear relationship
between synaptic weight and the neuron output mean firing rate. Fig. 3.9(a) shows
the results when neuron and synapse biases are optimised for only 1 kHz input spike
frequency. While this figure shows a very close match between the neuron output fre-
quency, and an expected linear behaviour, Fig. 3.9(b) shows the results when the biases
are optimised for three different input frequencies.
All presented measurements in this chapter show the high degree of programmabil-
ity the IFMEM chip possesses. It was shown that both neuron and synapse properties
Page 64
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
Figure 3.9. Synapse-neuron output response properties for high input frequencies. (a) The
synapse-neuron bias parameters are optimised only for 1 kHz. (b) Regular spike trains
with pre-synaptic input frequencies of 1, 3, and 5 kHz are applied to the synapse-
neuron circuit, for different synaptic weight settings. The synapse-neuron biases are
optimised to have an almost linear relationship for these three different input frequen-
cies (Azghadi et al. 2013d).
can be controlled, by optimising and calibrating their bias parameters, to achieve any
required behaviour from the neuron. Note that for optimising various biases that con-
trol different behaviours shown above and to reach any specific properties, the on-chip
Page 65
3.5 Chapter Summary
Figure 3.10. Neuron input-output response properties for various synaptic weights. Neuron
output frequency versus frequency of incoming spikes, representing either a very high
firing rate of a single source, or multiple sources at lower firing rates (Azghadi et al.
2014c).
bias generator circuit (Delbrück et al. 2010) was programmed using a microcontroller,
which is integrated on the host PCB (see Fig. 3.5).
It was shown that the neurons and synapses implemented on the IFMEM chip are bio-
physically realistic and can provide biologically plausible time-constants if required. It
was also demonstrated how the neuron-synapse circuits on the chip can be tuned to re-
spond appropriately for different ranges of input firing rates. These features along with
its programmability, make the IFMEM chip a very useful platform for implementing
various synaptic plasticity rules and for different applications, such as pattern classifi-
cation, and general purpose programmable neural learning systems.
Page 66
Chapter 3 Programmable Neuromorphic Circuits for Spike-based Neural Dynamics
Next chapter presents how the IFMEM device is used to realise various types of spike-
based learning algorithms, based on either spike-timing relationships (e.g., STDP mech-
anisms), or spike rate-based ones (e.g. Bienenstock-Cooper-Munro (BCM) type rules).
It also shows that how the IFMEM chip can be employed to classify complex rate-based
patterns, using the TSTDP learning algorithm.
The implementations of various synaptic plasticity rules and also using them for a clas-
sification task, build the knowledge required for the main focus of this thesis, which
is VLSI implementation of STDP rules, and using them in similar engineering applica-
tions.
Page 67
Page 68
Chapter 4
T
HIS chapter describes the implementation of STDP rules and a
pattern classification neural network on the IFMEM neuromor-
phic system. It is shown that both PSTDP and TSTDP rules can be
implemented on this neuromorphic setup, demonstrating the expected be-
haviours seen in biological experiments. This chapter shows how the STDP
window can be generated using the silicon neurons and synapses available
on the system. It also shows how the STDP rule is used for generating a
competitive Hebbian learning behaviour observed in computational STDP
experiments. Furthermore, the TSTDP learning algorithm is implemented
on the chip. In order to test this implementation, it is utilised to reproduce
a rate-based BCM learning behaviour. Obtained results show the useful-
ness of the TSTDP learning algorithm for generating the rate-based BCM
learning behaviour. Finally, the implemented TSTDP learning mechanism
is utilised to train a simple feedforward spiking neural network to classify
some complex rate-based patterns. Obtained results show the high perfor-
mance of the TSTDP rule in the targeted classification task. The experiments
carried out in this chapter provide a comprehensive view of the STDP rules
and their properties and features, which are essential when designing VLSI
STDP synapses in the following chapters.
Page 69
4.1 Introduction
4.1 Introduction
As discussed in previous chapters, hardware implementations of spiking neural net-
works offer promising solutions for computational tasks that require compact and
low power computing technologies. As these solutions depend on both the specific
network architecture and the type of learning algorithm used, it is important to de-
velop spiking neural network devices that offer the possibility to reconfigure their net-
work topology and to implement different types of learning mechanisms. The previ-
ous chapter presented a neuromorphic multi-neuron VLSI device, IFMEM chip, with
on-chip programmable event-based hybrid analog/digital circuits. The event-based
nature of the input/output signals allows the use of Address-Event Representation in-
frastructures for configuring arbitrary network architectures, while the programmable
synaptic efficacy circuits allow the implementation of different types of spike-based
learning mechanisms (Moradi and Indiveri 2014, Azghadi et al. 2014c). This chap-
ter focuses on the learning aspects of the IFMEM system, and shows how different
Spike-Timing Dependent Plasticity learning rules can be implemented on-line, when
the VLSI device is interfaced to a workstation. It will also be demonstrated how, after
training, the VLSI device can act as a compact stand-alone solution for binary classifi-
cation of correlated complex rate-based patterns.
After implementing the pair-based STDP rule and showing its associated window and
replicating the bimodal behaviour in the weight distribution due to the competitive
nature of STDP synapses, the chapter continues to explore the triplet-based STDP rule,
which is the focus of this thesis. It is shown that this rule could reproduce rate-based
Page 70
Chapter 4 Timing-based Synaptic Plasticity Utilised for Pattern Classification
Reported results in this chapter are mainly presented in The ACM Journal on Emerging
Technologies in Computing Systems (Azghadi et al. 2014c).
where ∆t = tpost − tpre is the timing difference between a single pair of pre- and post-
synaptic spikes. According to this model, the synaptic weight will be potentiated if
a pre-synaptic spike arrives in a specified time window (τ+ ) before the occurrence
of a post-synaptic spike. Analogously, depression will occur if a pre-synaptic spike
occurs within a time window (τ− ) after the post-synaptic spike. These time windows
are not usually longer than 50 ms. The magnitude of potentiation/depression will
be determined as a function of the timing difference between pre- and post-synaptic
spikes, their temporal order, and their relevant amplitude parameters (A+ and A− ).
Fig. 4.1 demonstrates the conventional antisymmetric learning window associated with
the pair-based STDP rule. For generating this window, first the neuron was set to fire
spikes in response to a regular pre-synaptic spike train with the rate of 50 Hz. In this
case the neuron shows a behaviour similar to the one shown in Fig. 3.8(a). Then out-
going spikes from the post-synaptic neuron were recorded. Next, the weight of the
associated synapse was altered off-chip according to the PSTDP rule shown in Eq. 4.1.
Figure 4.1 shows the resulting STDP weight changes that occurred due to the random
time difference among pre- and post-synaptic spikes.
The figure suggests that the post-synaptic neuron spikes in a regular way similar to the
pre-synaptic spikes applied to the synapse with PSTDP. It should be noted that during
Page 71
4.2 Spike Timing Dependent Plasticity (STDP)
Figure 4.1. PSTDP learning window generated on the IFMEM neuromorphic device. A
regular pre-synaptic spike train was applied to the neuron and the weights were modified
according to the timing differences between nearest neighbour spikes in the pre- and
post-synaptic spike trains. Note that, ∆w in this figure determines the amount of weight
change computed off-chip and according to the time differences between the spikes
applied to the synapse (pre-synaptic spike) and those generated by the silicon neuron
(post-synaptic spike). The STDP learning rule parameters are shown in Table 4.1.
all experiments performed in this thesis, the nearest neighbour (in contrast to all-to-
all) spike interaction, in which only the immediate preceding and succeeding adjacent
spikes are considered for weight modifications, is adopted. The synaptic parameters
used for the STDP window experiment are shown in Table 4.1. One can change the
amplitude as well as the time constants of the learning window using the parameters
shown in this Table.
It is already verified that Hebbian learning has two substantial requirements to be de-
veloped. The first requirement is to control synaptic efficacy through some activity-
dependent synaptic plasticity rules such as STDP, while the second requirement is a
competition mechanism among synapses (Song et al. 2000). It is shown that, under
specific circumstances, synapses, which their weights are governed by STDP, compete
to control the spiking activity of their post-synaptic neuron. This competition leads to
Page 72
Chapter 4 Timing-based Synaptic Plasticity Utilised for Pattern Classification
Table 4.1. STDP parameters for producing STDP learning window on the IFMEM chip. The
four required parameters of Eq. 4.1, for producing the STDP window (Song et al. 2000)
are shown here.
divergence of synaptic weights into two distinguished groups. The first group is com-
posed of strong synapses, which their input spikes have been more correlated so they
became strong due to STDP. By contrast, the second group includes weak synapses,
whose input spikes have been less correlated and therefore, they became weaker in
result of STDP (Song et al. 2000). The competition and the resulting stable synaptic
weight distribution, do not arise unless the following conditions are satisfied: (i) Im-
posing a hard boundary on the strength of individual synapses, and (ii) Setting synap-
tic parameters in a way that, synaptic weakening through STDP slightly outweighs
synaptic strengthening, i.e. A+ τ+ < A− τ− .
First, we calibrate the silicon neuron on the chip, in a way that it is reasonably excited
so that it can respond to its input spikes by firing action potentials. This is possible
through increasing the injection current applied to the neuron. There are excessive
control parameters for setting the dynamics of the silicon neuron including a parameter
for controlling its spiking threshold, a parameter for adapting its spiking frequency, as
well as a parameter for setting its refractory periods that also play fundamental roles
in the spiking activity of the neuron—please refer to Moradi and Indiveri (2014) for
further information.
Second, we set all 32 digital synaptic weights, which are 5-bit asynchronous SRAM
cells, and therefore have 32 efficacy states, to their mid values of 16.
Page 73
4.2 Spike Timing Dependent Plasticity (STDP)
Table 4.2. STDP parameters for producing competitive Hebbian learning behaviour on the
IFMEM chip. The four required parameters of Eq. 4.1, for producing the competitive
Hebbian learning behaviour (Song et al. 2000) are shown here.
Next, we apply 32 independent Poissonian spike trains with firing rates of 50 Hz to all
32 synapses. These spike trains will be time-multiplexed on a single DPI synapse, and
it produces an integration of 32 synaptic currents that are proportional to each of the
32 synaptic weights. The integrated current then generates the EPSC current, which in
turn is applied to the tuned post-synaptic silicon neuron. The neuron then generates
action potentials in response to the synaptic currents it receives.
The timing of the post-synaptic spikes generated by the silicon neuron and the timing
of pre-synaptic spikes applied to each of the 32 synapses then govern the magnitude
of changes to the digital weights stored in the SRAM cells affiliated with each synapse,
according to the STDP rule presented in Eq. 4.1.
Fig. 4.2 demonstrates how synaptic weights in the mentioned setup evolve over time
to reach a homeostatic state, in which synaptic weights are approximately either weak-
ened or strengthened. At time = 0 s, all digital synaptic weights are set to their middle
value (i.e. w[4...0] = (10000)2 = 16), so that a high firing rate of the silicon neuron is
achieved. Then, the 32 synaptic weights are modified by STDP rule, which is imple-
mented off-chip and updates synaptic weights stored in the SRAM cells, based on the
timing difference between the current post-synaptic spike, and pre-synaptic spikes ar-
rived immediately before or after this post-synaptic spike, in each of the 32 synapses.
The synaptic parameters for this experiment are shown in Table 4.2.
After updating the weights off-chip, the modified weights are written back to their re-
lated SRAM cells through AER communication system. In response to these changes
weights start to modify and distribute across the whole range of 32-state weight spec-
trum (see time = 100 s). Eventually the weight of the 32 synapses diverge into two
groups, one includes stronger synapses and the other one contains weaker synapses.
This divergence is mainly because those synapses whose spike trains are more corre-
lated get stronger due to STDP, while their stronger weights acts as a positive feedback
Page 74
Chapter 4 Timing-based Synaptic Plasticity Utilised for Pattern Classification
Figure 4.2. Synaptic weights evolve to reach an equilibrium state, when modified by STDP
learning rule. Here, 32 synaptic weights (weight bins) each one with 32 digital states,
are altered by STDP over time. The top figure shows that all 32 synaptic weights are
set to 16 in the beginning at time = 0 s, i.e. the fraction of weights in weight bin
16 is 32. The other figures show the evolution of weights over time to reach a steady
state at time = 1000 s. The synaptic weights stay almost fixed thereafter, and the
post-synaptic neuron firing rate held in an almost direct relation to the mean firing rate
of pre-synaptic spike trains (Azghadi et al. 2014c).
and help their weights gets even stranger. On the other hand, those synapses that re-
ceive less correlated spike trains get weaker, in this STDP competition. This feature can
be used in an unsupervised form of Hebbian learning based on the correlation among
input spikes.
Page 75
4.2 Spike Timing Dependent Plasticity (STDP)
At time = 1000 s, the weights reach an equilibrium state, in which synaptic weights
do not alter anymore and the spiking activity of the post-synaptic neurons is propor-
tionate to the mean firing rate of pre-synaptic spikes. This is an interesting feature of
STDP, which let the neuron reach a steady state.
It should be noted that, in the presented experiment, the synaptic weights are bounded
between 1 and 31, and each weight is rounded either upward or downward to its clos-
est digital weight value after each synaptic weight update. In addition, further emu-
lations suggest that the initial distribution of the 32 synaptic weights does not have a
significant impact on the distribution of final weights.
Although BCM is an inherently rate-based rule and depends on the activities of pre-
and post-synaptic neurons, recent studies have shown that timing-based triplet STDP
learning rule can reproduce BCM-like functionality (Gjorgjieva et al. 2011, Pfister and
Gerstner 2006). Here, it is demonstrated how this rate-based functionality can be
realised by our software-hardware system, by using the triplet STDP learning rule
(Pfister and Gerstner 2006, Gjorgjieva et al. 2011) to update the 5-bit synaptic weight
values of the IFMEM chip.
where ∆w = ∆w+ for t = tpost and if t = tpre then the weight change is ∆w = ∆w− .
Here, A2+ , A2− , A3+ and A3− are potentiation and depression amplitude parameters,
∆t1 = tpost(n) − tpre(n) , ∆t2 = tpost(n) − tpost(n−1) − ǫ and ∆t3 = tpre(n) − tpre(n−1) − ǫ, are
the time differences between combinations of pre- and post-synaptic spikes. Here, ǫ is
a small positive constant which ensures that the weight update uses the correct values
occurring just before the pre- or post-synaptic spike of interest, and finally τ− , τ+ , τx
and τy represent time constants (Pfister and Gerstner 2006).
It has been shown (Pfister and Gerstner 2006) that for Poisson distributed spike trains
Eq. 4.2 can be approximated as:
hdw/dti = − A2− τ− ρpre ρpost − A3− τ− τx ρ2pre ρpost + A2+ τ+ ρpre ρpost + A3+ τ+ τy ρ2post ρpost
(4.3)
Page 76
Chapter 4 Timing-based Synaptic Plasticity Utilised for Pattern Classification
where ρpre and ρpost represent the mean firing rates of the pre- and post-synaptic spike
trains, respectively.
Generally, the BCM theory suggests that the synaptic weight changes have a linear
relationship with the pre-synaptic, and a non-linear relationship with the post-synaptic
mean firing rates (Bienenstock et al. 1982). Therefore, a general description of the BCM
rule can be written as:
where φ is a function that satisfies the conditions φ(ρpost > θ, θ ) > 0, φ(ρpost < θ, θ ) <
0 and φ(0, θ ) = 0. Essentially, if the post-synaptic firing rate, ρpost , is below the thresh-
old θ, then dw/dt is negative and the synaptic weight is depressed. Conversely, the
synaptic weight is potentiated if the post-synaptic firing rate is larger than the thresh-
old θ, and it is left unchanged if φ = 0, i.e., if ρpost = θ (Pfister and Gerstner 2006).
The Eqs. 4.3 and 4.4 can be mapped together, if two conditions are satisfied. The first
condition requires having a linear relationship between the pre-synaptic firing activity,
ρpre , and the synaptic weight change, hdw/dti, as shown in Eq. 4.4. This condition is
satisfied if A3− = 0, in the triplet STDP equation (Eq. 4.3). This will lead to a minimal
version of the TSTDP rule presented in Pfister and Gerstner (2006), which has been
shown to account for various synaptic plasticity neuroscience experiments, including
those dealing with higher order spike trains (Wang et al. 2005). The second condition
requires that the sliding threshold θ, that determines the frequency, in which depres-
sion turns to potentiation, is proportional to the expectation of the pth power of the
post-synaptic firing rate (ρpost ) (Pfister and Gerstner 2006, Bienenstock et al. 1982). This
second condition can be satisfied if the threshold of the BCM rule is defined as
p p
θ = hρpost i( A2− τ− + A2+ τ+ )/ρ0 A3+ τ+ τy . (4.5)
Given this equation, the sliding threshold effect of the BCM rule is proportional to the
post-synaptic firing rate, with the proportionality factor set by the STDP rule param-
eters. Previous studies have shown the possibility of mimicking the effects of BCM
rule through TSTDP (Azghadi et al. 2013a). However, similar to the experiments per-
formed in Gjorgjieva et al. (2011), they have used independent pre- and post-synaptic
spike trains with mean firing rate of ρpre and ρpost , respectively.
Page 77
4.2 Spike Timing Dependent Plasticity (STDP)
To implement BCM via the triplet STDP rule in the IFMEM chip setup, we used a
single synapse, connected to a post-synaptic silicon neuron and changed its efficacy
using the STDP rule of Eq. (2.3). At the beginning of the experiment, the initial weight
of the synapse is set to its maximum value of 31 (i.e. w[4...0] = (11111)2 ). This high
synaptic weight makes the post-synaptic neuron fire at a high rate, proportional to
the pre-synaptic firing rate (Azghadi et al. 2013a). The pre-synaptic spike train here is a
Poisson spike train, similar to the spike trains used in previous studies (Gjorgjieva et al.
2011, Azghadi et al. 2013a). Using the AER protocol, we transmitted the software gener-
ated Poisson pre-synaptic spike train to the targeted post-synaptic silicon neuron, via a
synapse with an efficacy proportional to its weight stored in the corresponding SRAM
cell. The software pre-synaptic spike train, and the spike train produced by the silicon
neuron, are then used to calculate the amount of weight changes in the corresponding
synaptic efficacy, according to a minimal model of triplet STDP (Gjorgjieva et al. 2011).
[Hz]
Figure 4.3. The BCM rule is implemented through TSTDP rule on the IFMEM neuro-
morphic chip. The sliding threshold feature of the rate-based BCM rule is replicated
through Triplet STDP rule, implemented on the IFMEM chip (Azghadi et al. 2014c).
Figure 4.3 shows the total amount of weight change in response to Poisson spike trains
of 20 s length, for a range of pre-synaptic spike rates from 0 Hz up to 100 Hz. In this
figure, the sliding threshold feature of the BCM learning rule is regenerated through
changing the amount of one of the parameters of the TSTDP learning rule, i.e. A3+ . Ac-
cording to Eq. 4.5, with increase in A3+ parameter, the threshold decreases and slides
Page 78
Chapter 4 Timing-based Synaptic Plasticity Utilised for Pattern Classification
Table 4.3. Optimised TSTDP model parameters for generating BCM-like behaviour on the
IFMEM chip. The eight required parameters of Eq. 4.2, for producing a BCM-like
behaviour (Pfister and Gerstner 2006) are shown here. Note that x indicates ’don’t
care’.
Parameter name A2+ A2− A3+ A3− τ+ (ms) τ− (ms) τy (ms) τx (ms)
Value 0 0.0068 see Fig. 4.3 0 16.8 33.7 114 x
toward lower post- synaptic firing rates. Please note that, in the presented experiment,
the silicon neuron parameters, as well as the synaptic weight parameters in its corre-
sponding physical synapse, i.e. the differential pair integrator, are calibrated in a way
that pre- and post-synaptic neuron are in a relatively linear relationship (Moradi and
Indiveri 2014, Azghadi et al. 2013d). In this figure, each data point corresponds to the
mean of the weight changes over 10 trials, and the error bar represents the standard
deviation of the weight change over these trials. This amount of weight change can
then be discretised and written back into the SRAM. The STDP parameters that have
been used in this experiment are shown in Table 4.3.
In this Section, classification of complex rate-based patterns is targeted using the TSTDP
learning rule. For implementing the targeted classification task, the TSTDP learning
rule, with its parameters tuned for exhibiting BCM behaviour (see Fig. 4.3) are used.
This section demonstrates how the TSTDP rule implemented on the IFMEM device can
perform classification of binary patterns with high levels of correlations.
The neural classifier implemented on the chip is composed of one neuron and 30
synapses, which are arranged in a single layer perceptron-like architecture. The goal is
to train the perceptron synaptic weights, via the TSTDP algorithm, to learn to distin-
guish two input patterns, UP and DOWN, in an unsupervised fashion. After training,
the hardware perceptron should be able to respond with a high firing rate to pattern
UP, and a low one to pattern DOWN. This is a similar experimental scenario, to the
semi-supervised learning scenario utilised in an identical classification task performed
using spiking neural networks (Giulioni et al. 2009).
Page 79
4.3 Classification of Complex Correlated Patterns
The two UP and DOWN patterns can have various degrees of correlations. The correla-
tion determines the amount of overlap in the input synapses used, and the similarity in
the output response of the neuron. When there is no correlation, one pattern is applied
to 15 random synapses and the other pattern is applied to the remaining 15 synapses
(no overlap).
The two patterns are defined as follows. The pattern UP stimulates 15 synapses with
Poisson spike trains that have a high mean firing rate of 300 Hz, while pattern DOWN
comprises 15 Poisson spike trains with a low mean firing rate of 20 Hz. Therefore, in
the case of zero correlation, the two patterns are likely to produce different outputs (de-
pending on the values of the synaptic weights) even before learning. However, for the
case of non-zero correlations, a random subset of N input synapses are always stimu-
lated by high mean firing rate spike trains of 300 Hz, while the rest of the synapses are
assigned to the two UP and DOWN patterns. For instance, if the number of correlated
synapses is 10, 10 randomly synapses are stimulated by Poisson spike trains of 300 Hz,
and the remaining 20 synapses will be reserved for the UP and DOWN patterns. In
this case, pattern UP (DOWN) is presented as 10 high (low) rate spike trains that are
applied to 10 random synapses from the 20 synapses, and pattern DOWN (UP) is pre-
sented to the remaining 10 synapses. In this case, because of the N common high input
synapses, the two patterns will have closer mean firing rates, and therefore their classi-
fication becomes more challenging. Therefore, in the beginning of learning phase, the
output frequency range of the perceptron cannot be distinguished between the two
patterns and as a result, learning is required to classify the two patterns.
The training phase is composed of several trials. In each trial, one of the two patterns,
UP or DOWN is randomly applied to the 30 input synapses, with a set degree of cor-
relation, and with a new distribution of Poisson spikes. The two patterns have equal
probability to be selected. For each trial, the synaptic weights are modified according
to the TSTDP. In our experiment the synaptic weights reach a steady state and do not
change significantly after about 20 trials, in which the input spike trains lasted 10 s
each.
Figure 4.4 shows how the distribution of the neuron output firing rates, in response
to the two different patterns, changes with learning, after 1, 5, 10, and 20 trials. The
output neuron rates were collected over 20 classification runs, with each run compris-
ing 20 learning trials, while 20% correlation is set among the two patterns. In each
run the synaptic weights are initialised to random 5-bit values, the definition of UP
Page 80
Chapter 4 Timing-based Synaptic Plasticity Utilised for Pattern Classification
Figure 4.4. Distribution of the neuron output frequencies during different stages of learning.
In the beginning of the learning phase, when initial weights are random, the neuron
cannot distinguish between the two patterns. During the learning trials, the synapses
are being modified and the neuron begins to effectively discriminate between the two
patterns, UP and DOWN, from trial 20. In this experiment the correlation is equal to
20 %, i.e., there are 6 inputs that are common to the two patterns that always receive
high firing rates (Azghadi et al. 2014c).
and DOWN patterns is changed, and a new random order of UP and DOWN patterns
applied across trials, is defined.
As Fig. 4.4 shows after one stage of learning, trial 1, the neuron is still unable to distin-
guish between the two patterns, even though the weights have changed once, in result
of applying either pattern UP or DOWN to the network and updating the synaptic
weights according to the TSTDP learning rule. As the learning phase proceeds, the
Page 81
4.3 Classification of Complex Correlated Patterns
neuron starts to fire with higher rates for pattern UP, compared to the rates for pat-
tern DOWN. Figure 4.4 shows that after 20 trials, the neuron becomes significantly
specialised to the two patterns and fire with a higher rates for pattern UP, and with a
lower rate for pattern DOWN.
As expected, the utilised TSTDP learning rule tends to decrease the weights of the
synapses targeted by the DOWN pattern, while it tends to increase the weights of
both the UP and correlated (overlapping) synapses. After learning, the neuron will
therefore fire with high firing rates when stimulated with UP patterns, and low firing
rates when stimulated by DOWN patterns. While after a few trials (e.g. see second
and third panels of Fig. 4.4) the neuron already performs above chance levels, many
trials (20 in our experiments) are required to unambiguously classify the two patterns.
Regarding the classification performance, our results show that the implemented clas-
sification network performs robustly and holds for also large amount of correlation,
i.e. more than 50 % correlation, in the input patterns. In terms of classification accu-
racy, we consider a DOWN pattern correctly classified if the neuron output frequency
is less than a set threshold in response to that pattern; and similarly, an UP pattern
is correctly classified, if the neuron response to such pattern has a firing rate higher
than the threshold. In our experiments, the classifier has 100 % correct performance,
even with correlation levels of 87 % (i.e., 26 overlapping synapses), if the classification
threshold is adaptive (e.g., if it is set just below to the minimum frequency in response
to the UP patterns). What changes however is the difference in the responses to the
two patterns. Figure 4.5 shows how this difference decreases as the correlation among
the input patterns increases.
Figure 4.5 depicts that with increase in the correlation level between the two patterns,
the number of common active synapses between the two patterns increases and there-
fore, the neuron firing rate for the two patterns becomes closer. As a result, the differ-
ence in the output frequency for pattern UP and DOWN will decrease.
In contrary to an adaptive threshold, one might consider selecting a fixed threshold for
the whole range of correlations. In this case, the classification accuracy of the imple-
mented classifier will still be 100%, for patterns with up to 47% of correlations (corre-
lation level = 14). However, the accuracy drops to 50% afterwards, because in the case
of a fixed threshold, only UP patterns are correctly classified for correlation level great
than 14.
Page 82
Chapter 4 Timing-based Synaptic Plasticity Utilised for Pattern Classification
Figure 4.5. The performance of the classifier implemented on the IFMEM chip. Here ∆F =
min − F max ) /F min , where F min and F max
( FUP DOWN UP UP DOWN are the minimum and the maximum
frequencies for pattern UP and DOWN, respectively, for all 20 runs at the end of learning
in trial 20. (Azghadi et al. 2014c).
This chapter presented how the programmable neuromorphic chip that was presented
in previous chapter can be used in a hybrid software-hardware system to implement
different types of spike-timing dependent plasticity learning rules. It was also demon-
strated how these rules can reproduce competitive Hebbian learning and rate-based
behaviours, even with the limitations of the hardware implementation (5-bit resolu-
tion for the weights, mismatch of the analog subthreshold circuits, etc.). Finally it was
described how the hybrid software-hardware learning setup proposed can be used to
Page 83
4.4 Chapter Summary
The device and setup proposed in this chapter, therefore represents a useful real-time
low-power computing platform for exploring the effectiveness of different types of
spike-based learning algorithms, validating their performance at run-time on real-time
custom analog/digital hardware, and implementing robust perceptron-like neural net-
works to carry out real-time classifications tasks. If the task can be solved after training
the weights of the neural network, without requiring continuous or on-line training,
then the platform proposed represents a stand-alone compact and low-power alterna-
tive to standard full-digital computing solutions (no PC is required in the loop).
This chapter along with the previous one, provided us with a good knowledge on
the modelling, implementation and behaviour of both PSTDP and TSTDP rules. This
knowledge and experience is quite helpful while designing these rules in VLSI, spe-
cially when these synaptic plasticity rules are employed in a synapse and integrated
with a silicon neuron. The experiments performed in these two chapters are also es-
sential if using the VLSI versions of the STDP circuits in a neuromorphic system for
specific tasks such as pattern classification is targeted.
After gaining knowledge on the structure of the STDP rules and the behaviours these
rules show, we can now start designing these rules in VLSI and use them to show var-
ious behaviours already observed using the IFMEM chip. However, before designing
new circuits, a review of previous VLSI designs for various synaptic plasticity rules
is required to build our knowledge on the state-of-the-art synaptic plasticity rules in
silicon. The next chapter, is dedicated to discussion and review of various synaptic
plasticity models in VLSI.
Page 84
Chapter 5
Spike-based Synaptic
Plasticity Rules in Silicon
T
HIS chapter reviews Very Large Scale Integration (VLSI) circuit
implementations of various synaptic plasticity rules, ranging from
phenomenological ones (i.e. timing-based, rate-based, or hybrid
rules) to biophysically realistic ones (e.g. based on calcium dependent mod-
els). It discusses the application domains, weaknesses and strengths of the
various representative approaches proposed in the literature and provides
deeper insight into the challenges that engineers face when designing and
implementing synaptic plasticity rules in order to utilise them in real-world
applications. The chapter also proposes and discusses various counter ap-
proaches to tackle the challenges in neuromorphic engineering. The review
performed in this chapter helps build knowledge useful for the design of
new VLSI circuits for synaptic plasticity rules while considering the chal-
lenges, applications and effective design methods and techniques.
Page 85
5.1 Introduction
5.1 Introduction
For over a century, there has been considerable effort in attempting to find answers
to the question: “how does learning and memory take place in the brain?” Although
there is still no general agreement, neuroscientists agree on some general rules and hy-
potheses to answer this question (Hebb 2002, Cooper et al. 2004, Sjöström et al. 2008).
It is agreed that learning and memory in the brain are governed mainly by complex
molecular processes, which give rise to a phenomenon called synaptic plasticity. The
actions of synaptic plasticity can manifest themselves through alterations in the effi-
cacy of synapses that allow networks of cells to alter their communication. Hebbian
learning, postulated by Donald Hebb in 1949 (Hebb 2002), is the foremost recognised
class of synaptic plasticity rules. It has formed the foundation of a number of other
plasticity rules (Gerstner and Kistler 2002). Synaptic plasticity rules can be categorised
into two general groups i.e. Short Term Synaptic Plasticity (STSP) and Long-Term
Synaptic Plasticity (LTSP). While STSP plays a fundamental role in decoding and pro-
cessing neural signals on short time scales (Zucker and Regehr 2002, Buonomano 2000),
it is LTSP that is responsible for learning and memory in the brain (Sjöström et al. 2008).
This type of plasticity produces long lasting depression or potentiation in the synaptic
weights. Specifically long-term plasticity can produce an increase in synaptic efficacy,
which results in Long Term Potentiation (LTP) of the synapse, or it can produce a de-
crease in the synaptic weight, resulting in Long Term Depression (LTD). It is widely
believed that these long term processes are the basis for learning and memory in the
brain. From an engineering perspective, it is important to understand how these LTSP
mechanisms can be translated into physical models that can implement adaptive learn-
ing capabilities in artificial SNNs.
There are multiple approaches that address this problem, which depend on the target
application domain for the artificial SNN (e.g., for basic research, for providing tools to
computational neuroscientists, or for practical real-time engineering applications). For
example, some of these approaches mimic the mechanisms occurring in real synapses
and model them at a phenomenological level, while other approaches take into account
many details of the real plasticity mechanism to model biological synapses as faithfully
as possible.
Page 86
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Then these hypotheses can be implemented in software or hardware and tested with
real-world stimuli to verify their values in addressing real-world challenges. While
software implementations are ideal for exploring different hypotheses and testing dif-
ferent models, dedicated hardware implementations are commonly used to implement
efficient neural processing systems that can be exposed to real-world stimuli from the
environment and process them in real-time, using massively parallel elements that
operate with time constants that are similar to those measured in biological neural
systems. This approach is followed for both attempting to get a deeper understand-
ing of how learning occurs in physical systems (including the brain), and for realising
efficient hardware systems that can be used to carry out complex practical tasks, rang-
ing from sensory processing to surveillance, robotics, or brain-machine interfaces. The
synaptic plasticity models developed by neuroscientists are typically translated into
electronic circuits and implemented using conventional VLSI technologies. Currently,
many of these models form the foundations for developing VLSI “neuromorphic sys-
tems” (Mead 1990, Indiveri and Horiuchi 2011). This chapter reviews and discusses
the most representative implementations of synaptic plasticity models presented in
the literature and compare them in terms of complexity and usefulness for various
applications.
The remainder of this chapter is organised as follows. Section 5.2 describes a num-
ber of fundamental building blocks useful for implementing those synaptic plasticity
rules that were discussed in Chapter 2. In Section 5.3, we review a range of representa-
tive implementations of synaptic plasticity rules, designed using the aforementioned
building blocks. Section 5.4 addresses the challenges that neuromorphic engineers
encounter when designing these systems. Section 5.5 discusses the previously men-
tioned synaptic plasticity designs and approaches and points out their benefits and
limitations. Section 5.6 demonstrates examples of how synaptic plasticity rules can
be employed in practical applications. Finally, Section 5.7 summarises our concluding
remarks for this chapter.
Presented results, review and discussion in this chapter are mainly presented in The
Proceedings of the IEEE (Azghadi et al. 2014b).
Page 87
5.2 Building Blocks for Implementing Synaptic Plasticity Rules in VLSI
The storage of synaptic weight values, and of other state variables that need to be
memorised in models of synaptic plasticity requires memory elements in VLSI. Typi-
cally, analog values are stored using capacitors. In VLSI technology, capacitors can be
implemented using Metal-Oxide-Semiconductor Capacitors (MOSCAPs), or multiple
layers of poly-silicon separated by an insulator (typically silicon-dioxide). These solu-
tions usually offer the most compact and convenient way of storing variables, but they
have the limitation of being leaky: as the charge stored in these devices tend to slowly
leak away due to imperfect insulator used in building these devices. Alternative ways
of storing analog variables involve the use of floating-gate devices (Ramakrishnan et al.
2011), or of dedicated analog-to-digital converters (ADCs) and digital memory circuits,
such as Static Random Access Memory (SRAM) elements (Moradi and Indiveri 2011,
Azghadi et al. 2013d). Considering the required time constants, the targeted network
and its desired application, these approaches are more/less bulky and/or convenient,
compared to the storage on VLSI capacitors, which is not applicable for long-time stor-
age. Another issue that should be taken into account when selecting the storage tech-
nique for synaptic weights is the required precision needed for a given application
(Pfeil et al. 2012). This issue is discussed in Section 5.4.8.
Neuromorphic engineers are interested in the subthreshold domain for two essential
reasons. The first reason is the exponential relationship between the drain current, ID
Page 88
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Figure 5.1. NMOS transistor in subthreshold. (a) Symbol for a NMOS transistor. (b) The
drain-source current, Ids , of a NMOS device in its subthreshold region of operation is a
summation of two currents with opposite directions. (c) Current-voltage characteristic
of the NMOS transistor, which shows significantly different behaviour for above and
below threshold (Liu et al. 2002).
Page 89
5.2 Building Blocks for Implementing Synaptic Plasticity Rules in VLSI
potential, respectively (Liu et al. 2002). Fig. 5.1(b) shows that, the drain-source current
shown in Eq. 5.1 is a summation of two currents in opposite directions, one is called
forward current, I f , which is a function of the gate-source voltage, and flows from the
drain to the source, and the other current, Ir , the reverse current, flows from the source
to the drain
Ids = I0 eκn Vg /UT −Vs /UT − I0 eκn Vg /UT −Vd /UT = I f − Ir . (5.2)
If Vds > 4UT ≈ 100 mV, as the energy band diagram in Fig. 5.1(b) shows, because of
the larger barrier height (in contrast to the Vds < 4UT state, where barrier heights are
almost equal), the concentration of electrons at the drain end of the channel will be
much lower than that at the source end, and therefore the reverse current, from source
to drain, Ir becomes negligible, and the transistor will operate in the subthreshold sat-
uration regime. Therefore, there will be a pure exponential relationship between Vgs
and Ids as
Page 90
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Differential pairs (DPs) are electronic components widely utilised in neural analog cir-
cuit design (Liu et al. 2002, Douglas et al. 1995). A DP in its basic form consists of
three transistors, two of which are used for receiving the input voltages at their gates
and the other one for biasing the pair by a constant current source—see Fig. 5.2(a).
As shown in Fig. 5.2(c), a DP sets a sigmoidal relationship between differential input
voltages and the currents flowing across each of the two differential transistors. The
sigmoidal function is crucial to artificial neural networks and has been useful in de-
scribing the activities of populations of neurons (Wilson and Cowan 1972). This makes
the differential pair an interesting and useful building block for neuromorphic engi-
neers. Differential pairs can be used for various applications including spike integra-
tion for a synapse circuit (Bartolozzi and Indiveri 2007), and a rough voltage difference
calculator (Mayr et al. 2010). They are also the heart of Operational Transconductance
Amplifier (OTA)—see Fig. 5.2(b).
The OTA is another essential building block not only in neuromorphic engineering,
but also in general analog integrated circuit design (Liu et al. 2002, Douglas et al. 1995,
Razavi 2002). It is usually used to perform voltage mode computation and produces
an output as a current. This analog component is commonly employed as a voltage-
controlled linear conductor. However, in its simplest form the OTA is not really lin-
ear and usually sets a sigmoidal function between differential voltage inputs and the
output current—see Fig. 5.2(c). In various VLSI implementations of neuromorphic
synapses and synaptic plasticity rules, the OTA has been used in different roles (Cruz-
Albrecht et al. 2012). In some cases, it has been used to act as an active resistor when
forming a leaky integrator (Koickal et al. 2007, Mayr et al. 2010), and sometimes to act
as a low-cost comparator (Mayr et al. 2010). In addition, a number of neuromorphic
designers have carried out some changes to the basic structure of the OTA (Rachmuth
and Poon 2008, Mayr et al. 2010, Cruz-Albrecht et al. 2012) to increase its symmetry,
dynamic range and linearity and at the same time decrease the offset. As a result, the
OTA has greater stability against noise and process variation, and gains better ability
to mimic the desired neural function (Liu et al. 2002, Rachmuth and Poon 2008, Rach-
muth et al. 2011, Mayr et al. 2010).
Page 91
5.2 Building Blocks for Implementing Synaptic Plasticity Rules in VLSI
Figure 5.2. Differential Pair (DP) and Operational Transconductance Amplifier (OTA). (a)
A basic Differential Pair (DP) circuit consists of three transistors. (b) The Operational
Transconductance Amplifier (OTA) circuit converts the difference between its two input
voltages to a corresponding current at its output. This circuit has been extensively used
in the implementation of various neuromorphic devices (Liu et al. 2002, Rachmuth and
Poon 2008, Rachmuth et al. 2011, Mayr et al. 2010). (c) The DP sets a sigmoidal
relationship between differential input voltages and the currents flowing across each
of the two differential transistors. This is a useful behaviour for implementing similar
sigmoidal behaviour, observed in neural systems (Liu et al. 2002).
Page 92
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Figure 5.3. Synaptic potential (decay) circuit. (a) Synaptic potential (decay) circuit diagram.
(b) Synaptic potential module. The output of this module is a decaying signal, which
its time constant and amplitude are controlled by Itau and Iamp , respectively. The decay
starts once a pre/post spike arrives.
a number of PSTDP circuits including Bofill-I-Petit and Murray (2004), Azghadi et al.
(2012b), and Azghadi et al. (2013a). It can also be utilised in implementing many synap-
tic plasticity circuit, where there is a need for controllable decay dynamics. This circuit,
which acts as a leaky integrator controls both the amplitude of the generated potential
signal as well as its time constant.
There is another instance of the leaky integrator, in which only the time constant is con-
trollable and the required amplitude of the potentiation/depression may be realised
with another circuit/transistor. Two different arrangements of this leaky integrator are
shown in Fig. 5.4. In these circuits, the dawn of the signal determined by the arrival of
a spike, and the time constant is controlled by the voltage applied (Vtau ) to the gate of
a PMOS/NMOS transistor.
In addition to the above mentioned important building blocks for analog neural de-
signs, there are other essential circuits such as current mirrors, source-followers and
current-mode Winner Take All (WTA) circuits (Liu et al. 2002, Razavi 2002), which are
extensively used in neuromorphic analog designs including those reviewed in the next
Section.
Page 93
5.3 Neuromorphic Implementation of Synaptic Plasticity Rules
Figure 5.4. Leaky integrator circuit for producing required decay dynamics with adjustable
time constants. (a) Leaky integrator for driving a PMOS transistor. (b) Leaky inte-
grator for driving a NMOS transistor. (c) Leaky integrator module symbol.
One of the first VLSI designs utilising spike-based learning mechanisms, which is
closely related to many current designs for STDP, was proposed by Hafliger et al.
(1997), even before a large body of biological evidence for STDP was published. The
Page 94
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
learning part of the synapse circuit proposed in Hafliger et al. (1997) adjusts the synap-
tic weight stored on a capacitor, when a post-synaptic spike is received. The direction
and magnitude of the adjustment i.e. charging or discharging the capacitor, which re-
spectively corresponds to potentiation or depression of the synapse, is determined by
the post-synaptic pulse width as well as pre-synaptic spiking dynamics compared to
an adjustable threshold.
In 2000, another instance of a learning rule that functions based on both the occurrence
of the pre-synaptic spike, and the membrane potential of the post-synaptic neuron was
proposed by Fusi et al. (2000). This new learning rule is called Spike Driven Synaptic
Plasticity (SDSP). In the SDSP rule as described in Section 2.6.1, the dynamics of the
voltages produced in the neuron depends on the membrane potential, Vmem , of the
neuron. So the SDSP rule changes the synaptic weight according to the time of pre-
synaptic and membrane potential of the post-synaptic neuron. This membrane poten-
tial itself depends on the frequency of post-synaptic spikes generated by the neuron.
Figure 5.5 shows a brief view of the neuron and synapse structure implemented in
VLSI to realise the SDSP rule. Figure 5.5(a) shows schematic diagram of a VLSI learn-
ing neuron with an array of SDSP synapses. Multiple instances of synaptic circuits
output currents into the I&F neuron’s membrane capacitance (Indiveri et al. 2011). The
I&F neuron integrates the weighted sum of the currents and produces sequences of
spikes at the output.
Figure 5.5(b) shows that for implementing the SDSP synapse, a Differential Pair In-
tegrator (DPI) (Bartolozzi and Indiveri 2007) along with a bistability circuit, are the
main components, and the rest of the required components are only needed once per
neuron. This figure shows the synapse with pre-synaptic weight update module. An
AER asynchronous logic block receives input spikes and generates the pre and ∼ pre
(the inverse of pre) pulses. An amplifier in a positive-feedback configuration, forms a
bistability circuit that slowly drives the weight voltage VWi toward one of the two sta-
ble states Vwlow or Vwhi . The transistors driven by the pre and ∼ pre pulses, together
′ and V ′ signals, implement the weight update. The
with those controlled by the VUP DN
DPI block represents a current-mode low-pass filter circuit that generates an output
synaptic current Isyn with biologically plausible temporal dynamics. This current is
then sourced into the Vmem node of the I&F circuit.
In addition, Fig. 5.5(c) demonstrates the neuron with post-synaptic weight control
module. An I&F neuron circuit, integrates the input synaptic currents and produces a
Page 95
5.3 Neuromorphic Implementation of Synaptic Plasticity Rules
spike train at the output. A DPI filter generates the VCa signal, encoding the neuron’s
mean firing rate. Voltage and current comparator circuits determine whether to update
the synaptic weights of the afferent synapses, by increasing/decreasing their values.
Figure 5.5. Implementation of the SDSP learning rule. (a) Schematic diagram of a VLSI
learning neuron with an array of SDSP synapses. (b) Synapse with pre-synaptic weight
update module. (c) Neuron with post-synaptic weight control module.
Spike-based learning circuits (Fusi et al. 2000, Mitra et al. 2009, Giulioni et al. 2009) alter
the synaptic efficacy on pre-synaptic or post-synaptic spike arrival. They do not take
into account the exact timing difference between pre-synaptic and post-synaptic spikes
to induce synaptic weight changes. Therefore, spike-timing based learning circuits can
be categorised as another type of synaptic plasticity learning circuits.
Page 96
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Many spike-timing based circuits have been implemented by different groups and un-
der various design strategies in VLSI (Bofill-I-Petit and Murray 2004, Cameron et al.
2005, Indiveri et al. 2006, Koickal et al. 2007, Tanaka et al. 2009, Azghadi et al. 2011c, Bam-
ford et al. 2012b, Azghadi et al. 2012b, Azghadi et al. 2012a, Azghadi et al. 2013a). These
circuits are classified into several design categories, and are reviewed in different sub-
sections as follows.
One of the first designs for PSTDP, which is the conventional form of timing-dependent
plasticity, was first proposed by Bofill et al. (2001). In this design a transistor that oper-
ates in its subthreshold (weak inversion) region is utilised to control the amount of cur-
rent flowing into/out of the synaptic weight capacitor. The direction of the changes in
the voltage of the weight capacitor is determined by another circuit that generates the
required signals according to the timing differences of pre- and post-synaptic spikes.
This circuit utilises the same peak voltage, time constants and decays for both potentia-
tion and depression dynamics required in the PSTDP rule (shown in Eq. 2.2), however,
these values might be required to be different in various contexts. As a result, Bofill-
I-Petit and Murray (2004) presented a modified version of their original design and
proposed a new circuit, where potentiation and depression dynamics have their own
decay constants as well as peak values. In their design, they have employed additional
circuitry to form the required pre- and post-synaptic pulses for their PSTDP circuit. By
adding a few transistors to the main circuit required for potentiation and depression,
they also made their circuit capable of weight dependent synaptic weight modification.
Fig. 5.6(a) shows a version of the STDP circuit presented in Bofill-I-Petit and Murray
(2004). In this design, two transistors (Mp and Md) that operate in their subthresh-
old (weak inversion) region are utilised to control the amount of current flowing in-
to/out of the synaptic weight capacitor, CW . The voltages that control these transistors
are Vpot (potential for potentiation) and Vdep (potential for depression). These poten-
tials produced by two instances of the synaptic potential circuit presented in Fig. 5.3.
This design uses currents for controlling circuit bias parameters that correspond to the
PSTDP learning rule parameters presented in Eq. 2.2. Simulation results for generating
STDP learning window using this circuit are also presented in Fig. 5.6(b). This figure
demonstrates the exponential decay behaviour in the learning window, which is in
Page 97
5.3 Neuromorphic Implementation of Synaptic Plasticity Rules
Figure 5.6. Pair-based STDP circuit with synaptic potential blocks. (a) This PSTDP rule
circuit that is a modified version of the design proposed in Bofill-I-Petit and Murray
(2004) is presented in Azghadi et al. (2012b). (b) This is the exponential learning
window generated by Matlab and the PSTDP circuit under various process corners.
Similar protocols and time constants to Bi and Poo (1998) are employed.
accordance with the PSTDP rule formula presented in Eq. 2.2. This exponential be-
haviour is reached by biasing Mp and Md, in their subthreshold regions of operation.
Another well-known PSTDP circuit is the symmetric design proposed by Indiveri et al.
(2006). This circuit has two branches of transistors, as shown in Fig. 5.7(a). The upper
branch is responsible for charging the weight capacitor, if a pre-synaptic spike precedes
a post-synaptic one in a determined time, and the bottom branch is for discharging the
capacitor if the reverse spike order occurs, also within a predetermined time. The po-
tentiation and depression timings in this design are set by two leaky integrators, in
which their decays are set by two bias voltages, Vtp and Vtd , for potentiation and de-
pression time constants, respectively. In addition, the amplitude of the potentiation
Page 98
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Figure 5.7. Pair-based STDP circuit with leaky integrator blocks. (a) This is a different
representation of the PSTDP circuit presented in Indiveri et al. (2006). (b) The STDP
learning window generated by the PSTDP circuit shown in (a) for various potentiation
and depression time constants. A similar protocol to Bi and Poo (1998), which uses
60 pairs of pre- and post-synaptic spikes at a rate of 1 Hz, is employed. Note that
simulations are carried out in accelerated time, by a factor of 1000, compared to real
biological time. (c) A similar learning window was measured from the multi-neuron chip
presented in Indiveri et al. (2006) in biologically plausible time and under the PSTDP
experimental protocol utilised in Markram et al. (1997).
and depression are set by VA+ and VA− , respectively, and the pulse width that is an-
other important factor for the amplitude values, is kept fixed and equal to 1 µs in the
shown simulation in Fig. 5.7(b), which shows the STDP learning window in an acceler-
ated time scale. In addition, Fig. 5.7(c) shows chip measurement results for STDP learn-
ing window, in biologically plausible time (Indiveri et al. 2006, Azghadi et al. 2014c).
It is shown in a previous study that the circuit demonstrated in Fig. 5.7(a) can be min-
imised to decrease the number of transistors, and shrink the layout size for this cir-
cuit (Azghadi et al. 2011b). When considering the need for replicating this circuit in
every synapse in a large-scale neural network, this little area saving can be significant.
Page 99
5.3 Neuromorphic Implementation of Synaptic Plasticity Rules
The minimised circuit is shown in Fig. 5.8. Our simulation results show that this circuit
is able to reproduce a window similar to that of figure 5.7(b), that is generated by the
original PSTDP circuit shown in Fig. 5.7(a).
Figure 5.8. Minimised pair-based STDP circuit. This is the PSTDP circuit presented in
Azghadi et al. (2011b).
One of the other STDP designs that has been utilised in a VLSI spiking neural network
chip as part of the FACETS project is the design proposed by Schemmel et al. (2006).
In this design the STDP circuit that is local to each synapse has a symmetric structure.
The voltage potentials for potentiation or depression correspond to the quantity of
charge stored on synaptic capacitors, which are discharged by a fixed rate, determined
by a set of three diode connected transistors working in their subthreshold region of
operation to cause an exponential decay. These capacitors later determine the amount
of change in the synaptic weight corresponding to the time difference between the pre-
and post-synaptic spikes.
Another PSTDP circuit that was reported in 2006, is the design proposed by Arthur
and Boahen (2006). This symmetric analog design utilises a Static Random Access
Memory (SRAM) cell for storing a binary state of the synapse weight, which is either
high (potentiated) or low (depressed). This circuit uses leaky integrators in order to
implement the required dynamic for the plasticity, similar to the designs proposed in
Indiveri et al. (2006) and Azghadi et al. (2013a) and different from the design proposed
Page 100
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
in Schemmel et al. (2006). Upon the arrival of a spike, the plasticity potentials are gen-
erated. They start decaying linearly thereafter, and if a complementary spike arrives in
their decay intervals, its time difference with its complementary spike, determines the
required level of potentiation or depression.
All aforementioned subthreshold STDP circuits, except the circuit reported in Bofill-I-
Petit and Murray (2004), implement the simple conventional form of the STDP rule,
known as additive STDP, in which the current synaptic weight has no effect on the
synaptic plasticity mechanism (Song et al. 2000). Recently, a new analog VLSI de-
sign (Bamford et al. 2012b) for weight-dependent PSTDP (Kistler and Hemmen 2000,
Guetig et al. 2003) was proposed. This design exploits the MOS transistor physical
constraints and not any extra circuitry—as used in Bofill-I-Petit and Murray (2004)—
to implement a weight-dependent PSTDP model (Bamford et al. 2012b). The design
employs switched capacitors to implement the required leaky integrators needed for
potentiation and depression.
OTA-based Circuits
In addition to the aforementioned analog STDP designs, there are other analog im-
plementations of PSTDP in the literature that use the Operational Transconductance
Amplifier (OTA) as their main building block, and have been employed in specific ap-
plications. The circuit proposed by Koickal et al. (2007), implements PSTDP in a sym-
metric fashion and by employing four instances of OTAs. Similar to the previously
mentioned STDP designs, for generating the potentials for plasticity, there is a need
for leaky integrators. In all STDP designs mentioned so far, these integrators are built
using an RC network, where the resistor is implemented using the transistor’s chan-
nel resistance, which is function of the gate to source voltage of the transistor. Here
however, OTAs have been used as active resistors to generate the required long time
constants for STDP, which are essential in their olfaction chip. The resistors are formed
by having a feedback from the output of the OTA to its inverting input and a high
resistance is reached by reducing the transconductance of the OTAs.
Another STDP circuit that utilises OTAs for implementing PSTDP, is the design pro-
posed by Tanaka et al. (2009). They proposed two OTA-based circuits for implementing
symmetric and asymmetric STDP. Tanaka’s design uses the resistance of the transistor
channels instead of OTAs for generating the required time constants needed by the
STDP rule, however in a similar manner to Koickal et al. (2007) it still employs two
Page 101
5.3 Neuromorphic Implementation of Synaptic Plasticity Rules
OTAs for charging and discharging the weight capacitor in case of LTP and LTD, re-
spectively. Recently a low power design for synapses with STDP was proposed in
Cruz-Albrecht et al. (2012). This design utilises OTAs and have transistors all biased
in the deep subthreshold, while transistors have high threshold voltages, above the
circuit power supply, to build low power neuron and synapses. The synapse with
PSTDP weight update ability, is implemented using five OTAs from which three have
enable/disable capability for disabling the synapse when no weight change is required
(Cruz-Albrecht et al. 2012) and therefore having only leakage currents in the order of
100 fA, hence lower power consumption. The three OTAs also utilise a source degener-
ation technique to increase the linearity and dynamic range. The other two OTAs used
in the synapse circuit are simple OTA circuits that consist of a differential pair and a
set of current mirrors to generate the respective output current. The output current in
both types of OTAs employed in this design are governed by the tail current of the dif-
ferential pair, which itself is generated by a set of diode connected transistors stacked
on top of each other to form the bias circuitry (Cruz-Albrecht et al. 2012).
The Floating Gate (FG) technology is one of the popular approaches for implementing
STDP-based learning mechanism that updates the charge on floating gates according
to the STDP rule. By utilising FG, one can implement a spike-based learning rule such
as STDP using only one transistor and some extra circuitry which can be shared for
several Single Transistor Learning System (STLS) as it is named in this context (Gordon
and Hasler 2002, Ramakrishnan et al. 2011). The other design approach with long-term
weight storage capability, is the device-based design proposed by Zhang et al. (2010).
They have proposed an ionic/Si hybrid device along with some peripheral circuitry to
implement a Neural Phase Shifter (NPS), which represents the synaptic weight change
as changes in the phase shift of their proposed NPS versus the time differences between
pre- and post-synaptic spikes.
Digital Circuits
In addition to the above mentioned analog implementations of the STDP rule, this rule
has been implemented several times using digital circuitries (Cassidy et al. 2007, Bel-
hadj et al. 2009, Cassidy et al. 2011, Soleimani et al. 2012). Cassidy et al. first proposed
Page 102
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
a FPGA implementation of a SNN that was successfully used in a few neural experi-
ments including the demonstration of bimodal weight distribution behaviour (see Sec-
tion 4.2.1 for a similar example on the IFMEM device) as a feature of synapses with
STDP learning (Song et al. 2000, Cassidy et al. 2007). In 2009, neuromorphic researchers
proposed three various approaches to implement STDP on FPGA and tested them in
a SNN composed of analog neurons and STDP synapses that transfers their weight to
the neurons through a pulse width modulation (PWM) technique (Belhadj et al. 2009).
Cassidy and his colleagues again in 2011 proposed some other digital architectures for
implementing STDP on FPGA that outperform the previous FPGA designs proposed
in Belhadj et al. (2009), in terms of FPGA power and area consumption (Cassidy et al.
2011). Apart from these FPGA-based digital implementations of STDP, there are some
general digital neuromorphic architectures, where a SNN can be easily configured with
required learning strategy e.g. STDP. Researchers at IBM have reported implementa-
tion of a neurosynaptic core that is composed of digital Integrate and Fire (IF) neurons,
addressable axons, and a crossbar arrays of SRAMs that are programmable and act as
binary synapses. This design has been implemented in a 45 nm Silicon-On-Insulator
(SOI) process and consumes little power (Merolla et al. 2011, Arthur et al. 2012).
In recent years, along side the IBM digital neuromorphic architectures another neu-
romorphic architecture inspired by the parallelism of the human brain has also been
developed and tested (Furber et al. 2013, Painkras et al. 2013). This neuromorphic ar-
chitecture, called Spinnaker, is a hardware platform that utilises off-the-shelf ARM
processors to form a biomimetic massively parallel simulator for spiking neural net-
works.
Analog/Digital Circuits
Another popular approach to implement SNNs including required neural and synap-
tic dynamics, is the mixed signal VLSI design strategy that involves analog circuits
along with digital ones that are integrated to implement the required network com-
bination (Pfeil et al. 2012). For example, the neuromorphic architecture presented in
Moradi and Indiveri (2014) follows mixed signal design strategies in which synapses
are analog circuits (see Fig. 3.3), but with digital weights, stored on an asynchronous
SRAM block (Moradi and Indiveri 2011, Moradi and Indiveri 2014), that can be up-
dated off-chip and according to any expected plasticity rule, including timing- and
Page 103
5.3 Neuromorphic Implementation of Synaptic Plasticity Rules
rate-based rules e.g. STDP and BCM (Azghadi et al. 2014b). Experimental results pre-
sented in Azghadi et al. (2013d) show how this programmable neuromorphic design
can be tuned to exhibit biologically plausible response properties.
In addition to the spike-based and spike-timing based synaptic plasticity circuits men-
tioned so far, there is another circuitry proposed by Mayr et al. (2010) that uses a hy-
brid learning rule composed of both timing and rate of spikes to alter the synaptic
weight. This phenomenological rule was already introduced in Section 2.6.1. For gen-
erating the required exponential decay dynamics for both u(t) and g(t) (see Eq. 2.17),
an OTA-based design approach has been utilised. Similar to the PSTDP design re-
ported in Koickal et al. (2007), this design exploits a balanced OTA with negative feed-
back, which acts as a large resistor in the required leaky integrators. However, this
design uses an active source degeneration topology to further improve the dynamic
range and linearity of the integrator. These integrators are needed for both membrane
potential, u(t), which decays linearly back to the resting potential after a post-synaptic
pulse duration is finished (hyperpolarization dynamic), as well as the post-synaptic
current, g(t), which decays exponentially toward zero after a pre-synaptic spike ar-
rival. The time constants in these integrators can be tuned by changing the resistance
of the OTA-based resistor that in turn can be altered by calibrating the leak bias current
in the design. Beside these exponential decay dynamics, the rule needs subtraction and
multiplication for changing the synaptic weight. These functions are approximated us-
ing a differential pair and its tail current (Mayr et al. 2010).
Unlike all aforementioned neuromorphic VLSI designs of synaptic plasticity rules that
are based on a rough approximation of the mechanisms that occur in the neuron and
synapses, there are a few designs in the literature that go further and implement a
detailed account of synaptic plasticity mechanism including the whole chemical and
ionic interactions in the synaptic cleft. These designs are much more complex than the
previous timing-, spike- or BCM-based designs, since they take into account detailed
biophysical interactions when inducing synaptic weight changes.
Page 104
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
The two major biophysical designs available in the literature, that are able to demon-
strate both PSTDP and BCM-like behaviour are the designs proposed in Meng et al.
(2011) and Rachmuth et al. (2011). These designs are based on two similar biophys-
ical rules that were already discussed briefly in Section 2.6.2. The first design im-
plements an elaborate biophysical synaptic plasticity model, which is based on the
general biophysical processes taking place in the synapse (see Section 2.6.2). This im-
plementation utilises current-mode design technique in order to implement the tar-
geted biophysical rule that describes the detailed dynamic of the synaptic ion chan-
nels (Meng et al. 2011, Rachmuth and Poon 2008).
Recently the same group has published another iono-neuromorphic VLSI design which
explores a similar approach for implementing both Spike Rate Dependent Plasticity
(SRDP) and STDP using a unique biophysical synaptic plasticity model as briefly ex-
plained in Section 2.6.2. Identical to their first design, they used current-mode design
technique to implement the required ion channel dynamics (Rachmuth et al. 2011).
In the spiking neural network software approach, the required neural architecture
and its basic building blocks are implemented on conventional Von Neumann com-
puting architectures. The main advantages of this method are, shorter design time
compared to the physical implementation of neuromorphic systems, as well as the re-
configurability they offer compared to the non-reconfigurable design in most of phys-
ical design approaches. Furthermore, the input/outputs to software-based designs
are technically noise-free and easy to deal with, compared to analog/digital hardware
Page 105
5.4 Challenges in Neuromorphic Engineering
outputs that are not so easy to use. On the other hand though, as Poon and Zhou
(2011) state, the software-based approaches have significant shortcomings such as very
high power consumption and very large real estate, which are major obstacles in the
way of implementing a brain-scale neuromorphic system. An example of a software-
based neuromorphic system implemented on digital machines, is the realisation of
cat cortex on an IBM’s super computer that took 147,456 CPUs and 144 TB of mem-
ory (Ananthanarayanan et al. 2009). Although this supercomputer is much faster than
the brain in performing computation, since the computation takes place in a sequen-
tial order, compared to the massively parallel but slow computations that take place
in the brain, it is still far behind the brain parallel processing capability. That is the
situation, where a hardware-based approach becomes superior because of its parallel
nature, similar to that of the brain.
Page 106
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
affects on digital systems. Interestingly, the challenges and constraints faced by neuro-
morphic engineers for implementing synaptic learning, are similar to the ones encoun-
tered in biological learning, such as lack of long-time weight storage (Lisman 1985,
O’Connor et al. 2005) and limited wiring. Generally, the main challenges and obsta-
cles for implementing a large scale neuromorphic system can be summarised by the
following:
We know that the brain consists of billions of neurons and trillions of synapses, each
of which consumes much less power compared to their silicon counterparts (Poon
and Zhou 2011). Recently new integrated neural circuitry with a low power struc-
ture was proposed that consumes even less power per spike compared to a biological
neuron (Cruz-Albrecht et al. 2012). Although it is a large step toward having a low
power spiking neural system, it is very naive to think we are close to a neural system
with a power consumption close to the brain, since this work does not consider the
interconnection and communication among spikes and its required power. It also does
not take into account the required complexity in the neural and synaptic structures,
which is sometimes necessary for specific applications. In addition, the power con-
sumption of a synapse or neuron heavily depends on its model parameters and their
values that can change the weight modification pattern and at the same time lead to
high or low power consumption. The other fact that should be considered is the spike
pulse width utilised in the neuromorphic design, that can have significant affect on
both functionality and power consumption of the system (Azghadi et al. 2014a). There-
fore, an effective approach for decreasing the power consumption of a neuromorphic
system is to optimise the neuron and synapse circuit bias parameters, as well as the
structure of the spike pulses, in a way that while having the required functionality
minimises the power consumption (Azghadi et al. 2014a).
One of the other effective ways for implementing low-power neuromorphic circuits is
to design them using transistors operating deep in the subthreshold domain, which
results in extremely low currents (e.g., in the order of pA). In addition, this approach
allows the design of circuits that operate with low supply voltages, which is another
strategy for reducing power consumption (Cruz-Albrecht et al. 2012). However, oper-
ating in the subthreshold region of operation and using lower supply voltages result
in greater susceptibility to process variation and noise.
Page 107
5.4 Challenges in Neuromorphic Engineering
Post-Fabrication Calibration
Calibration of neural components that adjusts the circuit bias parameters, from those
used in the design procedure, compensates for the variations due to the fabrication
process. Although this approach can be used in small-scale neuromorphic circuits
(Grassia et al. 2011, Azghadi et al. 2013a, Carlson et al. 2013, Carlson et al. 2014), and
even shared biases might be useful for larger scale neuromorphic systems, for very
large-scale neuromorphic systems with millions of neural circuits, this approach is im-
practicable (Poon and Zhou 2011). Furthermore, the calibration technique cannot ac-
count for detrimental effects of temperature and supply voltage variations, which are
other sources of variations discussed later in this Section.
The use of analog circuit design techniques that minimise the effect of process varia-
tions and device mismatch is another approach used for tackling fabrication imperfec-
tions (Poon and Zhou 2011, Rachmuth et al. 2011). This design approach has been suc-
cessfully utilised in an iono-neuromorphic design presented in Rachmuth et al. (2011).
Page 108
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Living with the variation and considering a mechanistic way to reduce its effect on
the expected behaviour of the network is another instance of the approaches taken
against variation. For example, Neftci and Indiveri have proposed an approach for
compensating device mismatch by exploiting Address Event Representation (AER) in
multi-neuron chips (Neftci and Indiveri 2010), which is freely available in the network
and does not require adding extra circuitry or processing as mentioned for the first and
second classical approaches. Another approach that exploits learning and adaptation
in SNNs, is to utilise short term (Bill et al. 2010) or long term plasticity to alleviate pro-
cess variation effects. For example, Cameron et al. exploit a PSTDP circuit to reduce
the effect of process variation (Cameron et al. 2005). A similar approach that employs
PSTDP in order to alleviate variations in performance of similar neurons—due to pro-
cess variation and mismatch—in a SNN is also proposed in Bamford et al. (2012a). All
previous approaches have used some techniques to reduce the effect of process vari-
ations, however, some neuromorphic engineers interestingly exploited variations in
their neural circuit structure.
Sheik et al. have shown that the unwanted device mismatch and variations can be
exploited in neuromorphic systems to implement axonal delays, which are required
Page 109
5.4 Challenges in Neuromorphic Engineering
and useful for neural computations (Sheik et al. 2012a). In another work, neuromorphic
engineers exploited mismatch to perform vector-matrix multiplication in a extreme
learning machine (Yao et al. 2013). Despite all these methods, one may keep in mind
that using the process variation and mismatch such as the idea presented in Sheik et al.
(2012a), or trying to reduce the effect of variations and mismatch is challenging due to
mismatch inherent stochasticity.
All the mentioned approaches so far are only proposed to tackle the process varia-
tion and device mismatch. Although this is the most significant source of variation in
analog neuromorphic models of synaptic plasticity rules, other types of variations, i.e.
supply voltage and temperature variations must also be considered, especially when
a large-scale neuromorphic system is targeted. The following subsection discusses the
effect of variations on the design of a large-scale neuromorphic system in more details.
Note that VLSI chips are susceptible to variations in many parameters and behave
differently under various cases of variations. These include (i) fabrication process pa-
rameter variation such as deviation in the threshold voltage and channel length from
the exact value specified for the fabrication process (this issue was discussed in de-
tails in the previous subsection); (ii) supply voltage variations from the ideal supply
voltage required for the device operation; and (iii) temperature variations from the
ideal temperature needed for the normal operation of the devices and interconnects.
Each of these variation sources has its own corners, beyond those the device might not
function properly. An instance of these corners are shown for process parameters in
Fig. 5.6(b), where the STDP learning window is produced for various device corners,
and fortunately for all of them it is close to the required behaviour. Since there are
three simultaneous sources (i.e. PVT) of variations in an analog VLSI system, these
variations should be coupled together in order to form various PVT variation corners,
in which the device has its best, typical or worst characteristic. Apart from devices,
variation also affects the characteristics of the interconnects, that have their own cor-
ners, which are usually different from those of the device. More directly, the device and
interconnects potentially have worst performance at different corners. Considering
these corners when designing the targeted application is essential, as the design might
be dominated by device corners, interconnect corners, or a mixture of both (Weste and
Harris 2005).
Page 110
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Page 111
5.4 Challenges in Neuromorphic Engineering
potentials generated by the silicon neurons and share wires, creating therefore “virtual
axons”. The most common protocol that is used in the neuromorphic community to
accomplish this is based on the Address Event Representation (AER) (Deiss et al. 1994,
Boahen 2000). In this representation, the action potentials generated by a particular
neuron are transformed into a digital address that identifies the source neuron, and
broadcast asynchronously on a common data bus. By using asynchronous arbiters and
routing circuits (Moradi et al. 2013) it is therefore possible to create large-scale neural
networks with reconfigurable network topologies. These networks can be distributed
within the same chip e.g. among multiple neural cores (Arthur et al. 2012), or across
multiple chips (Merolla et al. 2013, Chicca et al. 2007).
Although there are a number of neuromorphic systems that deal with a relatively high
number of analog neurons, designing large-scale neuromorphic systems is still a very
complex task. One of the major obstacles on the way is the lack of an Electronic Design
Automation (EDA) tool, that can facilitate the design procedure, while taking into ac-
count the targeted design requirement. There are promising recent accomplishments
that exploit existing EDA tool-chains for automating the design of neuromorphic cir-
cuits. For examples see Imam et al. (2012) and Mostafa et al. (2013) for designing the
asynchronous logic circuits that make up the arbiters and routers described above.
However there is a need for a new generation of EDA tools that are optimised for
neuromorphic architectures with hybrid analog/digital circuits, asynchronous logic
circuits, and networks characterised by very large fan-in and fan-out topologies.
The complex behaviour of neural circuitries including neurons and synapses are con-
trolled by many parameters including synapse potentiation and depression time con-
stants and amplitudes, neuron spiking thresholds, spiking frequency adaptation, and
refractory period parameters. For controlling silicon neurons and synapses, these pa-
rameters should be presented as small-scale and high accuracy voltages and currents
to silicon neurons and synapses. Generating these bias voltages and currents, which
usually span over a wide range, usually needs a specific dedicated VLSI circuit that
Page 112
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
The challenges mentioned in the previous subsections are engaged with typical neuro-
morphic systems and are not specific to synaptic plasticity circuits. However, a specific
challenge on the way of implementing required synaptic plasticity rules and integrat-
ing them into network of neurons, is the synaptic weight storage method, which is
discussed in more details in the following subsection.
As there is a direct relationship between the stability of the voltage stored on a capac-
itor and its capacity, some previous STDP designs have used bulky capacitors (Bofill-
I-Petit and Murray 2004, Bamford et al. 2012b, Cruz-Albrecht et al. 2012, Azghadi et al.
2013a), which takes up a large portion of the precious silicon real estate, in order to
preserve the weight value at least for the period of time required for their desired
experiments. This does not appear compatible with the goal of neuromorphic engi-
neering, which ultimately aims to integrate a large scale neural system with billions of
synaptic circuits. Therefore a number of other approaches have been sought, in order
Page 113
5.4 Challenges in Neuromorphic Engineering
to address this obstacle on the way of realising long-term plasticity in silicon. These
approaches are briefly reviewed as follows.
Since the main reason for instability of the voltage stored on a weight capacitor is the
leakage current, one may increase the stability by reducing the leakage current. Us-
ing reverse-biased transistors in the path of charging/discharging weight capacitors,
reduces leakage currents and therefore increases the weight stability on that capaci-
tor. This approach was first proposed by Linares-Barranco and Serrano-Gotarredona
(2003). Recently, it has been successfully exploited in Bamford et al. (2012b) for stor-
ing the weight for a longer period of time in the order of hundreds of milliseconds.
In order to reverse bias the transistors in the design, as it is proposed in Bamford et al.
(2012b), the Gnd and Vdd are shifted few hundreds of millivolts toward Vdd and Gnd,
respectively. By reducing the supply voltage slightly or increasing the ground voltage
level, the transistor back gate will be in both cases at higher voltages, resulting in an
increase in the threshold voltage that leads to reduced leakage current.
Page 114
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Bistability Mechanism
Another approach for synaptic weight stabilisation, which has been used in a number
of synaptic plasticity circuits and for various learning rules, is a bistability mechanism
that is based on the idea of having the long-term state of a synapse either potentiated
or depressed. As shown in Fig. 5.5(b), in this approach, an amplifier with positive
feedback is utilised to drive the synaptic weight stored on the weight capacitor and
updated by the desired synaptic plasticity rule in the short-term, slowly either upward
or downward depending on the current value of the synaptic weight that is above
or below a predetermined threshold (Chicca et al. 2003, Indiveri et al. 2006, Mitra et al.
2009). Furthermore, some other neuromorphic circuits use the same approach but with
storing the weight on a SRAM i.e. only two stable states (Arthur and Boahen 2006), or
mapping and storing the modified weight on a multi-stage analog memory (Hafliger
and Kolle Riis 2003, Hafliger 2007). Using this method, the synaptic weight stored on
the capacitor is updated whenever there are spikes, but as soon as there is no spike, the
weight is driven toward a high or low analog value, depending on the current synaptic
weight on the capacitor, which can be either potentiated, above a certain threshold, or
depressed, below that threshold, respectively.
Bistability mechanism has experimental support, as well as benefits over the use of
large weight capacitors, in large neuromorphic systems. Considering a large network
of neurons and synapses, on long time scales, synaptic efficacy can be assumed to have
only two high (potentiated) or low (depressed) values. This assumption is compatible
with experimental data (Bliss and Collingridge 1993, Petersen et al. 1998). In addition,
from a theoretical perspective, it has been argued that the performance of associative
networks is not necessarily degraded if the dynamic range of the synaptic efficacy
is restricted even into two stable states (Amit and Fusi 1994). Furthermore, bistable
synapses can be implemented in a small area compared to having large-scale capacitors
for preserving the synaptic weights for longer periods of time (Indiveri et al. 2006). Due
to these benefits, this technique is a suitable approach to be used in all of our reviewed
synaptic plasticity circuits including the STDP and TSTDP circuits.
Despite the usefulness of the bistability mechanism for short term learning and weight
changes, for permanent storage of synaptic weights, which are quantised to two high-
/low states using the bistable technique, there is a need for a non-volatile storage tech-
nique. A number of these storage techniques have been discussed in the following.
Page 115
5.4 Challenges in Neuromorphic Engineering
This approach has been followed in a few ways. In one of the pioneering works on
neural networks presented in 1989, the approach was to serially and periodically re-
fresh the analog weights stored on the capacitor with the weight stored in the memory
(Eberhardt et al. 1989). This approach however needs bulky Digital-to-Analog Convert-
ers (DACs) and Analog-to-Digital Converters (ADC). In addition, designing these de-
vices in the subthreshold low-current regime is a crucial task, as there is a high demand
for low power consumption and small silicon area in neuromorphic applications (Poon
and Zhou 2011). Moradi and Indiveri have used a single current-mode DAC, available
beside each synapse integrated circuit, in order to convert 5-bit digitally stored synap-
tic weights in asynchronous SRAMs, to a current that drives the synapse integrator
(Moradi and Indiveri 2011, Moradi and Indiveri 2014). Therefore, the synaptic weights
here are considered as virtual synapses and their weights come into effect whenever
they receive a spike from the AER system (Moradi and Indiveri 2011, Moradi and
Indiveri 2014). This approach utilises a time multiplexing technique and therefore only
uses one DAC per several synapse memory cell. Whilst this saves silicon area, it causes
extra communication. A similar approach of using virtual synapses with digitised
synaptic weights, has been employed by other neuromorphic engineers, in order to
tackle both synaptic weight storage and also reduce area usage (Vogelstein et al. 2007b).
In Pfeil et al. (2012), the authors discuss the issue of digitising weight on the PSTDP rule
and show that considering other constraints of neuromorphic designs, increasing the
weight storage resolution is not necessarily useful for PSTDP.
Floating Gate technology (FG) is another possible approach for nonvolatile storage of
synaptic weights. It has been exploited extensively in neuromorphic systems to im-
plement Hebbian-based and STDP learning rules (Ramakrishnan et al. 2011, Gordon
and Hasler 2002, Holler et al. 1989). As already mentioned in Section 5.3.2, this storage
technique leads to a compact single transistor implementation of STDP, which saves
significant silicon area. However, the drawback for this approach is the severe mis-
match that occurs in the tunnelling and/or injection processes. Also the requirement
for a special purpose CMOS process and extra control circuitry are other drawbacks of
this approach.
Page 116
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
Memristor
The memristor as the fourth circuit element (Chua 2011) possesses invaluable charac-
teristics including non-volatility, low power, and high density (Strukov et al. 2008, For-
tuna et al. 2009, Eshraghian 2010, Eshraghian et al. 2012) which are the features have
always being sought for implementing large scale neuromorphic systems. Therefore,
memristors may be a possible solution for solving the problem of synaptic weight stor-
age (Jo et al. 2010, Zamarreño-Ramos et al. 2011, Pershin and Di Ventra 2012, Wang et al.
2014b, Sheri et al. 2014). It can also be integrated with CMOS (Eshraghian et al. 2011)
in order to form a non-volatile synapse circuit (Jo et al. 2010, Indiveri et al. 2013). These
hybrid CMOS/memristor synapse circuits then can be utilised to implement both com-
putational and detailed biophysical synaptic plasticity learning rules that are quite use-
ful for neural computation (Afifi et al. 2009, Ebong and Mazumder 2012, Azghadi et al.
2013d). Although the memristor has significant strengths, its weaknesses such as need
for read operation, the accuracy of device programming, the device yield, and its vari-
ation and non-linearity should be considered as well.
5.5 Discussion
In this chapter, after a brief review of some basic circuit building blocks useful for
synaptic plasticity models, the design and implementation of those models in VLSI
were discussed under various design strategies. Then, the main challenges neuro-
morphic engineers encounter when designing neuromorphic systems were presented.
Furthermore, several methods to address those challenges were discussed. In this Sec-
tion, first the goals for VLSI implementation of various synaptic plasticity rules are dis-
cussed, and then the performance of the reviewed VLSI designs in reaching these goals
as well as an efficient implementation are reviewed. Understanding these goals and
challenges, as well as the effective and useful strategies for designing various synaptic
plasticity rules is essential while new designs for other synaptic plasticity models are
proposed and discussed in the remainder of this thesis.
Shouval has argued in a recent article What is the appropriate description level for synaptic
plasticity? (Shouval 2011). He mentioned that the suitable and required level of descrip-
tion when considering synaptic plasticity is not yet known, but it essentially depends
on what we attempt to reach. He stated that in some sense a simple plasticity rule such
Page 117
5.5 Discussion
as TSTDP theory of synaptic plasticity might provide a sufficient level of plasticity de-
scription, if only reproducing a set of experiments, including higher order phenomena,
is needed. However, he also indicated that further research is needed to understand
what is the required level of synaptic plasticity description and how detailed the bio-
physical synaptic plasticity models ought to be for various applications.
The same argument holds for the application domain of synaptic plasticity rules, spe-
cially those that are implemented in silicon. Before designing any synaptic plasticity
rule in VLSI, one should first consider, what are the goals, application area, and needs
for learning. Then, they should find the required level of synaptic plasticity descrip-
tion, ideal for their applications and proportionate to their other design requirements,
such as area and power consumption. At this step the designers must set a trade-off
between various design aspects, such as the required complexity of the network and
synaptic plasticity mechanism, the needed level of weight storage precision, and the
limitations in power consumption and silicon area. It is at this stage that the designer
selects an appropriate synaptic plasticity rule from the continuum of rules, depending
on the design specification and required synaptic details. Some therefore might choose
simple phenomenological rules such as STDP (Azghadi et al. 2013a), while other opt
for more detailed biophysical rules (Rachmuth et al. 2011).
Many of the mentioned designs in Section 5.3, have the simple research goal to imple-
ment, verify and understand desired synaptic plasticity rules such as pair-based STDP
(Indiveri et al. 2006), triplet-based STDP (Azghadi et al. 2013a), BCM (Azghadi et al.
2012a, Azghadi et al. 2013a), BCM-like LCP (Mayr et al. 2010), and biophysical ion
channel-based plasticity models (Rachmuth et al. 2011, Meng et al. 2011). In order to un-
derstand and verify a synaptic plasticity rule, reproducing biological experiments and
replicating the experimental data using various models (VLSI circuits) is quite help-
ful. For instance, many of the VLSI implementation of STDP rules (Bofill-I-Petit and
Murray 2004, Indiveri et al. 2006, Schemmel et al. 2006, Tanaka et al. 2009, Koickal et al.
2007, Bamford et al. 2012b, Azghadi et al. 2013a) have shown to be able to reproduce
the STDP learning window, which represents both potentiation and depression as they
occur in biology (Bi and Poo 1998, Markram et al. 1997). However, the way these cir-
cuits reproduce the window and other experimental results and the performance of
the circuits are different.
If a plasticity circuit can show a close match to experimental data, therefore it can
be of great help for neuroscience studies. Typically, the VLSI designs for conventional
Page 118
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
PSTDP rule are able to reproduce the STDP learning window (Bofill-I-Petit and Murray
2004, Indiveri et al. 2006, Schemmel et al. 2006, Tanaka et al. 2009, Koickal et al. 2007,
Bamford et al. 2012b). They are also able to show BCM-like characteristics (Azghadi et al.
2012a), similar to the BCM behaviours that were reproduced using computational
STDP models (Pfister and Gerstner 2006, Izhikevich 2003). However, the PSTDP rule
circuits are not able to reproduce more complicated experimental data. Therefore,
modified and extended version of these circuits, which are based on updated STDP
computational models such as Triplet-based STDP (TSTDP) learning model (Pfister
and Gerstner 2006), will be needed.
Besides these synaptic plasticity models and VLSI designs, sometimes it is desirable to
replicate the dynamics of synapses in detail, and study the effect of neuromodulators
and intracellular signalling. In these cases, all the ion channels and synapse dynamics
should be taken into account in the targeted VLSI implementation. These detailed
implementations are potentially useful also for various neuroprosthesis, neural-based
control, and machine learning tasks (Rachmuth et al. 2011).
Apart from the level of complexity and the required level of synaptic details, there are
some other factors that affect the complexity, hence the area and power consumption of
neuromorphic synaptic plasticity rules in VLSI. One of these factors is the weight stor-
age techniques (see Section 5.4.8) and the required level of precision when storing the
weight. This approach is also related to the targeted application and the goal of the im-
plemented VLSI synaptic plasticity rule (Pfeil et al. 2012). In general, designers should
always consider the trade-off between the level of synaptic weight precision they re-
quire for their target applications and the resources they can afford. For instance, neu-
romorphic engineers have shown that for synchrony detection application in spiking
neural networks, a 4-bit synaptic weight precision is viable (Pfeil et al. 2012). On the
other hand, a number of other studies show that if a suitable learning mechanism is
used, then even binary weight precision controlled by the bistability mechanism is suf-
ficient for the task of classifying complex rate-based patterns (Mitra et al. 2009).
Table 5.1 summarises the key properties of some neuromorphic systems for learning
and synaptic plasticity applications. Note that the estimated area and power con-
sumption data in this Table only reflect the reported data in the related papers. These
numbers are dependent on many parameters including the synaptic plasticity rule im-
plemented, the synaptic plasticity parameters, the weight storage techniques, and the
network stimulation patterns and protocols. Since in some papers, the exact power
Page 119
5.5 Discussion
consumption and area requirement of the synapse is not available, the total power and
area of the chip are divided by the number of synapses and neurons on the chip, to
calculate a rough value of the size and power requirement of the synapse. Also, note
that the calculated estimated area for each synapse, encompasses both the synapse cir-
cuit as well as its synaptic plasticity circuit, which may be reported or implemented
separately in the related papers.
The feedforward network with STDP learning presented in Bofill-I-Petit and Murray
(2004) successfully implements the targeted synchrony detection, but it consumes sig-
nificant power and occupies a large area. The high power consumption is due to power
hungry biasing current distribution network designed to minimise mismatch between
synapses. In addition, the area of the designed STDP circuit is significantly large due
to huge capacitors of the order of several pFs.
The implementation of an array of neurons with bistable STDP synapses (Indiveri et al.
2006), is the next design that has better power and size performance compared to the
first mentioned design (Bofill-I-Petit and Murray 2004). Furthermore, this neuromor-
phic system utilises the AER communication protocol and therefore is reconfigurable,
in contrary to the hard-wired network structure presented in Bofill-I-Petit and Murray
(2004). The next two neural networks with STDP synapses, mentioned in Table 5.1, are
also configurable. This feature helps to customise the neural network topology where
there is a need for various studies and applications, such as the designs Indiveri et al.
(2006) and Schemmel et al. (2006), which have been used to show STDP learning win-
dow, LTP and LTD characteristics. In terms of silicon real estate required for the STDP
circuit, the design in Schemmel et al. (2006) has a compact structure that occupies an
area of 50 µm2 for the STDP circuit and 100 µm2 for the synapse including STDP, DAC
and memory cell for storing the synaptic weight. Power consumption information for
this FACETS accelerated-time neuromorphic architecture is not listed. The neuromor-
phic learning chip presented in Arthur and Boahen (2006) also uses STDP and on-chip
SRAM cells to store a binary state of the synaptic weight updated by the STDP circuit.
Considering the number of neurons and synapses in this architecture and the overall
area of the chip presented in Arthur and Boahen (2006), which is 10 mm2 , this design
that has been used for learning patterns, also has a compact synapse size, on par with
the FACETS project chip (Schemmel et al. 2006).
The next design reviewed in Table 5.1 is an adaptive olfactory neural network with on-
chip STDP learning (Koickal et al. 2007). There is no power consumption information
Page 120
Table 5.1. Synaptic plasticity circuit comparison.
Chapter 5
Learning network (Neuron, Synapse) Estimated power (Supply) Estimated area (Design Tech.) Weight storage (Precision) Plasticity Application
Feedforward (Bofill-I-Petit and Murray 0.34 mW (5 V) 19902 µm2 (0.6 µm) Capacitor (Analog)† STDP Synchrony detection
2004) (5, 40)
(Re)configurable (Indiveri et al. 2006) Not available 4495 µm2 (0.8 µm) Capacitor (Bistable)‡ STDP General purpose
(32, 256)
(Re)configurable (Schemmel et al. Not available 100 µm2 (0.18 µm) SRAM (4 bits) STDP The FACETS project
2006)** (384, 98304)
(Re)configurable (Arthur and Boahen Not available 440 µm2 (0.25 µm) SRAM (1 bit) STDP Learning patterns
2006) (1024, 21504)
Adaptive olfactory (Koickal et al. 2007) Not available 72000 µm2 (0.6 µm) Capacitor (Analog)† STDP Odor classification
(9, 81)
Hopfield Feedback (Tanaka et al. 2009) 250 µW (3.3 V) 5000 µm2 (0.25 µm) Capacitor (Analog)† STDP Associative memory
(5, 10)
available in the paper. In addition, the exact area occupied by neurons and synapses
on the chip has not been reported. However, considering the die area of the fabricated
olfactory chip, the OTA-based synapse circuit with STDP occupies an area larger than
the area required for the design mentioned in Schemmel et al. (2006).
Comparing to all previously mentioned STDP-based learning circuits and systems, the
neuromorphic learning network presented in Seo et al. (2011), with 256 neurons and
64K synapses, that only consumes 8 nW of power and occupies roughly 13 µm2 per
synapse in the UVT chip, is the most efficient neuromorphic design. It is shown in
Seo et al. (2011) that this design can be configured for various cognitive tasks such as
pattern recognition and classification as well as associative memory.
All neuromorphic systems mentioned so far have used STDP as the learning mecha-
nism in their networks. However, as already mentioned, other synaptic plasticity rules
Page 122
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
have also been implemented and tested in neuromorphic systems for applications and
synaptic plasticity experiment replications. One of the first designs that used a differ-
ent rule than STDP for a classification task, was the design presented in Mitra et al.
(2009) that employs SDSP learning algorithm for synaptic plasticity. The area of this
design is comparable to the area required for the STDP learning rule, implemented
in previous designs. The authors have also shown the significant performance of the
implemented neural network with SDSP learning in classifying complex rate-based
patterns (Mitra et al. 2009).
Another neuromorphic system that implements a different synaptic plasticity rule rather
than STDP is the design presented in Mayr et al. (2010) and Mayr et al. (2013). This de-
sign implements a BCM-like voltage-dependent rule called LCP (See 2.6.1) to replicate
synaptic plasticity experiments beyond STDP such as triplet (Froemke and Dan 2002)
and frequency-dependent STDP (Sjöström et al. 2001). Considering the higher ability
in replicating synaptic plasticity experiments compared to STDP, this circuit has higher
complexity. However, the presented design in Mayr et al. (2013) is in par with most of
the STDP designs presented so far in both power and area requirements.
There are also a few biophysical VLSI neuromorphic designs available in the litera-
ture that take into account details of synaptic plasticity phenomena and implement
its underlying mechanism with a high degree of similarity to biological synapses, in
silicon (Meng et al. 2011, Rachmuth et al. 2011). This similarity results in the specific
ability of these synapses to account for both SRDP and STDP experiments and repli-
cate intracellular dynamics of the synapse, where simple previous synapses with STDP
fail. It also leads to large silicon area requirement for these circuits, while their reported
power consumption is reasonable comparing to most of the other VLSI synaptic plas-
ticity designs presented in Table 5.1.
In addition to the custom made hardware systems that opt to implement a specific type
of learning (synaptic plasticity) rule and use it in a specifically designed and structured
spiking neural network for an application or neuromorphic research, general neural ar-
chitectures, such as the Spinnaker (Furber et al. 2013) can be instructed, using software,
to implement any desired spiking neural network (whether simple or complex) with
any learning rule of choice. In Spinnaker system, the targeted neural network is nu-
merically simulated in multiple core processors and the synaptic weights are stored in
shared Dynamic Random Access Memory (DRAM). This neural architecture utilises
Page 123
5.6 Applications of Neuromorphic Circuits with Synaptic Plasticity
asynchronous design strategy for global routing in its design, so that the power con-
sumption of the design can potentially be improved. It also uses low-power micro-
processors and DRAMs to reduce the power consumption of the system. However,
implementing a specific synaptic plasticity rule in this general neural architecture con-
sumes more power than a typical custom VLSI design of that rule, due to its software
based approach.
Since working with live creatures and measuring experimental data from biological
sources is time-consuming and challenging, maybe one of the first applications for
a neuromorphic system that contains both neurons and synapses with any desired
synaptic plasticity rule, is its use for experimental neuroscientists. These scientists can
use a neuromorphic system, which acts according to a desired synaptic plasticity rule
and neural combination, and therefore experiment with various features and charac-
teristics in that system. For example, the biophysically inspired iono-neuromorphic
circuits proposed in Meng et al. (2011) and Rachmuth et al. (2011), provide useful in-
sight into how the calcium level alters in a real synapse.
Furthermore, since it is widely believed that synaptic plasticity underlies learning and
computational power in the brain (Gerstner and Kistler 2002, Shouval 2011), various
mechanisms that have direct or hypothetical relation to the synaptic plasticity experi-
ments, are being used as the learning part of a spiking neural network, to perform vari-
ous cognitive and machine learning tasks (Mitra et al. 2009, Oliveri et al. 2007, Masque-
lier and Thorpe 2010, Ebong and Mazumder 2012, Neftci et al. 2013, Schmuker et al.
2014, Khosla et al. 2014).
Page 124
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
It is known that, the spiking behaviour and the activity of the pre- and post-synaptic
neurons cause the synapses in the network to adapt themselves to these activities
(i.e. learn). These activities that are coded in the form of spikes, represent the input
to the network. It is therefore, absolutely essential to first have the correct spike coding
structure to effectively represent data to the neural network, and then it is critical to
adapt the synapses in a proper way, which is efficient for learning the current type of
inputs to the network. This means that the learning mechanism, i.e. the synaptic plas-
ticity rule, can heavily depend on the structure of input to the network, which in turn
depends on the application. Sometimes neuroscientists modify a rule or even combine
a number of rules to use them for their intended applications. It means that after a
careful study of the nature of the input and the required process on that to reach the
desired output, they decide on the structure of the learning method.
The study presented in D’Souza et al. (2010) shows an example of a learning method,
that couples STDP and Spike Frequency Adaptation (SFA) technique for updating
synaptic weights, to enable learning in a perceptron like structure. This work proposes
an effective platform for sensory guided processing, where two sources of auditory
and visual sensory inputs, result in changes in the perceptron neuron spiking activity.
It is shown that visual inputs can act as teacher in their used perceptron learning mech-
anism, while auditory inputs are used for updating the synaptic weights and learning
the input auditory patterns (D’Souza et al. 2010). Another example is the neuromorphic
architecture developed for object recognition and motion anticipation using a modified
version of STDP (Nere et al. 2012).
The previous three examples show that depending on the needs for the application,
and with mathematical and computational analysis, modification to synaptic plasticity
rules can be useful for performing tasks, which cannot be carried out with the simple
form of the plasticity rules such as STDP, SFA, and BCM. Therefore, the nature and
Page 125
5.6 Applications of Neuromorphic Circuits with Synaptic Plasticity
needs of an application and its inputs has a direct impact on the synaptic plasticity
mechanism and hence on its VLSI implementation.
Perhaps the conventional form of STDP, which is according to the formulation shown
in Song et al. (2000) and is in agreement with PSTDP experiments in Bi and Poo (1998),
is the most examined type of synaptic plasticity rule that has been exploited for learn-
ing, in various applications ranging from pattern recognition (Masquelier et al. 2008),
to dataset classification (Oliveri et al. 2007), and to topographic mapping formation
(Bamford et al. 2010). The pair-based STDP has been also utilised for many learning
tasks including receptive field development through cortical reorganisation (Young et al.
2007), motion anticipation (Nere et al. 2012), unsupervised learning of visual features
(Masquelier and Thorpe 2007), learning cross-modal spatial transformations (Davison
and Frégnac 2006), object recognition (Masquelier and Thorpe 2010), odour data clas-
sification (Hsieh and Tang 2012), associative memory type of learning using STDP
(Tanaka et al. 2009), temporal synchrony detection (Bofill-I-Petit and Murray 2004),
robot navigation control (Arena et al. 2007, Arena et al. 2009) and associative memory, as
well as variability and noise compensation tasks (Arthur and Boahen 2006). Although
some of these learning applications such as the last five mentioned works, have been
successfully implemented as part of a neuromorphic system, many of the other synap-
tic plasticity rules that have been modelled based on biological experiments performed
in vivo and in vitro, are yet to be explored by neuromorphic engineers for other appli-
cations. Examples of these plasticity rules that have not been explored for any applica-
tion are the hybrid rules proposed in Mayr and Partzsch (2010), Clopath and Gerstner
(2010), and Graupner and Brunel (2012) as well as biophysical-based rule proposed
in Rachmuth et al. (2011), Shouval et al. (2002) and Meng et al. (2011). In addition to the
spike timing-based rules, other spike-based rules such as the SDSP (Brader et al. 2007)
rule is shown to be useful in other applications such as supervised learning for real-
time pattern classification (Mitra et al. 2009).
Page 126
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
level of performance for the targeted application; (iv) the possible need for modifying
the structure of available synaptic plasticity rules for better performance, lower im-
plementation complexity, or input data processing. As an example, here we review a
neuromorphic learning network and answer the above mentioned questions about it.
As already discussed, one of the most effective implementations of a VLSI SNN capable
of learning to preform a real-world task is the design presented in Seo et al. (2011). This
neuromorphic system is composed of 256 neurons and 256×256 synapses, in a crossbar
array structure to be used for various applications including an associative memory
task. The above mentioned questions are answered regarding this system.
(i) The input to this system can be set as 256 spike trains, each one corresponding to
a neuron in the network. These 256 spike trains encode the information embedded
in the input pattern and present it to the network of neurons. The network changes
its weights according to a PSTDP algorithm, in the training phase, when patterns are
presented to the network for learning. In the test phase, the neurons are presented with
a partial version of the original pattern, and the network through its weights reflects
the learned complete pattern, as output spikes.
(ii) The complexity of the targeted task and the number of patterns that can be learned
using this neuromorphic system is directly related to the complexity of the network, i.e.
its reconfigurability and neuron and synapses count. Since in the present network only
256 neurons with binary synapses are used, as the results in Seo et al. (2011) show, the
network can only learn 0.047 patterns per neuron in an associative memory task. It is
also shown that, if synapses with 4-bit precision are used instead of binary synapses,
the learning capacity of the hardware network increases up to 0.109 patterns per neu-
ron.
(iii) The spiking network implemented in this work has a highly reconfigurable struc-
ture with on-chip probabilistic PSTDP learning, thanks to its crossbar architecture and
transposable synapse SRAM cells, which make PSTDP possible. Therefore, it can re-
alise various network topologies and perform different cognitive tasks. Obviously, for
implementing more complex tasks and learning a high number of patterns, a large-
scale network with high-precision synapses is needed. Since this design is a basic
building block for a scalable neuromorphic system, this extension can be carried out
easily. The performance of the associative memory task presented for this network—
see (Seo et al. 2011)—shows that for this application, simple binary PSTDP synapses
integrated with digital integrate and fire neurons are enough. However, one can use
Page 127
5.7 Chapter Summary
other synaptic plasticity rules with higher complexity, such as those reviewed in this
thesis including TSTDP, SDSP, LCP and even biophysically-based rules, to reach better
performance or perform more complex tasks. However, the higher cost of these more
complex learning circuits should be considered.
(iv) In addition to the main chip that contains 64K probabilistic binary PSTDP synapses
and 256 neurons, three different variants of this chip were investigated to follow differ-
ent targets such as area, power consumption, and learning capability. It is shown that
the system with higher learning capability consumes the highest amount of power and
occupies the largest silicon real estate among all designs.
Page 128
Chapter 5 Spike-based Synaptic Plasticity Rules in Silicon
The presented review and discussion in this chapter provide us with a deep insight to
the field of designing VLSI circuits for synaptic plasticity rules as part of a neuromor-
phic system that tends to be used in an engineering application. This insight is quite
helpful while designing new circuits for unexplored synaptic plasticity rules such as
the TSTDP rule, which is the focus of this thesis. The following chapters present new
circuit designs for the TSTDP rule, and show how these designs are able to reproduce
the outcomes of a variety of synaptic plasticity experiments ranging from timing-based
experiments to rate-based experiments, while they follow various design strategies
and targets.
Page 129
Page 130
Chapter 6
A
S discussed in Chapter 2, classical model of STDP is incapable
of reproducing synaptic weight changes similar to those seen
in specific biological experiments that investigate the effects
of either higher order spike trains, or the simultaneous effects of the rate
and timing of spike pairs on synaptic plasticity. Significantly, a previously
described TSTDP rule succeeds in reproducing all of these synaptic plas-
ticity experiments. In this chapter, first, synaptic weight changes using a
number of widely used PSTDP circuits are investigated. The investigations
show that the class of PSTDP circuits fails to mimic the outcomes of the
mentioned complex biological experiments. Then, two new TSTDP VLSI
circuits, which are able to reproduce the outcomes of many complex ex-
periments, are presented. To the best of our knowledge, the presented cir-
cuits in this chapter are the first VLSI implementations of the TSTDP rule.
The new STDP VLSI circuits significantly improve upon previous circuits.
These new circuit capabilities in closely mimicking the outcomes of various
biological experiments, may play a significant role in future neuromorphic
VLSI systems with increased learning ability.
Page 131
6.1 Introduction
6.1 Introduction
Spike Timing-Dependent Plasticity (STDP) is an unsupervised synaptic plasticity rule
that induces changes in individual synaptic weights, based on the timing difference
between pre- and post-synaptic spikes (Song et al. 2000, Bi and Poo 1998). The classical
STDP model employs a pair of spikes (pre-post or post-pre) as the trigger for changes
in synaptic plasticity (Song et al. 2000, Bi and Poo 1998). However, if the repetition
frequency of spike pairs is increased, this model fails to correctly reproduce synap-
tic weight changes as observed in physiological experiments (Sjöström et al. 2001).
Furthermore, it is not able to account for experiments using triplet or quadruplet of
spikes (Wang et al. 2005, Pfister and Gerstner 2006). An explanation for these shortcom-
ings is that traditional pair-based STDP does not account for known nonlinear interac-
tions between successive spikes when more complex spike patterns are used (Froemke
and Dan 2002). In order to resolve the short-comings of the classical pair-based model,
a simple yet elegant STDP model was proposed by Pfister and Gerstner (2006), where
synaptic weight changes based on triplets of spikes was developed.
This chapter proposes two new VLSI implementations for TSTDP. These implementa-
tions are based on two previous PSTDP circuits already presented in Bofill-I-Petit and
Murray (2004) and Indiveri et al. (2006). It is shown that both previous PSTDP designs
lack the ability to account for the experiments presented in previous experimental pa-
pers (Sjöström et al. 2001, Froemke and Dan 2002, Wang et al. 2005). Furthermore, It
is demonstrated that both proposed TSTDP circuits succeed in closely reproducing all
the experimentally observed biological effects.
Page 132
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
The remainder of this chapter is organised as follows. In Section 6.2, two sample
PSTDP circuits are presented and described. Section 6.3 introduces the new TSTDP
circuits and describes how they are related to the TSTDP learning rule. Section 6.4 pro-
vides information about experimental setup including, experimental protocols, data
sets and error functions. Experimental results are provided in Section 6.5, where it is
shown that the TSTDP circuit successfully mimic many experiments, while the class
of PSTDP circuits, including the two example circuits presented here fails. Section 6.6
compares the two proposed designs and also compares them to previous works that
implemented other synaptic plasticity rules with similar or less ability. The chapter
finishes by conclusions in Section 6.7.
The results shown in this chapter are presented in The 21st Japanese Neural Network Soci-
ety Annual Conference (Azghadi et al. 2011d), as well as in The 7th IEEE International Con-
ference on Intelligent Sensors, Sensor Networks and Information Processing (Azghadi et al.
2011c), and The 2012 IEEE International Joint Conference on Neural Networks (Azghadi et al.
2012b, Azghadi et al. 2012a).
Before discussing the structure of the circuits implementing the PSTDP rule, looking at
the PSTDP model that was already described in Section 2.6.1 is useful. In the following,
first the PSTDP model is presented and then the circuit models are shown. Note that
different parts of both of these circuits are mapped to different terms in the PSTDP
model to facilitate understanding the circuits.
In the PSTDP plasticity model, potentiation occurs when a pre-synaptic spike precedes
a post-synaptic spike; otherwise depression occurs, where weight changes can be gov-
erned by a temporal learning window. The classical STDP temporal learning window
can be expressed as (Song et al. 2000)
∆w+ = A+ e( −τ+∆t ) if ∆t > 0
∆w = (6.1)
∆w− = − A− e( τ∆t− ) if ∆t ≤ 0,
Page 133
6.2 VLSI Implementation of Pair-based STDP
where ∆t = tpost − tpre is the time difference between a single pair of post- and pre-
synaptic spikes, τ+ and τ− are time constants of the learning window, and finally A+
and A− represent the maximal weight changes for potentiation and depression, re-
spectively.
A VLSI implementation of the PSTDP rule must account for various parts of the rule
shown in Eq. 6.1. As already discussed in Section 5.3.2, there are various VLSI imple-
mentations of PSTDP rule available in the literature. In this chapter, two simple PSTDP
circuits as representatives of the class of PSTDP circuit models are selected and tested
for generating various synaptic plasticity experiments to verify their abilities. The two
chosen circuits are the PSTDP circuit proposed by Indiveri et al. (2006), and the current
mode circuit presented by Bofill-I-Petit and Murray (2004). These circuits are chosen
due to their different voltage- and current-mode structures, so that two different tech-
niques are verified in our investigations. In addition, both of these designs are simple
and have been proven as silicon devices, which are used for some learning applica-
tions. Besides, they use two different types of leaky integrators (see Section 5.2.3)
and produce different STDP learning windows (Bofill-I-Petit and Murray 2004, Indi-
veri et al. 2006). In the following section, it is shown how these circuits implement an
approximation of the PSTDP model using CMOS transistors.
The implementation by Indiveri et al. (2006) was adopted, due to its low power and
small area. Fig. 6.1(a) depicts the Indiveri’s pair-based STDP circuit schematic and
Fig. 6.1(b) demonstrates its resulting temporal learning window for various τ+ and
τ− (Vtp , Vtd ). The timing of pre- and post-synaptic spikes are used to induce weight
changes across Cw . This circuit results in a learning window which captures the essen-
tial features of STDP, where there are two distinct regions, one for potentiation where
∆t ≥ 0 and depression for ∆t < 0.
When a pre-synaptic pulse, Vpre , or a post-synaptic pulse (V post ) occurs, Vpot (Vdep ) will
be set to zero (Vdd ). Note that Vpot (Vdep ) then changes linearly over time to reach Vdd
(zero), and represents the required time constants τ+ (τ− ). These time constants can be
set by changing the gate voltage of the corresponding transistor, i.e. Vtp (Vtd ). Fig. 6.1(b)
demonstrates the variation of the learning window for different values of Vtp (Vtd ), i.e.
τ+ (τ− ). So, if a Vpre (V post ) pulse occurs during time determined by its corresponding
Page 134
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
time constant, τ− (τ+ ), the output capacitor will be discharged (charged) by a current
that is proportional to the value of Vdep (Vpot ) and VA− (VA+ ).
Figure 6.1. Indiveri’s PSTDP circuit model. (a) Schematic circuit diagram of Indiveri et al.
(2006). (b) The learning window of the circuit based on our simulations in an accelerated
time scale.
The PSTDP circuit presented by Indiveri et al. (2006) cannot reproduce the required
exponential behaviour seen in Eq. 6.1, while the circuit proposed by Bofill-I-Petit and
Murray (2004) shown in Fig. 6.2 can.
The PSTDP circuit presented in Bofill-I-Petit and Murray (2004) implements a weight-
dependent STDP rule, in which the current synaptic weight has an impact on the
amount of potentiation/depression. In order to make this PSTDP circuit compatible
Page 135
6.2 VLSI Implementation of Pair-based STDP
Figure 6.2. Bofill and Murray’s PSTDP circuit model. This circuit that is presented in Bofill-I-
Petit and Murray (2004) has a weight-dependency part, shown in the dashed-box, which
depending on the current weight controls the amount of weight change.
with the normal PSTDP rule, i.e. weight independent, which is also the case for the
TSTDP rule, some modifications are needed. These modifications lead to the circuit
shown in Fig. 6.3(a). The modifications are as follows: (i) Since the classical PSTDP
model (Eq. 6.1) is weight independent, in the modified circuit, the weight dependency
part (shown in the dashed box in Fig. 6.2) is omitted. (ii) Potentiation and depression
in the modified circuit are represented through an increase or a decrease in the amount
of charges stored on the weight capacitor, respectively, which is in contrast to the cir-
cuit presented in Bofill-I-Petit and Murray (2004). (iii) Also, in order to simplify the
circuit, preLong and postLong pulses which should be generated by an additional cir-
cuitry, were replaced with Vpre and Vpost. These signals represent the input pre- and
post-synaptic pulses in the modified circuit. (iv) Furthermore, using bias voltages for
time constants control results in significant variation in time constants under various
3σ process corners. So, in order to make the circuit more robust against this condition,
the bias voltages were represented as the gate-to-source voltage of a number of diode
connected transistors that are biased by current sources (M3 and M14). Fig. 6.3(b)
demonstrates that using this approach results in a slight change in time constants at
all different process corners when compared with MATLAB simulations. Beside time
constants, the amplitude constants are also implemented as current sources (Ipot and
Idep ) to be less prone to process variations compared to the original circuit shown in
Fig. 6.2, and can later be used to fine tune the time constants when needed. This ap-
proach suggests that the time constants, as well as the amplitude parameters are more
robust against device mismatch.
Page 136
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
Figure 6.3. Modified PSTDP circuit. (a) The Bofill and Murray’s circuit shown in Fig. 6.2 was
modified (see the text for more detail). (b) STDP learning window generated by Matlab
simulation and also using different transistor process corners for the modified PSTDP
circuit shown in part (a). Note that similar protocols and time constants to Bi and Poo
(1998) are employed.
Before discussing the structure of the circuits that implement the TSTDP rule, looking
at the TSTDP model that was already described in Section 2.6.1 is useful. In the fol-
lowing, first the TSTDP model is presented and then the proposed circuit models are
shown.
Page 137
6.3 VLSI Implementation of Triplet-based STDP
As already mentioned, previous studies illustrated that classical STDP model fails to
reproduce the experimental outcomes involving higher order spike patterns such as
triplets and quadruplets of spikes (Wang et al. 2005, Pfister and Gerstner 2006). Further-
more, it fails to account for the observed dependence of synaptic plasticity on the rep-
etition frequency of pairs of spikes (Sjöström et al. 2001). To resolve these issues, pair-
based STDP was extended to include spike triplets that resulted in a spike-triplet-based
STDP learning rule which could sufficiently reproduce previously reported physiolog-
ical experiments. Based on the triplet synaptic learning rule presented in Pfister and
Gerstner (2006), the triplet synaptic modification rule can be written as
− ∆t − ∆t − ∆t
∆w+ = A+ e( τ+ 1 ) + A+ e( τy 2 ) e( τ+ 1 )
2 3
∆w = ∆t1 − ∆t3 ∆t (6.2)
( ) ( 1)
∆w− = − A2− e τ− − A3− e( τx ) e τ− ,
where ∆w = ∆w+ if t = tpost and ∆w = ∆w− if t = tpre . Here, A2+ , A2− , A3+ and A3− are
constants, ∆t1 = tpost − tpre , ∆t2 = tpost(n) − tpost(n−1) − ǫ and ∆t3 = tpre(n) − tpre(n−1) −
ǫ, are time difference between combinations of pre- and post-synaptic spikes, τ− , τ+ , τx
and τy are time constants, and finally ǫ is a small positive value which selects the con-
tribution to the weight change just before the final spike of the presented triplet (Pfister
and Gerstner 2006). Hence, triplet-based model induces weight change in proportion
to eight parameters (in comparison to four parameters for classical pair-based model);
four potentiation parameters (A2+ , τ+ , A3+ , and τy ) and four depression parameters
(A2− , τ− , A3− , and τx ).
The TSTDP model is not just a simple change in the degree of freedom of the PSTDP
model, but it tries to overcome some deficiencies of the PSTDP model. In Pfister and
Gerstner (2006), it is addressed that the TSTDP model, removed two main problems
of the PSTDP formula. These problems and how the TSTDP solves them are as fol-
lows: (i) As PSTDP considers just pairs of spikes, for any value of A+ > 0, if a pre-
synaptic spike precedes a post-synaptic one, it brings about potentiation, while ac-
cording to Sjöström et al. (2001), at low repetition frequencies, there is no potentiation.
In the TSTDP model, this deficiency can be solved by setting A2 + to a small value or in
the case of minimal TSTDP rule, to zero. So it makes the potentiation very small, which
can be neutralised by a bit of depression, or it can be zero. (ii) Considering biological
experiments in Sjöström et al. (2001), for ∆t > 0, potentiation will increase with the in-
crease in frequency. However, this behaviour cannot be generated by PSTDP, as when
Page 138
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
the frequency of pairs of spikes increases, it causes the pairs to interact with each other,
so it causes no significant potentiation. This problem can be solved again by correctly
tuning the TSTDP parameters. In this case, A3 + should be strong enough to make the
potentiation win over depression and so to have depression in high frequencies (Pfister
and Gerstner 2006).
A VLSI implementation of the TSTDP rule must account for all the mentioned details as
well as the various parts of the rule shown in Eq. 6.2. In this chapter, two new designs
for TSTDP are proposed, which account for the mentioned details and various parts
of the TSTDP rule. These circuits are tested for generating various synaptic plasticity
experiments to verify their abilities. The two proposed circuits are built upon the two
example PSTDP circuits shown in Section 6.2.
Unlike the pair-based model, in the triplet model, a pre-synaptic (post-synaptic) spike
further to having an affect on its successive post-synaptic (pre-synaptic) spike can also
have an affect on its consecutive pre-synaptic (post-synaptic) spike(s). In the proposed
triplet circuit, two more pulses, Vpost(n−1) and V pre(n−1) , are used in addition to V post(n)
and Vpre(n) , as shown in Fig. 6.4.
These extra pulses result in the required nonlinearity in the triplet-based model (Pfister
and Gerstner 2006). The circuit works as follows: upon the arrival of a post-synaptic
pulse, V post(n) , the M5, M10 and M18 transistor switches turn on. Then M10 sets a de-
potentiating voltage Vdep1 to Vdd . This voltage then starts decaying linearly with time
which can result in depression, if a pre-synaptic pulse, Vpre(n) arrives during the time
Vdep1 is decaying to zero (τ− time constant). In this situation, Cw will be discharged
through M7-M9 by a current that is limited by the M7 bias voltage (VA− ).
2
In contrast to M10, which can result in depression after receiving a post-synaptic pulse,
M5 and M18 can lead to two different potentiations. The first one can occur if M5 turns
on during time constant of Vpot1 (τ+ ). This potentiation will be through M4-M6 and is
proportional to the bias voltage at M6 (VA+ ). The second potentiation term can charge
2
Cw through M16-M19 and is proportional to VA+ if M18 is on at the required time, i.e.
3
when Vpot1 and Vpot2 have still kept M16 and M17 on, respectively. This is the term
that distinguishes triplet from pair-based STDP, as there is no such term in pair-based
STDP. Similarly, upon the arrival of a pre-synaptic pulse, Vpre(n) , a potentiating voltage
Page 139
6.3 VLSI Implementation of Triplet-based STDP
Figure 6.4. Proposed voltage-mode full TSTDP circuit. This figure shows a symmetric voltage-
mode design for triplet-based STDP rule (Azghadi et al. 2011d).
Vpot is set to zero and starts increasing linearly in time which can result in potentiation
when a V post(n) pulse arrives within the τ+ time constant. In addition, two possible
depressions proportional to A2− and A3− can take place, if this pre-synaptic pulse is in
the interval area of effect of Vdep1 and Vdep2 , i.e. in τ− and τx time constants. It is worth
mentioning that the required time constants in the proposed circuit, τ− , τ+ , τx and τy ,
are adjusted by altering their corresponding bias voltages, Vtd1 , Vtp1 , Vtd2 and Vtp2 .
Fig. 6.5 presents the proposed current-mode circuit for the TSTDP model. In this cir-
cuit, there are eight parameters that can be tuned by controlling eight bias currents as
follows: the first four currents including Idep1 , Ipot1 , Idep2 and Ipot2 represent the ampli-
tude of synaptic weight changes for post-pre, pre-post, pre-post-pre and post-pre-post
combinations of spike triplets, respectively. Another control parameter for these ampli-
tude values in the circuit is the pulse width of the spikes which was kept fixed during
all experiments in this chapter (1 µs). In addition to these amplitude parameters, four
more currents control the required time constants in the model for post-pre, pre-post,
Page 140
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
pre-post-pre and post-pre-post combinations of spike triplets, and can be adjusted us-
ing Itd1 , Itp1 , Itd2 and Itp2 respectively.
Figure 6.5. Proposed current-mode full TSTDP circuit. This figure shows a current-mode
circuit for the TSTDP rule (Azghadi et al. 2012b).
The proposed circuit works as follows: upon the arrival of a post-synaptic pulse,
Vpost(n) , M2, M8 and M22 switch on. At this time, Idep1 can charge the first depres-
sion capacitor, Cdep1 , through M2 to the voltage of Vdep1 . After finishing Vpost(n) , Vdep1
starts decaying linearly through M4 and with a rate proportional to Itd1 . Now, if a
pre-synaptic pulse, Vpre(n) arrives at M6 in the decaying period of Vdep1, namely when
M5 is still active, the weight capacitor, CW , will be discharged through M5-M6 tran-
sistors and a depression occurs due to the occurrence of a pre-synaptic pulse in the
interval of affect of a post-synaptic spike (post-pre combination of spikes). Addition-
ally, if a pre-synaptic spike arrives at M13, soon before the present post-synaptic spike
Page 141
6.4 Experimental Setup
at M8, the weight capacitor can be charged through M7-M8 transistors and a potentia-
tion happens. This potentiation happens because the current post-synaptic spike is in
the time of affect of a pre-synaptic spike (pre-post combination of spikes). The amount
of potentiation depends on Vpot1 , which itself can be tuned by the relevant amplitude
parameter Ipot1 . Also, the activation interval of M11 can be modified by changing the
related time constant parameter Itp1 . Furthermore, another contribution to potentia-
tion can occur if a previous post-synaptic pulse, Vpost(n−1) , arrives at M27 soon enough
before the current post-synaptic happens at M8 and also before a pre-synaptic pulse
happens at M32 (this is the same pulse as for M13). In this situation, the weight capac-
itor can be charged again through M7-M8 and by an amount proportional to Vpot2 and
Vpot3 . This is a triplet interaction in the proposed circuit that leads to the required non-
linearity mentioned in the triplet learning rule, appears. A similar description holds for
the situation when a pre-synaptic pulse occurs at M6, M13 and M21 transistors. But
this time one potentiation and two depression events can happen if the appropriate
situation is provided.
The first two parts of this current-mode TSTDP circuit (on the top left and the top right)
are identical to the PSTDP circuit presented in Fig. 6.3(a). Also, the two bottom parts of
this circuit carry out the triplet terms interactions. This circuit is in correspondence to
the full triplet learning rule presented in Pfister and Gerstner (2006) which takes into
account all four possible potentiation and depression interactions. However, as it is
shown in following sections, only some of the terms are really necessary to reproduce
the expected biological experiments. This is referred to as minimal triplet learning rule
in Pfister and Gerstner (2006), which makes the required circuit simpler and smaller.
In the next section, the two proposed circuits for the TSTDP model and the mentioned
PSTDP circuits are verified for reproducing the outcomes of several biological exper-
iments. These circuits are used in a specific experimental setup, which is explained
below.
The presented simulations of the two proposed TSTDP circuits as well as the two men-
tioned PSTDP circuits were carried out using parameters for a 0.35 µm standard CMOS
Page 142
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
technology in HSpice. The widths and lengths used for all transistors in presented de-
signs are 0.7 µm and 0.35 µm, respectively. The capacitor values for these circuits are as
follows. Indiveri’s PSTDP and the proposed voltage-mode TSTDP circuits weight ca-
pacitors Cw = 1 pF. Bofill and Murray’s PSTDP circuit and the proposed current-mode
TSTDP circuits weight capacitors Cw = 10 pF, and other capacitors in these designs are
set to 10 fF.
The synaptic weight capacitors in the proposed designs occupy a large portion of the
silicon area, as it is the case for almost all synaptic circuits. Therefore, one of the con-
cerns here is to reduce this capacitor value in order to make the area of the design
smaller. There are a number of approaches to reduce the size of these capacitors as
discussed in Section 5.4.8. In addition, our simulation results also show that, it is pos-
sible to scale the capacitor value (and hence, its size) to a significantly smaller value
and optimise the circuit biases for the amount of weight changes stored on this smaller
capacitor. However, the minimum fitting error (discussed in the next section) might
significantly increase.
It should be noted that, during all experiments in this thesis, the nearest spike interac-
tion, which considers the interaction of a spike only with its two immediate succeeding
and immediate preceding nearest neighbours, was used (see Section 2.6.1). In addition,
the simulations presented in this chapter are all carried out in an accelerated time scale
of 1000 compared to real biological time scale. It means that in the performed simu-
lations, each ms corresponds to one second of actual biological time. For the sake of
simplicity while comparing the results from the proposed circuits, to the one from bio-
logical experiments, all the simulations are scaled back to the real time. The approach
of time-scaling has been used extensively in many previous neuromorphic studies such
as Schemmel et al. (2006), Tanaka et al. (2009) and Mayr et al. (2010).
Page 143
6.4 Experimental Setup
This data set is composed of 10 data points that are obtained from Table 1 of Pfister and
Gerstner (2006) that represents experimental weight change, ∆w, for two different ∆t’s,
and as a function of the frequency of spike pairs under a pairing protocol in the visual
cortex. The second experimental data set that was utilised, originates from hippocam-
pal cultures experiments from Wang et al. (2005), which examined pairing, triplet and
quadruplet protocols effects on synaptic weight change. This data set consists of 13
data points obtained from Table 2 of Pfister and Gerstner (2006). This data set shows
the experimental weight change, ∆w, as a function of the relative spike timing ∆t, ∆t1 ,
∆t2 and T under pairing, triplet and quadruplet protocols in hippocampal cultures.
Identical to Pfister and Gerstner (2006) that tests its proposed triplet model, as well as a
PSTDP model simulation results against the experimental data and reports their differ-
ences as Normalised Mean Square Error (NMSE) for each data set, here the mentioned
PSTDP and TSTDP circuit simulation results are verified under same conditions. The
mentioned NMSE (Pfister and Gerstner 2006) is calculated using the following equa-
tion (presented already in Section 2.5):
!2
1
p ∆wiexp − ∆wicir
p i∑
E= , (6.3)
=1
σi
where ∆wiexp , ∆wicir and σi are the mean weight change obtained from biological ex-
periments, the weight change obtained from the circuit under consideration, and the
standard error mean of ∆wiexp for a given data point i, respectively; p represents the
number of data points in a specified data set (can be 10 or 13).
In order to minimise the resulting NMSE for a circuit, there was a need to adjust the
parameters and time constants to minimise the resulting NMSE. In the following sub-
sections, the circuit simulation results and applied bias currents, and voltages for set-
ting the required parameters, in order to have the minimum achieved NMSEs for each
circuit under test, are reported.
The following section shows the experimental circuit results after optimising circuit
biases for the four different PSTDP and TSTDP circuits presented in the previous sec-
tions.
Page 144
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
In order to test the accuracy of Indiveri’s PSTDP circuit, Fig. 6.1(a), first the circuit was
checked for producing the frequency-dependent pairing experimental data, i.e. first
data set. The circuit parameters were optimised to minimise the NMSE for reproducing
the 10 data points available in the first data set. Obtained results shown in Fig. 6.6
clearly demonstrates that this PSTDP circuit fails to mimic the experimental data. The
final set of optimised parameters for the shown results are those reported in Table 6.1
and the minimal resulting NMSE was 10.03. This error is close to the error reported in
the best case of parameters obtained from classical pair-based STDP model simulations
reported in Pfister and Gerstner (2006) where NMSE ∼ = 7.5—data obtained from Fig. 6
of Pfister and Gerstner (2006).
Figure 6.6. Indiveri’s PSTDP circuit fails to reproduce the outcomes of frequency-dependent
pairing experiments. Frequency-dependent pairing protocol (see Section 2.5.2) is
applied to the circuit. Here, ρ is the repetition rate of pre- and post-synaptic spike
pairs. Note that there is no experimental data available at ρ = 30 Hz.
In addition, further simulations on pairing, quadruplet and triplet protocols were con-
ducted. Again, we optimised the parameters of the VLSI implementation of pair-based
STDP so that the NMSE was minimal across the entire data set, i.e. for all three pro-
tocols, we employed similar VA+ , VA− , τ+ and τ− values. Obtained results show that
the classical VLSI implementation for pair-based STDP, like its mathematical model,
fails to reproduce the experimental data obtained using quadruplet, Fig. 6.7(a), and
triplet protocols, Fig. 6.7(b)-(c). The NMSE in this case was 11.3, which is close to the
Page 145
6.5 Experimental Circuit Results
Table 6.1. Indiveri’s PSTDP circuit bias voltages for mimicking two different data sets and
their resulting NMSEs. The first data set includes 10 data points from pairing fre-
quency experiments presented in Sjöström et al. (2001) and the second data set consists
of 13 data points from pairing experiments (2 points), quadruplet experiments (3 points)
and two different triplet experiments (each one with 4 points) presented in Wang et al.
(2005).
Data set VA+ (V) VA− (V) Vtp (V) Vtd (V) NMSE
First 2.36 0.69 2.87 0.29 10.03
Second 2.28 0.68 2.8 0.25 11.36
Table 6.2. Bofill and Murray’s PSTDP circuit bias currents for mimicking two different data
sets and their resulting NMSEs. Similar data sets to Table 6.1 are used.
Data set Ipot (nA) Idep (nA) Itp (pA) Itd (pA) NMSE
First 150 150 24 18 7.26
Second 410 190 20 5 10.76
optimal value obtained from the pair-based model in Pfister and Gerstner (2006)—
NMSE ∼= 10.5, data obtained from Fig. 6 in Pfister and Gerstner (2006).
Simulation results for the frequency-dependent pairing experiments, using the circuit
presented in Fig. 6.3(a) is demonstrated in Fig. 6.8. This figure shows how this PSTDP
circuit, similar to the previous PSTDP circuit, fails to reproduce the observed experi-
mental results in visual cortex reported in Sjöström et al. (2001). The minimal NMSE
obtained in this situation was 7.26, which is consistent with the reported minimal
achieved error using computer simulation of the PSTDP rule in Fig. 6A of Pfister and
Gerstner (2006). The four required bias currents for controlling the model parameters
are reported in Table 6.2.
Simulation results of the second data set also suggest that this PSTDP circuit fails to
reproduce experimental results observed in hippocampal cultures. Fig. 6.9(a) shows
Page 146
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
Figure 6.7. Indiveri’s PSTDP circuit fails to reproduce the outcomes of triplet and quadru-
plet experiments. (a) Quadruplet spikes were applied to the circuit, under the quadru-
plet protocol, and the biases were optimised to minimise the NMSE to fit the 3 quadru-
plet experimental data points (with error bars) shown in (a), as well as the 8 triplet
data bars shown in (b) and (c). (b) Pre-post-pre triplet of spikes were applied to the
circuit according to the triplet protocols. (c) Same as (b), but for the post-pre-post
spike triplets.
the circuit simulation results along with experimental data for the quadruplet proto-
col, while Fig. 6.9(b)-(c) represent the results under triplet protocols for pre-post-pre
and post-pre-post combinations of spike triplets, respectively. The minimal NMSE ob-
tained in this situation was 10.76, again consistent with the reported results in Fig. 6B
of Pfister and Gerstner (2006). The four required bias currents for controlling the model
parameters are reported in Table 6.2.
The simulation results shown in Figs. 6.10 and 6.11 demonstrate that the proposed
VLSI triplet-based circuit has a significantly improved weight change prediction capa-
bility in comparison to its pair-based counterparts. Like pair-based circuit experiments,
Fig. 6.10 shows the total weight change induced by a pairing protocol for various pulse
Page 147
6.5 Experimental Circuit Results
Figure 6.8. Bofill and Murray’s PSTDP circuit fails to reproduce the outcomes of frequency-
dependent pairing experiments. Frequency-dependent pairing protocol (see Section
2.5.2) is applied to the circuit. Here, ρ is the repetition rate of pre- and post-synaptic
spike pairs. Note that there is no experimental data available at ρ = 30 Hz.
Table 6.3. Proposed voltage-mode TSTDP circuit bias voltages for mimicking two different
data sets and their resulting NMSEs. Similar data sets to Table 6.1 are used.
Data set VA2+ (V) VA2− (V) Vtp1 (V) Vtd1 (V) VA3+ (V) VA3− (V) Vtp2 (V) Vtd2 (V) NMSE
First 2.49 0.59 2.49 0.59 2.36 0.25 2.44 2.6 0.82
Second 2.49 0.66 2.4 0.59 2.3 0.25 2.45 2.7 3.46
repetition rates. As can be seen from the figure, a better match between the experimen-
tal data and simulations was observed. The NMSE achieved was 0.82, which is far
better than the NMSE for the pair-based case and much closer to the NMSE = 0.22,
obtained through analytical calculation of the triplet-based model, given in Table 3 of
Pfister and Gerstner (2006). The optimised bias voltages to reach this NMSE are shown
in Table 6.3.
In addition to the bias optimisation performed on the circuit, to approximate the first
data set, the circuit bias parameters were also optimised to approximate the outcome
of triplet, and quadruplet experiments. Achieved results for these experiments are
shown in Fig. 6.11. The minimal obtained NMSE for this case was 3.46 that is close to
the one achieved using the optimised parameters for the TSTDP model (NMSE = 2.9)
presented in Pfister and Gerstner (2006). The optimised bias voltages to reach this
NMSE are shown in Table 6.3.
Page 148
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
Figure 6.9. Bofill and Murray’s PSTDP circuit fails to reproduce the outcomes of triplet,
and quadruplet experiments. (a) Quadruplet spikes were applied to the circuit, under
the quadruplet protocol, and the biases were optimised to minimise the NMSE to fit the
3 quadruplet experimental data points (with error bars) shown in (a), as well as the 8
triplet data bars shown in (b) and (c). (b) Pre-post-pre triplet of spikes were applied to
the circuit according to the triplet protocols. (c) Same as (b), but for the post-pre-post
spike triplets.
Note that in the presented simulations for the proposed voltage-mode design, the full
TSTDP circuit, which is in accordance to the full TSTDP model is used. However, as
discussed in Pfister and Gerstner (2006), the rule can be minimised without having sig-
nificant effect on the performance of the rule in generating the required experiments.
Similarly, in the case of the proposed circuits, simulation results for replicating the
shown experiments using minimal versions of the circuit (and minimal circuit), show
similar performance in reproducing the outcomes of the targeted experiments. In the
case of the minimal model, the corresponding circuit that contains 26 transistors, is
minimised to a circuit with 16 transistors for reproducing the visual cortex experiments
(first date set). In this case M4-M6, M13-M15 and M20-M23 are removed from the full
TSTDP circuit shown in Fig. 6.1. In addition, the circuit is minimised to 19 transis-
tors for the hippocampal experiments (second data set), since M13-M15 and M20-M23
transistors are removed from the full TSTDP circuit.
Page 149
6.5 Experimental Circuit Results
Figure 6.10. Proposed voltage-mode TSTDP circuit mimics the outcomes of frequency-
dependent pairing experiments. Frequency-dependent pairing protocol (see Section
2.5.2) is applied to the circuit. Here, ρ is the repetition rate of pre- and post-synaptic
spike pairs. Note that there is no experimental data available at ρ = 30 Hz.
In order to test the proposed current-mode TSTDP circuit under the mentioned pro-
tocols and using the two data sets, firstly the full TSTDP circuit was employed. This
circuit is shown in Fig. 6.5 and consists of four distinct parts each of them related to
one of the pair or triplet combination of spikes. However, as stated in Pfister and Ger-
stner (2006), only some of these combinations are really necessary and play significant
roles in synaptic weight change under different protocols. Therefore, the full TSTDP
circuit was changed to two minimal TSTDP circuits in correspondence to two minimal
TSTDP rules in Pfister and Gerstner (2006). In these circuits, the inconsequential parts
of the proposed full-triplet circuit are removed to have the minimal circuits. This is in
the contrary to the voltage-mode design simulations that utilised the full TSTDP circuit
and optimised all circuit biases to reach a minimal NMSE.
As it can be extracted from the last line of Table 3 in Pfister and Gerstner (2006), the
minimal TSTDP rule that is capable of reproducing the expected visual cortex weight
change experiments (the first data set), sets pre-post and pre-post-pre spike combina-
tion amplitude parameters to zero. This means that this rule neither require the pre-
post interactions of spikes, nor the pre-post-pre interactions to take part in synaptic
weight modification. Therefore, these parts are not needed also in the corresponding
minimal TSTDP circuit. This circuit composed of 19 transistors (exclude the parts in
Page 150
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
Figure 6.11. Proposed voltage-mode TSTDP circuit mimics the outcomes of triplet, and
quadruplet experiments. (a) Quadruplet spikes were applied to the circuit, under
the quadruplet protocol, and the biases were optimised to minimise the NMSE to fit
the 3 quadruplet experimental data points (with error bars) shown in (a), as well as
the 8 triplet data bars shown in (b) and (c). (b) Pre-post-pre triplet of spikes were
applied to the circuit according to the triplet protocols. (c) Same as (b), but for the
post-pre-post spike triplets.
the dashed and dotted boxes in the circuit presented in Fig. 6.5) and it can reproduce
very similar results to the full TSTDP circuit which contains 34 transistors (Fig. 6.5).
The minimum NMSE obtained for the first data set and using this first minimal TSTDP
circuit was 0.64, which is near the minimum NMSE obtained by means of computer
simulations of minimal TSTDP model, NMSE = 0.34, obtained from Table 3 of Pfister
and Gerstner (2006). The five required bias currents for controlling the model parame-
ters are reported in Table 6.4.
Furthermore, the minimum obtained NMSE for the second data set using the second
minimal TSTDP circuit is 2.25. The achieved results are shown in Fig. 6.13(a)-(c). The
second minimal TSTDP circuit is composed of the whole top parts and the right bottom
part of the full TSTDP circuit presented in Fig. 6.5—see the last line of Table 4 in Pfister
and Gerstner (2006). The obtained NMSE using this circuit is slightly better than the
NMSE obtained using minimal TSTDP model and by means of computer simulations,
Page 151
6.6 Discussion
Figure 6.12. Proposed current-mode TSTDP circuit mimics the outcomes of frequency-
dependent pairing experiments. Frequency-dependent pairing protocol (see Section
2.5.2) is applied to the circuit. Here, ρ is the repetition rate of pre- and post-synaptic
spike pairs. Note that there is no experimental data available at ρ = 30 Hz.
Table 6.4. First TSTDP circuit bias currents and its resulting NMSE. The first data set includes
10 data points from pairing frequency experiments presented in Sjöström et al. (2001).
Data set Idep1 (nA) Itp1 (nA) Itd1 (pA) Ipot2 (µA) Itp2 (pA) NMSE
First 300 40 40 1.5 50 0.64
NMSE = 2.9, extracted from Table 4 of Pfister and Gerstner (2006). The six required
bias currents for controlling the model parameters are recorded in Table 6.5.
6.6 Discussion
In this chapter, a current-mode (Azghadi et al. 2012b) and a voltage-mode (Azghadi et al.
2011c) VLSI design, were proposed to implement the TSTDP learning rule. Although
it is shown that both circuits are able to account for various synaptic plasticity ex-
periments, the presented results in this chapter suggest that these two circuits have
different performance in reproducing the outcomes of various experiments. From the
NMSEs obtained for each of the circuits and for two different data sets, it is clear that
the current-mode design has a better performance in reproducing the experiments out-
comes. This is mainly because of the fact that the current-mode design produces a bet-
ter exponential behaviour, which is closer to the TSTDP model. On the other hand,
Page 152
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
Table 6.5. Second TSTDP circuit bias currents and its resulting NMSE. The second data set
consists of 13 data points from pairing experiments (2 points), quadruplet experiments
(3 points) and two different triplet experiments (each one with 4 points) presented in
Wang et al. (2005).
Data set Ipot1 (nA) Idep1 (nA) Itp1 (pA) Itd1 (pA) Ipot2 (nA) Itp2 (pA) NMSE
Second 160 130 28 20 400 10 2.25
the voltage-mode design has a simpler structure and uses a smaller number of tran-
sistors and capacitors, compared to the current-mode design. Therefore, each of the
proposed circuit models for TSTDP rule has its own pros and cons in terms of area,
power consumption, or synaptic accuracy. The voltage-mode design presents a better
area and power performance, but lacks in terms of accuracy, while the current-mode
design presents a high accuracy at the cost of more area and power.
Previous computational studies show that PSTDP (Izhikevich and Desai 2003) and
TSTDP (Pfister and Gerstner 2006) rules under specific circumstances can reproduce
BCM-like learning behaviour with a sliding threshold feature (Bienenstock et al. 1982,
Cooper et al. 2004). As part of the investigations in this thesis, the proposed voltage-
mode TSTDP circuit was examined to generate the BCM-like learning behaviour. Ob-
tained results presented in Azghadi et al. (2011a) demonstrate that the proposed circuit
can successfully generate this behaviour. In another study (Azghadi et al. 2012a), in
order to compare the performance of both PSTDP and TSTDP circuits for producing
the BCM-like behaviour, the voltage-mode PSTDP circuit proposed by Indiveri et al.
(2006) as well as the proposed voltage-mode TSTDP circuit were stimulated according
to a Poissonian protocol (see Section 2.5.6) to reproduce the required behaviour. Simu-
lation results demonstrate that the TSTDP circuit significantly produces the threshold-
based behaviour of the BCM. Also, the results suggest that the PSTDP circuit is able to
account for the BCM-like behaviour (Azghadi et al. 2012a).
In addition to the two proposed TSTDP designs, which have capabilities beyond the
PSTDP circuits, there is a previous VLSI implementation proposed by Mayr et al. (2010)
that is capable of reproducing the mentioned biological experiments (except the quadru-
plet protocol which has not been shown) similar to the proposed triplet circuits. In
terms of functionality, these implementations are different, since the design of Mayr et al.
(2010) implements the BCM-like rule, requiring the voltage of the neuron to take part
Page 153
6.6 Discussion
Figure 6.13. Proposed current-mode TSTDP circuit mimics the outcomes of triplet, and
quadruplet experiments. (a) Quadruplet spikes were applied to the circuit, under
the quadruplet protocol, and the biases were optimised to minimise the NMSE to fit
the 3 quadruplet experimental data points (with error bars) shown in (a), as well as
the 8 triplet data bars shown in (b) and (c). (b) Pre-post-pre triplet of spikes were
applied to the circuit according to the triplet protocols. (c) Same as (b), but for the
post-pre-post spike triplets.
Page 154
Chapter 6 First VLSI Designs for Triplet-based Spike Timing Dependent Plasticity
but also they can be tuned to mimic those experimentally observed behaviour with
a small error, while Mayr et al. (2010) just depict the behaviour and not the required
values observed in biological experiments. Also, the proposed circuits would require
smaller silicon real estates and presents lower power consumption when compared to
the circuit presented in Mayr et al. (2010).
Next chapter proposes a new TSTDP circuit that improves the synaptic ability com-
pared to the designs presented in this chapter and results in significantly lower error,
while reproducing the outcomes of many experiments, some of which cannot be repli-
cated by the circuits presented in this chapter.
Page 155
Page 156
Chapter 7
High-performance TSTDP
VLSI Design
T
HIS chapter introduces circuit design and implementation of a
new high-performance VLSI design for the TSTDP rule that out-
performs the other TSTDP VLSI designs in several aspects. It is
shown in this chapter, how different terms in the TSTDP synaptic plasticity
equation, are implemented to have a very close fit to the model. This results
in the proposed design to have significantly lower synaptic plasticity pre-
diction error, in comparison with previous designs for TSTDP and PSTDP
rules. In addition, it is shown that the new proposed design can success-
fully account for a number of new experiments, including experiments in-
volved with various spike triplet combinations, where the previous TSTDP
designs do not show acceptable performance and cannot mimic the exper-
iments effectively. This chapter also discusses some of the main challenges
in designing the proposed TSTDP circuit such as power consumption, sil-
icon real estate and process variations. We show that it is possible to miti-
gate the effect of process variations in the proposed circuit. In addition, the
power consumption and area of the proposed design are also investigated
and discussed in this chapter. The proposed circuit has been fabricated as a
proof of principle. Performed chip measurement results testify the correct
functionality of the fabricated circuit in performing triplet-based synaptic
weight modification.
Page 157
7.1 Introduction
7.1 Introduction
As already discussed in Chapter 2, neuro-physiological experiments have illustrated
that plastic changes to synapses can occur via spike-timing, varying the frequency of
inputs to the neuron, or changes to internal concentration of calcium in the neuron’s
spine apparatus (Bi and Poo 1998, Sjöström et al. 2001, Wang et al. 2005). Many theoret-
ical and experimental studies have focused on studying changes to synaptic strength
caused by STDP (Gerstner et al. 1996, Bi and Poo 1998, Song et al. 2000, Froemke and
Dan 2002, Wang et al. 2005, Pfister and Gerstner 2006, Iannella and Tanaka 2006, Ian-
nella et al. 2010). At the same time, there have been attempts at translating such rules
to VLSI circuit implementations (Bofill-I-Petit and Murray 2004, Indiveri et al. 2006,
Tanaka et al. 2009, Rachmuth et al. 2011, Azghadi et al. 2011a, Azghadi et al. 2011b,
Azghadi et al. 2011c, Azghadi et al. 2011d, Azghadi et al. 2012a, Azghadi et al. 2012b).
These attempts represent the crucial technological steps in developing smart VLSI
chips with adaptive capabilities similar to that of the mammalian brain. The long term
aim is to have VLSI circuits that can learn to adapt to changes and result in modifying
their functionality to improve their performance. The realisation of such adaptive VLSI
circuits will have widely varying applications ranging from artificial bionic prostheses
through to improved autonomous navigation systems.
The main contribution of this chapter is to significantly advance previous VLSI imple-
mentations of triplet-based STDP and introduce a new synaptic analog circuit that pos-
sesses some critical capabilities that have not been demonstrated in previous VLSI im-
plementations. The proposed circuit not only can replicate known outcomes of STDP,
including the effects of input frequency, but also it is capable of mimicking BCM-like
behaviour (Bienenstock et al. 1982). It improves the synaptic weight change modifica-
tion ability and results in less error while curve fitting the experimental data. In ad-
dition, the proposed circuit captures important aspects of both timing- and rate-based
synaptic plasticity that is of great interest for researchers in the field of neuromorphic
engineering, specifically to those who are involved in experiments dealing with learn-
ing and memory in-silico.
Page 158
Chapter 7 High-performance TSTDP VLSI Design
and describes the effects of process variation and transistor mismatch on the proposed
design, and suggests a tuning mechanism to overcome the performance degradation
in the presence of physical variations. Section 7.6 describes the VLSI implementation
of the proposed high-performance design and presents chip measurement results. Sec-
tion 7.7, provides a discussion of advantages as well as limitations of the proposed
design and suggests ways and their costs, in order to reduce the limitations of the
novel presented circuit. Section 7.8 gives concluding remarks of the chapter.
The results shown in this chapter are presented mainly in Neural Networks (Azghadi et al.
2013a).
In order to have a better understanding of the structure of the proposed circuit in this
chapter, the TSTDP synaptic plasticity rule is again mentioned here. However, a dif-
ferent arrangement of the rule presented in Eq. 2.3 has been utilised, which facilitates
understanding the structure of the new proposed circuit. This new arrangement of the
TSTDP rule (Pfister and Gerstner 2006) is given by
− ∆t2
− ∆t1
∆w+ = e( τ+ ) A+ + A+ e( τy )
2 3
∆w = ∆t1 − ∆t (7.1)
∆w− = −e( τ− ) A− + A− e( τx 3 ) ,
2 3
where ∆w = ∆w+ for t = tpost and if t = tpre then the weight change is ∆w =
∆w− . A2+ , A2− , A3+ and A3− are potentiation and depression amplitude parameters,
∆t1 = tpost(n) − tpre(n) , ∆t2 = tpost(n) − tpost(n−1) − ǫ and ∆t3 = tpre(n) − tpre(n−1) − ǫ,
are the time differences between combinations of pre- and post-synaptic spikes. Here, ǫ
is a small positive constant which ensures that the weight update uses the correct val-
ues occurring just before the pre or post-synaptic spike of interest, and finally τ− , τ+ , τx
and τy are time constants (Pfister and Gerstner 2006). Prior to this TSTDP model, there
was another rule proposed by Froemke and Dan (2002) which considers higher order
temporal patterns (quadruplets) of spikes to induce synaptic modification. Both of
these rules tend to explore the impact of higher order spike patterns on synaptic plas-
ticity. In this study, the proposed analog circuit aims to mimic the model presented in
Eq. 7.1.
Page 159
7.3 High-performance Circuit for TSTDP Rule
The new high-performance circuit produces a close fit to the outcomes of the TSTDP
rule. Fig. 7.1 presents the proposed circuit implementation of the full TSTDP model.
In the full TSTDP model, there are eight parameters that can be tuned in the proposed
circuit, by controlling eight bias currents as follows: Idep1 , Ipot1 , Idep2 and Ipot2 represent
the amplitude of synaptic weight changes for post-pre (A2− ) and pre-post (A2+ ) spike
pairs, and pre-post-pre (A3− ) and post-pre-post (A3+ ) combinations of spike triplets,
respectively. Another control parameter for these amplitude values in the circuit is
the pulse width of the spikes, which was kept fixed during all experiments in this
chapter (1 µs). In addition to these amplitude parameters, the required time constants
in the model for post-pre (τ− ), pre-post (τ+ ), pre-post-pre (τx ) and post-pre-post (τy )
spike patterns, can be adjusted using Itd1 , Itp1 , Itd2 and Itp2 respectively (see Eq. 7.1 and
Fig. 7.1).
The proposed circuit works as follows: upon the arrival of a pre-synaptic pulse, Vpre(n) ,
M9 and M15 are switched on. At this time, Ipot1 can charge the first potentiation capac-
itor, Cpot1 , through M9 to the voltage of Vpot1 . After finishing Vpre(n) , Vpot1 starts decay-
ing linearly through M11 and with a rate proportional to Itp1 . Now, if a post-synaptic
pulse, Vpost(n) arrives at M13 in the decaying period of Vpot1 , namely when M12 is still
active, the weight capacitor, CW , will be discharged through M12-M13 transistors and
a potentiation occurs because of the arrival of a post-synaptic pulse in the interval of
effect of a pre-synaptic spike (pre-post combination of spikes). Additionally, if a post-
synaptic spike has arrived at M19, soon before the current pre-synaptic spike at M15,
the weight capacitor can be charged through M14-M15 transistors and a depression
happens. This depression happens because the present pre-synaptic spike is in the
time of effect of a post-synaptic spike (post-pre combination of spikes). The amount
of depression depends on Vdep1 , which itself can be tuned by the relevant amplitude
parameter Idep1 . Also, the activation interval of M18 can be modified by changing the
related time constant parameter Itd1 . Furthermore, another contribution to depression
can occur if a previous pre-synaptic pulse, Vpre(n−1) , has arrived at M26 soon enough
before the current pre-synaptic happens at M15 and also before a post-synaptic pulse
happens at M19. In this situation, the weight capacitor can be charged again through
M14-M15 by an amount proportional to an effect of both Vdep2 and Vdep1, simulta-
neously. This triplet interaction leads to the required non-linearity mentioned in the
triplet learning rule.
Page 160
Chapter 7 High-performance TSTDP VLSI Design
A similar description holds for the situation when a post-synaptic pulse occurs at M13
and M19 transistors. But this time one depression will take place as the result of charg-
ing the weight capacitor up through M14-M15 and because of an arriving post-synaptic
spike at M19 before a pre-synaptic spike at M15. Besides, two potentiation events can
happen if an appropriate situation is provided to discharge the weight capacitor be-
cause of a pre-post or a post-pre-post combination of spikes. Note that, in this im-
plementation, the synaptic strength is inversely proportional to the voltage stored on
the weight capacitor, CW . However, for the sake of simplicity when comparing the
achieved results to experimental data, the weights are shown in a consistent way to
biological data, i.e. potentiation with positive strength and depression with negative
strength.
Upon examination of the TSTDP expression (Eq. 7.1), there are four different parts that
need to be implemented, in order to satisfy the equation as accurately as possible. The
proposed circuit (Fig. 7.1) is composed of four leaky integrators which are arranged in
a way that form the required addition and multiplications in the formula in a simple
manner. Furthermore, in order to have the exponential behaviour required for the
TSTDP rule, M5, M12, M18 and M25 are biased in the subthreshold region of operation.
The most left part of the circuit implements the potentiation triplet component of the
rule using a simple leaky integrator and the resulting current produced by this part
(Ipot−trip ) is given by
− ∆t2
( τy )
Ipot−trip = A3+ e , (7.2)
where Ipot2 represents A3+ , Itp2 can control τy and finally ∆t2 = tpost(n) − tpost(n−1) − ǫ
controlled by M2 and M13. Next, Ipot−trip is added up to Ipot1 current which repre-
sents A2+ in the TSTDP formula (Eq. 7.1). Hence, the amount of current going to M8
transistor is given by
− ∆t2
( τy )
IM8 = A2+ + A3+ e . (7.3)
Page 161
Page 162
This current then goes to the second leaky integrator on the second left box in Fig. 7.1
and will result in Ipot−total passing through M12 and M13 and discharging the weight
capacitor, CW , hence causes a potentiation equal to ∆w+ . The amount of this current
which is in result of the contribution of both triplet and pair-based spike patterns, can
be written as
− ∆t1 − ∆t2
( τ+ )
( τy )
Ipot−total = e A2+ + A3+ e , (7.4)
where Itp1 can control τ+ and finally ∆t1 = tpost(n) − tpre(n) is controlled by M9 and
M13.
The same approach applies for the depression part of Eq. 7.1. There are two leaky
integrators (the blue boxes in Fig. 7.1), each one is responsible for building an expo-
nential current and the final current (Idep−total ) which will be mirrored through M14
and M17 into the weight capacitor and result in charging the weight capacitor and
hence depression. This is the full TSTDP circuit which realises the full-TSTDP rule
(Eq. 7.1). However, according to the analytical calculations and numerical simulations
presented in Pfister and Gerstner (2006), some parts of the full TSTDP rule may be
omitted without a significant effect on the efficiency of the rule when replicating bi-
ological experiments. Pfister and Gerstner called these new modified rules, minimal
triplet rules.
According to the first minimal TSTDP rule, when generating the biological experi-
ment outcomes for the visual cortex data set presented in Sjöström et al. (2001), the
triplet contribution for depression, as well as the pairing contribution of the potentia-
tion parts of Eq. 7.1 can be dismissed (i.e. A3− = 0 and A2+ = 0) and the outcome will
be quite similar to using the full TSTDP rule—Table 3 in Pfister and Gerstner (2006).
Furthermore, the second minimal TSTDP rule which considers a zero value for A3−
(Eq. 7.1) has quite similar consequences to the full TSTDP rule and allows reproducing
the hippocampal culture data set experimental data presented in Wang et al. (2005).
As the rules are simplified, the full TSTDP circuit also can be minimised. This minimi-
sation can be performed by removing those parts of the circuit that correspond to the
omitted parts from the full TSTDP model. These parts are M23-M29 transistors which
can be removed when Idep2 = 0 (i.e. A3− = 0). Also Ipot1 can be set to zero, as it repre-
sents A2+ that is not necessary for the first minimal triplet rule. The resulting minimal
circuit based on these assumptions is shown in Fig. 7.2 with two separate parts for
potentiation and depression.
Page 163
Page 164
The potentiation part (a) which is composed of two leaky integrators is responsible for
voltage decrements across the weight capacitor (potentiation), in case of pre-post or
post-pre-post of spike patterns in the required timing periods. This part receives two
inputs backpropagated from the post-synaptic neuron (Vpost(n−1) , and Vpost(n) ), and
another input forwarded from a pre-synaptic neuron (Vpre(n) ). As there can be several
synapses on each post-synaptic neuron, this part of the minimal circuit which receives
inputs from different pre-synaptic neurons, needs to be replicated for every synapse.
However, the depression part of the minimal circuit, part (b), just receives an input
from the post-synaptic neuron and hence can be replicated once per neuron. That is
why we represent the potentiation and depression inversely to the charge stored on
the weight capacitor. As the number of neurons is significantly lower than the number
of synapses, this area saving can result in a significantly smaller area for a large neuro-
morphic system with TSTDP synapses. A similar approach was also utilised by Bofill-
I-Petit and Murray (2004).
In order to validate the functionality of the proposed TSTDP circuit, 12 different pat-
terns of spikes including spike pairs (four patterns), spike triplets (six patterns) and
spike quadruplets (two patterns) were utilised. These patterns were applied to the cir-
cuit and recorded weight changes were compared to their corresponding experimental
Page 165
7.4 Simulation Results
data. All simulation results show a good match to their related experimental data. The
first and second simulations were performed using two different data sets and for dif-
ferent experimental protocols. The optimisation scheme and the data fitting method
used here were that of Pfister and Gerstner (2006). The required experimental proto-
cols, different sets of data, the data fitting method as well as the achieved simulation
results, are explained and presented in the following subsections.
Additionally, for the third set of simulations, the proposed circuit was examined for
generating weight changes using all six possible spike triplet patterns presented in
Froemke and Dan (2002). Furthermore, the circuit was also used to reproduce the
weight changes produced by the rate-based BCM rule under a Poissonian protocol.
The achieved results for these two simulations, the triplet and Poissonian protocols are
also explained in the following subsections.
The experimental protocols used to stimulate the proposed triplet circuit are those ex-
plained in Section 2.5. The results presented in this section are due to the different
protocols to the proposed circuit.
Pairing Protocol
Fig. 7.3 shows that the proposed minimal triplet circuit can reproduce the exponential
learning window produced by both PSTDP and TSTDP models, under the conven-
tional pairing protocol described in Section 2.5.1 and adopted in many experiments (Bi
and Poo 1998, Wang et al. 2005). This exponential learning window can also be repro-
duced using a number of the previously described PSTDP circuits e.g. Bofill-I-Petit and
Murray (2004).
However, it has been illustrated in Sjöström et al. (2001) that altering the pairing rep-
etition frequency affects the total change in weight of the synapse. As it is shown
in Azghadi et al. (2011c) and Azghadi et al. (2012b) and was discussed in Chapter 6,
PSTDP circuits are not capable of reproducing such biological experiments that inves-
tigators examine the effect of changes in pairing frequency on synaptic weight. How-
ever, Fig. 7.4 illustrates how the proposed TSTDP circuit can readily reproduce these
experiments.
Page 166
Chapter 7 High-performance TSTDP VLSI Design
Figure 7.3. Exponential learning window produced by the proposed minimal TSTDP circuit.
The learning window was produced based on the pairing protocol for different transistor
process corners. The required bias currents taken for the triplet circuit correspond to
the hippocampal culture data set (Table 7.1). Experimental data and error bars are
extracted from Wang et al. (2005).
Triplet Protocol
There are two types of triplet patterns that are used in the hippocampal experiments,
which are also adopted in this chapter to compute the prediction error as described
in Section 2.5.3. Figures 7.5(a)-(b) show how the proposed minimal triplet circuit pro-
duces a close fit to the triplet experiments reported in Wang et al. (2005).
Quadruplet Protocol
Fig. 7.6 shows the weight changes produced by the proposed minimal TSTDP circuit
under quadruplet protocol conditions. This protocol is explained in detail in Sec-
tion 2.5.5. Identical to Pfister and Gerstner (2006), in all quadruplet experiments in
this chapter, ∆t = −∆t1 = ∆t2 = 5 µs. Note that none of the previously proposed
PSTDP circuits are capable of showing such a close fit, neither to triplet, nor to quadru-
plet experiments (Azghadi et al. 2011c, Azghadi et al. 2012b).
Page 167
7.4 Simulation Results
Figure 7.4. Weight changes produced by the proposed minimal TSTDP circuit under a
frequency-dependent pairing protocol. This figure demonstrates weight changes in
a pairing protocol as a function of the pairing frequency, ρ. The synaptic weight changes
are reproduced by the proposed minimal TSTDP circuit for different transistor process
corners. Experimental data points and error bars are extracted from Sjöström et al.
(2001)—no data point at ρ = 30 Hz. The required bias currents taken for the triplet
circuit correspond to the visual cortex data set (Table 7.1).
Page 168
Chapter 7 High-performance TSTDP VLSI Design
Figure 7.5. Weight changes produced by the proposed minimal TSTDP circuit under triplet
protocol for two different spike triplet combinations. (a) This shows weight changes
induced by the triplet protocol for the pre-post-pre combination of spikes. The synaptic
weight changes are reproduced by the proposed minimal TSTDP circuit for different
transistor process corners. (b) This shows the same case as (a), but for the post-
pre-post combination of spikes. The required bias currents taken for the triplet circuit
correspond to the hippocampal culture data set (Table 7.1).
was utilised originates from hippocampal culture experiments, which examine pair-
ing, triplet and quadruplet protocols effects on synaptic weight. This data set consists
of 13 data points obtained from Table 2 of Pfister and Gerstner (2006) including (i) two
data points and error bars for pairing protocol in Fig. 7.3, (ii) three data points and er-
ror bars for quadruplet protocol in Fig. 7.6, and (iii) eight data points and error bars for
triplet protocol in Figures 7.5(a) and (b). This data set shows the experimental weight
change, ∆w, as a function of the relative spike timing ∆t, ∆t1 , ∆t2 and T under pairing,
triplet and quadruplet protocols in hippocampal culture.
Identical to Pfister and Gerstner (2006) that test their proposed triplet model simu-
lation results against the experimental data using a Normalised Mean Square Error
(NMSE) for each of the data sets, and similar to the error measurements in experiments
performed in Chapter 6, the proposed circuit is verified by comparing its simulation
results with the experimental data and ensuring a small NMSE value. The NMSE is
calculated using the following equation:
Page 169
7.4 Simulation Results
Figure 7.6. Weight changes produced by the proposed minimal TSTDP circuit under quadru-
plet protocol. The synaptic weight changes are reproduced by the proposed minimal
TSTDP circuit for different transistor process corners. The required bias currents taken
for the triplet circuit correspond to the hippocampal culture data set (Table 7.1). Ex-
perimental data points and error bars are after Wang et al. (2005).
!2
1
p ∆wiexp − ∆wicir
NMSE = ∑ , (7.5)
p i =1 σi
where ∆wiexp , ∆wicir and σi are the mean weight change obtained from biological ex-
periments, the weight change obtained from the circuit under consideration, and the
standard error mean of ∆wiexp for a given data point i, respectively; p represents the
number of data points in a specified data set (can be 10 or 13).
In order to minimise the resulting NMSEs for the circuit and fit the circuit output to
the experimental data, there is a need to adjust the circuit bias parameters and time
constants. This is an optimisation process of the circuit bias currents which results in
reaching a minimum NMSE value and so the closest possible fit to the experimental
data. In the following subsection, the optimisation method used to tune the circuit
bias currents is introduced.
Page 170
Chapter 7 High-performance TSTDP VLSI Design
Table 7.1. Minimal TSTDP circuit bias currents and the resulted NMSEs for the two data
sets. The visual cortex data set includes 10 data points from pairing frequency exper-
iments presented in Sjöström et al. (2001) and the hippocampal data set consists of
13 data points from pairing experiments (2 points), quadruplet experiments (3 points)
and two different triplet experiments (each one with 4 points) presented in Wang et al.
(2005).
In order to minimise the NMSE function mentioned above and achieve the highest
analogy to the experimental data, the circuit bias currents which tunes the required
parameters from the model should be optimised as it is the case for TSTDP model pa-
rameters (Eq. 7.1). For this purpose, Matlab and HSpice were integrated in a way to
minimise the NMSE resulted from circuit simulations using the Matlab built-in func-
tion fminsearch. This function finds the minimum of an unconstrained multi-variable
function using a derivative-free simplex search method. Table 7.1 demonstrates bias
currents achieved from the mentioned optimisation method in order to reach the min-
imum NMSE for the two sets of data: the visual cortex data set and the hippocampal
culture data set. The minimum obtained NMSEs for the visual cortex and hippocampal
data sets are also presented in Table 7.1. These results are consistent with the obtained
NMSEs using TSTDP model reported in Pfister and Gerstner (2006).
In addition to the above mentioned experiments that have been carried out in Pfister
and Gerstner (2006), the proposed design has been additionally tested for all possible
combination of spike triplets. Applied protocol and more explanation on these experi-
ments are provided in the following subsection.
Apart from reproducing the behaviour of the TSTDP model proposed by Pfister and
Gerstner (2006), the proposed circuit is also able to reproduce the observed weight
modifications for other combinations (rather than pre-post-pre or post-pre-post) of
Page 171
7.4 Simulation Results
spikes triplets which have not been explored in Pfister and Gerstner (2006), but have
been used in another set of multi-spike interaction experiments performed by Froemke
and Dan (2002). In these experiments, six different combinations of spike triplets in-
duce synaptic weight changes. These changes in Froemke and Dan (2002) have been
calculated according to a suppressive model described in Section 2.6.1.
The simulation protocol (for suppressive STDP model) as described in Froemke and
Dan (2002) is as follows; a third spike is added either pre- or post-synaptically to the
pre-post spike pairs, to form a triplet. Then this triplet is repeated 60 times at 0.2 Hz to
induce synaptic weight changes. Here, the same protocol has been used to stimulate
the proposed minimal TSTDP circuit. In this protocol, there are two timing differences
shown as ∆t1 = tpost − tpre which is the timing difference between the two most left
pre-post or post-pre spike pairs, and ∆t2 = tpost − tpre which is the timing difference
between the two most right pre-post or post-pre spike pairs.
Although the proposed circuit implements the triplet model presented in Pfister and
Gerstner (2006), and not the suppressive model in Froemke and Dan (2002), obtained
results shown in Figs 7.7(a)-(b) demonstrate qualitative regional agreement with the
reported results in Froemke and Dan (2002). Nonetheless, there is a direct contrast
between our circuit results and their results in the post-pre-post case of spike patterns.
Page 172
Chapter 7
High-performance TSTDP VLSI Design
Figure 7.7. Synaptic weight changes in result of the extra triplet protocol and using the proposed minimal TSTDP circuit. (a) Synaptic
weight changes in result of the extra triplet protocol for pre-post-post (top right triangle), post-post-pre (bottom left triangle) and
post-pre-post (right bottom square) combination of spikes. (b) Synaptic weight changes in result of the extra triplet protocol for pre-
post-pre (top left square), pre-pre-post (top right triangle) and post-pre-pre (left bottom triangle) combination of spikes. The required
bias currents taken for the triplet circuit correspond to the hippocampal culture data set (Table 7.1).
Page 173
7.4 Simulation Results
In this figure, each data point at each post-synaptic frequency, ρpost , demonstrates the
average value of weight changes for ten different realisations of post-synaptic and pre-
synaptic Poissonian spike trains. In addition, each error bar shows the standard de-
viation of the weight changes over these ten trials. The demonstrated results were
produced using the bias currents which correspond to the visual cortex data set (Table
7.1). In the circuit, Vpost(n−1) , Vpost(n) , V pre(n) and Vpre(n) are Poissonian spike trains
where ρpost , ρpost , ρpre and ρpre denote their average firing rates, respectively. The
three different curves presented in Fig. 7.8 display three different weight modification
thresholds. In the original BCM rule, these thresholds are related to the post-synaptic
Page 174
Chapter 7 High-performance TSTDP VLSI Design
Figure 7.8. The proposed TSTDP circuit can generate BCM-like behaviour. The required
bias currents for the circuit correspond to those used for the visual cortex data set
(Table 7.1). The three different curves show the synaptic weight changes according to
three different synaptic modification thresholds demonstrating the points where LTD
changes to LTP, which is controlled by the current Ipot2 . The threshold is adjustable
using the TSTDP rule parameters. In order to move the sliding threshold toward left
or right, the Ipot2 parameter can be altered as depicted in this figure. The rate of
pre-synaptic spike trains, ρpre , is 10 Hz in all experiments. Each data point shows the
mean value of the weight changes for 10 different trials and the error bars depict the
standard deviations of the weight changes for each value of ρpost for these trials.
firing rate, ρpost . Based on Pfister and Gerstner (2006), the modification threshold for
the all-to-all spike interactions can be expressed as
E ( A− τ A+ τ )
2 − 2 +
D
p
θ = ρpost p , (7.6)
(ρ0 A3+ τ+ τy )
D E
p
where is the expectation over the statistics of the pth power of the post-synaptic
ρpost
D E
p p
firing rate and ρ0 = ρpost for large time constants (10 min or more). However,
for the nearest-spike model which is the case for the proposed TSTDP circuit, it is
not possible to derive a closed form expression for the modification threshold based
p
on ρpost , however for post-synaptic firing rate up to 100 Hz, a similar behaviour to
what Eq. 7.6 presents is inferable from the simulation results (supplementary materials
of Pfister and Gerstner (2006)). The three different curves in Fig. 7.8 are the results of
three different values for Ipot2 currents which correspond to three different values of
Page 175
7.4 Simulation Results
A3+ . This simulation suggests that the proposed circuit not only can reproduce timing-
based experimental outcomes, but also can reproduce some rate-based synaptic weight
modifications.
To analyse how BCM-like behaviour emerges from TSTDP, we need to go through the
same analysis used by Pfister and Gerstner (2006). In this circumstance, the triplet
learning rule can be recast into a simpler form by considering the statistical properties
of TSTDP weight changes which leads to the following time averaged equation,
dw
= − A2− τ− ρpre ρpost + A2+ τ+ ρpre ρpost
dt
− A3− τ− τx ρ2pre ρpost + A3+ τ+ τy ρ2post ρpre ,
(7.7)
where ρpre and ρpost are the pre- and post-synaptic firing rates, respectively. The other
parameters in the above equation τ− , and τ+ , are time constants for the pair-based
contribution and τx , and τy are the corresponding time constants for the triplet-based
contribution of the original triplet learning rule by Pfister and Gerstner (2006).
By considering the mapping of Eq. 7.7 into a mathematically similar functional form of
the BCM rule, shown in Eq.2.6—following the method as described in Pfister and Ger-
stner (2006)—one can simply set A3− = 0 and for simplicity, keep A2− and A2+ constant
in Eq. 7.7. This gives rise to the following expression
dw
= − A2− τ− ρpre ρpost + A2+ τ+ ρpre ρpost
dt
+ A3+ τ+ τy ρ2post ρpre . (7.8)
The above equation, given an appropriate choice of parameter values, can mimic BCM-
like nonlinear weight change dynamics by keeping ρpre fixed and altering the value of
the ρpost ; under these conditions, one can numerically illustrate that the weight changes
as a function of increasing post-synaptic frequency, has a similar profile to the weight
changes of the original BCM rule as described by Eq. 2.6.
However, one must keep in mind an important aspect of the original BCM experi-
ments (Kirkwood et al. 1996, Cooper et al. 2004) in order not to introduce any mis-
conceptions about the original BCM rule. This aspect (excluding neuromodulatory
Page 176
Chapter 7 High-performance TSTDP VLSI Design
effects) is that the original experiments were conducted using increasing pre-synaptic
frequency of inputs (Kirkwood et al. 1996). It is a well-known and undisputed fact that
neurophysiological experiments have shown that pre-synaptic activity typically drives
post-synaptic responses, and changes in post-synaptic firing rate only occurs as a re-
sult of changes to input activity. Put simply, changes in post-synaptic firing cannot be
considered independent from changes in pre-synaptic activity, they are functionally re-
lated. Hence, in a more precise physiological terms, the firing rate of the post-synaptic
neuron really needs to be considered as a function of its pre-synaptic inputs. A more in-
formative analysis of the weight dynamics of the triplet rule should take this fact about
pre- and post-synaptic firing rate, i.e. ρpost = F(ρpre ), into account. Hence changing
the post-synaptic firing rates should really be driven by changes in pre-synaptic firing
rates, as they do in any neurophysiological setting; in this manner one can deduce a
more informative link between the plasticity model and the original BCM rule. Chang-
ing ρpost while keeping the pre-synaptic firing rate ρpre fixed, needs to be viewed with
caution as it represents a misinterpretation in the application of the original stimulus
protocol used in LTD/LTP experiment, despite leading to BCM-like weight changes.
As a check that our circuit can reproduce BCM-like behaviour which is driven by pre-
synaptic (rather than post-synaptic) activity, we have repeated our circuit simulations
but made the naive assumption that post-synaptic firing rate is a linear function of the
pre-synaptic firing rate, i.e. ρpost = Aρpre and for the sake of simplicity we let A = 1,
i.e ρpost = ρpre . Despite such a crude approximation, the circuit successfully was able
to mimic BCM-like behaviour where weight changes were pre-synaptically driven, as
illustrated in Fig. 7.9. In this figure, each data point shows the mean value of the weight
changes for 10 different trials and the error bars depict the standard deviations of the
associated weight changes.
Additionally, Matlab simulations were conducted using both the linear Poisson neuron
model and the Izhikevich model, in order to assess whether such models can reproduce
pre-synaptically driven BCM-like changes to synaptic strength. We found that in the
case of increasing the pre-synaptic activity, the resulting synaptic weight changes fol-
lowed a BCM-like profile where for low pre-synaptic activity, there was no alteration
to synaptic weight; for moderate levels of pre-synaptic activity, gave rise to depres-
sion (LTD) and for further increases in (pre-synaptic) activity led to potentiation (LTP).
Such a pre-synaptically driven BCM-like profile of synaptic change occurs for each
Page 177
7.5 Mismatch and Variation
Figure 7.9. The proposed TSTDP circuit can generate pre-synaptically driven BCM-like
weight changes. In this simulation, post-synaptic firing rate is a linear function of
pre-synaptic firing rate.
above stated neuron model and the results of these simulations are presented in Ap-
pendix A. These preliminary Matlab simulations were pursued in order to inform us
whether combining a circuit based model of a neuron with our TSTDP circuit will lead
to a circuit implementation capable of both timing and rate-based synaptic plasticity
changes.
Page 178
Chapter 7 High-performance TSTDP VLSI Design
In order to minimise the effect of variations and mismatch, the analog VLSI signal pro-
cessing guidelines proposed by Vittoz (1985) can be adopted. It should be acknowl-
edged that a complete elimination of mismatch and variations is not possible.
The proposed circuit uses a number of transistors operating in the subthreshold region
and also includes current mirrors. Therefore, it is expected that this circuit will be
susceptible to process variations. In order to show that the proposed design is process
tolerant, two types of analysis were performed here. First, the proposed design was
simulated using the worst case process corners of the AMS 0.35 µm CMOS model. The
simulation results shown in Figs. 7.3 to 7.6 demonstrate that under the process corners,
the proposed circuit functions within expectation (reasonable bounds) and can show
the expected behaviour in all cases. These figures show that there are slight variations
in the amplitudes, which can be adjusted by retuning the circuit’s bias currents. This
robustness suggests that the physical implementation of the proposed design would
be also robust and work within the expected design boundaries—Chapter 4 of Weste
and Harris (2005).
Furthermore, since the proposed design utilises current mirrors to copy currents and
set the required amplitudes and time constants of the TSTDP model, the effect of tran-
sistors mismatch on the circuit performance must be considered. Therefore as the sec-
ond variation analysis, the proposed circuit was examined against process variation
and transistor mismatch. For this purpose, a 1000 Monte Carlo (MC) runs were per-
formed on the proposed circuit in order to test its operational robustness.
The process variation scenario is as follows. All the circuit transistors go under a lo-
cal variation which changes their absolute parameter values in the typical model. The
process parameter that was chosen to go under these variations is the transistor thresh-
old voltage, which is one of the most important process parameters especially in the
proposed design consisting of transistors operating in the subthreshold region of op-
eration (Seebacher 2005). The parameters vary according to the AMS C35 MC process
statistical parameters. The threshold voltages of PMOS and NMOS transistors varied
according to a one sigma Gaussian distribution, which can change the threshold volt-
ages by up to 30 mV. This variation can be much less if larger transistor aspect ratios be
used for the design. According to Pelgrom’s law (Pelgrom et al. 1989), the variation has
a relation with transistors aspect ratio. According to the transistor sizing in the pro-
posed circuit, a σ(∆Vth ) equal to 10 mV might be faced. Under such a circumstance,
Page 179
7.5 Mismatch and Variation
a 1000 MC runs were performed on the proposed circuit for three different cases, as
described below.
As the first case of MC analysis, the circuit was stimulated by a pairing protocol to re-
produce the exponential STDP learning window in the presence of the mentioned local
variation. The circuit bias currents correspond to those used for the typical model and
hippocampal data set reported in Table 7.1. Note that Fig. 7.10 shows a 1000 MC analy-
sis performed on the proposed circuit. This figure also shows that the proposed circuit
is less susceptible to process variation and the overall LTP/LTD exponential behaviour
can be preserved. However, the strength of the proposed circuit is in its controllabil-
ity using the bias currents. The observed variations in the design can be alleviated
by means of readjusting the circuit bias currents. This tuning can be conducted even
after the circuit is fabricated. If the fabricated circuit performance changes from the ex-
pected characteristics, the circuit bias currents, which serve as inputs to the fabricated
chip, can be retuned to reach the required behaviour.
[V]
Figure 7.10. STDP learning windows produced in 1000 MC runs using the optimised bias
parameters for the hippocampal data set. Each curve represents the weight change
obtained from the minimal TSTDP circuit that is stimulated according to the pairing
protocol to reproduce the STDP window. In each run, a random variation of transistors
threshold voltages occurs. The circuit bias parameters are those used for the typical
transistor model for the hippocampal data set, which are reported in Table 7.1. The
inset figure shows the STDP experimental data extracted from Bi and Poo (1998).
Note the similarity between the simulation results and the experimental data.
Page 180
Chapter 7 High-performance TSTDP VLSI Design
The second analysis was performed under similar process variation conditions to the
first case, but this time the circuit was stimulated by the frequency-dependent pair-
ing protocol to reproduce the visual cortex data set and the resulting NMSEs were
computed for 1000 MC runs. The circuit bias currents correspond to those used for
the typical model, which are reported in Table 7.1. The obtained results are shown
in Fig. 7.11. Furthermore, as the third case, the same analysis was carried out for the
hippocampal data set and bias parameters presented in Table 7.1. Achieved results
are demonstrated in Fig. 7.12. Figures 7.11 and 7.12 show significant variations in the
value of NMSE compared to the typical transistor parameters that the circuit bias cur-
rents (see Table 7.1) were optimised for. Despite these deviations in the NMSE values
under process variations, they are easily treatable by retuning the bias currents.
Figure 7.11. NMSEs obtained to reproduce the visual cortex data set in 1000 MC runs,
using the optimised bias parameters for this data set. Each run represents a
NMSE value obtained from the minimal TSTDP circuit that is stimulated according
to the freqneucy-dependent pairing protocol to mimic the visual cortex data set. The
circuit bias parameters are those used for the typical transistor model for the visual
cortex data set, which are reported in Table 7.1. In each run, a random variation of
transistors threshold voltages occurs.
In the case of the visual cortex data set, the worst NMSE was almost 78 that is much
larger than a minimal NMSE obtained using the typical model, NMSE = 0.33. Also,
in the case of hippocampal data set, the worst NMSE is about 306, which is again
significantly bigger than the NMSE obtained using the typical model, NMSE = 1.74.
These major deviations can be significantly reduced by retuning circuit bias currents
Page 181
7.5 Mismatch and Variation
and optimising them to get a minimal NMSE, in the presence of process variation. It
means that, some bias tuning should be performed on the circuit to reach a minimal
NMSE comparable to the design target.
Figure 7.12. NMSEs obtained to reproduce the hippocampal data set in 1000 MC runs,
using the optimised bias parameters for this data set. Each run represents a
NMSE value obtained from the minimal TSTDP circuit that is stimulated according
to the triplet, qudruplet, and pairing protocols to mimic the hippocampal data set.
The circuit bias parameters are those used for the typical transistor model for the
hippocampal data set, which are reported in Table 7.1. In each run, a random variation
of transistors threshold voltages occurs.
As an example, the worst case of NMSE for the hippocampal data set (NMSE = 306.4)
is in the case of some big changes in the threshold voltages around 30 mV. In the pres-
ence of these parameter variations in the design, all circuit bias currents were adjusted
again and a new minimum NMSE was obtained. The achieved NMSE, which is equal
to 1.92, is consistent with the design expectations. The retuned circuit bias currents in
this case are given in Table 7.2. Using these new bias currents, the required behaviour
for the pairing experiment (Fig. 7.3), the quadruplet experiment (Fig. 7.6), as well as the
triplet experiment (Figures 7.5(a) and (b)) were well observed. The same approach was
considered for the visual cortex data set, and the worst NMSE(= 78) changed to an ac-
ceptable NMSE = 0.47 that can faithfully represent the required frequency-dependent
behaviour in the pairing visual cortex experiment shown in Fig. 7.4.
Both worst case and MC analysis performed on the circuit show the robustness and
the controllability of the design in the presence of physical variations. Hence, despite
Page 182
Chapter 7 High-performance TSTDP VLSI Design
Table 7.2. Retuned TSTDP circuit bias currents and the resulted NMSEs in the presence
of the worst case variation in 1000 MC runs. The table shows the NMSEs in the
presence of the worst case variation in 1000 MC runs shown in Figures 7.11 and 7.12.
NMSEs were equal to 78 and 306.4 for the visual cortex and the hippocampal data sets,
respectively, but they were brought back to the shown NMSEs by readjusting the circuit
bias currents from the values shown in Table 7.1.
the fact that the proposed design has some susceptibility to process variations, a post-
fabrication calibration is possible through retuning the bias currents of the design to
achieve a minimal NMSE to faithfully reproduce the needed learning behaviour.
Page 183
7.6 TSTDP Chip Measurement Results
Figure 7.13. The MN256R1 multi-neuron chip under the microscope. (a) This figure shows
the complete chip including neurons and plastic synapses. The TSTDP circuit block
is shown inside a red dashed box. (b) This photo micrograph shows further detail of
the TSTDP circuit block and its five large capacitors. Note that the relevant circuit
is inside the area surrounded by the red dashed box.
The chip was designed and fabricated using a 1-poly 6-metal 0.18 µm AMS CMOS
process. The utilised design kit was Hit-Kit 4.01, and the supply voltage is 1.8 V. The
layout of the implemented TSTDP circuit is shown in Fig. 7.14. The required silicon
area for the implemented circuit is 165 µm ×60 µm, from which almost 75 percent is
Page 184
Chapter 7 High-performance TSTDP VLSI Design
occupied by the five capacitors. In addition to the four time constants capacitors i.e.
Cpot1 , Cdep1 , Cpot2 and Cdep2 , as shown in Fig. 7.1, another capacitor is the Cw capacitor,
which has the same size as the time constant capacitors.
Figure 7.14. The layout of the TSTDP circuit implemented on the MN256R1 chip. The
layout shows various parts of the TSTDP circuit, including five large capacitors with
the size of 500 fF in order to reach biologically plausible time constants in the order of
ms, in contrary to the circuit simulations that were performed in an accelerated time
scale and therefore four of the capacitors related to the TSTDP time constants were
smaller and only the weight capacitor was large for retaining the synaptic weight for
longer time. The occupied area by the implemented TSTDP circuit is 165 µm × 60 µm.
Performed chip measurement results show that the implemented TSTDP device func-
tions as expected and correctly follows the weight update pattern of the TSTDP rule.
As an example, Fig. 7.15 depicts measurement results from the TSTDP circuit, while
it is stimulated with four required signals, namely pre, post, pre(n-1) and post(n-
1). These signals are generated as pulses with 1 ms width. Each signal is repeated
with a delay of 20 ms for a period of time. The time difference among pre and post-
synaptic spikes is set to be 10 ms. This time difference can be both increased and
decreased, which therefore results in different synaptic weight changes amplitudes,
which is shown as the fourth signal in Fig. 7.15. The figure shows that the synaptic
weight, W, is first depressed when a pre-synaptic spike arrives. This is due to a pre-
post-pre (the first pre in this pattern is not shown in the figure) spike triplet, which
results in depression, as expected. Next the synaptic weight is increased (potentiated),
which is in result of a post-pre-post (the first pre in this pattern is not shown in the
figure) spike combination, as expected. Finally, the synaptic weight is shown to be
depressed again at the arrival of the second pre signal shown in the figure. This de-
pression is in result of the pre-post-pre triplet, which is shown in the figure. Note that,
Page 185
7.6 TSTDP Chip Measurement Results
the circuit also needs a delayed version of pre- and post-synaptic spikes, pre(n-1) and
post(n-1), to function correctly. The figure only shows the post(n-1) signal, which is a
duplicate of post spike train, with 1 ms delay. Note that the circuit also needs to re-
ceive pre(n-1), which is a delayed version of pre spike train with 1 ms delay, to function
according to the TSTDP equation (see Eq. 7.1).
Figure 7.15 shows stronger potentiation than depression. As already discussed in Sec-
tion 7.2, in the TSTDP rule, the magnitude of potentiation and depression are deter-
mined by eight synaptic parameters including pair- and triplet-based potentiation and
depression time constants and magnitude parameters. The same set of parameters are
available in the implemented circuit as eight input analog biases that are set by a pro-
grammable bias generator device. In the shown measurement, the circuit is set in a
way to show stronger potentiation than depression.
Figure 7.15. Measurement results of the fabricated TSTDP circuit. This figure demonstrtaes
synaptic weight changes due to pre-post-pre, post-pre-post and pre-post-pre spike com-
binations from left to right. The synaptic weight signal, W, shows the voltage changes
across the weight capacitor. As expected, a pre-post-pre combination of spikes results
in depression, while a post-pre-post spike triplet results in potentiation. The magnitude
of weight change at the time of each spike is controlled by pair and triplet potentia-
tion and depression time constants and amplitude parameters (see circuit description
in Section 7.3 and equation 7.1) that are applied to the circuit as 8 different analog
biases.
Page 186
Chapter 7 High-performance TSTDP VLSI Design
The MN256R1 is composed of 256 low power IF neurons. The implemented TSTDP
learning circuit can be interfaced to 256 low power IF neurons on the MN256R1 chip,
using the available AER system already discussed in Sections 3.2 and 3.3. The AER
system is also useful to generate the required pre(n-1) and post(n-1) spike trains which
are delayed versions of pre- and post-synaptic spike trains.
7.7 Discussion
As it is shown in this chapter, the proposed VLSI design for the TSTDP learning rule
produces a very close fit to the experimental data. Having a minimal fitting error has
been the primary goal of the proposed design. Although this goal has been success-
fully satisfied, considering the challenges mentioned in Section 5.4, this design faces
some challenges when integrated in a large-scale SNN, to be used in a learning or
computation application. One of the first challenges that is generic to almost all VLSI
implementations of synaptic plasticity models is the silicon real estate used by the de-
sign and its weight storage technique. In the proposed design, a large capacitor has
been utilised to store the synaptic weights for long period of times in the orders of
hundreds of ms. However, in a large-scale neuromorphic system having such large
capacitors for storing the synaptic weight is not practicable. A number of methods to
tackle the weight storage technique problem was discussed in Section 5.4.8.
A counter approach that is appropriate in the proposed design is to use a smaller ca-
pacitor along with a bistability circuit (Indiveri et al. 2006). However, this technique
has its own problem, as it confines the synaptic weights to two final steady states and
therefore, the analog nature of the weight is compromised. In addition, integrating the
TSTDP circuit with a memristive memory element is also another option to save silicon
area (Azghadi et al. 2013d). This approach has its own problems such as high variabil-
ity though. Besides, the use of compact CMOS memory elements to store the synaptic
weight (Seo et al. 2011) is another promising solution for storing the synaptic weight.
Either of the mentioned methods are useful to decrease the silicon are required for
weight storage, however none of them decrease the large area required for the TSTDP
circuit itself that contains 37 transistors, and four capacitors. In the next chapter, a
new design will be proposed that minimise the number of transistors used to realise
the TSTDP rule, and decrease the size of the weight capacitor by a factor of 10, while
loosing just a little of synaptic weight change prediction ability.
Page 187
7.7 Discussion
In addition to the weight storage, another challenge the proposed design faces to be
used in large-scale systems is its susceptibility to process variation, similar to all analog
VLSI designs in micro/nano meter design regime. Although the presented MC simula-
tion analysis and the result from fine-tuned circuits (presented in Section 7.5) show that
it is possible to minimise the effect of process variations, fine tuning of the circuit bias
parameters in a large spiking neural network of neurons and proposed triplet synapses
sounds to be not practical. Hence, a variation aware design technique is required that
take into account the process variation while designing the desired neuronal circuit.
The technique utilised to alleviate variations and mismatch in our design is using the
rules of transistor matching proposed by Vittoz (1985). However, in the worst case
variation condition, this approach is not adequate.
An interesting approach available in the literature that is useful for the proposed TSTDP
circuit when it is integrated in a large-scale neuromorphic system is the use of already
available AER framework in the system to fine-tune the circuit. This approach has been
successfully utilised in previous works for both small-scale (Neftci and Indiveri 2010)
and large-scale SNNs (Choudhary et al. 2012).
Another approach available in the literature, which is not useful for our proposed cir-
cuit design, is the design technique proposed and well utilised by Rachmuth et al.
(2011) and Meng et al. (2011) in neuromorphic modelling of ion channel and ionic dy-
namics. This variation aware design technique exploits source degeneration and neg-
ative feedback methods to increase the dynamic range of input voltages of transistors
and make them robust against mismatch errors that happen mainly because of the low
input voltage dynamic range in traditional subthreshold current mode circuits (Poon
and Zhou 2011). The exploitation of the source degeneration and negative feedback de-
sign techniques can be another alternative to build a network of neurons with TSTDP
synapses, which are not susceptible to process variations.
Apart from the above mentioned challenges, power consumption of the proposed de-
sign is another essential feature of a neuromorphic learning circuit, which should be
considered. The simulation results show that the proposed design consumes almost
60 pJ energy per spike, which is rather high comparing to some other synaptic plas-
ticity circuit implementations as presented in Table 5.1. Note that in a large-scale neu-
romorphic system that contains billions of synaptic plasticity circuit, saving a little of
energy per circuit is critical. Next chapter proposes a new TSTDP design that not only
Page 188
Chapter 7 High-performance TSTDP VLSI Design
is more compact, but also is more energy efficient than the current design. This be-
comes possible with the cost of losing a bit of weight modification accuracy.
A high performance synaptic plasticity circuit for the TSTDP rule was proposed in
this chapter. The presented results show that this new circuit is able to very closely
fit many biological experiments including rate-based (Bienenstock et al. 1982, Pfister
and Gerstner 2006), timing-based (Bi and Poo 1998, Wang et al. 2005), and hybrid rate
and timing-based experiments (Sjöström et al. 2008). It was also shown that the circuit
is able to closely match the experiments for the extra triplet patterns of a suppres-
sive STDP model presented in Froemke and Dan (2002). Furthermore, the rate-based
behaviour of the proposed timing-based circuit was examined under various experi-
mental protocols to stimulate the circuit both pre- and post-synaptically to mimic the
results presented in computational model experiments—post-synaptically driven—
(Izhikevich 2003, Pfister and Gerstner 2006) as well as the pre-synaptically driven bio-
logical experiments (Kirkwood et al. 1996).
In addition to its biological accuracy and plausibility, the circuit was investigated in
terms of process variation. As reported in the chapter, although the inherent analog
VLSI process variation may result in incorrect circuit operation, it is possible to retune
the circuit to perform its expected function. However, designing a circuit with the
same capabilities, but less prone to variation will be of high interest, since in this case
the functionality is not severely affected by the variations, so the circuit may work as
expected without the need for retuning.
Besides, the investigation on the power consumption of the circuit shows that, despite
having a high performance when reproducing the biological experiments, it consumes
relatively high energy, compared to some of the low energy designs available in the
literature (Cruz-Albrecht et al. 2012). Besides, it was discussed that this design that
utilises a large capacitor for storing the synaptic weight occupies large silicon area, a
problem that is generic to most of the synaptic plasticity learning circuits (Bofill-I-Petit
and Murray 2004, Indiveri et al. 2006, Koickal et al. 2007). Therefore, there is a need for
a new design that not only reproduces all the mentioned biological experiments with
an acceptable accuracy, but also is very low energy and compact.
Page 189
7.8 Chapter Summary
In the next chapter, a novel ultra low-energy, and compact design is presented that
improves the previous designs in all essential aspects, but in result its synaptic weight
modification ability is slightly compromised.
Page 190
Chapter 8
T
HIS chapter introduces a new accelerated-time circuit that has
several advantages over its previous neuromorphic counterparts,
which were discussed in previous chapters, in terms of compact-
ness, power consumption, and capability to mimic the outcomes of bio-
logical experiments. The proposed circuit is investigated and compared
to other designs in terms of tolerance to mismatch and process variation.
Monte Carlo (MC) simulation results show that the proposed design is
much more stable than its previous counterparts in terms of vulnerability to
transistor mismatch, which is a significant challenge in analog neuromor-
phic design. All these features make the proposed circuit an ideal device
for use in large scale spiking neural networks, which aim at implementing
neuromorphic systems with an inherent capability that can adapt to a con-
tinuously changing environment, thus leading to systems with significant
learning and computational abilities.
Page 191
8.1 Introduction
8.1 Introduction
The brain processes large amounts of data in real-time in the presence of noise, while
consuming little power. It also takes little space and has extraordinary processing fea-
tures. The ultimate goal for neuromorphic engineers is to develop a cybernetic sys-
tem, which closely mimics the capabilities of the brain. To reach this goal, as already
mentioned, understanding and implementing in silico the main components of cortical
networks, i.e. neurons and synapses, is a crucial first step.
Identical to neuron models, there are a variety of synaptic plasticity models. Some
of these models embrace certain features of real biological synapses, however they
tend to be complex in their (mathematical) formulation. On the other hand, other
models have been mathematically formulated to replicate the outcomes of a subset
of known experiments. Their representation is typically simpler in form allowing, in
some cases, reduced problematic translation into silicon. Generally, the main purpose
of such simplified rules is to mimic, as accurately as possible, the outcomes of various
experimental synaptic plasticity protocols (Mayr and Partzsch 2010, Morrison et al.
2008).
In this chapter, a new VLSI implementation of a malleable synaptic circuit that is capa-
ble of mimicking the outcomes of various synaptic plasticity experiments, is proposed.
It is demonstrated that the new design has a compact structure and possesses low
power features, which are required for VLSI implementations of large-scale spiking
neural networks. In addition, the robustness of the proposed circuit is verified against
transistor mismatch and process variations. The results show that the new circuit is a
reliable design in terms of transistor mismatch. These features make this new design
an ideal learning component that may benefit various VLSI synaptic plasticity systems.
The proposed circuit is of potential interest for future large scale neuromorphic circuits
Page 192
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
with significantly high numbers of neurons and synapses, where low power consump-
tion, compactness, accuracy and mismatch tolerance are absolutely essential.
The results shown in this chapter are presented mainly in PLOS ONE (Azghadi et al.
2014a), as well as in The 2013 IFIP/IEEE International VLSI/SOC conference (Azghadi et al.
2013c), and in The 20th IEEE International Conference on Electronics, Circuits and Sys-
tems (Azghadi et al. 2013b).
where the synaptic weight can be decreased (depressed) if a pre-synaptic spike oc-
curs, or can be increased (potentiated) at the time when a post-synaptic spike arrives.
Here, A2+ , A3+ and A2− , A3− are the potentiation and depression amplitude parame-
ters, respectively. In addition, ∆t1 = tpost(n) − tpre(n) , ∆t2 = tpost(n) − tpost(n−1) − ǫ
and ∆t3 = tpre(n) − tpre(n−1) − ǫ, are the time differences between combinations of pre-
and post-synaptic spikes, while ǫ is a small positive constant, which ensures that the
weight update uses the correct values occurring just before the pre or post-synaptic
spike of interest. In Eq. 8.1, τ− and τx are depression time constants, while τ+ and τy
are potentiation time constants (Pfister and Gerstner 2006).
In addition, it was pointed out that since the TSTDP rule utilises higher order temporal
patterns of spikes, it is shown to be able to account for the outcomes of several exper-
imental protocols including the frequency-dependent pairing experiments performed
in the visual cortex (Sjöström et al. 2001), or triplet, and quadruplet spike experiments
performed in the hippocampus (Wang et al. 2005). Note that, the PSTDP rule fails to
reproduce the outcomes of these experiments (see Chapter 6). This is due to a linear
summation of the effect of potentiation and depression in the PSTDP rule, while the
underlying potentiation and depression contributions in the TSTDP rule, do not sum
linearly (Froemke and Dan 2002).
Page 193
8.2 Minimal Representation of Triplet STDP Model
− ∆t − ∆t − ∆t
∆w+ = A+ e( τ+ 1 ) + A+ e( τy 2 ) e( τ+ 1 )
2 3
∆w = ∆t1 (8.2)
( )
∆w− = − A2− e τ− .
This model is able to account for quadruplet, triplet, and pairing (window) experi-
ments as shown in Pfister and Gerstner (2006) and Azghadi et al. (2013a). In addition to
the capability of simultaneously approximation of triplet, quadruplet and STDP win-
dow experiments with the same set of synaptic parameters, another minimal version
of TSTDP rule, is also capable of reproducing the results of the frequency-dependent
pairing experiments performed in the visual cortex (Sjöström et al. 2001). The minimal
model for this experiment can be shown as
− ∆t − ∆t
∆w+ = A+ e( τy 2 ) e( τ+ 1 )
3
∆w = ∆t (8.3)
( 1)
∆w = − A2− e τ− ,
−
which is simpler and utilises a lower number of synaptic parameters, and therefore
needs a new set of parameters, in comparison with the previous minimal model for
hippocampal experiments.
Besides the ability of reproducing timing-based experiments, the TSTDP rule has the
capability to demonstrate BCM-like behaviour. The BCM learning rule is an exper-
imentally verified (Dudek and Bear 1992, Wang and Wagner 1999) spike rate-based
synaptic plasticity rule, proposed in Bienenstock et al. (1982). Unlike STDP, which is a
spike-timing based learning rule, synaptic modifications resulting from the BCM rule
depends on the rate (activity) of the pre- and post-synaptic spikes (Bienenstock et al.
1982).
In the following section, a novel VLSI design for TSTDP rule is proposed that have
fewer number of transistors, smaller area, and lower power consumption, than all pre-
viously published circuits, yet with all their synaptic capabilities. These features make
this design an ideal learning component for large-scale neuromorphic circuits. It is
shown that the proposed circuit is able to faithfully reproduce the outcomes of many
biological experiments, when examined under experimental protocols mentioned in
Section 2.5.
Page 194
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
− ∆t2
− ∆t1
∆w+ = e( τ+ ) A+ + A+ e( τy )
2 3
∆w = ∆t1 − ∆t (8.4)
( )
∆w− = −e τ− A− + A− e( τx 3 ) .
2 3
The new TSTDP circuit is demonstrated in Fig. 8.1. This symmetric circuit operates
as follows: When a pre-synaptic spike, Vpre(n) , is received at the gate of M6 at tpre(n) ,
Vpot1 reaches ground resulting in switching on M8, and then starts to increase linearly
toward Vdd . The rate of this increase is determined by Vtp1 that is applied to the gate of
M5, and corresponds to the pairing potentiation time constants, τ+ , which is present
in both pairing and triplet potentiation terms as shown in the first line of Eq. 8.1.
Figure 8.1. Proposed circuit for the full TSTDP rule. This circuit corresponds to the modified
full TSTDP rule shown in Eq. 8.4. The minimal circuit that corresponds to the first
minimal TSTDP model shown in Eq. 8.2 does not include transistors M1-M4 shown
in the red dashed-box. Furthermore, the minimal TSTDP circuit that corresponds to
the second minimal TSTDP model shown in Eq. 8.3, does not include the M1-M4
transistors, nor the M7 transistor, shown in the blue dashed-box.
In fact, Vpot1 is a triangular voltage, which is controlled by the leaky integrator com-
posed of the output conductance of M5 and the gate capacitor of M8, to control the
Page 195
8.3 Proposed Low Energy and Compact STDP Circuit
existence of the potentiation in the first place and allows a current, Ipot , to flow through
the potentiation branches (M7-M9 and/or M15-M16-M8-M9) at the time of arrival of a
post-synaptic spike at M9, tpost(n) .
The linear increase of Vpot1 , which starts at tpre(n) , and leads to charging the weight
capacitor through M8 once tpost(n) arrives, is approximately proportional to
− ∆t1
( τ+ )
e ,
where ∆t1 = tpost(n) − tpre(n) and τ+ approximates by Vtp1 . This term is repeated twice
in the first line of Eq. 8.1, and can be factorised as it is shown in the first line of Eq. 8.4.
Furthermore, the addition term shown in the second term of first line of Eq. 8.4 that de-
termines the amount of potentiation as a result of both pair and triplet interactions, is
approximated through a sum of two currents that charge the weight capacitor, Cw , and
represent synaptic weight potentiation. The first current is controlled by the control-
lable voltage VA+ , while the second one is determined by both the second potentiation
2
dynamic Vpot2 , as well as the controllable voltage VA+ . When a post-synaptic spike
3
arrives at M18, Vpot2 reaches ground and after the post-synaptic pulse duration is fin-
ished, it starts to increase linearly toward Vdd . The rate of this increase is determined
by Vtp2 that is applied to the gate of M17, and corresponds to the triplet potentiation
time constants, τy . Therefore, the current flowing through M15-M16 can be an approx-
imation of
− ∆t2
( τy )
A3+ e ,
where ∆t2 = tpost(n) − tpost(n−1) . The current flowing through M15-M16 transistors
accumulates with the current flowing through M7 transistor (which is controlled by
gate voltage VA+ ) and forms the total current that is approximately proportional to
2
− ∆t2
( τy )
A2+ + A3+ e ,
and it represents an approximation of the second term of the first line of Eq. 8.4.
The same dynamic operates in the depression half of the proposed circuit, in which
currents flow away from the weight capacitor, Cw , and represent synaptic weight de-
pression. In this part, current sinks away from the weight capacitor through M10-
M12, if there has been a pre-synaptic action potential that arrives at M10, in a spec-
ified time window defined by Vtd1 (which corresponds to τ− ), after a post-synaptic
Page 196
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
spike arrives at M13. The amount of this current is first determined by the time dif-
ference between the pre- and post-synaptic spikes (∆t1 ) and then by the controllable
voltage, VA− . Therefore, this current approximates
2
∆t
( τ 1)
A2− e − ,
where ∆t1 = tpost(n) − tpre(n) . This is the pairing depression current that flows away
from the weight capacitor and results in depression due to post-pre spike pairs.
In addition, another current that can discharge the capacitor and results in depression,
will flow through M10-M11-M3-M4, if two conditions are satisfied. First, if there has
been a previous pre-synaptic spike, Vpre(n−1) , in a specified time window, set by Vtd2
(which corresponds to τx ), before the current pre-synaptic spike, Vpre(n) , arrives at M10
gate. And second, if a post-synaptic spike arrived at M13 gate in a specified time
window set by Vtd1 before the current and after the previous pre-synaptic spikes. The
magnitude of this current is first controlled by the time difference between the pre-
and post-synaptic spikes (∆t1 ), second with the time difference between the Vpre(n) and
Vpre(n−1) spikes, (∆t3 ), and then by controllable voltage, VA− . Therefore, this current
3
approximates
∆t1 − ∆t3
A3− e τ− e τx ,
where ∆t1 = tpost(n) − tpre(n) and ∆t3 = tpre(n) − tpre(n−1) . This is the triplet depression
current that flows away from the weight capacitor and results in depression due to
pre-post-pre spike triplet.
If the above two currents accumulate together, they form the depression term of both
Eqs. 8.1 and 8.4 that are equal as follows,
∆t ∆t1 − ∆t3 ∆t − ∆t3
( τ 1) ( τ 1)
− A2− e − − A3− e τ− e τx = −e − A2− + A3− e τx ), (8.5)
where the negative sign indicates that the current is depressive and that it flows away
from the weight capacitor.
Note that the above explanations contain assumptions that approximate the TSTDP
rule using our proposed circuit. However, from a circuit analysis point of view, if M3-
M4, M7-M12, and M15-M16 operate in the subthreshold regime (Liu et al. 2002), the
Page 197
8.4 Experimental Results
analytical expressions for Ipot and Idep , which are potentiation and depression currents,
respectively, can written as follows
I0
Ipot (t) = −(κ/UT )Vpot1 ( t− tpre(n) ) −(κ/UT )VA +
e +e 2+
(8.6)
I0
−(κ/UT )Vpot1 ( t− tpre(n) ) −(κ/UT )Vpot2 ( t− tpost(n−1)) −(κ/UT )VA ,
e +e +e 3+
I0
Idep (t) = −(κ/UT )Vdep1 ( t− tpost(n)) −(κ/UT )VA +
e +e 2−
(8.7)
I0
−(κ/UT )Vdep1 ( t− tpost(n)) −(κ/UT )Vdep2 ( t− tpre(n−1)) −(κ/UT )VA
e +e +e 3−
where tpre(n) and tpost(n) are current pre- and post-synaptic spike times respectively,
while tpre(n−1) and tpost(n−1) are the times at which the previous pre- and post-synaptic
spikes have arrived. Therefore, the voltage change in synaptic weight, shown as Vw in
Fig. 8.1, is approximated as:
+ Ipot (tpost(n) )
∆Vw = ∆tspk
Cpot1
∆Vw = (8.8)
Idep (tpre(n) )
∆Vw− = ∆tspk
C dep1
where ∆tspk are the width of pre- and post-synaptic spike pulses, and Cpot1 and Cdep1
are the parasitic capacitance at the gate of M8, and M11, respectively. Please note that,
in the proposed circuit, similar to the TSTDP model, whenever a pre-synaptic spike
arrives at tpre(n) , a depression can happen, while a potentiation can happen whenever
a post-synaptic spike arrives.
Below, experimental results of the proposed circuit are presented. In addition, the cir-
cuit is compared with previous synaptic plasticity circuits in terms of power consump-
tion, area and ability in reproducing the outcomes of various biological experiments.
This section provides information about the experimental setup, under which simula-
tions are performed. These simulations are carried out in order to verify the perfor-
mance of the proposed circuit and compare its performance with published synaptic
plasticity circuits in the literature.
Page 198
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
This chapter presents experimental results of two minimal TSTDP circuits that corre-
spond to the two minimal TSTDP models mentioned in previous paragraph. Accord-
ing to the minimal rules shown in both Eqs. 8.2 and 8.3, the depression contribution
of the spikes triplet interactions can be neglected without having a significant effect
on the circuit performance in reproducing the targeted biological experiments. The
triplet depression part in the full TSTDP circuit shown in Fig. 8.1, is the four transis-
tors surrounded in the red-dashed box. Therefore, the minimal TSTDP circuit, is the
one shown in Fig. 8.1 minus the part enclosed in the red-dashed box, i.e only 14 tran-
sistors are needed to regenerate all desired biological experiments, as it is shown in
Fig. 8.2. This is the first minimal TSTDP circuit.
In addition, the numerical simulation results suggest that, for generating the frequency-
dependent pairing experiments, as well as the BCM experiment, further to the triplet
depression part of the circuit, the pairing potentiation part is not also necessary and
can be removed (see Section 2.6.1). Therefore, in the case of second minimal TSTDP
rule, shown in Eq. 8.3, A2+ can be set to zero. As a result, one more transistor that
is shown in the blue dashed-box can be also removed from the proposed full TSTDP
circuit (Fig. 8.1) and therefore only 13 transistors are required for generating the men-
tioned pairing and BCM experiments (Pfister and Gerstner 2006). The resulting circuit
is shown in Fig. 8.3. This is the second minimal TSTDP circuit (Azghadi et al. 2013b).
Page 199
8.4 Experimental Results
Figure 8.2. First minimal TSTDP circuit. This circuit corresponds to the first minimal TSTDP
model shown in Eq. 8.2. Therefore, this circuit does not include the triplet depression
section of the full TSTDP circuit shown in Fig. 8.1.
Since there are two versions of the minimal TSTDP rule each corresponding to a mini-
mal TSTDP circuit, two sets of simulations are performed using the proposed minimal
circuits. Each simulation set considers a specific set of data from the experiments. The
first experimental data set that is utilised originates from hippocampal culture experi-
ments that examine pairing, triplet and quadruplet protocols effects on synaptic weight
change (Wang et al. 2005). This first data set consists of 13 data points obtained from
Table 2 of Pfister and Gerstner (2006). These data points include (i) two data points
and error bars for pairing protocol (ii) three data points and error bars for quadru-
plet protocol, and (iii) eight data points and error bars for triplet protocol. This data
set shows the experimental weight changes, ∆w, as a function of the relative spike tim-
ing ∆t, ∆t1 , ∆t2 and T under pairing, triplet and quadruplet protocols in hippocampal
culture.
Page 200
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
Figure 8.3. Second minimal TSTDP circuit. This circuit corresponds to the second minimal
TSTDP model shown in Eq. 8.3. Therefore, this circuit does not include the triplet de-
pression section of the full TSTDP circuit shown in Fig. 8.1, nor its pairing potentiation
part (Azghadi et al. 2013b).
The second data set originates from experiments on the visual cortex, which investi-
gated how altering the repetition frequency of spike pairings affects the overall synap-
tic weight change (Sjöström et al. 2001, Sjöström et al. 2008). This data set is composed
of 10 data points—obtained from Table 1 of Pfister and Gerstner (2006)—that repre-
sent experimental weight change, ∆w, for two different ∆t’s, and as a function of the
frequency of spike pairs under a frequency-dependent pairing protocol in the visual
cortex. Note that, the two mentioned data sets are those that were also used in the ex-
periments performed in the previous chapters, when other instances of PSTDP as well
as TSTDP circuits were verified for showing various synaptic plasticity experimental
outcomes.
Page 201
8.4 Experimental Results
The minimised circuits are simulated in HSpice using the 0.35 µm C35 CMOS pro-
cess by AMS. All transistors in the both designs (shown in Figs. 8.2 and 8.3) are set to
1.05 µm wide and 0.7 µm long. The weight capacitor value is set to 1 pF. It should be
noted that the circuits are simulated in an accelerated time scale of 1000 times com-
pared to real time, with all pulses having a 1 µs pulse width. This is the same approach
that has been utilised in previous chapters and by previous synaptic plasticity circuit
implementations such as Schemmel et al. (2006), Tanaka et al. (2009), Schemmel et al.
(2010), Mayr et al. (2010) and Wijekoon and Dudek (2012). For the sake of simplic-
ity when comparing simulation results to the biological experimental data, all shown
results are scaled back to real time.
Identical to the TSTDP computational simulations (Pfister and Gerstner 2006) and pre-
vious TSTDP circuit studies (Azghadi et al. 2011c, Azghadi et al. 2013a), which test
the triplet model/circuit simulation results against the experimental data using a Nor-
malised Mean Square Error (NMSE) for each of the data sets, the proposed circuit is
verified by comparing its simulation results with the experimental data and ensuring
a small NMSE value. The NMSE is calculated using Eq. 2.1.
In order to minimise the resulting NMSEs for the circuit and fit the circuit output to
the reported experimental data in the literature, there is a need to adjust the circuit
bias parameters and time constants. This is an optimisation process of the circuit bias
voltages, which results in reaching a minimum NMSE value and so the closest possible
fit to the experimental data.
In order to minimise the NMSE function and achieve the highest analogy to the exper-
imental data, the circuit bias voltages, which tunes the required parameters from the
Page 202
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
models should be optimised. For this purpose, Matlab and HSpice were integrated in a
way to minimise the NMSE resulted from circuit simulations using the Matlab built-in
function fminsearch. This function finds the minimum of an unconstrained multi-
variable function using a derivative-free simplex search method. This is the same op-
timisation function that was also used in the simulation results of the previous TSTDP
circuits.
The first simulation that is performed using the first proposed minimal TSTDP cir-
cuit shown in Fig. 8.2, is reproducing the STDP learning window that demonstrates
spike timing-dependent potentiation and depression, under pairing protocol (see Sec-
tion 2.5.1). Fig. 8.4 shows how the proposed circuit can successfully perform the tim-
ing dependent weight modifications. This figure shows the normalised experimental
data extracted from (Bi and Poo 1998) in blue. It suggests that the proposed circuit
behaviour under a pairing (window) protocol can approximate the experimental data
generated with the same protocol. Besides the blue experimental data, two other ex-
perimental values for ∆t = 10 ms and ∆t = −10 ms are shown with their standard
error mean represented by black bars. These points are the first two points of the 13
data points of the aforementioned first (hippocampal) data set. These two points, were
utilised to test and optimise the bias voltages of the first minimal TSTDP circuit. This
is a similar approach to the method used in Pfister and Gerstner (2006). The circuit
bias parameters for generating the STDP window are those corresponding to the hip-
pocampal data set as shown in Table 8.1.
Quadruplet Experiment
The second simulation is performed using the first minimal TSTDP circuit and under
quadruplet protocol. Fig. 8.5 demonstrates how the proposed circuit approximates the
timing dependent weight modifications close to those for quadruplet experiment. In
these results, the black data points are extracted from Wang et al. (2005), and the black
deviation bars and data points are those that were used in Pfister and Gerstner (2006)
Page 203
8.4 Experimental Results
Figure 8.4. STDP timing window experiment in the hippocampal region can be approxi-
mated using the first minimal TSTDP circuit. Synaptic weight changes are pro-
duced under pairing protocol. The circuit bias parameters for generating the window
approximation are those corresponding to the hippocampal data set shown in Table 8.1.
The first experimental data set shown in black contains two data points with their
standard error mean extracted from Pfister and Gerstner (2006), and the second exper-
imental data set is part of the normalised experimental data extracted from Bi and Poo
(1998).
for quadruplet experiments. The circuit bias parameters for generating the quadruplet
approximation are those corresponding to the hippocampal data set shown in Table 8.1.
Triplet Experiment
The third experiment that is performed on the first minimal TSTDP circuit is the triplet
experiment performed in the hippocampal region and reported in Wang et al. (2005)
and Pfister and Gerstner (2006). Fig. 8.6 demonstrates how the proposed circuit ap-
proximates the timing dependent weight modifications close to those for triplet exper-
iments. In the shown results, the black data and deviation bars are those that were used
in Wang et al. (2005) and Pfister and Gerstner (2006) for triplet experiments. The circuit
bias parameters for generating the triplet approximation are those corresponding to
the hippocampal data set as shown in Table 8.1.
Page 204
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
Figure 8.5. Quadruplet experiment in the hippocampal region can be approximated using the
first minimal TSTDP circuit. Synaptic weight changes are produced under quadru-
plet protocol. The circuit bias parameters for generating the quadruplet approximation
are those corresponding to the hippocampal data set as shown in Table 8.1. The exper-
imental data shown in black were extracted from Wang et al. (2005) and Pfister and
Gerstner (2006).
Simulation results show that the TSTDP circuit can distinguish between the pre-post-
pre and post-pre-post spike combinations and show analogy to the experiments. How-
ever, the simulation results using the computational pair-based STDP model shown in
Pfister and Gerstner (2006), as well as the results generated using different PSTDP cir-
cuits (Azghadi et al. 2011c, Azghadi et al. 2012b), demonstrate that the pair-based STDP
models and circuits do not have the ability to distinguish among triplet combinations.
Considering Figs. 8.4 to 8.6, the first proposed minimal TSTDP circuit, can reach a
good approximation of pairing, quadruplet, and triplet experiments, using a shared
optimised set of bias voltages. Using these bias voltages a NMSE = 2.04 is obtained,
when considering the 13 data points in the hippocampal data set. This is better than the
minimal NMSE obtained using the minimal TSTDP computational model, as presented
in Pfister and Gerstner (2006).
Page 205
8.4 Experimental Results
Table 8.1. Optimised biases for the minimal TSTDP circuits and two data sets. The presented
values show optimised bias parameters for the minimal TSTDP circuits in order to reach
the minimal NMSEs for the targeted set of data and experiments. The hippocampal
(first) set of optimised bias parameters generate the results shown for pairing, quadruplet
and triplet experiments. The visual cortex (second) set of optimised bias parameters are
used to reach the minimal NMSE in frequency-dependent pairing experiment.
Data set VA+ (V) VA− (V) VA+ (V) Vtp1 (V) Vtd1 (V) Vtp2 (V) NMSE
2 2 3
Besides the above experiments, which are similar to the experiments performed by
Pfister and Gerstner (2006), the proposed minimal circuit is additionally tested for all
possible combination of spike triplets under the same protocol that used by Froemke
and Dan (2002).
As already mentioned, in 2002 Froemke and Dan proposed a suppression model for
higher order spike trains and performed some experiments using the aforementioned
extra triplet protocol. Their proposed suppression model can account for the required
non-linearity in STDP experiments, when considering higher order of spike combi-
nations. Fig. 8.7 shows that the first minimal TSTDP circuit, under the extra triplet
protocol, and using the same set of parameters that were optimised for hippocam-
pal experiments (shown in Table 8.1), is able to account for a similar behaviour to the
experiments performed by Froemke and Dan in 2002 and for extra triplet patterns.
Nonetheless, there is slight contrast between the achieved results using the TSTDP cir-
cuits and those produced under the suppressive model and reported in Froemke and
Dan (2002). The result shown in the right bottom square of Fig. 8.7(a), which presents
synaptic weight changes due to post-pre-post, demonstrates potentiation. This is in
total agreement to the result shown in Fig. 8.6(b), which also shows potentiation for
post-pre-post spike combination. However, the suppressive model results show a de-
pression for this spike combination—Fig. 3b in Froemke and Dan (2002). Pfister and
Gerstner (2006) discussed that this difference is due to the nature of the suppressive
model, which gives rise to a depression when a post-pre-post spike triplet occurs, while
Page 206
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
Figure 8.6. Triplet experiments in the hippocampal region can be approximated using the
first minimal TSTDP circuit. Synaptic weight changes are produced under the triplet
protocol. The circuit bias parameters for generating the triplet approximation are those
corresponding to the hippocampal data set as shown in Table 8.1. The experimental
data, shown in black and their standard deviations extracted from Wang et al. (2005) and
Pfister and Gerstner (2006). (a) Simulation and experimental results for the pre-post-
pre combination of spike triplets with various timings. (b) Simulation and experimental
results for the post-pre-post combination of spike triplets with various timings.
clearly it leads to a potentiation in the TSTDP model. In a later study, Froemke et al.
(2006) revised their model in order to address this issue.
Page 207
8.4 Experimental Results
Figure 8.7. Extra triplet experiments using the suppression STDP model can be approxi-
mated using the first minimal TSTDP circuit. Synaptic weight changes in result
of extra triplet protocol for (a) pre-post-post (top right triangle), post-post-pre (bot-
tom left triangle) and post-pre-post (right bottom square) and (b) for pre-post-pre (top
left square), pre-pre-post (top right triangle) and post-pre-pre (left bottom triangle)
combination of spikes produced by the first minimal TSTDP circuit. The circuit bias
parameters for generating the synaptic weight changes shown in this figure correspond
to the hippocampal bias set shown in Table 8.1.
Page 208
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
As shown in Fig. 8.8, the optimised biases (shown in Table 8.1) for the second min-
imal circuit can closely approximate the outcomes of experiments under frequency-
dependent pairing protocol. The minimal obtained NMSE for this experiments was
0.39, which is close to the numerical simulation result of 0.34 reported in Pfister and
Gerstner (2006). It is worth mentioning that the second minimal TSTDP circuit has
only one transistor more than the simple PSTDP circuit proposed in Indiveri et al.
(2006), but it has the ability to reproduce the frequency-dependent pairing experi-
ments, while all neuromorphic PSTDP circuits, even with much higher number of
transistors—see (Bofill-I-Petit and Murray 2004, Tanaka et al. 2009, Bamford et al. 2012b)
for example—fail to replicate these experiments (Azghadi et al. 2011c).
Page 209
8.4 Experimental Results
Figure 8.8. Frequency-dependent pairing experiment in the visual cortex region can be ap-
proximated using the second minimal TSTDP circuit. Simulation results are pro-
duced under frequency-dependent pairing protocol (Azghadi et al. 2013b). The circuit
bias parameters for generating the synaptic weight changes shown in this figure corre-
spond to the visual cortex (second) set of bias parameters shown in Table 8.1. The
experimental data shown in black are extracted from Sjöström et al. (2001) and Pfister
and Gerstner (2006).
equal to 10 Hz, and the trains with this spiking rate, are regenerated for each data
point. Each data point shows the mean value of the weight changes for 10 various
post-synaptic Poissonian spike trains and the error bars depict the standard deviations
of the weight changes for each data points over 10 runs. In this experiment, similar
to the experiment performed in Pfister and Gerstner (2006), the frequency of the post-
synaptic spike, ρpost is swept over a range of frequencies from 0 Hz up to 50 Hz, while
the pre-synaptic spiking frequency, ρpre , is kept fixed at 10 Hz.
Although Pfister and Gerstner have used this methodology to show that their model
is able to reproduce a BCM-like behaviour, in the original BCM experiments reported
in Kirkwood et al. (1996), the synaptic weight changes were measured whilst the pre-
synaptic and not the post-synaptic spike rate was swept (Cooper et al. 2004). In order to
check that the proposed circuit could reproduce BCM-like behaviour, which is driven
by pre-synaptic activity, the circuit simulation was repeated. Similar to the experiments
presented in Chapter 7, we made this simple assumption that post-synaptic firing rate
Page 210
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
is a linear function of the pre-synaptic firing rate, i.e. ρpost = Aρpre and for the sake
of simplicity we let A = 1, i.e ρpost = ρpre . Despite such a crude approximation, the
circuit is successfully able to mimic BCM-like behaviour where weight changes were
pre-synaptically driven, as illustrated in Fig. 8.10. In this figure, each data point shows
the mean value of the weight changes for 10 different trials using random Poissonian
pre- and post-synaptic spike trains for each trial, and the error bars depict the standard
deviations of the associated weight changes over these 10 trials.
Figure 8.9. Post-synaptically driven BCM-like behaviour with sliding threshold feature can
be approximated using the second minimal TSTDP circuit. Simulation results are
produced under Poissonian protocol for BCM. The circuit bias parameters for generating
the synaptic weight changes shown in this figure correspond to the visual cortex (second)
set of bias parameters shown in Table 8.1. In this simulation, the pre-synaptic frequency,
ρpre , was kept fixed at 10 Hz, and the post-synaptic frequency, ρpost , was swept (see
the text for more details).
All these experiments suggest that the proposed timing-based circuit has sufficient
ability to replicate the outcome of other synaptic plasticity experiments, for BCM-like
behaviour. In the next section we discuss and compare the proposed circuit and its
counterparts from various circuit design as well as biological plausibility perspectives.
Page 211
8.5 Synaptic Plasticity Circuit Comparison
Figure 8.10. Pre-synaptically driven BCM-like behaviour with sliding threshold feature can
be approximated using the second minimal TSTDP circuit. Simulation results
are produced under Poissonian protocol for BCM. The circuit bias parameters for
generating the synaptic weight changes shown in this figure correspond to the visual
cortex (second) set of bias parameters shown in Table 8.1. In this simulation, the
pre-synaptic frequency, ρpre , was swept, while the neuron is linear and ρpre = ρpost
(see the text for more details).
Page 212
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
This table demonstrates that all PSTDP and TSTDP circuits are able to account for a
BCM-like behaviour. However, simulation results presented in Azghadi et al. (2012a)
suggest that, using a TSTDP circuit, a better BCM-like behaviour is attainable and since
there are more parameters available in the circuit, there will be a higher degree of
control over the sliding threshold of the BCM rule. In addition, there is no evidence,
if any of the circuits proposed in Mitra et al. (2009) or Meng et al. (2011) are capable of
showing a BCM-like behaviour with sliding threshold feature.
The table also summarises the ability of the proposed TSTDP circuit in reproducing
other required experiments. Although a number of other synaptic plasticity circuits
that are shown in the table, are also capable of qualitatively generating the required
experiments (Mayr et al. 2010, Rachmuth et al. 2011), they need changes in their synap-
tic parameters or in their initial implementations, in order to be able to mimic biological
experiments closely and with a small error. The table shows that the TSTDP designs
proposed in Azghadi et al. (2011c), Azghadi et al. (2012b) and Azghadi et al. (2013a)
as well as the proposed design in this chapter are able to account for all experiments
using shared set of bias parameters. This is a useful feature of the synaptic plasticity
circuit, to be able to reproduce as many experimental outcomes as possible, using a
single set of parameters, and by means of least changes to the hardware. As a result,
this new plasticity circuit can be used in developing large-scale networks of spiking
neurons with high synaptic plasticity abilities.
Page 213
8.5 Synaptic Plasticity Circuit Comparison
```
``` Experiment
``` STDP window Pairing frequency Triplet Quadruplet BCM
Plasticity Circuit ```
Ö Ö Ö
`
PSTDP (Bofill-I-Petit and Murray 2004) X X
PSTDP (Cameron et al. 2005) X Ö Ö Ö X
PSTDP (Indiveri et al. 2006) X Ö Ö Ö X
PSTDP (Schemmel et al. 2006) X Ö Ö Ö X
PSTDP (Arthur and Boahen 2006) X Ö Ö Ö X
PSTDP (Koickal et al. 2007) X Ö Ö Ö X
PSTDP (Tanaka et al. 2009) X Ö Ö Ö X
PSTDP (Ramakrishnan et al. 2011) X Ö Ö Ö X
PSTDP (Cassidy et al. 2011) X Ö Ö Ö X
PSTDP (Bamford et al. 2012b) X Ö Ö Ö X
PSTDP (Cruz-Albrecht et al. 2012) X Ö Ö Ö X
SDSP (Mitra et al. 2009) X* X** X** X** X**
Voltage-based BCM (Mayr et al. 2010) X X X X* X
Iono-neuromorphic (Meng et al. 2011) X* X** X** X** X**
Iono-neuromorphic (Rachmuth et al. 2011) X* X** X** X** X
TSTDP (Azghadi et al. 2011c) X X X X X
TSTDP (Azghadi et al. 2012b) X X X X X
TSTDP (Azghadi et al. 2013a) X X X X X
New low energy and compact TSTDP circuit X X X X X
are large and power hungry such as the designs presented in Bofill-I-Petit and Mur-
ray (2004), Mayr et al. (2010), Meng et al. (2011), Rachmuth et al. (2011), Azghadi et al.
(2012b) and Azghadi et al. (2013a). Some other designs such as the synaptic plasticity
circuits presented in Indiveri et al. (2006), Tanaka et al. (2009), Bamford et al. (2012b)
and Cruz-Albrecht et al. (2012), have improved power and area features, but do not
have most of the required biological abilities. Therefore, a circuit with low power and
area consumption and at the same time with high synaptic plasticity capabilities is re-
quired. The design presented in this chapter aims at reaching these goals. This design
Page 214
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
has high synaptic weight modification ability, while it is low power and occupies small
silicon area.
Since the proposed design only uses a small number of transistors to reach its required
synaptic plasticity features, compared to many previous designs with inferior or equal
synaptic capability, the area and power consumption in this design are lower than
all previous designs with similar capabilities, and close to other designs with much
lower synaptic strength. Table 8.3 compares the proposed design, with a number of the
previous synaptic plasticity designs available in the literature, in terms of complexity
(required number of transistors and capacitors), which has a direct relation with the
needed silicon area, and their estimated power consumption.
Power consumption of a synaptic plasticity circuit is directly linked to its synaptic bias-
ing parameters such as its synaptic time constants e.g. Vtp1 , Vtd1 , Vtp2 , Vtd2 , as well as its
synaptic amplitude parameters e.g. VA+ , VA− , VA+ , VA− . In addition, consumed power
2 2 3 3
is in a direct relation with the supply power, as well as the spike pulse width. There-
fore, in order to have a fair comparison among synaptic plasticity circuits, they should
all be compared under similar conditions. The presented results in the last six rows
of Table 8.3, depict the simulation results for various circuits under similar conditions.
The synaptic parameters, for all these synaptic circuits are firstly optimised to reach
the best NMSEs for the hippocampal data set. The optimisation process determines
the value of synaptic biasing parameters, which significantly influence the power con-
sumption of these circuits. For instance, the high power consumption observed in the
TSTDP circuit proposed in Azghadi et al. (2013a) is due to large time constants required
for reaching a small NMSE = 1.74, which results in transistors being on for longer pe-
riod of time and this leads to high power consumption. Table 8.3 reports the energy
consumption per spike for a number of the mentioned designs. The energy consump-
tion is measured on both pre-synaptic and post-synaptic spikes. Due to differences
in depression and potentiation biasing parameters, different energy consumptions are
measured for pre- and post-synaptic spikes, but the larger one is reported in Table 8.3.
The energy consumption per spike for the first three designs in Table 8.3, are extracted
from related papers. These circuits are PSTDP circuits, which do not possess the high
Page 215
Table 8.3. Area and power comparison for various synaptic plasticity circuits.
* The biases are optimised for the hippocampal (first) data set to reach minimal NMSEs and then the energy consumptions are measured.
** The PSTDP and TSTDP designs presented in the last six rows of this table are all simulated using a 3.3 V supply voltage, while other
designs use equal or lower supply voltages.
*** This design has been implemented in a 90 nm CMOS process with a supply voltage of 0.6 V.
hhhh
hhhhComparison Measure
hhh
hhh
hhhh Transistor No. Capacitor No. Energy per spike** NMSE*
Plasticity Circuit hhhh
Page 216
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
biological plausibility available in TSTDP circuits including the low power TSTDP de-
sign presented in this chapter. Although two of these designs are low power and con-
sumes very low energy per spike, they require a high number of transistors/capacitors
that require large silicon area. Note that in the best case, the NMSE of these designs that
implement the same STDP rule as the design presented and simulated in Azghadi et al.
(2011c) and Azghadi et al. (2012b), will be >10, which is not acceptable as a fitting error.
In addition, there is no energy consumption information available for the other three
designs shown in the fourth to sixth rows of the table. Two of these designs are
biophysically-based synaptic plasticity circuits, which are bulky detailed VLSI circuits
implemented with more than 100 transistors, and the other one that implements the
voltage-based BCM rule, imposes an inevitable interference with the neuron circuit
and also needs more than 100 transistors for the design (Mayr et al. 2010, Mayr and
Partzsch 2010).
Considering both area and power consumption, under similar conditions to other
synaptic plasticity circuits, Table 8.3 suggests that the proposed design outperforms
all other designs in terms of energy consumption, silicon real estate, and biological
accuracy.
This allows the proposed design to be a suitable learning and computational compo-
nent for large scale and low power neuromorphic circuits with high biological capabil-
ity. However, one should keep in mind that, any analog VLSI design will be affected
by the mismatch due to fabrication imperfections. Therefore, besides area and energy
consumption, mismatch may also be taken into account when considering design of
an analog synaptic plasticity circuit for learning and computational purposes.
As already mentioned in Chapter 5, apart from power consumption and silicon area,
transistor mismatch is another challenge that is always associated with all analog VLSI
Page 217
8.5 Synaptic Plasticity Circuit Comparison
designs, especially designs for synaptic plasticity circuits (see Section 5.4.2). The func-
tionality of these circuits are dependent on the synaptic parameters and changes in
the values of these parameters, which can occur due to process variations, result in
deviations from the synaptic circuit expected behaviour. These deviations can bring
about degradation of synaptic plasticity capability. The mismatch may be taken into
account from two different design perspectives. First, is a mismatch that occurs be-
tween the targeted design and the implemented design, and results in the physically
implemented transistor to be different from the designed one. Second, is a mismatch
that occurs among the transistors all over the fabricated design. These transistors sup-
pose to have similar behaviour and functionality inter- or intra-chip. The design of
large neuromorphic circuits become challenging due to these mismatches.
Transistor mismatch becomes more challenging when the transistor works in its sub-
threshold region of operation. This is due to the changes to the threshold of the tran-
sistor, which affect its subthreshold current characteristics. Due to the exponential be-
haviour and also low power consumption of transistors in their subthreshold regime,
many spiking neural circuits, including neurons and synaptic weight change compo-
nents are implemented in this region. In addition, many neuromorphic VLSI designs
employ mismatch susceptible components such as current mirrors and differential
pairs in their current- or voltage-mode structures. Therefore, these neural systems are
seriously susceptible to device mismatch (Azghadi et al. 2012b, Azghadi et al. 2013a,
Mayr et al. 2010, Poon and Zhou 2011).
Apart from the techniques to reduce the mismatch and/or alleviate its effect, in order to
have a process tolerant design, it is essential to use less components susceptible to mis-
match including current mirrors (Azghadi et al. 2012b, Azghadi et al. 2013a), differen-
tial pairs (Douglas et al. 1995), and OTAs (Cruz-Albrecht et al. 2012, Koickal et al. 2007).
Page 218
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
The low energy and compact design proposed in this chapter does not use any of these
components. Hence, it is less susceptible to process variations than many previous
designs.
Fig. 8.11 shows the variation in NMSE for the first (Hippocampal) data set, if a rigorous
case mismatch scenario occurs in the fabrication. In the applied scenario, all transis-
tors in the design independently go under a 1000 Monte Carlo (MC) threshold voltage
variation, with three standard deviations from their typical process technology thresh-
old voltage. This may cause deviations in the threshold voltage of any transistors up
to 30 mV. This level of variation in the thresholds of transistors is very unlikely to oc-
cur. This variation scenario was used in a previous design proposed in Azghadi et al.
(2013a), where under the same protocol the worst case NMSE can go up to 306 (See
Fig. 7.12). Therefore, the proposed design is much more robust compared to the pre-
vious designs and that is because of not using of process variation susceptible circuit
modules, such as current mirrors, which are extensively used in the previous designs
(See Fig. 7.1, as well as Fig. 6.5). Note that the circuit bias parameters for all 1000 MC
runs are fixed and correspond to the parameters for Hippocampal experiments shown
in Table 8.1. However, as the results presented in Azghadi et al. (2013a) show, the
bias parameters can be justified again and bring the circuit back to a significantly low
NMSE.
Identical to the mismatch analysis performed in Fig. 8.11, the proposed TSTDP cir-
cuit is subjected to another variation analysis, this time using the second minimal
TSTDP circuit and while stimulated under the frequency-dependent pairing proto-
col (see Fig. 8.8), in order to measure the variation effect. Fig. 8.12 represents 1000
MC runs, and the NMSE deviation, for the mismatch scenario explained earlier. The
NMSE obtained using the new proposed circuit is significantly smaller than that of the
designs presented in Azghadi et al. (2012b) and Azghadi et al. (2013a) and shown in
Chapters 6 and 7.
According to Figs. 8.11 and 8.12, in both cases of mismatch analysis, more than 60% of
NMSEs are very close to the best reached NMSEs in simulations. In addition, even the
worst NMSEs shown in these figures that are due to severe unlikely mismatch, are still
better than PSTDP circuit NMSEs even without considering variation in them.
Furthermore, it should be noted that, the applied variation scenario considers inde-
pendent changes in the design. This means that the threshold voltage of every sin-
gle transistor in the design changes independently, which is not likely in the case of
Page 219
8.5 Synaptic Plasticity Circuit Comparison
Figure 8.11. Transistor mismatch effects on the first minimal design. This figure shows 1000
Monte Carlo (MC) runs. In each run, the threshold voltage of all transistors are in-
dependently varied, based on a three-sigma deviation. The NMSE in each MC run
shows the fitting error of the design, which is affected by transistors threshold devia-
tions. Simulation results are produced under pairing, triplet and quadruplet protocols
and using the first minimal TSTDP circuit. The circuit bias parameters correspond to
those for the hippocampal region shown in Table 8.1.
closely positioned transistors in the proposed compact design. Considering this fact, a
mismatch tolerant synaptic circuit design is expected after fabrication. However, these
independent changes can happen globally and in the replicates of the proposed plastic-
ity circuit across the chip, in the case of a large scale neuromorphic design. This means
that shared fine-tuning for various sets of synaptic circuits, which are positioned in a
close neighbourhood on the chip, could be an effective way of tackling the mismatch
problem (Gao et al. 2012).
In general, Figs. 8.11 and 8.12, suggest that the proposed circuit is not heavily affected
by process variation, and an acceptable synaptic behaviour compatible with several
synaptic plasticity protocols is expected after fabrication. This feature along with low
power consumption, small area requirement, and high biological accuracy, make the
proposed circuit an ideal synaptic plasticity component that can be utilised in large-
scale neuromorphic systems. These systems will have higher capability to mimic more
biological experiments, while enjoying a compact structure, which consumes little
Page 220
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
Figure 8.12. Transistor mismatch effects on the second minimal design. This figure shows
1000 Monte Carlo (MC) runs. In each run, the threshold voltage of all transistors are
independently varied, based on a three-sigma deviation. The NMSE in each MC run
shows the fitting error of the design, which is affected by that run deviated transis-
tors thresholds. Similar to Fig. 8.8, simulation results are produced under frequency-
dependent pairing protocol and using the second minimal TSTDP circuit. The circuit
bias parameters correspond to those for the visual cortex region shown in Table 8.1.
8.6 Discussion
Despite the performance advantages that the proposed circuit presents, it has a number
of limitations that need to be considered when integrating it within a network config-
uration. As Fig. 8.1 demonstrates, in order to induce weight changes using the triplet
circuit, current pre- or post-synaptic spike, i.e. Vpre(n) or Vpost(n) , as well as the imme-
diate previous pre- or post-synaptic spike, i.e. Vpre(n−1) or Vpost(n−1) , are needed. This
results in the need for introducing a delay into the design that provides the circuit with
a delayed version of pre- and post-synaptic spike trains. Note that this is a limitation
that all previous TSTDP circuits also have.
Page 221
8.6 Discussion
In our simulation setup, we have delayed the input pre- and post-synaptic spike trains,
generated in software, for one spike width of 1 µs, and produced the required delayed
spike trains, i.e. Vpre(n−1) and Vpost(n−1) . However, in the physical implementation of
TSTDP circuits, the mentioned delay element should be combined with either neuron
or synapse circuit, in order to produce the required delayed spike trains. Since the
density of neurons is significantly lower than that of synapses in a neuromorphic sys-
tem, it is therefore preferred to integrate the required delay element into the neuron
design, hence saving precious silicon real estate and power. Another viable method
for implementing a delay into the system is to delay the spike while transmitting it
via an Address Event Representation (AER) protocol in the system. Since in the AER,
only spike time stamps are transferred, the spike time for any specified value, can be
easily delayed. Because the AER is an unavoidable part of any neuromorphic system,
it is beneficial to use AER instead of any extra circuitry (whether part of the neuron or
synapse) for introducing the required delay times into the system.
Another limitation in the proposed circuit is the use of a large weight capacitor, in order
to retain the synaptic weight for required period of times, needed for adopted exper-
imental protocols. Although this capacitor is much smaller than the weight capacitor
used in the design proposed in Chapter 7, considering a large-scale neuromorphic sys-
tem, there is a need to further minimise the size of capacitor. The utilised capacitor can
be implemented using Metal Oxide Semiconductor Capacitors (MOSCAPs), which ap-
proximately consumes up 20 × 20 µm2 of silicon real estate. Therefore, compared to
the full TSTDP circuit body that is composed of 18 transistors all with 1.05 µm width
and 0.35 µm length, the capacitor takes up about 90 % of the whole area required for
the TSTDP circuit.
In a recent study we have shown that a similar version of the proposed low energy
and compact circuit can use a 50 fF capacitor instead of the large 1 pF one, while re-
taining its ability to reproduce the STDP learning window, and the triplet and quadru-
plet experimental data (Azghadi et al. 2013c). This becomes possible if a modified
version of the experimental protocols is used. This modified protocol considers only
one pair, triplet or quadruplet of spikes, instead of the original protocols that use 60
spike sets with a frequency of 1 Hz (Pfister and Gerstner 2006). The design presented
in Azghadi et al. (2013c), cannot account for the frequency-dependent pairing exper-
iments, or other complicated experiments shown in this chapter, and is suitable only
for experiments with high spike frequencies.
Page 222
Chapter 8 Compact Low Energy Neuromorphic Circuit for Triplet STDP
On the contrary, the utilised experimental protocols in this chapter introduce 60 pairs,
triplet, or quadruplet of spikes with frequency of 1 Hz, into the TSTDP circuit, and the
resulting weight change is the summation of the weight changes of all these 60 spike
sets. Therefore, the synaptic weight change after each of these spike sets should be
strongly preserved during the rest period before the arrival of the next spike set, or
for longer times when there is no spike. As already discussed in Section 5.4.8, due to
the capacitor leakage, the synaptic weight stored on the capacitor, will leak away re-
sulting in the learnt weight will be eventually altered/lost. This is the reason why a
1 pF capacitor in the design in this chapter is employed to minimise this loss. Similarly,
many of the previous designs (Indiveri et al. 2006, Bofill-I-Petit and Murray 2004, Bam-
ford et al. 2012b), which only able to produce synaptic weight changes for the STDP
protocol, with only one spike pair, also utilised large capacitors, for the same reason.
However, with large capacitors, and even accelerated time, the leakage current still
has a significant effect on the stored synaptic weight value. In the performed simula-
tions throughout this thesis, the voltage difference between the synaptic weight values
stored on the capacitor, at the start of the experiments and just after the experiment
is finished, is reported. During the experiment, the leakage is not significant and can
be compensated for, using the parameter tuning performed for the STDP circuit under
test. However, after the experiment is finished, namely when there is no spike com-
ing, the updated weight stored on the capacitor will leak away in less than a second.
For an example, see the STDP measurement results from a similar accelerated-time
neuromorphic chip reported in Wijekoon and Dudek (2012).
One of the possible approaches that can be employed beside any STDP circuit, in-
cluding the proposed TSTDP circuits in this thesis is the use of a bistability circuit as
discussed in Section 5.4.8. However, even with the use of a bistable mechanism, the
final synaptic weight ought to be in a nonvolatile storage element for later use. There-
fore, there is always need for long-term synaptic weight storage. There exist a number
of nonvolatile weight storage methods in neuromorphic engineering such as (i) mem-
ory cells (Azghadi et al. 2013d), (ii) floating gate (Ramakrishnan et al. 2011), and (iii)
memristive devices (Zamarreño-Ramos et al. 2011, Azghadi et al. 2013d), which could
be utilised for this task. For further details see Section 5.4.8.
Page 223
8.7 Chapter Summary
Page 224
Chapter 9
T
HIS chapter summarises the presented study and gives conclud-
ing remarks on the research carried out in this thesis. It highlights
the original contributions the proposed research makes to the field
of neuromorphic engineering, and discusses how the research conducted in
this thesis extends the state of the art to elevate the ongoing research for re-
alising a large-scale neuromorphic system with capabilities close to that of
the brain. The chapter also provides ideas for future research to further
boost neuromorphic engineering. It also states the author’s outlook of the
filed of neuromorphic engineering and learning in spiking neural networks.
Page 225
9.1 Introduction
9.1 Introduction
This thesis discusses the design, implementation, application, and challenges of vari-
ous spike-based synaptic plasticity rules in silicon, especially those that have not been
explored yet (Azghadi et al. 2014c). It provides the reader with an insight on the pre-
vious and current states of VLSI synaptic plasticity circuits that have been utilised
in different applications and proposes new VLSI designs and implementations for a
novel STDP learning rule, that has not been presented in previous studies. The thesis
also shows how this timing-based rule is able to give rise to a rate-based learning be-
haviour observed in previous studies (Pfister and Gerstner 2006). Furthermore, for the
first time this timing-based rule is utilised to carry out a pattern classification task in
a neuromorphic system (Azghadi et al. 2014b). The original contributions presented in
different chapters of this thesis, which extend the state-of-the-art neuromorphic engi-
neering research for implementing a high performance VLSI spiking neural network,
capable of performing engineering tasks, are discussed in the following sections.
Page 226
Chapter 9 Conclusion, Future Work, and Outlook
• Pair-based STDP rule was successfully implemented using silicon neurons and
programmable synapses on the IFMEM chip. In order to test the correct func-
tionality of the implemented STDP learning rule, the spiking neural network on
the IFMEM chip was utilised to generate the well-known STDP learning window
presented in both biological experiments (Bi and Poo 1998, Wang et al. 2005) as
well as in computational studies (Song et al. 2000, Pfister and Gerstner 2006). The
produced window correctly follows both experimental and computational data.
Page 227
9.2 Implementing Spike Timing- and Rate-Based Synaptic Plasticity Rules on the
IFMEM Device for Pattern Classification
• The triplet-based STDP learning algorithm was implemented on the IFMEM neu-
romorphic hardware. Following computational experiments presented in Izhike-
vich (2003) and Pfister and Gerstner (2006), a rate-based BCM learning behaviour
was produced using the TSTDP learning rule, implemented on the IFMEM de-
vice (Azghadi et al. 2014b). The results show the excellent agreement between the
implemented neural network outputs and the outcomes of BCM computational
experiments (Pfister and Gerstner 2006).
• A perceptron like neural network was set up on the IFMEM device and the
synapses and neurons were programmed and tuned in a way that the device acts
as a pattern classification tool. The utilised learning algorithm for the performed
pattern classification task was the triplet-based STDP learning rule, which is
shown to be useful for learning and classification (Gjorgjieva et al. 2011) of rate-
based patterns. Obtained results show the high performance of the TSTDP rule in
real-time classification of complex correlated rate-based patterns (Azghadi et al.
2014b).
The preformed research in this part provides good view of the STDP and TSTDP
rules and their properties and features, which are essential when designing VLSI
STDP synapses. The above mentioned original contributions were described in
detail in Azghadi et al. (2014c).
• The use of the AER representation for receiving inputs, computing with spikes,
and transmitting signals in output, makes the IFMEM device an ideal compu-
tational platform for building embedded neuromorphic event-based computa-
tional systems that process events generated by neuromorphic sensory systems
(Liu and Delbrück 2010). Therefore, in a future work the programmed IFMEM
Page 228
Chapter 9 Conclusion, Future Work, and Outlook
chip that has been already adapted with the TSTDP (or can be adapted to any
other learning rule), can be interfaced to a neuromorphic sensory system such as
the Dynamic Vision Sensor (DVS) presented in Lichtsteiner et al. (2008).
• In the performed experiments, only PSTDP, TSTDP and BCM learning rules were
implemented and tested using the IFMEM device. However, in future stud-
ies, any synaptic plasticity rule of choice including complex and detailed bio-
physically grounded rules (Shouval 2011) as well as other simple or complicated
phenomenological rules (Clopath and Gerstner 2010, Graupner and Brunel 2012,
Uramoto and Torikai 2013) can be implemented and tested on the IFMEM de-
vice. This will provide us with a good comparison of various synaptic plastic-
ity rules performance in carrying out different applications while they are in-
terfaced to silicon neurons and have been utilised in a network configuration.
However, a current limitation of the IFMEM device is its limited number of neu-
rons and synapses (32 neurons and 1K synapses), which restricts the extension
of the required neural network size. To address this limitation, currently new
programmable neuromorphic devices are being developed in the NCS group of
INI. For example see the characteristics of the newly developed MNR256R1 chip
briefly described in Chapter 7.
The PSTDP learning algorithm has been implemented by various groups and under
different design strategies (Bofill-I-Petit and Murray 2004, Cameron et al. 2005, In-
diveri et al. 2006, Koickal et al. 2007, Tanaka et al. 2009, Bamford et al. 2012b). In
Page 229
9.3 Spike-based Synaptic Plasticity in Silicon: Design, Implementation, Application
and Challenges
this thesis, for the first time a number of these PSTDP designs were tested if they
are able to generate several synaptic plasticity experiments including the pairing ex-
periment for generating STDP learning window (Bi and Poo 1998, Wang et al. 2005),
triplet (Wang et al. 2005), quadruplet (Wang et al. 2005) and frequency-dependent pair-
ing experiments (Sjöström et al. 2001). In addition, two various PSTDP learning cir-
cuits, as representative for the class of PSTDP circuits, are also tested for their abil-
ities to mimic a rate-based BCM learning behaviour with sliding threshold feature
(Bienenstock et al. 1982, Cooper et al. 2004).
Original Contributions
• The previous VLSI implementation of the PSTDP rule presented in Indiveri et al.
(2006) was simplified to reduce area and power consumption. The result of this
study is presented in Azghadi et al. (2011b). This design is unable to account
for the mentioned triplet, quadruplet, and frequency-dependent pairing experi-
ments, since it also implements the PSTDP rule, which according to the compu-
tational studies cannot account for these experiments (Pfister and Gerstner 2006).
• For the first time a PSTDP learning circuit was utilised to generate a rate-based
BCM learning behaviour under a Poissonian protocol described in Section 2.5.6.
The results presented in Azghadi et al. (2012a) demonstrate that the PSTDP learn-
ing circuit, similar to the PSTDP learning computational model (Izhikevich and
Desai 2003), can generate a rate-based BCM learning behaviour.
Future Work
• The mentioned higher order spike experimental protocols, i.e. triplet, quadruplet
and frequency-dependent pairing protocols, were all used to stimulate the previ-
ous and current PSTDP VLSI circuits (Azghadi et al. 2011c, Azghadi et al. 2012b)
and show that these circuits are unable to account for several biological exper-
iments, due to their weakness in processing higher order spike combinations.
Page 230
Chapter 9 Conclusion, Future Work, and Outlook
However, in future studies, one might investigate the ability of some other spike-
based synaptic plasticity models/VLSI designs in reproducing the outcomes of
various timing-based (Wang et al. 2005, Pfister and Gerstner 2006) as well as hy-
brid rate/timing-based biological experiments (Sjöström et al. 2001). One of the
main synaptic plasticity rules discussed and reviewed in this thesis is the SDSP
learning model (Fusi et al. 2000, Brader et al. 2007), which has two various VLSI
implementations presented in Fusi et al. (2000) and Mitra et al. (2009). A future
research direction is to investigate the ability of this rule and its variant VLSI im-
plementations for mimicking the outcome of complicated timing- and rate-based
experiments.
The first VLSI designs for the triplet-based STDP learning circuit were proposed in this
thesis. These circuits were devised to overcome the deficiencies of PSTDP circuits in
synaptic plasticity experiments. Below sections provide a summary of original contri-
butions made in this relation.
Original Contributions
• The first VLSI design for the TSTDP learning rule was proposed. The new pro-
posed voltage-mode circuit presented in Azghadi et al. (2011d), is able to account
for many biological experiments, where the previous PSTDP circuits clearly fail.
It was first shown that this circuit is able to mimic the outcomes of a wide range
of synaptic plasticity experiments including timing-based, hybrid rate/timing-
based, and rate-based synaptic plasticity experiments (Azghadi et al. 2011d).
• In another study (Azghadi et al. 2011c), the proposed voltage-mode TSTDP and
a previous voltage-mode PSTDP VLSI design proposed by Indiveri et al. (2006)
were optimised and simulated under same experimental protocols and condi-
tions, to reproduce the outcome of various synaptic plasticity experiments. The
comparison of the results show that the TSTDP design significantly outperforms
the PSTDP design in closely mimicking the experimental data (Azghadi et al.
2011c).
Page 231
9.3 Spike-based Synaptic Plasticity in Silicon: Design, Implementation, Application
and Challenges
• Furthermore, the proposed voltage-mode TSTDP circuit was used (Azghadi et al.
2011a) to mimic a similar behaviour to the outcomes of a rate-based BCM exper-
iment (Izhikevich and Desai 2003, Pfister and Gerstner 2006). The achieved re-
sults demonstrate that this circuit closely mimics the sliding threshold behaviour
of the BCM rule (Azghadi et al. 2011a). In addition, the performance of the previ-
ous voltage-based PSTDP circuit presented in Indiveri et al. (2006) in reproducing
the BCM-like behaviour was also compared to the proposed TSTDP circuit. The
comparison shows that the TSTDP circuit has higher ability in mimicking the
required BCM-like behaviour (Azghadi et al. 2012a).
Page 232
Chapter 9 Conclusion, Future Work, and Outlook
complicated new experiments. The new proposed TSTDP circuit has a signifi-
cantly better performance in mimicking the various experiments in comparison
to the two previously developed TSTDP circuits, and shows lower data fitting
error (i.e. NMSE) as discussed in Chapter 7. It is shown how the new TSTDP
circuit is able to account for all previously mentioned experiments with higher
performance compared to the previous PSTDP and TSTDP designs, and also can
closely mimic the experimental data in new experiments such as experiments in-
volved with various spike triplet combinations, as well as pre-synaptic and post-
synaptic driven rate-based BCM-like experiments (Azghadi et al. 2013a), where
the previous TSTDP and PSTDP designs do not show suitable performance and
cannot mimic the experiments effectively.
• The new TSTDP circuit uses subthreshold transistors to reach the required ex-
ponential behaviour that is needed in the TSTDP computational model (Pfister
and Gerstner 2006). Therefore, the circuit is inherently prone to process varia-
tions and device mismatch due to the imperfect fabrication processes (Poon and
Zhou 2011, Azghadi et al. 2014c). In order to investigate the susceptibility of the
proposed design against device mismatch, the design underwent a severe de-
vice mismatch verification in 1000 MC simulation runs. The presented results
in Azghadi et al. (2013a) show that although the circuit is susceptible to process
variation, the effect of variations can be mitigated through a post-fabrication cal-
ibration technique to bring the circuit back to its desired behaviour even in the
presence of severe variations (Azghadi et al. 2013a).
• Since this new design enjoys a high-performance structure and is able to effi-
ciently reproduce all the required synaptic plasticity experimental data, it was
chosen to be fabricated in silicon. A proof of concept TSTDP circuit device was
successfully fabricated and tested as part of this thesis. The chip measurement
results presented in Section 7.6, show the correct functionality of the circuit that
was fabricated in an AMS 0.18 µm CMOS technology.
• The presented simulation results for the high-performance TSTDP circuit are per-
formed with the circuit including a 10 pF weight capacitor, which occupies a very
large portion of the proposed circuit. This large capacitor is needed to maintain
the synaptic weight value for long period of time, required to replicate the ex-
perimental data, under the same circumstances as those utilised in the original
Page 233
9.3 Spike-based Synaptic Plasticity in Silicon: Design, Implementation, Application
and Challenges
biological experiments (Pfister and Gerstner 2006). In order to address the prob-
lem of the very large capacitor, a new compact TSTDP circuit that uses a small
50 fF weight capacitor was proposed (Azghadi et al. 2013c). Due to the small ca-
pacitor used in this design, it is quite suitable for short term plasticity or when
processing inputs with high frequency. However, for lower spike frequencies
and long term plasticity, the latest updated value across the capacitor will be lost
during the circuit operation. In addition, this design is not able to account for the
hybrid rate/timing-based experiments presented in Sjöström et al. (2001), due to
its limited capacitor size (Azghadi et al. 2013c).
• The mentioned limitations in the design presented in Azghadi et al. (2013c) re-
sulted in utilising a large capacitor of 1 pF size. This design only uses a 1 pF
capacitor and 18 transistors, and in comparison with the previous TSTDP design
presented in Azghadi et al. (2013a), which uses 37 transistors, a 10 pF weight
capacitor, as well as four 100 fF time constant capacitors, needs a significantly
smaller silicon area. The presented results in Azghadi et al. (2013c) show that the
new compact circuit can account for the hybrid rate/timing-based experiments
presented in Sjöström et al. (2001).
• Further investigations on this new compact circuit show that a 14-transistor min-
imal version of this TSTDP circuit is able to account for the full set of experi-
ments reviewed in Section 2.5. This minimal circuit which is developed based
on the minimal TSTDP rule (Pfister and Gerstner 2006) has only two transis-
tors more than its smallest PSTDP counterpart presented in Indiveri et al. (2006).
The proposed design is not only more compact than all other previous TSTDP
(Azghadi et al. 2011c, Azghadi et al. 2012b, Azghadi et al. 2013a) and many of
the previous PSTDP designs (Bofill-I-Petit and Murray 2004, Cameron et al. 2005,
Koickal et al. 2007, Tanaka et al. 2009, Bamford et al. 2012b) in the literature, but
also it consumes lesser power than all these designs (Azghadi et al. 2014a). For
further details please refer to Table 8.3.
• The proposed compact circuit is also investigated and compared to other designs
in terms of tolerance to mismatch and process variation (Azghadi et al. 2014a).
Monte Carlo simulation results show that the proposed design, due to its cir-
cuit structure, is much more stable than its previous counterparts (Azghadi et al.
Page 234
Chapter 9 Conclusion, Future Work, and Outlook
2011d, Azghadi et al. 2012b, Azghadi et al. 2012b) in terms of vulnerability to tran-
sistor mismatch, which is a significant challenge in analog neuromorphic design
(Azghadi et al. 2014a).
Future Work
In a future work, wide dynamic range devices can be utilised to implement the
circuit and therefore minimise the mismatch effect (Rachmuth et al. 2011). In ad-
dition, off-chip event-based mismatch compensation strategies (Choudhary et al.
2012) are other viable methods useful for utilising the proposed synaptic plastic-
ity element in a large-scale neuromorphic system.
Page 235
9.3 Spike-based Synaptic Plasticity in Silicon: Design, Implementation, Application
and Challenges
storage. However, each of these approaches has its own limitations and advan-
tages. The proposed synaptic circuits employed the accelerated time design strat-
egy to minimise the effect of leakage on the synaptic weight capacitor. This ap-
proach has been utilised in many previous neuromorphic designs such as the
design presented in Schemmel et al. (2006), Tanaka et al. (2009), Schemmel et al.
(2010), Mayr et al. (2010), and Wijekoon and Dudek (2012). Although this ap-
proach can be useful in applications were high synaptic plasticity update speed
is required, it has the limitation of requiring higher bandwidth for spike com-
munication (Schemmel et al. 2010). In addition, a neuromorphic system utilising
this technique cannot simply be interfaced to sensory systems with biologically
plausible time constants.
• The synaptic weight in the proposed circuit is updated with the arrival of each
spike which leads to charging/discharging the weight capacitor, according to the
TSTDP rule. However, if there is no spike coming, or when the learning phase has
finished, the final synaptic weight must be stored for later use. Therefore, there is
an essential need for a non-volatile memory element to store the latest weight.
Many neuromorphic systems utilise memory cells and DACs to store and re-
store the synaptic weight when required (Arthur and Boahen 2006, Seo et al. 2011,
Pfeil et al. 2012, Azghadi et al. 2014b). In addition, Ramakrishnan et al. (2011) have
used a floating gate device to store the synaptic weight in a non-volatile fashion
on their single transistor synaptic device. Furthermore, in a recent study we have
used the non-volatile characteristic of memristor to implement a programmable
DAC (Azghadi et al. 2013d).
In future research, the TSTDP learning algorithm can be investigated using simi-
lar techniques for storing the synaptic weight. Therefore, one may utilise SRAM
cells along with ADC circuits and record the latest weight in the memory once
there is no more input spikes or after the learning phase finished or need to be
Page 236
Chapter 9 Conclusion, Future Work, and Outlook
stopped (Mitra et al. 2009). Also, one might utilise memristor features to imple-
ment the TSTDP rule, in a similar way to that utilised in Zamarreño-Ramos et al.
(2011), for implementing PSTDP. Furthermore, the TSTDP rule can be imple-
mented using two instances of the single transistor floating gate synaptic device
presented in Ramakrishnan et al. (2011).
• The developed knowledge gained through the course of the presented project
resulted in a reliable VLSI chip including a prototype TSTDP circuit that has been
tested and shown desired behaviour while being stimulated with artificial input
spikes. This circuit is now ready to be interfaced to silicon neurons and other
spike-based devices including neuromorphic sensory systems (Lichtsteiner et al.
2008, Liu and Delbrück 2010), for engineering applications such as the pattern
classification task carried out in the IFMEM device (Azghadi et al. 2014b).
9.4 Outlook
Page 237
9.4 Outlook
Another open challenge that is hindering progress in the design of large scale neuro-
morphic systems is the lack of appropriate EDA tools to assists neuromorphic design-
ers in the design, verification, and testing phases. As already mentioned in Chap-
ter 5, currently there are several promising design automation tools for generating
asynchronous logic circuits that are helpful for designing interconnecting circuits in
large-scale neuromorphic systems, but further developments for mixed analog/digital
design tools is needed. The area requirement for synaptic weight storage is another
challenge for large-scale neuromorphic systems. This can be addressed with the use of
newly developed resistive memory elements, which are integrable with CMOS tech-
nology, occupy small area, and consume little power (Indiveri et al. 2013). However,
these resistive elements are susceptible to variations and suffer from low yields, which
should be effectively addressed before utilising them in large-sale systems.
All these and other mentioned challenges are currently being addressed by an active
and enthusiastic research community. The small group of neuromorphic engineers that
was once limited to a dozen research laboratories around the world in the mid 90s is
now flourishing, with many more groups spread around the whole globe, and with
increasing support from both research funding organisations and strong industrial mi-
croelectronic groups.
In general, with the many efforts and initiatives that are being started in the field of
neuromorphic engineering, the future of this field is very promising, and the ongo-
ing research on implementations of learning mechanisms in neuromorphic systems is
likely to lead to systems that can be used in real-world applications in the near future.
Page 238
Appendix A
T
HE influence of high and low pre-synaptic rates on the BCM-
like behaviour produced using the proposed high-performance
TSTDP circuit that was presented in Chapter 7, is investigated.
It is also verified if the desired BCM-like behaviour is possible when a neu-
ron is integrated with the TSTDP synapse. The Izhikevich neuron model as
well as a linear Poisson neuron model are simulated along with a TSTDP
synapse and the resulting weight changes under the Poissonian protocol
mentioned in Section 2.5.6 are recorded. The simulation results show that,
in all cases, a well-shaped BCM-like behaviour with distinguishable sliding
thresholds can be achieved.
Page 239
A.1 Introduction
A.1 Introduction
According to the literature, STDP and BCM rules are related and a BCM-like weight
modification behaviour with sliding threshold that depends on the rate of the pre- or
post-synaptic spike trains, is an emergent property of the STDP rules such as PSTDP
(Izhikevich 2003) and TSTDP (Pfister and Gerstner 2006, Gjorgjieva et al. 2011). In
both cases of these timing-based plasticity rules, a BCM-like behaviour emerges when
synaptic weight modification changes are reported against changes in the post-synaptic
spike train rates. In this case, the sliding threshold of the BCM rule depends on the pa-
rameters of the utilised STDP model and can be modified accordingly as described
in Section 7.4.6 for TSTDP and in Izhikevich (2003) for PSTDP. The BCM protocol in
this case, which involves in sweeping the post-synaptic rate and recording the changes
in the synaptic weight accordingly, is mainly used in computational modelling of the
synaptic plasticity rules (Izhikevich 2003, Pfister and Gerstner 2006, Gjorgjieva et al.
2011). However, in the original BCM experiments performed by Kirkwood et al. (1996),
the pre-synaptic spike train rate is swept, while the post-synaptic firing rate is deter-
mined by the current synaptic weight and the dynamics of the neuron. Hence, in order
to test the response of the proposed circuit that implements a TSTDP model, while it is
pre-synaptically driven, a neuron model is required. Here, it is shown that using two
different neuron models, a BCM-like behaviour is achievable, when the TSTDP model
(circuit) is pre-synaptically (in contrary to the post-synaptically) driven.
Page 240
Appendix A Extra Investigations on the Proposed TSTDP Circuit
this figure, each data point at each post-synaptic frequency (ρpost ), is the average value
of the weight changes for ten different realisations of post-synaptic and pre-synaptic
Poissonian spike trains, where the error bar indicates the standard deviation.
Figure A.1. The proposed high-performance TSTDP circuit can generate BCM-like be-
haviour for various pre-synaptic spike rates. The three different curves show the
synaptic weight changes according to three different synaptic modification thresholds.
The thresholds that are controlled by the current Ipot2 , demonstrate the points where
LTD changes to LTP. The rate of pre-synaptic spike trains, ρpre , used in (a) and (b)
was 5 and 15 Hz, respectively. Each data point shows the mean value of the weight
changes for 10 different trials and the error bars depict the standard deviations of the
weight changes for each value of ρpost .
The demonstrated results were produced using the bias currents that correspond to the
visual cortex data set (see Table 7.1 for these values). The three different curves pre-
sented in Fig. A.1(a-b) display three different weight modification thresholds. These
curves are in the results of three different values for Ipot2 currents that correspond to
Page 241
A.2 Post-synaptically Driven BCM-like Behaviour
three different values of A3+ . These thresholds are related to the post-synaptic firing
rate, ρpost , for the rates up to 50 Hz, akin to previously reported results in Pfister and
Gerstner (2006). The simulation results show that if the mean pre-synaptic firing rate
decreases to 5 or increases to 15 Hz (in comparison to 10 Hz in the original experi-
ments), the post-synaptically driven BCM-like behaviour can be still preserved.
Figure A.2. Pre-synaptically driven BCM-like behaviour from Matlab simulations for the lin-
ear Poisson neuron model. This figure shows the synaptic weight changes produced
by the minimal TSTDP model (shown in Eq. 2.5), when integrated with a linear Pois-
sonian neuron. The three different curves represent three different BCM thresholds,
which are controlled by A3+ parameter of the TSTDP rule.
Figure A.3. Pre-synaptically driven BCM-like behaviour from Matlab simulations for the
Izhikevich’s neuron model. This figure shows the synaptic weight changes produced
by the minimal TSTDP model (shown in Eq. 2.5), when integrated with an Izhikevich
neuron. The three different curves represent three different BCM thresholds, which are
controlled by A3+ parameter of the TSTDP rule.
Page 242
Appendix A Extra Investigations on the Proposed TSTDP Circuit
Page 243
Page 244
Bibliography
A BARBANEL -H. D., H UERTA -R., AND R ABINOVICH -M. (2002). Dynamical model of long-term synaptic
plasticity, Proceedings of the National Academy of Sciences, 99(15), pp. 10132–10137.
A BBOTT-L. F., AND N ELSON -S. B. (2000). Synaptic plasticity: taming the beast, Nature Neuroscience, 3,
pp. 1178–1183.
A FIFI -A., AYATOLLAHI -A., AND R AISSI -F. (2009). Implementation of biologically plausible spiking neu-
ral network models on the memristor crossbar-based CMOS/nano circuits, European Conference
on Circuit Theory and Design, pp. 563–566.
A L -S ARAWI -S., A BBOTT-D., AND F RANZON -P. (1998). A review of 3-D packaging technology, IEEE
Transactions on Components, Packaging, and Manufacturing Technology, Part B: Advanced Pack-
aging, 21(1), pp. 2–14.
A MIT-D. J., AND F USI -S. (1994). Learning in neural networks with material synapses, Neural Compu-
tation, 6(5), pp. 957–982.
A NANTHANARAYANAN -R., E SSER -S. K., S IMON -H. D., AND M ODHA -D. S. (2009). The cat is out of the
bag: cortical simulations with 109 neurons, 1013 synapses, Proceedings of the Conference on High
Performance Computing Networking, Storage and Analysis, DOI: 10.1145/1654059.1654124.
A NDREOU -A. G., B OAHEN -K. A., P OULIQUEN -P. O., PAVASOVIC -A., J ENKINS -R. E., AND
S TROHBEHN -K. (1991). Current-mode subthreshold MOS circuits for analog VLSI neural systems,
IEEE Transactions on Neural Networks, 2(2), pp. 205–213.
A RENA -P., F ORTUNA -L., F RASCA -M., AND PATAN É -L. (2009). Learning anticipation via spiking net-
works: application to navigation control, IEEE Transactions on Neural Networks, 20(2), pp. 202–
216.
A RENA -P., F ORTUNA -L., F RASCA -M., PATAN É -L., AND S ALA -C. (2007). Integrating high-level sen-
sor features via STDP for bio-inspired navigation, IEEE International Symposium on Circuits and
Systems, pp. 609–612.
A RTHUR -J., M EROLLA -P., A KOPYAN -F., A LVAREZ -R., C ASSIDY-A., C HANDRA -S., E SSER -S., I MAM -
N., R ISK -W., R UBIN -D., M ANOHAR -R., AND M ODHA -D. (2012). Building block of a pro-
grammable neuromorphic substrate: A digital neurosynaptic core, The 2012 International Joint
Conference on Neural Networks, DOI: 10.1109/IJCNN.2012.6252637.
A RTHUR -J. V., AND B OAHEN -K. (2004). Recurrently connected silicon neurons with active dendrites for
one-shot learning, 2004 IEEE International Joint Conference on Neural Networks, Vol. 3, pp. 1699–
1704.
A RTHUR -J. V., AND B OAHEN -K. (2006). Learning in silicon: Timing is everything, Advances in Neural
Information Processing Systems, pp. 75–82.
Page 245
Bibliography
A ZGHADI -M. R., A L -S ARAWI -S., A BBOTT-D., AND I ANNELLA -N. (2013a). A neuromorphic VLSI de-
sign for spike timing and rate based synaptic plasticity, Neural Networks, 45, pp. 70–82.
A ZGHADI -M. R., A L -S ARAWI -S., A BBOTT-D., AND I ANNELLA -N. (2013b). Pairing frequency experi-
ments in visual cortex reproduced in a neuromorphic STDP circuit, 2013 IEEE International Con-
ference on Electronics, Circuits, and Systems, pp. 229–232.
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2011a). Emergent BCM via neu-
romorphic VLSI synapses with STDP, 5th Australian Workshop on Computational Neuroscience,
p. 31.
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2011b). Physical implementation of
pair-based spike-timing-dependent plasticity, EPSM-ABEC 2011 Conference, Vol. 34, art. no. 141.
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2012a). Design and implementa-
tion of BCM rule based on spike-timing dependent plasticity, Proc. IEEE 2012 International Joint
Conference on Neural Networks (IJCNN), DOI: 10.1109/IJCNN.2012.6252778.
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2012b). Efficient design of triplet
based spike-timing dependent plasticity, Proc. IEEE 2012 International Joint Conference on Neural
Networks (IJCNN), DOI: 10.1109/IJCNN.2012.6252820.
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2013c). A new compact analog VLSI
model for spike timing dependent plasticity, 2013 IFIP/IEEE 21st International Conference on Very
Large Scale Integration (VLSI-SoC), pp. 7–12.
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2014a). Tunable low energy, com-
pact and high performance neuromorphic circuit for spike-based synaptic plasticity, PLoS ONE,
9(2), p. art. no. e88326.
A ZGHADI -M. R., A L -S ARAWI -S., I ANNELLA -N., I NDIVERI -G., AND A BBOTT-D. (2014b). Spike-based
synaptic plasticity in silicon: Design, implementation, application, and challenges, Proceedings of
the IEEE, 102(5), pp. 717–737.
A ZGHADI -M. R., B ONYADI -M. R., AND S HAHHOSSEINI -H. (2007). Gender classification based on feed-
forward backpropagation neural network, Artificial Intelligence and Innovations 2007: from The-
ory to Applications, Springer US, pp. 299–304.
A ZGHADI -M. R., B ONYADI -M. R., H ASHEMI -S., AND M OGHADAM -M. E. (2008). A hybrid multipro-
cessor task scheduling method based on immune genetic algorithm, Proceedings of the 5th In-
ternational ICST Conference on Heterogeneous Networking for Quality, Reliability, Security and
Robustness, pp. 561–564.
A ZGHADI -M. R., K AVEHEI -O., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2011c). Novel VLSI
implementation for triplet-based spike-timing dependent plasticity, Proceedings of the IEEE Inter-
national Conference on Intelligent Sensors, Sensor Networks and Information Processing, pp. 158–
162.
A ZGHADI -M. R., K AVEHEI -O., A L -S ARAWI -S., I ANNELLA -N., AND A BBOTT-D. (2011d). Triplet-based
spike-timing dependent plasticity in silicon, The 21st Annual Conference of the Japanese Neural
Network Society, Japanese Neural Network Society, pp. P34–35.
Page 246
Bibliography
A ZGHADI -M. R., M ORADI -S., AND I NDIVERI -G. (2013d). Programmable neuromorphic circuits for
spike-based neural dynamics, 11th IEEE International New Circuit and Systems (NEWCAS) Con-
ference, DOI: 10.1109/NEWCAS.2013.6573600.
A ZGHADI -M. R., M ORADI -S., FASTNACHT-D., O ZDAS -M. S., AND I NDIVERI -G. (2014c). Pro-
grammable spike-timing dependent plasticity learning circuits in neuromorphic VLSI architec-
tures, ACM Journal on Emerging Technologies in Computing Systems, submitted.
B ADOUAL -M., Z OU -Q., D AVISON -A. P., R UDOLPH -M., B AL -T., F R ÉGNAC -Y., AND D ESTEXHE -A.
(2006). Biophysical and phenomenological models of multiple spike interactions in spike-timing
dependent plasticity, International Journal of Neural Systems, 16(02), pp. 79–97.
B AMFORD -S. A., M URRAY-A. F., AND W ILLSHAW-D. J. (2010). Synaptic rewiring for topographic map-
ping and receptive field development, Neural Networks, 23(4), pp. 517–527.
B AMFORD -S. A., M URRAY-A. F., AND W ILLSHAW-D. J. (2012a). Silicon synapses self-correct for both
mismatch and design inhomogeneities, Electronics Letters, 48(7), pp. 360–361.
B AMFORD -S., M URRAY-A., AND W ILLSHAW-D. (2012b). Spike-timing dependent plasticity with weight
dependence evoked from physical constraints, IEEE Transactions on Biomedical Circuits and Sys-
tems, 6(4), pp. 385–398.
B ARTOLOZZI -C., AND I NDIVERI -G. (2007). Synaptic dynamics in analog VLSI, Neural Computation,
19(10), pp. 2581–2603.
B ARTOLOZZI -C., M ITRA -S., AND I NDIVERI -G. (2006). An ultra low power current–mode filter for neu-
romorphic systems and biomedical signal processing, Proc. IEEE Biomedical Circuits and Systems
Conference, (BioCAS), pp. 130–133.
B ELATRECHE -A., M AGUIRE -L., AND M C G INNITY-M. (2006). Advances in design and application of
spiking neural networks, Soft Computing, 11(3), pp. 239–248.
B ELHADJ -B., T OMAS -J., B ORNAT-Y., D AOUZLI -A., M ALOT-O., AND R ENAUD -S. (2009). Digital map-
ping of a realistic spike timing plasticity model for real-time neural simulations, Proc. XXIV Con-
ference on Design of Circuits and Integrated Systems (DCIS), pp. 326–331.
B ENUSKOVA -L., AND A BRAHAM -W. C. (2007). STDP rule endowed with the BCM sliding threshold
accounts for hippocampal heterosynaptic plasticity, Journal of Computational Neuroscience, 22(2),
pp. 129–133.
B IENENSTOCK -E., C OOPER -L., AND M UNRO -P. (1982). Theory for the development of neuron selectiv-
ity: orientation specificity and binocular interaction in visual cortex, The Journal of Neuroscience,
2(1), pp. 32–48.
B I -G., AND P OO -M. (1998). Synaptic modifications in cultured hippocampal neurons: dependence on
spike timing, synaptic strength, and postsynaptic cell type, The Journal of Neuroscience, 18(24),
pp. 10464–10472.
B ILL -J., S CHUCH -K., B R ÜDERLE -D., S CHEMMEL -J., M AASS -W., AND M EIER -K. (2010). Compensat-
ing inhomogeneities of neuromorphic VLSI devices via short-term synaptic plasticity, Frontiers in
Computational Neuroscience, DOI: 10.3389/fncom.2010.00129.
Page 247
Bibliography
B LISS -T. V., AND C OLLINGRIDGE -G. L. (1993). A synaptic model of memory: long-term potentiation in
the hippocampus, Nature, 361(6407), pp. 31–39.
B OAHEN -K. A. (2000). Point-to-point connectivity between neuromorphic chips using address events,
IEEE Transactions on Circuits and Systems II: Analog and Digital Signal Processing, 47(5), pp. 416–
434.
B OFILL -A., M URRAY-A. F., AND T HOMPSON -D. P. (2001). Circuits for VLSI implementation of
temporally-asymmetric hebbian learning, Advances in Neural Information Processing Systems
(NIPS), Vol. 9, pp. 1091–1098.
B OFILL -I-P ETIT-A., AND M URRAY-A. (2004). Synchrony detection and amplification by silicon neurons
with STDP synapses, IEEE Transaction on Neural Networks, 15(5), pp. 1296–1304.
B ONYADI -M. R., A ZGHADI -S. M. R., AND H OSSEINI -H. S. (2007). Solving traveling salesman problem
using combinational evolutionary algorithm, Artificial Intelligence and Innovations 2007: from
Theory to Applications, Springer US, pp. 37–44.
B RADER -J., S ENN -W., AND F USI -S. (2007). Learning real-world stimuli in a neural network with spike-
driven synaptic dynamics, Neural Computation, 19(11), pp. 2881–2912.
B RETTE -R., AND G ERSTNER -W. (2005). Adaptive exponential integrate-and-fire model as an effective
description of neuronal activity, Journal of Neurophysiology, 94, pp. 3637–3642.
B UONOMANO -D. (2000). Decoding temporal information: A model based on short-term synaptic plas-
ticity, The Journal of Neuroscience, 20, pp. 1129–1141.
C AMERON -K., B OONSOBHAK -V., M URRAY-A., AND R ENSHAW-D. (2005). Spike timing dependent plas-
ticity (STDP) can ameliorate process variations in neuromorphic VLSI, IEEE Transactions on Neural
Networks, 16(6), pp. 1626–1637.
C ARLSON -K. D., D UTT-N., N AGESWARAN -J. M., AND K RICHMAR -J. L. (2013). Design space
exploration and parameter tuning for neuromorphic applications, Proceedings of the Ninth
IEEE/ACM/IFIP International Conference on Hardware/Software Codesign and System Synthe-
sis, CODES+ISSS ’13, IEEE Press, Piscataway, NJ, USA, pp. 20:1–20:2.
C ARLSON -K., N AGESWARAN -J., D UTT-N., AND K RICHMAR -J. (2014). An efficient automated parame-
ter tuning framework for spiking neural networks, Frontiers in Neuroscience.
C ASSIDY-A., A NDREOU -A. G., AND G EORGIOU -J. (2011). A combinational digital logic approach to
STDP, The 2011 International Symposium on Circuits and Systems, pp. 673–676.
C ASSIDY-A., D ENHAM -S., K ANOLD -P., AND A NDREOU -A. (2007). FPGA based silicon spiking neural
array, IEEE Biomedical Circuits and Systems Conference, pp. 75–78.
C HICCA -E., B ADONI -D., D ANTE -V., D’A NDREAGIOVANNI -M., S ALINA -G., C AROTA -L., F USI -S., AND
D EL G IUDICE -P. (2003). A VLSI recurrent network of integrate-and-fire neurons connected by
plastic synapses with long-term memory, IEEE Transactions on Neural Networks, 14(5), pp. 1297–
1307.
Page 248
Bibliography
C HICCA -E., W HATLEY-A., L ICHTSTEINER -P., D ANTE -V., D ELBR ÜCK -T., D EL G IUDICE -P., D OUGLAS -
R., AND I NDIVERI -G. (2007). A multi-chip pulse-based neuromorphic infrastructure and its ap-
plication to a model of orientation selectivity, IEEE Transactions on Circuits and Systems I, 5(54),
pp. 981–993.
C HOI -T. Y. W., S HI -B. E., AND B OAHEN -K. A. (2004). An on-off orientation selective address event rep-
resentation image transceiver chip, IEEE Transactions on Circuits and Systems I: Regular Papers,
51(2), pp. 342–353.
C HOUDHARY-S., S LOAN -S., F OK -S., N ECKAR -A., T RAUTMANN -E., G AO -P., S TEWART-T., E LIASMITH -
C., AND B OAHEN -K. (2012). Silicon neurons that compute, in A. Villa., W. Duch., P. Érdi., F. Ma-
sulli., and G. Palm. (eds.), Artificial Neural Networks and Machine Learning – ICANN 2012, Vol.
7552 of Lecture Notes in Computer Science, Springer Berlin / Heidelberg, pp. 121–128.
C HUA -L. (2011). Resistance switching memories are memristors, Applied Physics A, 102(4), pp. 765–
783.
C LOPATH -C., AND G ERSTNER -W. (2010). Voltage and spike timing interact in STDP—a unified model,
Frontiers in Synaptic Neuroscience, DOI: 10.3389/fnsyn.2010.00025.
C LOPATH -C., B ÜSING -L., VASILAKI -E., AND G ERSTNER -W. (2010). Connectivity reflects coding: a
model of voltage-based STDP with homeostasis, Nature Neuroscience, 13(3), pp. 344–352.
C OOPER -L., I NTRATOR -N., B LAIS -B., AND S HOUVAL -H. (2004). Theory of Cortical Plasticity, World
Scientific Pub Co Inc.
C RUZ -A LBRECHT-J. M., Y UNG -M. W., AND S RINIVASA -N. (2012). Energy-efficient neuron, synapse
and STDP integrated circuits, IEEE Transactions on Biomedical Circuits and Systems, 6(3), pp. 246–
256.
D AS G UPTA -B., AND S CHNITGER -G. (1994). The power of approximating: a comparison of activation
functions, Mathematical Research, 79, pp. 641–641.
D AVISON -A. P., AND F R ÉGNAC -Y. (2006). Learning cross-modal spatial transformations through spike
timing-dependent plasticity, The Journal of Neuroscience, 26(21), pp. 5604–5615.
D AYAN -P., AND A BBOTT-L. (2001). Theoretical Neuroscience: Computational and Mathematical Mod-
eling of Neural Systems, Taylor & Francis.
D EISS -S., D ELBR ÜCK -T., D OUGLAS -R., F ISCHER -M., M AHOWALD -M., M ATTHEWS -T., AND
W HATLEY-A. (1994). Address-event asynchronous local broadcast protocol, World Wide Web page.
http://www.ini.uzh.ch/˜amw/scx/aeprotocol.html.
D ELBR ÜCK -T., AND VAN S CHAIK -A. (2005). Bias current generators with wide dynamic range, Analog
Integrated Circuits and Signal Processing, 43(3), pp. 247–268.
D ELBR ÜCK -T., B ERNER -R., L ICHTSTEINER -P., AND D UALIBE -C. (2010). 32-bit configurable bias current
generator with sub-off-current capability, Proceedings of 2010 IEEE International Symposium on
Circuits and Systems (ISCAS), pp. 1647–1650.
D EMMING -A., G IMZEWSKI -J. K., AND V UILLAUME -D. (2013). Synaptic electronics, Nanotechnology,
doi:10.1088/0957-4484/24/38/380201.
Page 249
Bibliography
D OUGLAS -R., M AHOWALD -M., AND M EAD -C. (1995). Neuromorphic analogue VLSI, Annual Review
of Neuroscience, 18, pp. 255–281.
D’S OUZA -P., L IU -S.-C., AND H AHNLOSER -R. H. (2010). Perceptron learning rule derived from spike-
frequency adaptation and spike-time-dependent plasticity, Proceedings of the National Academy
of Sciences, 107(10), pp. 4722–4727.
D UDEK -S. M., AND B EAR -M. F. (1992). Homosynaptic long-term depression in area CA1 of hippocam-
pus and effects of N-methyl-D-aspartate receptor blockade, Proceedings of the National Academy
of Sciences, 89(10), pp. 4363–4367.
E BERHARDT-S., D UONG -T., AND T HAKOOR -A. (1989). Design of parallel hardware neural network
systems from custom analog VLSI ’building block’ chips, International Joint Conference on Neural
Networks, pp. 183–190.
E BONG -I., AND M AZUMDER -P. (2012). CMOS and memristor-based neural network design for position
detection, Proceedings of the IEEE, 100(6), pp. 2050–2060.
E RMENTROUT-B. (1996). Type I membranes, phase resetting curves, and synchrony, Neural Computa-
tion, 8(5), pp. 979–1001.
E SHRAGHIAN -K. (2010). Evolution of nonvolatile resistive switching memory technologies: The related
influence on hetrogeneous nanoarchitectures, Transactions on Electrical and Electronic Materials
(TEEM), 11(6), pp. 243–248.
E SHRAGHIAN -K., C HO -K.-R., K AVEHEI -O., K ANG -S.-K., A BBOTT-D., AND K ANG -S.-M. S. (2011).
Memristor MOS content addressable memory (MCAM): Hybrid architecture for future high per-
formance search engines, IEEE Transactions on Very Large Scale Integration (VLSI) Systems, 19(8),
pp. 1407–1417.
E SHRAGHIAN -K., K AVEHEI -O., C HAPPELL -J., I QBAL -A., A L -S ARAWI -S., AND A BBOTT-D. (2012).
Memristive device fundamentals and modeling: Applications to circuits and systems simulation,
Proceedings of the IEEE, 100(6), pp. 1991–2007.
FASNACHT-D., W HATLEY-A., AND I NDIVERI -G. (2008). A serial communication infrastructure for
multi-chip address event system, International Symposium on Circuits and Systems, (ISCAS), 2008,
IEEE, pp. 648–651.
F ITZ H UGH -R. (1961). Impulses and physiological states in theoretical models of nerve membrane, Bio-
physical Journal, 1(6), pp. 445–466.
F ORTUNA -L., F RASCA -M., AND X IBILIA -M. G. (2009). Chua’s Circuit Implementations, Yesterday, To-
day and Tomorrow, World Scientific.
F ROEMKE -R., AND D AN -Y. (2002). Spike-timing-dependent synaptic modification induced by natural
spike trains, Nature, 416(6879), pp. 433–438.
F ROEMKE -R., T SAY-I., R AAD -M., L ONG -J., AND D AN -Y. (2006). Contribution of individual spikes in
burst-induced long-term synaptic modification, Journal of Neurophysiology, 95(3), pp. 1620–1629.
F RYE -J., A NANTHANARAYANAN -R., AND M ODHA -D. S. (2007). Towards real-time, mouse-scale corti-
cal simulations, CoSyNe: Computational and Systems Neuroscience, p. 106.
Page 250
Bibliography
F URBER -S. B., L ESTER -D. R., P LANA -L. A., G ARSIDE -J. D., PAINKRAS -E., T EMPLE -S., AND B ROWN -
A. D. (2013). Overview of the spinnaker system architecture, IEEE Transactions on Computers,
62(12), pp. 2454–2467.
F USI -S., A NNUNZIATO -M., B ADONI -D., S ALAMON -A., AND A MIT-D. J. (2000). Spike-driven synaptic
plasticity: theory, simulation, VLSI implementation, Neural Computation, 12(10), pp. 2227–2258.
G AO -P., B ENJAMIN -B., AND B OAHEN -K. (2012). Dynamical system guided mapping of quantitative
neuronal models onto neuromorphic hardware, Journal of Neurophysiology, 59(10), pp. 2383–
2394.
G ERSTNER -W., AND K ISTLER -W. (2002). Spiking Neuron Models: Single Neurons, Populations, Plastic-
ity, Cambridge University Press.
G ERSTNER -W., K EMPTER -R., VAN H EMMEN -J., AND WAGNER -H. (1996). A neuronal learning rule for
sub-millisecond temporal coding, Nature, 383(6595), pp. 76–78.
G IULIONI -M., PANNUNZI -M., B ADONI -D., D ANTE -V., AND D EL G IUDICE -P. (2009). Classification of
correlated patterns with a configurable analog VLSI neural network of spiking neurons and self-
regulating plastic synapses, Neural Computation, 21(11), pp. 3106–3129.
G JORGJIEVA -J., C LOPATH -C., A UDET-J., AND P FISTER -J. (2011). A triplet spike-timing–dependent plas-
ticity model generalizes the Bienenstock–Cooper–Munro rule to higher-order spatiotemporal cor-
relations, Proceedings of the National Academy of Sciences, 108(48), pp. 19383–19388.
G OLDBERG -D. E. (1989). Genetic Algorithms in Search, Optimization, and Machine Learning, Vol. 412,
Addison-Wesley Reading Menlo Park.
G ORDON -C., AND H ASLER -P. (2002). Biological learning modeled in an adaptive floating-gate system,
IEEE International Symposium on Circuits and Systems, Vol. 5, pp. 609–612.
G RASSIA -F., B UHRY-L., L ÉVI -T., T OMAS -J., D ESTEXHE -A., AND S A ÏGHI -S. (2011). Tunable neu-
romimetic integrated system for emulating cortical neuron models, Frontiers in Neuroscience, DOI:
10.3389/fnins.2011.00134.
G RAUPNER -M., AND B RUNEL -N. (2012). Calcium-based plasticity model explains sensitivity of synap-
tic changes to spike pattern, rate, and dendritic location, Proceedings of the National Academy of
Sciences, 109(10), pp. 3991–3996.
G UETIG -R., A HARONOV-R., R OTTER -S., AND S OMPOLINSKY-H. (2003). Learning input correlations
through nonlinear temporally asymmetric Hebbian plasticity, The Journal of Neuroscience, 23(9),
pp. 3697–3714.
H AFLIGER -P. (2007). Adaptive WTA with an analog VLSI neuromorphic learning chip, IEEE Transac-
tions on Neural Networks, 18(2), pp. 551–572.
H AFLIGER -P., AND K OLLE R IIS -H. (2003). A multi-level static memory cell, Proceedings of the 2003
International Symposium on Circuits and Systems, Vol. 1, pp. I–25–I–28.
H AFLIGER -P., M AHOWALD -M., AND WATTS -L. (1997). A spike based learning neuron in analog VLSI,
Advances in Neural Information Processing Systems, 9, pp. 692–698.
Page 251
Bibliography
H AMILTON -T. J., AND TAPSON -J. (2011). A neuromorphic cross-correlation chip, 2011 IEEE Interna-
tional Symposium on Circuits and Systems (ISCAS), pp. 865–868.
H AMILTON -T. J., AND VAN S CHAIK -A. (2011). Silicon implementation of the generalized integrate-and-
fire neuron model, 2011 IEEE International Conference on Intelligent Sensors, Sensor Networks
and Information Processing (ISSNIP), pp. 108–112.
H AMILTON -T. J., J IN -C. T., VAN S CHAIK -A., AND TAPSON -J. (2008). An active 2-D silicon cochlea, IEEE
Transactions on Biomedical Circuits and Systems, 2(1), pp. 30–43.
H AMILTON -T. J., N ELSON -N. M., S ANDER -D., AND A BSHIRE -P. (2009). A cell impedance sensor based
on a silicon cochlea, IEEE Biomedical Circuits and Systems Conference, pp. 117–120.
H ASLER -J., AND M ARR -H. B. (2013). Finding a roadmap to achieve large neuromorphic hardware
systems, Frontiers in Neuroscience, DOI: 10.3389/fnins.2013.00118.
H AYKIN -S. (1994). Neural Networks: A Comprehensive Foundation, Prentice Hall PTR.
H EBB -D. (2002). The Organization of Behavior: A Neuropsychological Theory, Lawrence Erlbaum.
H ODGKIN -A. L., AND H UXLEY-A. F. (1952). A quantitative description of membrane current and its
application to conduction and excitation in nerve, The Journal of Physiology, 117(4), p. 500.
H OLLER -M., TAM -S., C ASTRO -H., AND B ENSON -R. (1989). An electrically trainable artificial neural
network (ETANN) with 10240 ’floating gate’ synapses, International Joint Conference on Neural
Networks, pp. 191–196.
H SIEH -H.-Y., AND TANG -K.-T. (2012). A spiking neural network chip for odor data classification, 2012
IEEE Asia Pacific Conference on Circuits and Systems (APCCAS), pp. 88–91.
H YNNA -K., AND B OAHEN -K. (2001). Space-rate coding in an adaptive silicon neuron, Neural Net-
works, 14(67), pp. 645–656.
I ANNELLA -N., AND TANAKA -S. (2006). Synaptic efficacy cluster formation across the dendrite via
STDP, Neuroscience Letters, 403(1-2), pp. 24–29.
I ANNELLA -N., L AUNEY-T., AND TANAKA -S. (2010). Spike timing-dependent plasticity as the origin
of the formation of clustered synaptic efficacy engrams, Frontiers in Computational Neuroscience,
DOI: 10.3389/fncom.2010.00021.
I MAM -N., A KOPYAN -F., A RTHUR -J., M EROLLA -P., M ANOHAR -R., AND M ODHA -D. (2012). A digital
neurosynaptic core using event-driven QDI circuits, 2012 18th IEEE International Symposium on
Asynchronous Circuits and Systems (ASYNC), pp. 25–32.
I NDIVERI -G. (2003). A low-power adaptive integrate-and-fire neuron circuit, Proceedings of the 2003
International Symposium on Circuits and Systems, Vol. 4, pp. 820–823.
I NDIVERI -G., AND H ORIUCHI -T. (2011). Frontiers in neuromorphic engineering, Frontiers in Neuro-
science, DOI: 10.3389/fnins.2011.00118.
I NDIVERI -G., C HICCA -E., AND D OUGLAS -R. (2006). A VLSI array of low-power spiking neurons and
bistable synapses with spike-timing dependent plasticity, IEEE Transactions on Neural Networks,
17(1), pp. 211–221.
Page 252
Bibliography
I NDIVERI -G., L INARES -B ARRANCO -B., H AMILTON -T., VAN S CHAIK -A., E TIENNE -C UMMINGS -R.,
D ELBR ÜCK -T., L IU -S., D UDEK -P., H ÄFLIGER -P., R ENAUD -S., S CHEMMEL -J., C AUWENBERGHS -
G., A RTHUR -J., H YNNA -K., F OLOWOSELE -F., S AIGHI -S., S ERRANO -G OTARREDONA -T.,
W IJEKOON -J., WANG -Y., AND B OAHEN -K. (2011). Neuromorphic silicon neuron circuits,
Frontiers in Neuroscience, DOI: 10.3389/fnins.2011.00073.
I NDIVERI -G., L INARES -B ARRANCO -B., L EGENSTEIN -R., D ELIGEORGIS -G., AND P RODROMAKIS -T.
(2013). Integration of nanoscale memristor synapses in neuromorphic computing architectures,
Nanotechnology, 24(38), p. art. no. 384010.
I NDIVERI -G., L IU -S., D ELBR ÜCK -T., AND D OUGLAS -R. (2009). Neuromorphic systems, Encyclopedia
of Neuroscience, pp. 521–528.
I NDIVERI -G., S TEFANINI -F., AND C HICCA -E. (2010). Spike-based learning with a generalized integrate
and fire silicon neuron, Proceedings of 2010 IEEE International Symposium on Circuits and Sys-
tems (ISCAS), pp. 1951–1954.
I ZHIKEVICH -E., AND D ESAI -N. (2003). Relating STDP to BCM, Neural Computation, 15(7), pp. 1511–
1523.
I ZHIKEVICH -E. M. (2001). Resonate-and-fire neurons, Neural Networks, 14(6), pp. 883–894.
I ZHIKEVICH -E. M. (2003). Simple model of spiking neurons, IEEE Transactions on Neural Networks,
14(6), pp. 1569–1572.
I ZHIKEVICH -E. M. (2004). Which model to use for cortical spiking neurons?, IEEE Transactions on
Neural Networks, 15(5), pp. 1063–1070.
J O -S. H., C HANG -T., E BONG -I., B HADVIYA -B. B., M AZUMDER -P., AND L U -W. (2010). Nanoscale mem-
ristor device as synapse in neuromorphic systems, Nano letters, 10(4), pp. 1297–1301.
K ASABOV-N., F EIGIN -V., H OU -Z.-G., C HEN -Y., L IANG -L., K RISHNAMURTHI -R., O THMAN -M., AND
PARMAR -P. (2014). Evolving spiking neural networks for personalised modelling, classification
and prediction of spatio-temporal patterns with a case study on stroke, Neurocomputing, 134,
pp. 269–279.
K ASABOV-N. K. (2014). Neucube: A spiking neural network architecture for mapping, learning and
understanding of spatio-temporal brain data, Neural Networks, 52, pp. 62–76.
K EMPTER -R., G ERSTNER -W., AND VAN H EMMEN -J. (1999). Hebbian learning and spiking neurons,
Physical Review E, 59(4), pp. 4498–4514.
K HOSLA -D., H UBER -D. J., AND K ANAN -C. (2014). A neuromorphic system for visual object recogni-
tion, Biologically Inspired Cognitive Architectures, DOI: 10.1016/j.bica.2014.02.001.
K IRKWOOD -A., R IOULT-M., AND B EAR -M. (1996). Experience-dependent modification of synaptic
plasticity in visual cortex, Nature, 381(6582), pp. 526–528.
K ISTLER -W. M., AND H EMMEN -J. L. (2000). Modeling synaptic plasticity in conjunction with the timing
of pre-and postsynaptic action potentials, Neural Computation, 12(2), pp. 385–405.
Page 253
Bibliography
K OICKAL -T., H AMILTON -A., TAN -S., C OVINGTON -J., G ARDNER -J., AND P EARCE -T. (2007). Analog
VLSI circuit implementation of an adaptive neuromorphic olfaction chip, IEEE Transactions on
Circuits and Systems I, 54(1), pp. 60–73.
K UDITHIPUDI -D., M ERKEL -C., S OLTIZ -M., R OSE -G., AND P INO -R. (2014). Design of neuromorphic
architectures with memristors, in R. E. Pino. (ed.), Network Science and Cybersecurity, Vol. 55 of
Advances in Information Security, Springer New York, pp. 93–103.
L AUGHLIN -S., AND S EJNOWSKI -T. (2003). Communication in neuronal networks, Science, 301(5641),
pp. 1870–1874.
L EE -S.-J., K AVEHEI -O., H ONG -Y.-K., C HO -T.-W., Y OU -Y., C HO -K., AND E SHRAGHIAN -K. (2010). 3-D
system-on-system (SoS) biomedical-imaging architecture for health-care applications, IEEE Trans-
actions on Biomedical Circuits and Systems, 4(6), pp. 426–436.
L ICHTSTEINER -P., P OSCH -C., AND D ELBR ÜCK -T. (2008). A 128 ×128 120 dB 15 µs latency asynchronous
temporal contrast vision sensor, IEEE Journal of Solid-State Circuits, 43(2), pp. 566–576.
L IKHAREV-K., AND S TRUKOV-D. (2005). CMOL: Devices, circuits, and architectures, Introducing Molec-
ular Electronics, pp. 447–477.
L INARES -B ARRANCO -B., AND S ERRANO -G OTARREDONA -T. (2003). On the design and characterization
of femtoampere current-mode circuits, IEEE Journal of Solid-State Circuits, 38(8), pp. 1353–1363.
L ISMAN -J., AND S PRUSTON -N. (2010). Questions about STDP as a general model of synaptic plasticity,
Frontiers in Synaptic Neuroscience, DOI: 10.3389/fnsyn.2010.00140.
L ISMAN -J. E. (1985). A mechanism for memory storage insensitive to molecular turnover: a bistable
autophosphorylating kinase, Proceedings of the National Academy of Sciences, 82(9), pp. 3055–
3057.
L IU -S.-C., AND D ELBR ÜCK -T. (2010). Neuromorphic sensory systems, Current Opinion in Neurobiol-
ogy, 20(3), pp. 288–295.
L IU -S.-C., D ELBR ÜCK -T., K RAMER -J., I NDIVERI -G., AND D OUGLAS -R. (2002). Analog VLSI: Circuits
and Principles, MIT Press, Cambridge, MA.
M ARKRAM -H., L ÜBKE -J., F ROTSCHER -M., AND S AKMANN -B. (1997). Regulation of synaptic efficacy
by coincidence of postsynaptic APs and EPSPs, Science, 275(5297), pp. 213–215.
M ARTIN -S., G RIMWOOD -P., AND M ORRIS -R. (2000). Synaptic plasticity and memory: an evaluation of
the hypothesis, Annual Review of Neuroscience, 23(1), pp. 649–711.
M ASQUELIER -T., AND T HORPE -S. J. (2007). Unsupervised learning of visual features through spike
timing dependent plasticity, PLoS Computational Biology, 3(2), p. art. no. e31.
M ASQUELIER -T., AND T HORPE -S. J. (2010). Learning to recognize objects using waves of spikes and
spike timing-dependent plasticity, International Joint Conference on Neural Networks (IJCNN),
DOI: 10.1109/IJCNN.2010.5596934.
M ASQUELIER -T., G UYONNEAU -R., AND T HORPE -S. J. (2008). Spike timing dependent plasticity finds
the start of repeating patterns in continuous spike trains, PLoS ONE, 3(1), p. art. no. e1377.
Page 254
Bibliography
M AYR -C., AND PARTZSCH -J. (2010). Rate and pulse based plasticity governed by local synaptic state
variables, Frontiers in Synaptic Neuroscience, DOI: 10.3389/fnsyn.2010.00033.
M AYR -C., N OACK -M., PARTZSCH -J., AND S CHUFFNY-R. (2010). Replicating experimental spike and
rate based neural learning in CMOS, Proceedings of IEEE International Symposium on Circuits
and Systems (ISCAS), pp. 105–108.
M AYR -C., PARTZSCH -J., N OACK -M., AND S CHUFFNY-R. (2013). Live demonstration: Multiple-
timescale plasticity in a neuromorphic system, 2013 IEEE International Symposium on Circuits
and Systems (ISCAS), pp. 666–670.
M C C ULLOCH -W. S., AND P ITTS -W. (1943). A logical calculus of the ideas immanent in nervous activity,
The Bulletin of Mathematical Biophysics, 5(4), pp. 115–133.
M EAD -C. (1990). Neuromorphic electronic systems, Proceedings of the IEEE, 78(10), pp. 1629–1636.
M ENG -Y., Z HOU -K., M ONZON -J., AND P OON -C. (2011). Iono-neuromorphic implementation of spike-
timing-dependent synaptic plasticity, 2011 Annual International Conference of the IEEE Engineer-
ing in Medicine and Biology Society, EMBC, pp. 7274–7277.
M EROLLA -P., AND B OAHEN -K. (2004). A recurrent model of orientation maps with simple and complex
cells, Advances in Neural Information Processing Systems 17 (NIPS 2004), pp. 1995–2002.
M EROLLA -P., A RTHUR -J., A KOPYAN -F., I MAM -N., M ANOHAR -R., AND M ODHA -D. S. (2011). A digital
neurosynaptic core using embedded crossbar memory with 45 pJ per spike in 45 nm, 2011 IEEE
Custom Integrated Circuits Conference (CICC), DOI: 10.1109/CICC.2011.6055294.
M EROLLA -P., A RTHUR -J., A LVAREZ -R., B USSAT-J.-M., AND B OAHEN -K. (2013). A multicast tree router
for multichip neuromorphic systems, IEEE Transactions on Circuits and Systems I: Regular Papers,
61(3), pp. 820–833.
M ITRA -S., F USI -S., AND I NDIVERI -G. (2009). Real-time classification of complex patterns using spike-
based learning in neuromorphic VLSI, IEEE Transactions on Biomedical Circuits and Systems, 3(1),
pp. 32–42.
M ITRA -S., I NDIVERI -G., AND E TIENNE -C UMMINGS -R. (2010). Synthesis of log-domain integrators for
silicon synapses with global parametric control, International Symposium on Circuits and Systems
(ISCAS), pp. 97–100.
M ORADI -S., AND I NDIVERI -G. (2011). A VLSI network of spiking neurons with an asynchronous static
random access memory, 2011 IEEE Biomedical Circuits and Systems Conference (BioCAS), pp. 277–
280.
M ORADI -S., AND I NDIVERI -G. (2014). An event-based neural network architecture with an asyn-
chronous programmable synaptic memory, IEEE Transactions on Biomedical Circuits and Systems,
8(1), pp. 98–107.
M ORADI -S., I MAM -N., M ANOHAR -R., AND I NDIVERI -G. (2013). A memory-efficient routing method
for large-scale spiking neural networks, 2013 European Conference on Circuit Theory and Design
(ECCTD), DOI: 10.1109/ECCTD.2013.6662203.
Page 255
Bibliography
M ORRIS -C., AND L ECAR -H. (1981). Voltage oscillations in the barnacle giant muscle fiber, Biophysical
Journal, 35(1), pp. 193–213.
M ORRISON -A., D IESMANN -M., AND G ERSTNER -W. (2008). Phenomenological models of synaptic plas-
ticity based on spike timing, Biological Cybernetics, 98(6), pp. 459–478.
M OSTAFA -H., C ORRADI -F., O SSWALD -M., AND I NDIVERI -G. (2013). Automated synthesis of asyn-
chronous event-based interfaces for neuromorphic systems, 2013 European Conference on Circuit
Theory and Design (ECCTD), DOI: 10.1109/ECCTD.2013.6662213.
N EFTCI -E., AND I NDIVERI -G. (2010). A device mismatch compensation method for VLSI neural net-
works, IEEE Biomedical Circuits and Systems Conference (BioCAS), pp. 262–265.
N EFTCI -E., B INAS -J., R UTISHAUSER -U., C HICCA -E., I NDIVERI -G., AND D OUGLAS -R. J. (2013). Syn-
thesizing cognition in neuromorphic electronic systems, Proceedings of the National Academy of
Sciences, 110(37), pp. E3468–E3476.
N ERE -A., O LCESE -U., B ALDUZZI -D., AND T ONONI -G. (2012). A neuromorphic architecture for object
recognition and motion anticipation using burst-STDP, PloS ONE, 7(5), p. art. no. e36958.
O’C ONNOR -D. H., W ITTENBERG -G. M., AND WANG -S. S.-H. (2005). Graded bidirectional synaptic
plasticity is composed of switch-like unitary events, Proceedings of the National Academy of Sci-
ences of the United States of America, 102(27), pp. 9679–9684.
O LIVERI -A., R IZZO -R., AND C HELLA -A. (2007). An application of spike-timing-dependent plasticity
to readout circuit for liquid state machine, International Joint Conference on Neural Networks,
pp. 1441–1445.
PAINKRAS -E., P LANA -L., G ARSIDE -J., T EMPLE -S., G ALLUPPI -F., PATTERSON -C., L ESTER -D., B ROWN -
A., AND F URBER -S. (2013). Spinnaker: A 1-W 18-core system-on-chip for massively-parallel neural
network simulation, IEEE Journal of Solid-State Circuits, 48(8), pp. 1–11.
P ELGROM -M., D UINMAIJER -A., AND W ELBERS -A. (1989). Matching properties of MOS transistors,
IEEE Journal of Solid-State Circuits, 24(5), pp. 1433–1439.
P ERSHIN -Y. V., AND D I V ENTRA -M. (2012). Neuromorphic, digital, and quantum computation with
memory circuit elements, Proceedings of the IEEE, 100(6), pp. 2071–2080.
P ETERSEN -C. C., M ALENKA -R. C., N ICOLL -R. A., AND H OPFIELD -J. J. (1998). All-or-none potentiation
at CA3-CA1 synapses, Proceedings of the National Academy of Sciences, 95(8), pp. 4732–4737.
P FEIL -T., P OTJANS -T. C., S CHRADER -S., P OTJANS -W., S CHEMMEL -J., D IESMANN -M., AND M EIER -
K. (2012). Is a 4-bit synaptic weight resolution enough?–constraints on enabling spike-
timing dependent plasticity in neuromorphic hardware, Frontiers in Neuroscience, DOI:
10.3389/fnins.2012.00090.
P FISTER -J., AND G ERSTNER -W. (2006). Triplets of spikes in a model of spike timing-dependent plasticity,
The Journal of Neuroscience, 26(38), pp. 9673–9682.
P ILLOW-J. W., S HLENS -J., PANINSKI -L., S HER -A., L ITKE -A. M., C HICHILNISKY-E., AND S IMONCELLI -
E. P. (2008). Spatio-temporal correlations and visual signalling in a complete neuronal population,
Nature, 454(7207), pp. 995–999.
Page 256
Bibliography
P OON -C., AND Z HOU -K. (2011). Neuromorphic silicon neurons and large-scale neural networks: chal-
lenges and opportunities, Frontiers in Neuroscience, DOI: 10.3389/fnins.2011.00108.
R ACHMUTH -G., AND P OON -C.-S. (2008). Transistor analogs of emergent iono-neuronal dynamics,
HFSP Journal, 2(3), pp. 156–166.
R ACHMUTH -G., S HOUVAL -H., B EAR -M., AND P OON -C. (2011). A biophysically-based neuromorphic
model of spike rate- and timing-dependent plasticity, Proceedings of the National Academy of
Sciences, 108(49), pp. E1266–E1274.
R AMAKRISHNAN -S., H ASLER -P., AND G ORDON -C. (2011). Floating gate synapses with spike-time-
dependent plasticity, IEEE Transactions on Biomedical Circuits and Systems, 5(3), pp. 244–252.
R AZAVI -B. (2002). Design of Analog CMOS Integrated Circuits, Tata McGraw-Hill Education.
R OSE -R., AND H INDMARSH -J. (1989). The assembly of ionic currents in a thalamic neuron I. The three-
dimensional model, Proceedings of the Royal Society of London. B. Biological Sciences, 237(1288),
pp. 267–288.
R OWCLIFFE -P., AND F ENG -J. (2008). Training spiking neuronal networks with applications in engineer-
ing tasks, IEEE Transactions on Neural Networks, 19(9), pp. 1626–1640.
S ATYANARAYANA -S., T SIVIDIS -Y., AND G RAF -H.-P. (1992). A reconfigurable vlsi neural network, IEEE
Journal of Solid-State Circuits, 27(1), pp. 67–81.
S CHEMMEL -J., B RUDERLE -D., G RUBL -A., H OCK -M., M EIER -K., AND M ILLNER -S. (2010). A wafer-
scale neuromorphic hardware system for large-scale neural modeling, Proceedings of IEEE Inter-
national Symposium on Circuits and Systems (ISCAS), pp. 1947–1950.
S CHEMMEL -J., G RUBL -A., M EIER -K., AND M UELLER -E. (2006). Implementing synaptic plasticity in a
VLSI spiking neural network model, The 2006 International Joint Conference on Neural Networks,
DOI: 10.1109/IJCNN.2006.246651.
S CHMUKER -M., P FEIL -T., AND N AWROT-M. P. (2014). A neuromorphic network for generic mul-
tivariate data classification, Proceedings of the National Academy of Sciences, DOI: 10.1073/p-
nas.1303053111.
S CHOLZE -S., S CHIEFER -S., PARTZSCH -J., H ARTMANN -S., M AYR -C. G., H ÖPPNER -S., E ISENREICH -H.,
H ENKER -S., V OGGINGER -B., AND S CHFFNY-R. (2011). VLSI implementation of a 2.8 Gevent/s
packet based AER interface with routing and event sorting functionality, Frontiers in Neuroscience,
DOI: 10.3389/fnins.2011.00117.
S EEBACHER -E. (2005). CMOS technology characterisation for analog/RF application, 11th Workshop
on Electronics for LHC and Future Experiments, CERN, pp. 33 – 41.
S EO -J.-S., B REZZO -B., L IU -Y., PARKER -B. D., E SSER -S. K., M ONTOYE -R. K., R AJENDRAN -B., T IERNO -
J. A., C HANG -L., AND M ODHA -D. S. (2011). A 45 nm CMOS neuromorphic chip with a scalable
architecture for learning in networks of spiking neurons, The 2011 IEEE Custom Integrated Circuits
Conference (CICC), DOI: 10.1109/CICC.2011.6055293.
Page 257
Bibliography
S ERRANO -G OTARREDONA -T., M ASQUELIER -T., P RODROMAKIS -T., I NDIVERI -G., AND L INARES -
B ARRANCO -B. (2013). STDP and STDP variations with memristors for spiking neuromorphic
learning systems, Frontiers in Neuroscience, DOI: 10.3389/fnins.2013.00002.
S HAH -N. T., Y EUNG -L. C., C OOPER -L. N., C AI -Y., AND S HOUVAL -H. Z. (2006). A biophysical basis
for the inter-spike interaction of spike-timing-dependent plasticity, Biological Cybernetics, 95(2),
pp. 113–121.
S HEIK -S., C HICCA -E., AND I NDIVERI -G. (2012a). Exploiting device mismatch in neuromorphic VLSI
systems to implement axonal delays, The 2012 International Joint Conference on Neural Networks
(IJCNN), DOI: 10.1109/IJCNN.2012.6252636.
S HEIK -S., C OATH -M., I NDIVERI -G., D ENHAM -S. L., W ENNEKERS -T., AND C HICCA -E. (2012b). Emer-
gent auditory feature tuning in a real-time neuromorphic VLSI system, Frontiers in Neuroscience,
DOI: 10.3389/fnins.2012.00017.
S HERI -A., H WANG -H., J EON -M., AND GEUN L EE -B. (2014). Neuromorphic character recognition sys-
tem with two PCMO memristors as a synapse, IEEE Transactions on Industrial Electronics, 61(6),
pp. 2933–2941.
S HERIDAN -P., AND L U -W. (2014). Memristors and memristive devices for neuromorphic computing,
Memristor Networks, Springer, pp. 129–149.
S HOUVAL -H. Z. (2011). What is the appropriate description level for synaptic plasticity?, Proceedings
of the National Academy of Sciences, 108(48), pp. 19103–19104.
S HOUVAL -H. Z., B EAR -M. F., AND C OOPER -L. N. (2002). A unified model of NMDA receptor-
dependent bidirectional synaptic plasticity, Proceedings of the National Academy of Sciences,
99(16), pp. 10831–10836.
S HOUVAL -H. Z., WANG -S. S.-H., AND W ITTENBERG -G. M. (2010). Spike timing dependent plasticity:
a consequence of more fundamental learning rules, Frontiers in computational neuroscience, DOI:
10.3389/fncom.2010.00019.
S J ÖSTR ÖM -P. J., R ANCZ -E. A., R OTH -A., AND H ÄUSSER -M. (2008). Dendritic excitability and synaptic
plasticity, Physiological Reviews, 88(2), pp. 769–840.
S J ÖSTR ÖM -P., T URRIGIANO -G., AND N ELSON -S. (2001). Rate, timing, and cooperativity jointly deter-
mine cortical synaptic plasticity, Neuron, 32(6), pp. 1149–1164.
S MITH -G. D., C OX -C. L., S HERMAN -S. M., AND R INZEL -J. (2000). Fourier analysis of sinusoidally
driven thalamocortical relay neurons and a minimal integrate-and-fire-or-burst model, Journal of
Neurophysiology, 83(1), pp. 588–610.
S MITH -L. S. (2006). Implementing neural models in silicon, Handbook of Nature-Inspired and Innova-
tive Computing, Springer, pp. 433–475.
S OLEIMANI -H., A HMADI -A., AND B AVANDPOUR -M. (2012). Biologically inspired spiking neurons:
Piecewise linear models and digital implementation, IEEE Transactions on Circuits and Systems I:
Regular Papers, 59(12), pp. 2991–3004.
Page 258
Bibliography
S ONG -S., M ILLER -K., AND A BBOTT-L. (2000). Competitive Hebbian learning through spike-timing-
dependent synaptic plasticity, Nature Neuroscience, 3, pp. 919–926.
S TRUKOV-D., S NIDER -G., S TEWART-D., AND W ILLIAMS -R. (2008). The missing memristor found, Na-
ture, 453(7191), pp. 80–83.
TANAKA -H., M ORIE -T., AND A IHARA -K. (2009). A CMOS spiking neural network circuit with sym-
metric/asymmetric STDP function, IEICE Transactions on Fundamentals of Electronics, Commu-
nications and Computer Sciences, E92-A(7), pp. 1690–1698.
U RAMOTO -T., AND T ORIKAI -H. (2013). A calcium-based simple model of multiple spike interactions in
spike-timing-dependent plasticity, Neural Computation, 25(7), pp. 1853–1869.
V ITTOZ -E. (1985). The design of high-performance analog circuits on digital CMOS chips, IEEE Journal
of Solid-State Circuits, 20(3), pp. 657–665.
V OGELSTEIN -R. J., M ALLIK -U., C ULURCIELLO -E., C AUWENBERGHS -G., AND E TIENNE -C UMMINGS -R.
(2007a). A multichip neuromorphic system for spike-based visual information processing, Neural
Computation, 19(9), pp. 2281–2300.
V OGELSTEIN -R. J., M ALLIK -U., V OGELSTEIN -J. T., AND C AUWENBERGHS -G. (2007b). Dynamically re-
configurable silicon array of spiking neurons with conductance-based synapses, IEEE Transactions
on Neural Networks, 18(1), pp. 253–265.
V REEKEN -J. (2003). Spiking neural networks, an introduction, Technical Report UU-CS, (2003-008),
pp. 1–5.
WANG -H., AND WAGNER -J. J. (1999). Priming-induced shift in synaptic plasticity in the rat hippocam-
pus, Journal of Neurophysiology, 82(4), pp. 2024–2028.
WANG -H., G ERKIN -R., N AUEN -D., AND B I -G. (2005). Coactivation and timing-dependent integration
of synaptic potentiation and depression, Nature Neuroscience, 8(2), pp. 187–193.
WANG -R. M., H AMILTON -T. J., TAPSON -J. C., AND VAN S CHAIK -A. (2014a). A mixed-signal imple-
mentation of a polychronous spiking neural network with delay adaptation, Frontiers in neuro-
science, DOI: 10.3389/fnins.2014.00051.
WANG -Z., Z HAO -W., K ANG -W., Z HANG -Y., K LEIN -J.-O., R AVELOSONA -D., AND C HAPPERT-C.
(2014b). Compact modelling of ferroelectric tunnel memristor and its use for neuromorphic simu-
lation, Applied Physics Letters, 104(5), pp. 053505:1–053505:5.
W ESTE -N., AND E SHRAGHIAN -K. (1994). Principles of CMOS VLSI design: a systems perspective, 2nd
edn, Addison-Wesley.
W ESTE -N., AND H ARRIS -D. (2005). CMOS VLSI Design: A Circuits and Systems Perspective, Addison-
Wesley.
W IJEKOON -J. H., AND D UDEK -P. (2012). VLSI circuits implementing computational models of neocor-
tical circuits, Journal of Neuroscience Methods, 210(1), pp. 93–109.
W ILSON -H. R. (1999). Simplified dynamics of human and mammalian neocortical neurons, Journal of
Theoretical Biology, 200(4), pp. 375–388.
Page 259
W ILSON -H. R., AND C OWAN -J. D. (1972). Excitatory and inhibitory interactions in localized popula-
tions of model neurons, Biophysical Journal, 12(1), pp. 1–24.
YAO -E., H USSAIN -S., B ASU -A., AND H UANG -G.-B. (2013). Computation using mismatch: Neuromor-
phic extreme learning machines, 2013 IEEE Biomedical Circuits and Systems Conference (BioCAS),
pp. 294–297.
Y OUNG -J. M., WALESZCZYK -W. J., WANG -C., C ALFORD -M. B., D REHER -B., AND O BERMAYER -K.
(2007). Cortical reorganization consistent with spike timing—but not correlation-dependent plas-
ticity, Nature Neuroscience, 10(7), pp. 887–895.
Y U -S., G AO -B., FANG -Z., Y U -H., K ANG -J., AND W ONG -H.-S. P. (2013). A low energy oxide-based
electronic synaptic device for neuromorphic visual systems with tolerance to device variation, Ad-
vanced Materials, 25(12), pp. 1774–1779.
Z AMARRE ÑO -R AMOS -C., C AMU ÑAS -M ESA -L. A., P ÉREZ -C ARRASCO -J. A., M ASQUELIER -T.,
S ERRANO -G OTARREDONA -T., AND L INARES -B ARRANCO -B. (2011). On spike-timing-dependent-
plasticity, memristive devices, and building a self-learning visual cortex, Frontiers in Neuroscience,
DOI: 10.3389/fnins.2011.00026.
Z HANG -L., L AI -Q., AND C HEN -Y. (2010). Configurable neural phase shifter with spike-timing-
dependent plasticity, IEEE Electron Device Letters, 31(7), pp. 716–718.
Z UCKER -R., AND R EGEHR -W. (2002). Short-term synaptic plasticity, Annual Review of Physiology, 64,
pp. 355–405.
Page 260
List of Acronyms
GA Genetic Algorithm
HH Hodgkin-Huxley
IFB Integrate-and-Fire-or-Burst
Page 261
List of Acronyms
Page 262
Index
Abstract synaptic plasticity models, 30 Frequency-dependent pairing experiments, 35,
Accelerated-time, 114, 122, 135, 183, 202 168
ADC, 88 Frequency-dependent pairing protocol, 27, 168
Address Event Representation (AER), 57, 70, 112,
120, 222 Hopfield network, 122
All-to-all, 38
IF neuron, 73, 95, 187
Arbiter, 55, 112
IFMEM chip, 53, 76
ARM processor, 103
Interconnection, 111
Artificial Neural Network (ANN), 2
Ion dynamics, 26, 97
Asynchronous logic, 112, 237
Ionic conductance, 90
Axon, 20
IPSC, 24
BCM, 8, 30, 39, 176 Izhikevich’s neuron model, 21, 177, 243
BCM-like behaviour, 30, 35
Bias generation, 112 Large-scale hardware, 25
Biophysical synaptic plasticity, 30, 46, 104, 123 Large-scale neural simulation, 25
Bistability circuit, 115 Leakage current, 114, 223
Bistability mechanism, 115 Leaky integrator, 91, 94, 100, 101
Brain machine interface, 5 Learning, 3, 21, 52
Linear Poisson neuron, 177, 243
Calcium, 42, 46 Local Correlation Plasticity, 118
Current mirror, 24, 93, 179, 218 Local Correlation Plasticity (LCP), 44
Current source synapse, 24 Log-domain integrator synapse, 24
Long-term memory, 6
DAC, 55, 116, 236
Long-Term Synaptic Plasticity (LTSP), 86
Decay, 92
Low-pass filter, 54, 95
Depolarisation, 43
LTD, 26, 27, 86, 177, 209
Depression, 33, 43
LTP, 26, 27, 86, 177, 209
Differential Pair (DP), 91, 102
DPI, 24, 53, 56, 95 McCulloch neuron, 2
Dynamic Random Access Memory (DRAM), 123 Membrane potential, 42, 90
Dynamic Vision Sensor, 229 Memory, 52
Memristor, 5, 117
EDA, 112
Micro-controller, 52
EPSC, 24, 74
Minimal TSTDP, 33, 34, 200, 201
Error function, 26, 133, 144
Mismatch, 109, 178
Experimental protocols, 25, 143, 166, 189
MOSCAP, 88, 222
Floating Gate (FG), 88, 102, 116, 223 MOSFET, 88
FPGA, 52, 57, 102 Multiplier synapse, 24
Page 263
Index
SDSP, 95
Weak inversion, 97
Sigmoid neuron, 2
Weight-dependent STDP (W-STDP), 97, 101, 122,
Sigmoidal function, 91
136
Silicon real estate, 111
Winner Take All (WTA), 93
Silicon synapse, 24
Silicon-On-Insulator (SOI), 103
Soma, 20, 21
Page 264
Biography
S. Mostafa Rahimi Azghadi was born in Mashhad, Iran,
in 1982. In 2006, he graduated from Sadjad University
of Mashhad, with a Bachelor’s Degree in Computer En-
gineering (majoring in hardware) with the first class rank.
He obtained his Master of Engineering in Computer Ar-
chitecture with first class honours, from Shahid Beheshti
University (SBU), Tehran, Iran, in 2009 under the supervi-
sion of Prof. Keivan Navi.
Page 265
Biography
Page 266