Linear Equation Solution ISCAS2020
Linear Equation Solution ISCAS2020
on a Configurable Platform
Aishwarya Natarajan and Jennifer Hasler
Georgia Institute of Technology, [email protected]
Authorized licensed use limited to: IEEE Xplore. Downloaded on October 03,2020 at 21:41:11 UTC from IEEE Xplore. Restrictions apply.
G1,1 V1 G2,1 V1 Gm,1 V1
V1 V1 V1
fixed fixed fixed
GND GND GND
G1,2 V2 G2,2 V2 Gm,2 V2
Vdd Vdd Vdd
V2 V2 V2
fixed fixed I2 fixed
I1 Im
GND GND GND
V1 V2 Vm
C1 C2 Cm
G1,m Vm G2,m Vm Gm,m Vm
Vm Vm Vm
GND GND GND
fixed fixed fixed
GND GND GND
Fig. 2. A potential continuous-time circuit architecture to solve systems of linear equations is shown. Linear matrix solution is built from Transconductance
Amplifiers (OTA), which are elements of the Computational Analog Block (CAB) on the SoC. OTAs with Floating-Gate (FG) bias currents make up the
individual conductances allowing for a general constraint matrix to solve. The input vectors are fed into the gate input of the 9 transistor OTA, which acts as
a current source, while the offset for the reference voltages can be controlled by DACs compiled on the Array.
The paper focuses on solving systems of linear equations The dynamics of (2) are seen along the eigenspace corre-
using analog computation. The discussion moves towards iter- sponding to A. We define the eigenvalue / eigenvector of A
ative methods for solving linear equations, good methods for as
analog and digital computation, and its potential configurable A = EΛE−1 (4)
circuit approaches. The discussion then moves towards analyz-
where Λ is a diagonal matrix of eigenvalues, and E are the
ing the algorithmic issues and analog numerical analysis issues
corresponding rows of normalized (power = 1) eigenvectors
of this approach. These techniques provide energy-efficient
corresponding to the particular eigenvector. By projecting (1)
continuous-time solutions. Even if an ideal solution exists, one
and (2) on the eigenvector basis in (4), we get
needs confidence in the overall accuracy and convergence time.
dy
x = Ey, τ + Λy = E−1 b, (5)
I. P HYSICAL C OMPUTATION L INEAR E QUATION dt
S OLUTIONS → I TERATIVE M ATRIX T ECHNIQUES This ODE requires positive Λ values requiring that A be
positive definite. Non positive definite matrices can be solved
Linear systems are also solved by iterative techniques, using through multiple transformations [6], [16], such as multiplying
coupled equations that converge x to the desired solution. (2) by AT resulting in a positive definite matrix [16]. There-
Iterative digital techniques are sometimes faster and have fore, this method is general enough for these solution methods.
fewer operations than digital Gaussian elimination, particularly
for sparse matrix solutions. As one should simply use physical II. L INEAR E QUATION S OLUTION USING
techniques to solve ODE / PDE applications [14], we focus T RANSCONDUCTANCE A MPLIFIERS
on physical solution systems of linear systems from different We solve these linear system of equations employing
applications. Transconductance Amplifiers (TA) in an architecture shown
A physical approach will be aligned with iterative matrix in Fig. 2. One can achieve a larger set of matrix equations
equations. Physical computing solutions of linear systems are utilizing TAs, enabling a richer set of potential matrices. The
iterative techniques, where the iterations result from coupled linear region of an OTA is expressed as
differential equations whose steady state represents the desired
output (x). One might modify (1) to build the differential Iout = G(V + − V − ) (6)
equation For an OTA programmed with subthreshold bias currents
dx (Ibias ), the resulting coupling between nodes (k,l) would be
τ + Ax = b, (2)
dt Gk,l = κIUbiasT
. The discussion holds when utilizing above
threshold currents with a modification of the underlying func-
where τ is the time-constant for the network. Each row could
tion for G.
have a different τ , although we present the simpler case for
Figure 2 shows the resulting general network to solve (1),
clarity. Digital (sampled time) approaches would be written as
where the l row is written as
x[n] = x[n − 1] + ǫ (b − Ax[n]) , (3) dVl Xm
C = Il − Gl,k Vk (7)
dt
effectively approximating the time derivative of x. These k=1
techniques are related to gradient-descent, Jacobi, and Gauss- and by comparing with (2), al,k ∝ Gl,k and bk ∝ Ik .
Seidel iterative matrix solutions [15]. The timeconstant, τ is the load capacitance divided by a
Authorized licensed use limited to: IEEE Xplore. Downloaded on October 03,2020 at 21:41:11 UTC from IEEE Xplore. Restrictions apply.
1.1 60
b Voltage (V)
Time (ms)
100nA 100nA 100nA 200nA 100
1.05
10
x2: 50.5mV
x2
x1
x1: 21.2mV
1 x4: 19.2mV
V = 1.05, Offsets = [-12.7 29.0 -33.7 49.9] mV
ref x3: 13.7mV
1
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35
Time (ms) Time (ms)
Fig. 3. Analysis for the convergence of A matrix with 200nA and 100nA as the diagonal and off-diagonal elements respectively, is shown. For a 40mV b
input applied as current sources, the x(t) solution is plotted and is within the linear range of the OTAs. The time constant is 61µs curve-fitting the exponential
curves from the step responses. One of the components along the eigenvalue that is larger by a factor of 5, converges 5 times faster in 12µs.
normalizing conductance, often related to a larger diagonal sizes. Additional A matrices can be programmed and dynamics
element of A. For a particular normalizing subthreshold bias measured.
CUT
current, Ibias , τ = κI . One could have a different τ per
ref
row due to different conductance modeling and capacitances, Figure 3 analyzes the convergence and resulting eigenvalue
and potentially tuned to optimize convergence. Current sources timeconstants from the step responses. The dynamics are
set b inputs where positive current sources would go to Vdd , studied by plotting the matrix solution outputs, x(t) which
and negative current sources would go to GND. These current are the steady-state responses. In case of the matrix with
sources could also be implemented as additional TAs with 200nA diagonal elements, the effective time constant is 61µs
controlling input voltages. from the three eigenvalues, obtained through curve-fitting the
exponential curves from the step responses. Since one of the
III. M EASURED DYNAMICS ON THE S O C FPAA eigenvalues is larger by a factor of 5, it converges 5 times
This approach is implemented in an SoC FPAA device [13]. faster, in 12µs. The time constant is 47µs for the diagonal
The linear equations are solved by 9-transistor TA elements matrix with 1µA as the 1 element and 10nA as the 0 elements.
in Computational Analog Blocks (CAB) [13]. The A matrix The effective time constant is related to the minimum
is set by the conductances of the individual OTAs, that in eigenvalues, resulting in the largest value for settling to a
turn are set by Floating-Gate (FG) bias currents. A TA is solution. Other terms will settle faster, and depending on the
used to convert the input voltage (b) signal to a current. The particular problem and method, the τ for each row could be
reference voltage(s) can be controlled by DACs compiled on adapted to optimize convergence. Larger eigenvalues converge
the FPAA [13]. A 4x4 linear system solution is compiled and faster. Smaller eigenvalues that correspond to the smaller
programmed on the FPAA using 16 TAs to implement the 4x4 changes in the solution (signal power), have little changes
A matrix, and 4 TAs to implement the b matrix. The sign of overall, and therefore, not in the required precision. Hence,
each A matrix element determines the TA input sign, where it is not important to wait for them. One could visualize the
positive values are input in the - input and the reference to the adaptation of τ to accommodate for λ, because, if (2) is stable
+ input, while negative values are input in the + input and the for one set of τ , it is stable for all positive τ .
reference to the - input. The outputs, x(t) can be scanned and
buffered out to be measured. Another way to understand the ODE of the linear equation
Figure 3 shows the dynamics of the matrix solutions, solver as a dynamical system or fully coupled filters, is
demonstrated through a 4x4 network. The b vectors are inputs typically as a set of low-pass filters. Figure 4 shows the
fed to the gate input of the TA structure. In particular, the response of the network to a chirp input, sweeping between
inputs are step functions. They are step inputs, starting from frequencies from 1Hz to 20kHz over a time duration of 1ms.
zero (= fixed potential) and moving to a single input value for These approaches give a new way of approaching the solution
b. The matrix solution outputs, x(t) , show the dynamics when of linear equations as effectively filter design, consistent with
new inputs are applied to different conductance matrices. The an approach typically seen in control system implementations.
time constant is roughly the diagonal TA bias current and the The dynamics of the output at one tap show the higher
FPAA routing capacitive load, and is related to any mismatch frequency components being attenuated, which is consistent
in the resulting components, both for small and large signal with low pass behavior.
Authorized licensed use limited to: IEEE Xplore. Downloaded on October 03,2020 at 21:41:11 UTC from IEEE Xplore. Restrictions apply.
V. A NALOG N UMERICS OF I TERATIVE L INEAR E QUATION
1.26
S OLUTIONS
Input (V)
Authorized licensed use limited to: IEEE Xplore. Downloaded on October 03,2020 at 21:41:11 UTC from IEEE Xplore. Restrictions apply.
R EFERENCES
[1] J. Hasler, “Starting Framework for Analog Numerical Analysis for
Energy Efficient Computing,” Journal of Low Power Electronics Appli-
cations, vol. 7, no. 17, June 2017. pp. 1-22.
[2] J.J. Dongarra, P.Luszczek, and A. Petitet, “The LINPACK Benchmark:
past, present and future,” Concurrency and Computation: Practice and
Experience, Wiley, 2003, pp. 803-820.
[3] J.J. Hopfield, “Neural networks and physical systems with emergent
collective computational abilities,” Proceedings of National Academy of
Science, vol. 79, 1982, pp. 2554.
[4] J.J. Hopfield, “Neurons with graded responses have collective compu-
tational properties like those of two- state neurons,” Proceedings of
National Academy of Science, vol. 81, 1984, pp. 3088-3092.
[5] R. Umbehauen and A. Cichocki, MOS Switched-Capacitor and
Continuous-Time Integrated Circuits and Systems, Springer-Verlag,
1989.
[6] A. Cichocki and R. Umbehauen, “Neural networks for solving systems
of linear equations and related problems,” IEEE March 1992 CAS I, vol.
39, no. 2,
[7] N. Guo, Y. Huang, T. Mai, S. Patil, C. Cao, M Seok, S. Sethumadhavan,
and Y. Tsividis, “Energy-Efficient Hybrid Analog/Digital Approximate
Computation in Continuous Time,” IEEE Journal of Solid State Circuits,
2016.
[8] Y. Huang, N. Guo, M. Seok, Y. Tsividis, and S. Sethumadhavan,
“An Analog Accelerator for Linear Algebra,” Proceedings of the 43rd
International Symposium on Computer Architecture, Seoul, June 18,
2016. pp. 570-582.
[9] R. M. Walker, “An Analogue Computer for the Solution of Linear
Simulaneous Equations,” Proceedings of the IRE – Waves and Electrons
Section, 1949, pp. 1467-1473.
[10] S. K. Mitra, “Electrical Analog Computing Machine for Solving Linear
Equations and Related Problems,” Review of Scientific Instruments vol.
26, 1955. pp.453-457.
[11] K. P. Lanneau and Lindsay I. Griffin, “Analogue Computer for Solving
Simultaneous Equations,” US Patent 2,911,146 Filed May 27, 2953,
issued, Nov. 3, 1959.
[12] Sun, Z., Pedretti, G., Ambrosi, E., Bricalli, A., Wang, W., and Ielmini, D.
, “Solving matrix equations in one step with cross-point resistive arrays”,
Proceedings of the National Academy of Sciences, 2019, 116(10), 4123-
4128.
[13] S. George, S. Kim, S. Shah, J. Hasler, M. Collins, F. Adil, R, Wunderlich,
S. Nease, and S. Ramakrishnan, “A Programmable and Configurable
Mixed-Mode FPAA SoC,” IEEE Transactions on VLSI, vol. 24, no. 6,
2016, pp. 2253-2261.
[14] J. Hasler, “Opportunities in Physical Computing driven by Analog
Realization,” IEEE ICRC, San Diego, 2016.
[15] G. E. Golub and C.F. Van Loan, Matrix Computation, 2nd ed, John
Hopkins Press, 1989.
[16] B. Ulmann and D. Killat, “Solving systems of linear equations on analog
computers,” IEEE Kleinheubach Conference, 2019.
[17] C. Schlottmann, and P. Hasler, “A highly dense, low power, pro-
grammable analog vector-matrix multiplier: the FPAA implementation,”
IEEE Journal of Emerging CAS, vol. 1, 2012, pp. 403-411.
[18] J. Hasler, “Analog Architecture Complexity Theory Empowering Ultra-
Low Power Configurable Analog and Mixed Mode SoC Systems,”
Journal of Low Power Electronics Applications, Jan. 2019. pp. 1-37.
[19] J. Hasler, “Large-Scale Field Programmable Analog Arrays,” IEEE
Proceedings, February 2020.
[20] K.-H. Kim, S. Gaba, D. Wheeler, J. M. Cruz-Albrecht, T. Hussain,
N. Srinivasa, and W. Lu, ”A Functional Hybrid Memristor Crossbar-
Array/CMOS System for Data Storage and Neuromorphic Applications”,
Nano Lett., vol. 12, 2012, pp. 389-395.
[21] S. H. Jo, and W. Lu, “Programmable Resistance Switching in Nanoscale
Two-Terminal Devices”, Nano Letters vol. 9, 2009, pp. 496-500.
[22] S.Y. Kung, VLSI Array Processors, Prentice Hall, 1988.
Authorized licensed use limited to: IEEE Xplore. Downloaded on October 03,2020 at 21:41:11 UTC from IEEE Xplore. Restrictions apply.