An Efficient Floating Point Adder For Low-Power Devices
An Efficient Floating Point Adder For Low-Power Devices
Corresponding Author:
Manjula Narayanappa
Department of Electronic and Communication, Dr. Ambedkar Institute of Technology
Bengaluru, Karnataka, India
Email: [email protected]
1. INTRODUCTION
Battery-operated portable electronic devices have increasingly become an indispensable part of
everyday life. The key behind this is the scaling ability of metal oxide silicon field effect transistors
(MOSFETS) seen in very large-scale integration (VLSI) due to which functionality per unit area has
increased which has brought the price of the devices down leading to wide usage. Due to scaling and an
increase in functionality per unit area, the power consumption has increased. The increase in power
consumption of the VLSI devices has not been matched by the improvement in the capacity of the battery.
Therefore, operation time per charge has come down causing inconvenience to the users. For this reason,
reducing the power consumption of portable devices has become a compelling design constraint. A large
portion of energy consumption is dominated by two components: dynamic power and leakage power. To
extend the battery life various technology-based, architecture-based, and circuit-based solutions that reduce
the sum of the two power components without sacrificing the performance have to be developed. At the
technology level, feature size scaling has continuously brought lower power circuits by reducing the supply
voltages. To retain performance, the threshold voltages of these circuits have also been reduced with
technology scaling. However, in recent technologies, the benefits of constant-field scaling have been
compromised by an exponential increase in the leakage current. On the architectural level, pipelining and
parallelism have helped in lowering the power consumption of digital circuits.
In the current complementary metal-oxide semiconductor (CMOS) technology, the benefits of
device scaling are impeded by the reliability issues due to the process variations, ageing effects and soft
errors. Leakage current, static power are increasingly adding to the concerns towards achieving low power
consumption. Hence, the device scaling which once used to offer advantages for low power applications is no
more attractive and hence, new architectures need to be evolved to achieve low power consumption. Design
of approximate computation blocks is one such potential solution [1].
Most of the modern graphics processors for multimedia and other applications have dedicated
digital signal processing blocks. These applications output an image, video or an audio signal and the limited
perception of human senses allows for an approximation of the computations involved in the demanding
digital signal processing (DSP) algorithms for these applications [2]. Even an analog computation that yields
good enough results instead of accurate results is also acceptable [3]. Addition is the most fundamental and
significant mathematical operations used in all signal/image processing applications [4], [5]. Deterministic
approximate logic or probabilistic imprecise arithmetic are normally employed for soft adders [6].
Various low-power design approaches using approximate computing have been introduced, such as
algorithmic noise tolerance [7], [8], non-uniform voltage over scaling [9], and significance-driven
computation [10], [11]. Verma et al. [12] have presented an innovative adder design known as the almost
correct adder (ACA), which offers exponentially faster performance compared to traditional adders.
They also proposed the variable latency speculative adder (VLSA) with a slight area overhead. Additionally,
some adder configurations meet real-time energy requirements by reducing complexity at the algorithmic
level [13], [14]. The lower part OR adder [15] relies on approximate logic with a distinct truth table
compared to a standard adder. The probabilistic full adder (PFA) [16]-[20] is based on probabilistic CMOS
technology, which is a platform for modeling nano-scale designs and reducing power consumption [21]-[23].
2. BACKGROUND
2.1. IEEE-754 floating point format
Floating point representation offers a wider dynamic range in comparison to fixed-point
representation for real numbers. However, floating point hardware is known for its complexity and
substantial power consumption. The predominant standard for floating point formats is IEEE 754-2008 [24],
which encompasses various basic and extended types. These formats include half precision (16-bits), single
precision (32-bits), double precision (64-bits), extended precision (80-bits), and quad precision (128-bits).
The typical IEEE floating point format, as depicted in Figure 1, features an exponent part with a bias of
2^(E-1)-1, where E denotes the number of exponent bits. Single precision and double precision formats are
the most commonly used in contemporary computer systems. You can find details regarding the exponent
and mantissa bits for IEEE-754 basic and extended floating point types in Table 1.
Table 1. Exponent and Mantissa bits for IEEE-754 basic and extended floating point types
Type Sign bit Exp. bits Mant. bits Total Mant. bits/total
Half 1 5 10 16 62.5%
Single 1 8 23 32 71.9%
Double 1 11 52 64 81.2%
Extended 1 15 64 80 80.0%
Quad 1 15 112 128 87.5%
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 2, July 2024: 253-261
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 255
detection of leading zeros a critical step in this process. Finally, rounding the normalized result is the last
operation before storing the result back. Special cases such as overflow, underflow, and not-a-number are
also detected and indicated by flags.
The modified approximate adder concept can also be used for the mantissa adder for approximate
computation. The mantissa adder will provide a larger scope, as the number of bits in the mantissa are higher
than the exponent and at the same time, the approximate design in the mantissa adder has a lower impact on
the error, because the mantissa part is less significant than the exponent part. Therefore, an inexact design of
a mantissa adder is more appropriate.
An efficient floating point adder for low-power devices (Manjula Narayanappa)
256 ISSN: 2089-4864
where, 𝐶𝑖𝑛 is the input carry and 𝑃𝑖 and 𝐺𝑖 are propagate and generate signals of the ith stage. If the carry
equation is split up into two segments, as in (2):
where, 𝑊 is the window size and, the first segment consists of W most significant (MS) bits and the second
segment consists of N-W least significant (LS) bits. The first part of the (2) is the approximate part, while the
second part is called the augmenting part. For approximate carry generation with a window size of W, the
output carry at the ith stage is compute using the approximate part only. Computing an approximate 𝐶𝑖+1 is
faster and consumes less hardware resources and hence lesser power as compared to computing precise carry.
An 8-bit adder is chosen as the basic building block for the floating point approximate adder in the proposed
design. Figure 4 shows the structure of conventional full adder. As shown in Figure 5, the 8 bits are
partitioned into two blocks; the MS block is of 4 bits, while the LS block is of 4 bits. The output carry of this
8-bit adder block is computed approximately using the approximate part in (2), 4-bit carry generator block as
shown in Figure 6 is used for generating the approximate carry for the 8-bit adder using the 4-MS bits.
In the proposed adder, for the LS, W-bit window sum and carry are computed as per (3). The
schematic of 1-bit full adder in the proposed configuration is given by Figure 7. The sum and carry for most
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 2, July 2024: 253-261
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 257
significant (8-W=4) bits are computed as for an exact adder are given in (4). The truth table for the proposed
sum and carry equations is given in Table 2.
𝐶𝑖+1 = 𝑎𝑖 𝑏𝑖 + 𝑐𝑖𝑛
𝑆𝑖 = 𝐶𝑖+1 (3)
Overall, 3 errors are introduced in sum computation and 1 error in carry computation. Assigning the
inverted carry out at each stage to the sum computed for that stage reduces the hardware for sum computation
block. This is a significant reduction in hardware requirements as compared to a conventional adder.
Utilizing the look ahead carry generation logic from 4 MS bits improves the timing performance of the
circuit by not depending on the sequential computation of carry at each bit. A total of 8 transistors are used
for 1-bit sum and carry generation.
4. ERROR METRICS
4.1. Error distance
The error distance (ED) between two binary numbers, 𝑎 (erroneous) and 𝑏 (correct), is defined as
the arithmetic distance between these two numbers. Where, 𝑖 and 𝑗 are the indices for the bits in 𝑎 and 𝑏,
respectively. Suppose for an 8-bit adder, the correct sum for a given set of operands is “1110 0101” and the
incorrect outputs are “11100100” and “11110101”. Then the two erroneous values “11100100” and
“11110101” have an ED of 1 and 16 respectively.
For a non-deterministic implementation, the output is probabilistic and usually follows a distribution
for a given input 𝑎𝑖 . In this case, the ED of the output (denoted by 𝑑𝑖 ) is defined as the weighted average of
EDs of all possible outputs to the nominal output. Assume that for a given input, the output has a nominal
value b, but it can take any value given in a set of vectors 𝑏𝑗 (1 ≤ 𝑗 ≤ 𝑟). The ED of the output is then given
by (6). Where 𝑝𝑗 is the output probability of 𝑏𝑗 (1 ≤ 𝑗 ≤ 𝑟).
𝑑𝑖 = ∑𝑖 𝐸𝐷(𝑏𝑗 , 𝑏) × 𝑝𝑗 (6)
𝑑𝑚 = ∑𝑖 𝑑𝑖 × 𝑞𝑖 (7)
5.2. Delay
Considering a conventional 8-bit adder, the delay in 8-bit computation is due to the ripple carry
effect, which takes 8 cycles. Assuming the delay in computation of 1-bit full adder result to be T, the delay in
generating the 8-bit adder is 8T. In the proposed adder, the total delay is equal to the delay for computation of
carry out from MS 4-bits, which is equal to 4T.
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 2, July 2024: 253-261
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 259
resulting in an overall reduction in power consumption. In the case of a traditional full adder, if the power
consumed for a 1-bit operation is normalized to 1, then the power consumed for a k-bit conventional full
adder amounts to k. However, in the context of the proposed 8-bit adder, the reduction in the number of
transistors leads to a decrease in the operating voltage, from 1.13 V in an accurate implementation to 1.04 V.
Consequently, this reduction in voltage contributes to an estimated decrease in power consumption.
1.042 1.042
𝐸8−𝑏𝑖𝑡 = 𝑊 × + (8 − 𝑊) = 4 × +4 (8)
1.132 1.132
Which is 7.5% lower than the conventional adder. Both the discrete cosine transform (DCT) and
inverse discrete cosine transform (IDCT) blocks operates at a lower supply voltage in case of approximate
adders than the exact mode. Here, DCT and IDCT operates at a supply voltage of 1.28 and 1.13 V in the
exact mode, respectively. The different supply operating voltages are demonstrated in Figures 9 and 10 for
different approximations and truncations considering varied bits. Table 3 demonstrates the percentage power
savings considering varied approximations and truncation against the base case. Approximation 3 saves the
maximum power.
1.15
1.1
1.05
Truncation Approximation 1 Approximation 2 Approximation 3
Technique
7 LSB’s 8 LSB’s 9 LSB’s
Figure 10. Operating voltages considering different bits for IDCT technique
Table 3. Percentage power savings for approximations over the base case
Technique 7 LSB’s 8 LSB’s 9 LSB’s
Truncation 48.22 56.23 61.24
Approximation 1 37.86 50.85 55.26
Approximation 2 41.21 49.13 53.84
Approximation 3 42.46 52.64 59.23
6. CONCLUSION
A novel approximate adder topology for single point floating point adder is presented in the paper.
The proposed design takes advantage of the fact that the lower significant bit addition can be approximate
and this will not be affecting the solution to a great extent, at the same time the power savings due to the
approximate computation will be significant. The proposed configurations has a lower propagation delay and
comparable error performance as compared to other architectures. With the proposed mantissa adder, which
is a hybrid of look, ahead carry adder for the carry generation and that of the approximate adder for the sum
generation gives a distinct advantage in terms of the power consumption as compared to the conventional full
adder.
REFERENCES
[1] J. Han and M. Orshansky, “Approximate computing: An emerging paradigm for energy-efficient design,” in 2013 18th IEEE
European Test Symposium (ETS), May 2013, pp. 1–6, doi: 10.1109/ETS.2013.6569370.
[2] R. Hegde and N. R. Shanbhag, “Soft digital signal processing,” IEEE Transactions on Very Large Scale Integration (VLSI)
Systems, vol. 9, no. 6, pp. 813–823, Dec. 2001, doi: 10.1109/92.974895.
[3] M. A. Breuer, “Let’s think analog,” in IEEE Computer Society Annual Symposium on VLSI: New Frontiers in VLSI Design
(ISVLSI’05), 2005, pp. 2–5, doi: 10.1109/ISVLSI.2005.48.
[4] V. Beiu, S. Aunet, J. Nyathi, R. R. Rydberg, and W. Ibrahim, “Serial addition: locally connected architectures,” IEEE
Transactions on Circuits and Systems I: Regular Papers, vol. 54, no. 11, pp. 2564–2579, Nov. 2007, doi:
10.1109/TCSI.2007.907885.
[5] S. Cotofana, C. Lageweg, and S. Vassiliadis, “Addition related arithmetic operations via controlled transport of charge,” IEEE
Transactions on Computers, vol. 54, no. 3, pp. 243–256, Mar. 2005, doi: 10.1109/TC.2005.40.
[6] J. Huang and J. Lach, “Exploring the fidelity-efficiency design space using imprecise arithmetic,” in 16th Asia and South Pacific
Design Automation Conference (ASP-DAC 2011), Jan. 2011, pp. 579–584, doi: 10.1109/ASPDAC.2011.5722256.
[7] S. Byonghyo, S. R. Sridhara, and N. R. Shanbhag, “Reliable low-power digital signal processing via reduced precision
redundancy,” IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 12, no. 5, pp. 497–510, May 2004, doi:
10.1109/TVLSI.2004.826201.
[8] G. V. Varatkar and N. R. Shanbhag, “Energy-efficient motion estimation using error-tolerance,” in Proceedings of the 2006
international symposium on Low power electronics and design-ISLPED ’06, Oct. 2006, pp. 113–118, doi:
10.1145/1165573.1165599.
[9] L. N. B. Chakrapani, K. K. Muntimadugu, A. Lingamneni, J. George, and K. V. Palem, “Highly energy and performance efficient
embedded computing through approximately correct arithmetic,” in Proceedings of the 2008 international conference on
Compilers, architectures and synthesis for embedded systems, Oct. 2008, pp. 187–196, doi: 10.1145/1450095.1450124.
[10] D. Mohapatra, G. Karakonstantis, and K. Roy, “Significance driven computation,” in Proceedings of the 2009 ACM/IEEE
international symposium on Low power electronics and design, Aug. 2009, pp. 195–200, doi: 10.1145/1594233.1594282.
[11] N. Banerjee, G. Karakonstantis, and K. Roy, “Process variation tolerant low power DCT architecture,” in 2007 Design,
Automation and Test in Europe Conference and Exhibition, Apr. 2007, pp. 1–6, doi: 10.1109/DATE.2007.364664.
[12] A. K. Verma, P. Brisk, and P. Ienne, “Variable latency speculative addition: a new paradigm for arithmetic circuit design,” in
2008 Design, Automation and Test in Europe, Mar. 2008, pp. 1250–1255, doi: 10.1109/DATE.2008.4484850.
[13] Y. V. Ivanov and C. J. Bleakley, “Real-time H.264 video encoding in software with fast mode decision and dynamic complexity
control,” ACM Transactions on Multimedia Computing, Communications, and Applications, vol. 6, no. 1, pp. 1–21, Feb. 2010,
doi: 10.1145/1671954.1671959.
[14] M. Shafique, L. Bauer, and J. Henkel, “enBudget: a run-time adaptive predictive energy-budgeting scheme for energy-aware
motion estimation in H.264/MPEG-4 AVC video encoder,” in 2010 Design, Automation and Test in Europe Conference and
Exhibition (DATE 2010), Mar. 2010, pp. 1725–1730, doi: 10.1109/DATE.2010.5457093.
[15] H. R. Mahdiani, A. Ahmadi, S. M. Fakhraie, and C. Lucas, “Bio-inspired imprecise computational blocks for efficient VLSI
implementation of soft-computing applications,” IEEE Transactions on Circuits and Systems I: Regular Papers, vol. 57, no. 4,
pp. 850–862, Apr. 2010, doi: 10.1109/TCSI.2009.2027626.
[16] S. H. Sreedhara, V. Kumar, and S. Salma, “Efficient big data clustering using adhoc fuzzy C means and auto-encoder CNN,” in
Inventive Computation and Information Technologies, 2023, pp. 353–368, doi: 10.1007/978-981-19-7402-1_25.
[17] M. S. K. Lau, L. Keck-Voon, C. Yun-Chung, and A. Bhanu, “A general mathematical model of probabilistic ripple-carry adders,”
in 2010 Design, Automation and Test in Europe Conference and Exhibition (DATE 2010), Mar. 2010, pp. 1100–1105, doi:
10.1109/DATE.2010.5456973.
[18] S. Borkar, T. Karnik, and V. De, “Design and reliability challenges in nanometer technologies,” in Proceedings of the 41st annual
Design Automation Conference, Jun. 2004, pp. 75–75, doi: 10.1145/996566.996588.
[19] W. Huan-Sheng and R. M. Mersereau, “Fast algorithms for the estimation of motion vectors,” IEEE Transactions on Image
Processing, vol. 8, no. 3, pp. 435–438, Mar. 1999, doi: 10.1109/83.748899.
[20] M. Shafique, L. Bauer, and J. Henkel, “3-tier dynamically adaptive power-aware motion estimator for h.264/AVC video
encoding,” in Proceeding of the thirteenth international symposium on Low power electronics and design-ISLPED ’08, 2008, pp.
147–152, doi: 10.1145/1393921.1393962.
[21] J. George, B. Marr, B. E. S. Akgul, and K. V. Palem, “Probabilistic arithmetic and energy efficient embedded signal processing,”
in Proceedings of the 2006 international conference on Compilers, architecture and synthesis for embedded systems, Oct. 2006,
pp. 158–168, doi: 10.1145/1176760.1176781.
[22] G. V. Varatkar and N. R. Shanbhag, “Energy-efficient motion estimation using error-tolerance,” in Proceedings of the 2006
international symposium on Low power electronics and design-ISLPED ’06, 2006, pp. 113–118, doi: 10.1145/1165573.1165599.
[23] D. Mohapatra, G. Karakonstantis, and K. Roy, “Low-power process-variation tolerant arithmetic units using input-based elastic
clocking,” in Proceedings of the 2007 international symposium on Low power electronics and design, Aug. 2007, pp. 74–79, doi:
10.1145/1283780.1283797.
[24] “IEEE standard for floating-point arithmetic,” IEEE Std 754-2008, Aug. 2008, doi: 10.1109/IEEESTD.2008.4610935.
[25] P. Behrooz, Computer arithmetic: algorithms and hardware designs, Oxford University Press, 2000.
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 2, July 2024: 253-261
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 261
BIOGRAPHIES OF AUTHOR
Siva S. Yellampalli obtained his M.S. and Ph.D. from Louisiana State University.
He is currently with VTU Extension Centre, UTL Technologies Ltd. He worked on a broad
range of research topics including VLSI, mixed signal circuits/systems development, micro-
electromechanical systems (MEMS), and integrated carbon nanotube based sensors. He has
published a book in the area of mixed signal design, and edited two books on carbon nano
tubes. He also published multiple journal papers and IEEE conference papers in these areas of
research. He can be contacted at this email: [email protected].