Double Precision IEEE-754 Floating-Point Adder Design Based on FPGA

Similar documents
High-Radix Implementation of IEEE Floating-Point Addition

Implementation of Double Precision Floating Point Multiplier on FPGA

Fig.1. Floating point number representation of single-precision (32-bit). Floating point number representation in double-precision (64-bit) format:

International Journal of Advanced Research in Electrical, Electronics and Instrumentation Engineering

Implementation of Double Precision Floating Point Multiplier Using Wallace Tree Multiplier

FPGA based High Speed Double Precision Floating Point Divider

FPGA IMPLEMENTATION OF FLOATING POINT ADDER AND MULTIPLIER UNDER ROUND TO NEAREST

Design and Optimized Implementation of Six-Operand Single- Precision Floating-Point Addition

Redundant Data Formats for Faster Floating Point Addition. Abstract

Implementation of Double Precision Floating Point Multiplier in VHDL

Design of Double Precision Floating Point Multiplier Using Vedic Multiplication

Architecture and Design of Generic IEEE-754 Based Floating Point Adder, Subtractor and Multiplier

HIGH SPEED SINGLE PRECISION FLOATING POINT UNIT IMPLEMENTATION USING VERILOG

THE INTERNATIONAL JOURNAL OF SCIENCE & TECHNOLEDGE

An FPGA Based Floating Point Arithmetic Unit Using Verilog

Design and Implementation of IEEE-754 Decimal Floating Point Adder, Subtractor and Multiplier

Implementation of Floating Point Multiplier Using Dadda Algorithm

ISSN Vol.02, Issue.11, December-2014, Pages:

Double Precision Floating-Point Arithmetic on FPGAs

VHDL IMPLEMENTATION OF FLOATING POINT MULTIPLIER USING VEDIC MATHEMATICS

Implementation of IEEE-754 Double Precision Floating Point Multiplier

FPGA Implementation of Single Precision Floating Point Multiplier Using High Speed Compressors

Run-Time Reconfigurable multi-precision floating point multiplier design based on pipelining technique using Karatsuba-Urdhva algorithms

A High Speed Binary Floating Point Multiplier Using Dadda Algorithm

Pipelined High Speed Double Precision Floating Point Multiplier Using Dadda Algorithm Based on FPGA

IJCSIET--International Journal of Computer Science information and Engg., Technologies ISSN

International Journal of Research in Computer and Communication Technology, Vol 4, Issue 11, November- 2015

Chapter 3: part 3 Binary Subtraction

Operations On Data CHAPTER 4. (Solutions to Odd-Numbered Problems) Review Questions

An Effective Implementation of Dual Path Fused Floating-Point Add-Subtract Unit for Reconfigurable Architectures

FLOATING POINT ADDERS AND MULTIPLIERS

EC2303-COMPUTER ARCHITECTURE AND ORGANIZATION

An Efficient Implementation of Floating Point Multiplier

Design and Simulation of Pipelined Double Precision Floating Point Adder/Subtractor and Multiplier Using Verilog

Implementation of Efficient Modified Booth Recoder for Fused Sum-Product Operator

Figurel. TEEE-754 double precision floating point format. Keywords- Double precision, Floating point, Multiplier,FPGA,IEEE-754.

VLSI Implementation of High Speed and Area Efficient Double-Precision Floating Point Multiplier

AN IMPROVED FUSED FLOATING-POINT THREE-TERM ADDER. Mohyiddin K, Nithin Jose, Mitha Raj, Muhamed Jasim TK, Bijith PS, Mohamed Waseem P

University, Patiala, Punjab, India 1 2

An FPGA based Implementation of Floating-point Multiplier

Prachi Sharma 1, Rama Laxmi 2, Arun Kumar Mishra 3 1 Student, 2,3 Assistant Professor, EC Department, Bhabha College of Engineering

International Journal Of Global Innovations -Vol.1, Issue.II Paper Id: SP-V1-I2-221 ISSN Online:

International Journal of Advanced Research in Computer Science and Software Engineering

Module 2: Computer Arithmetic

VHDL implementation of 32-bit floating point unit (FPU)

Review on Floating Point Adder and Converter Units Using VHDL

ADDERS AND MULTIPLIERS

Implementation of a High Speed Binary Floating point Multiplier Using Dadda Algorithm in FPGA

Chapter 4 Arithmetic Functions

An Implementation of Double precision Floating point Adder & Subtractor Using Verilog

A Library of Parameterized Floating-point Modules and Their Use

COMPUTER ARCHITECTURE AND ORGANIZATION. Operation Add Magnitudes Subtract Magnitudes (+A) + ( B) + (A B) (B A) + (A B)

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Digital Computer Arithmetic ECE 666

High Speed Single Precision Floating Point Unit Implementation Using Verilog

International Journal of Advanced Research in Computer Science and Software Engineering

DESIGN OF LOGARITHM BASED FLOATING POINT MULTIPLICATION AND DIVISION ON FPGA

IEEE-754 compliant Algorithms for Fast Multiplication of Double Precision Floating Point Numbers

CO212 Lecture 10: Arithmetic & Logical Unit

ARCHITECTURAL DESIGN OF 8 BIT FLOATING POINT MULTIPLICATION UNIT

CPE300: Digital System Architecture and Design

FPGA Implementation of Low-Area Floating Point Multiplier Using Vedic Mathematics

Implementation of IEEE754 Floating Point Multiplier

Chapter 10 Binary Arithmetics

A Novel Carry-look ahead approach to an Unified BCD and Binary Adder/Subtractor

By, Ajinkya Karande Adarsh Yoga

VTU NOTES QUESTION PAPERS NEWS RESULTS FORUMS Arithmetic (a) The four possible cases Carry (b) Truth table x y

COMPUTER ORGANIZATION AND DESIGN. 5 th Edition. The Hardware/Software Interface. Chapter 3. Arithmetic for Computers Implementation

A Pipelined Fused Processing Unit for DSP Applications

UNIT-III REGISTER TRANSFER LANGUAGE AND DESIGN OF CONTROL UNIT

CS Computer Architecture. 1. Explain Carry Look Ahead adders in detail

A Decimal Floating Point Arithmetic Unit for Embedded System Applications using VLSI Techniques

PROJECT REPORT IMPLEMENTATION OF LOGARITHM COMPUTATION DEVICE AS PART OF VLSI TOOLS COURSE

Binary Addition. Add the binary numbers and and show the equivalent decimal addition.

Implementation of Double Precision Floating Point Adder with Residue for High Accuracy Using FPGA

IJRASET 2015: All Rights are Reserved

COMPARISION OF PARALLEL BCD MULTIPLICATION IN LUT-6 FPGA AND 64-BIT FLOTING POINT ARITHMATIC USING VHDL

ISSN: X Impact factor: (Volume3, Issue2) Analyzing Two-Term Dot Product of Multiplier Using Floating Point and Booth Multiplier

Comparison of Adders for optimized Exponent Addition circuit in IEEE754 Floating point multiplier using VHDL

COMPUTER ARCHITECTURE AND ORGANIZATION Register Transfer and Micro-operations 1. Introduction A digital system is an interconnection of digital

10.1. Unit 10. Signed Representation Systems Binary Arithmetic

FPGA Implementation of Multiplier for Floating- Point Numbers Based on IEEE Standard

Assistant Professor, PICT, Pune, Maharashtra, India

Low Power Floating-Point Multiplier Based On Vedic Mathematics

Lecture Topics. Announcements. Today: Integer Arithmetic (P&H ) Next: continued. Consulting hours. Introduction to Sim. Milestone #1 (due 1/26)

Floating Point. The World is Not Just Integers. Programming languages support numbers with fraction

ISSN: (Online) Volume 2, Issue 10, October 2014 International Journal of Advance Research in Computer Science and Management Studies

Design and Implementation of Floating Point Multiplier for Better Timing Performance

CS 101: Computer Programming and Utilization

Number Systems Standard positional representation of numbers: An unsigned number with whole and fraction portions is represented as:

Organisasi Sistem Komputer

ECE 645: Lecture 1. Basic Adders and Counters. Implementation of Adders in FPGAs

Chapter 4. Operations on Data

A comparative study of Floating Point Multipliers Using Ripple Carry Adder and Carry Look Ahead Adder

C NUMERIC FORMATS. Overview. IEEE Single-Precision Floating-point Data Format. Figure C-0. Table C-0. Listing C-0.

Digital Logic & Computer Design CS Professor Dan Moldovan Spring 2010

Chapter 3: Arithmetic for Computers

A Quadruple Precision and Dual Double Precision Floating-Point Multiplier

1. NUMBER SYSTEMS USED IN COMPUTING: THE BINARY NUMBER SYSTEM

Chapter 10 - Computer Arithmetic

Pipelined Quadratic Equation based Novel Multiplication Method for Cryptographic Applications

Transcription:

Double Precision IEEE-754 Floating-Point Adder Design Based on FPGA Adarsha KM 1, Ashwini SS 2, Dr. MZ Kurian 3 PG Student [VLS& ES], Dept. of ECE, Sri Siddhartha Institute of Technology, Tumkur, Karnataka, India 1 Assistant professor, Dept. of ECE, Sri Siddhartha Institute of Technology, Tumkur, Karnataka, India 2 HOD, Dept. of ECE, Sri Siddhartha Institute of Technology, Tumkur, Karnataka, India 3 ABSTRACT:Because of dynamic representation capabilities and large spectrum of numbers can be represented with limited number of bits, floating-point numbers are being widely adapted in the fields of scientific applications. A floating-point arithmetic unit is specifically designed to carry out on floating-point numbers and is one of the most common part of any computing system in the area of binary applications. Floating-point additions are the most frequent floating-point operations and floating-point adders are therefore critically important components in signal processing and embedded platforms. This review paper presents the survey of related works of different algorithms/techniques which are important for implementation of double precision floating point adder with reduced latency based on FPGAs. As per the review paper the basic design deals with the floating point adder implementation, the proposed design handles delay optimization. KEYWORDS:Floating-Point Addition, IEEE-754, FPGA, Delay Optimization. I.INTRODUCTION In digital systems IEEE-754 [8] standard is used for representing the floating-point numbers and are of different types according to their mantissa bit length. There are half, single, double and quadruple precision binary numbers having a mantissa of bit length 16, 32, 64 and 128 respectively in accordance with IEEE-754 2008 standard format. Using double precision binary numbers large spectrum of numbers can be expressed with a limited number of bits and hence these are being most widely used in the area of scientific and engineering applications. Double precision representation is the IEEE-754 standard of 64-bit format with 1 bit for sign, 11 bits of exponent and 52 mantissa bits, thus provides greater dynamic range as shown in fig 1. Length = 1 Length = 11 Length = 52 Sign (S) Exponent (E) Significand Fig 1: General Representation of IEEE 754-2008 Double Precision Floating-Point Numbers In the implementation of high performance systems, modern FPGAs are considering as a great valuable quality hardware prototyping tools. Every year FPGAs are coming out with more features and are being established as most valuable tools in the implementation of high performance systems to perform computations at high clock frequencies. FPGAs have dedicated multipliers, memories and carry chains, most of the blocks are aiming at signal processing computations and even general processors can be incorporated. Although FPGAs have the advantage of reprogrammability over general purpose processors with the parallel processing advantage and speed of custom hardware improves fixed point computations but it is harder to implement high performance floating-point computations on FPGAs. One of the major challenge during the implementation is the requirement of the normalization in a floatingpoint addition. To handle arithmetic operations, floating-point unit is designed and it also performs functions such as exponential calculations. This paper presents the review of an implementation of the double precision floating point Copyright to IJAREEIE www.ijareeie.com 8431

adder which can perform the addition within two clock cycles with considerable reduction in latency. Over the fixed point and integer representation floating point-representation has its own advantages like it can support wide range of values. A.OPTIMIZATION TECHNIQUES [4] II.RELATED WORKS A detailed examination of optimization techniques enables designer to demonstrate how these methods can be grouped to achieve an overall fast floating-point adder implementation. In general, considerable reduction of latency by parallel paths requires balancing the delay of the parallel paths. In [4] such a balance was achieved by a gate level consideration of the design. A brief overview of the optimization techniques used in paper [4] are given below: A two path design with a nonstandard separation criterion. Instead of dividing based on magnitude of the exponent difference, defining a separation criterion that also considers whether the operation is effective subtraction and the significand difference. Here the advantage is that, alignment shift and normalization shift take place only in one of the paths and the full exponent difference is calculated only in one path and also this separation technique requires rounding to take place only in one path. IEEE rounding modes are reduced to three modes and use of injection based rounding. A simple design can be achieved by using unconditional pre-shifts for effective subtractions to reduce to two the number of binades that sum and difference of significand may belong to. Exponent s difference and significand s differences sign magnitude representation is derived from one s compliment representation of the difference. Sum and the incremented sum of the significands can be computed using the parallel-prefix adder. Using of recoding s to estimate the number of leading zeros in the non-redundant representation of a number represented as a borrow-save number. Post-normalization will take place before the rounding decision is ready. From the results of implementation in [4], one can see that the presentation of technology-independent analysis and optimization of the IEEE floating point addition implementation based on the logical effort hardware model and determined optimal gate size and optimal buffer insertion. The adder receives normalized numbers, supports all IEEE rounding modes, and registers the outputs correctly normalized rounded sum/difference in the format required by the IEEE standard. The presented algorithm was a two staged pipeline partitioned into two parallel paths called R-path and the N-path. The final result is selected between the outcomes of the two paths. The floating-point adder designed achieved a low latency by combining various optimization techniques. B.PIPELINED FPGA ARCHITECTURE [5] The advantage of the pipelining architecture mechanism is that, regardless of having a higher input and output sequential length, it offers an unmatched throughput by virtue of their assembly structure. Pipelined FPGA architectures significantly reduce the latency and some of the commercial (open source) floating-point cores assume the precision to be the critical parameter. To get optimal throughput rates, the number of pipelining stages also has to be considered as a parameter. Some of the floating-point cores use a common format with conversion to and from the IEEE-754 technical standard at interfaces to other resources in the system. Various tools have been developed to automate the optimization and generation of the floating-point units. C.PIPELINED PACKET-FORWARDING PARADIGM [2] Paper [2] presents the pipelined adder design and algorithm for a floating-point addition employing a packet forwarding pipeline paradigm. To handle data hazards in deeply pipelined floating-point pipeline, their proposed work and packet forwarding format constitute a new paradigm. The two algorithms addition and rounding form a phase of four stage execution pipeline with each stage responsible for implementation in a minimum clock period. The first two clock cycles perform addition and last two clock cycles perform the rounding. There are two format given to the adder one Copyright to IJAREEIE www.ijareeie.com 8432

operand in standard binary format (Fig 2) at the start of the cycle first and other operand is in the packet forwarding floating-point format, (Fig 3) which is divided into four parts: the sign bit, the exponent string, the principle part of the significand, and carry-round packet. The second operand s first three parts are input at the start of the cycle one and the carry-round packet is input at the start of the input two. According to the pipelined packet-forwarding paradigm, a floating-point adder is implemented by a pipeline of 4 stages as depicted in Fig.4. The first two stages perform the addition operation and the last two stages deal with rounding. The registered result at the output is in two formats one in packet forwarding floating-point format at the end of the cycle two and three to allow forwarding with an effective latency of two cycles and another, in the standard IEEE 754 at the end of cycle four. Fig 5 depicts an execution of successive dependent floating-point operations. The latency is 2 clock cycles and there is only one stall cycle between successive dependent operations. Sign S Exponent e 14 e 13 e 12 e 0 Significand 1 a 1 a 2 a 61 a 62 a 63 Fig 2. Standard IEEE-754 Floating-Point operand format Sign S Exponent e 14 e 13 e 12 e 0 Carry-Round Packet C + 62 C + C - 63 62 C- 63 Principal part 1 b 0 + b 0 - b 1 + b 0 - + + b 61 b 62 - - b 62 b 62 Fig 3. Packet-forwarding floating-point operand format Sign, exponent and Principal part input Carry-round input Standard input Sign, exponent and Principal output Carry-round output standard output A A2 R1 R2 Fig 4. Input-Output schedule of pipeline A1 A2 R1 R2 A1A2R1R2 M1 M2 R1 R2 Fig 5.Execution of successive dependent operations Copyright to IJAREEIE www.ijareeie.com 8433

D.FLAGGED PREFIX ADDITION [1] Paper [1] describes the reduced latency IEEE floating point standard architectures which uses flagged prefix addition to merge rounding with the significand addition. They presented architecture of a floating-point adder with latency of two cycle with potentially the same cycle time that also employs flagged prefix addition as the modified architecture for their three cycle latency and supports all rounding modes. A flagged prefix adder is a modified parallel prefix type carry look ahead adder which calculates the longest string of one s from the second LSB upwards and sets them as flag bits to indicate those bits which can be inverted for incrementing a two s compliment number. A three cycle floatingpoint adder using a flagged prefix integer adder is similar in design to the traditional three stage pipelined adder, in that rounding of the significand, which is normally done after the normalization shift is merged with the significand addition in the second pipeline stage. A two cycle floating-point adder has separate data paths depending on whether or not the data requires a large alignment shift or large normalization shift. N.Kikkeri, P.M Seidel (2007) reported an implementation of highly optimized double precision IEEE floating-point adder, which was based strongly on optimization techniques such as nonstandard dividing of the two paths, a simple rounding algorithm, unification of rounding cases for addition and subtraction, one s complement subtraction based sign magnitude computation of a difference, compound adders and fast circuits for approximate counting of leading zeros from borrow save representation. They reported fully gate-level verification and is important to increase the confidence in the resulting implementation as opposed to verification on high level abstractions where the representation of the inputs and the blocks may be different from the actual hardware. They used Verilog code for their design then compiled in altera qwartus2 to realize the implementation. The implemented design was having a successfully reduction in latency on an altera starix device and their design was implemented as a two stage pipeline to handle data hazards in deeply pipelining stages. [7]. Purna Ramesh Addanki, Venkata Nagartna Tilak Alapati and Mallikarjuna Prasad Avana (2013) presented a high speed floating-point double precision adder/subtractor and multiplier, which are implemented on a virtex-6 FPGA using Verilog language. Their proposed designs were compliant with IEEE 754 standard format and handles overflow, underflow cases and rounding mode. The IEEE standard specifies four rounding modes and the rounding odes are selected for various bit combinations of mode. Based on the changes in the rounding to the mantissa corresponding changes has to be made in the exponent path also. They showed that, their presented design was achieved high operating frequency with better accuracy and considerably good performance. [9]. III.PROPOSED DESIGN WORK Proposed implementation work is following the similar design method as in [1]. Floating-Point adder is partitioned into two stage pipeline as shown in Fig 6. Two stage pipeline is divided into two paths namely R and N paths, these are being selected based on the exponent difference of two operands. The new algorithm is being achieved at by bringing a few implementation changes in the algorithm of [1]. The two pipeline stages are going to execute in two different clock cycles. Let, two input operands which are in the IEEE-754 format, represented by their bit vectors as, (SA, EA, FA) and (SB, EB, FB) and let SOP defines the operation going to perform (0-for addition and 1-for subtraction). The required floating-point addition operation will be performed by the below formulae: Where, rnd(sum) = rnd(( 1). 2. FA + ( 1) (). 2. FB) S. EFF = SA SB SOP So, sum = ( 1). 2. (Fl + ( 1). Fs. 2 ) Where δ is the exponent difference and (Sl, El, Fl), (Ss, Es, Fs) are the small and large operands respectively. Copyright to IJAREEIE www.ijareeie.com 8434

Flip-Flops Significand s complement Exponent Difference Alignment 1 Swapping Alignment 2 Flip-Flops 2 nd cycle 1 st cycle Significand s Addition Rounding Decision Post- Normalization Normalization Alignment Output Result [0:64] A.OVERVIEW OFFIRST CLOCK CYCLE Fig. 6. Higher Level Representation of the Algorithm Fig.7 is the first stage of the pipelining mechanism.exponent difference will be calculated for two ranges: the medium exponent is in the range of [-63, 64], and the big exponent difference is in the range of [-, -64] and [65, ]. The outputs of the exponent difference module are given as follows:sign_med and MAG_MAD are the sign-magnitude representation of exponent difference (δ). If δ is in the medium exponent difference interval. ( 1).{ } = δ δ 1 if 64 δ 1 if 0 δ 63 "don t care otherwise In the big exponent difference interval range, the required alignment shift is at least 64 bit positions. SIGN_BIG is the sign bit of the exponent difference δ. IS_BIG is a flag defined by: 1, if δ 65 or δ 64 IS_BIG = 0, otherwise Copyright to IJAREEIE www.ijareeie.com 8435

An 11-bit adder is used to subtract the exponents and the carry out is used to identify if e1 is smaller than e2. If the result is negative, it has to be complemented and a 1 has to be added to it in order to get the required difference. The exponent difference module can be implemented by cascading a seven bit adder with five bit adder. The seven bit adder calculates the lazy 1s complement exponent difference in the medium interval. This difference is converted to sign and magnitude representations as shown in fig 6. Using cascading adders one can evaluate the exponent difference in parallel with finding out whether the exponent difference is in the big interval or range. Signal SIGN_BIG is the MSB of the lazy 1s complement exponent difference, IS_BIG signal is determining by OR-ing the bits in the positions [6:10] of magnitude of the lazy 1s complement exponent difference and explains why the medium interval is not symmetric around zero.one s complement box is responsible for computing the signals FAO, FBO, and S.EFF. The FAO and FBO signals are specified by FA[0: 52], FB[0: 52], if S. EFF = 0 FAO[0: 52], FBO[0: 52] = not(fa[0: 52]), not(fb[0: 52]), otherwise If the range of exponent difference is in the medium interval then pre-shift and align region computations are considered as relevant. If an effective subtraction takes place then the significands are pre-shifted. Based upon the signal sign_bigthe large operand is selected in the swap region and similarly based upon the signal sign_med small operand is selected for the medium exponent difference and based on sign_big large exponent difference range is selected as shown in fig.6.align 2 region deals with pre-shifting the minuend in case an effective subtraction takes place. Fig. 6. Block Level Representation of the 1 st Cycle of the Algorithm Copyright to IJAREEIE www.ijareeie.com 8436

B.OVERVIEW OF SECOND CLOCK CYCLE This is the second stage of the pipelining mechanism. Here the two significands which are pre-processed are added and the result is rounded in accordance with the IEEE-754 standard floating-point rounding algorithm.inputs to the second cycle are: the sign bit SL, exponent representation el, the significand strings FLP[-1:52] and FSOPA[-1:116], and the rounding mode. At the end rounding algorithm will be implemented and it is normalized. The output result will be a sixty four bit binary floating-point number. rnd(sum) = rnd(( 1). 2. FA + ( 1) (). 2. FB) Fig 7. Block Level Representation of the 2 nd Cycle of the Algorithm IV.RESULT AND CONCLUSION Survey of different algorithms/techniques which are essential for implementation of floating-point arithmetic is discussed. Overview of the proposed implementation work is shown. We have seen that the double precision floatingpoint unit compared to single precision floating-point provides high speed and gives more accuracy. FPGA based embedded systems have a higher advantage of lower computational aspects, and the double precision floating-point adder supports the IEEE standard format. REFERENCES [1] A. Beaumont-Smith, N. Burgess, S. Lefrere, C. Lim, "Reduced Latency IEEE Floating-Point Standard Adder Architectures," Proc. of 14th IEEE Symposium on Computer Arithmetic, pp. 35-43, 1999. [2] A. Nielsen, D. Matula, e.n. Lyu, G. Even, "IEEE Compliant Floating-Point Adder that Conforms with the Pipelined Packet-Forwarding Paradigm," IEEE Trans. on Computers, vol. 49, no. 1, pp. 33-47, Jan. 2000. [3] Paschalakis, S., Lee, P., Double Precision Floating-Point Arithmetic on FPGAs, In Proc. 2003 2 nd IEEE International Conference on Field Programmable Technology (FPT 03), Tokyo, Japan, 2003. Copyright to IJAREEIE www.ijareeie.com 8437

[4] Peter-Michael Seidel, Guy Even, "Delay-Optimized Implementation of IEEE Floating-Point Addition", IEEE Trans. on Computers, vol. 53, no. 2, pp. 97-113, Feb. 2004. [5] Akshay Sharma, Katherine Compton, Carl Ebeling and Scott Hauck, Exploration of Pipelined FPGA Interconnect Structures February 22 24, 2004, Monterey, California, USA. [6] Xilinx Inc., Floating-Point Operator v3.0. Product Specification, 2005. [7] N. Kikkeri, P.M. Seidel, "An FPGA Implementation of a Fully Verified Double Precision IEEE Floating-Point Adder", Proc. of IEEE International Conference on Application-specific Systems, Architectures and Processors, pp. 83-88, 9-11 July 2007. [8] IEEE Computer Society, IEEE Standard for Floating-Point Arithmetic, IEEE Std. 754 TM -2008 (IEEE Std 754-1985), Aug, 2008. [9] Purna Ramesh Addanki, Venkata Nagaratna Tilak Alapati and Mallikarjuna Prasad Avana, An FPGA Based High Speed IEEE 754 Double Precision Floating Point Adder\Subtractor and Multiplier Using Verilog, International journal of advanced science and technology. Vol. 52, March, 2013. [10] Xilinx; Virtex-4 User Guide. Copyright to IJAREEIE www.ijareeie.com 8438