Measuring Improvement When Using HUB Formats to Implement Floating-Point Systems under Round-to- Nearest

Similar documents
FPGA IMPLEMENTATION OF FLOATING POINT ADDER AND MULTIPLIER UNDER ROUND TO NEAREST

Floating Point Square Root under HUB Format

Floating Point. The World is Not Just Integers. Programming languages support numbers with fraction

Floating-Point Butterfly Architecture Based on Binary Signed-Digit Representation

Divide: Paper & Pencil

Floating-Point Data Representation and Manipulation 198:231 Introduction to Computer Organization Lecture 3

Floating-point representations

Floating-point representations

ECE232: Hardware Organization and Design

Number Systems CHAPTER Positional Number Systems

SONIA GONZALEZ-NAVARRO AND JAVIER HORMIGO Dept. Computer Architecture Universidad de Málaga (Spain)

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Digital Computer Arithmetic ECE 666

Number Systems and Computer Arithmetic

A Library of Parameterized Floating-point Modules and Their Use

Floating point. Today! IEEE Floating Point Standard! Rounding! Floating Point Operations! Mathematical properties. Next time. !

Floating Point (with contributions from Dr. Bin Ren, William & Mary Computer Science)

Computer (Literacy) Skills. Number representations and memory. Lubomír Bulej KDSS MFF UK

Signed Multiplication Multiply the positives Negate result if signs of operand are different

CO212 Lecture 10: Arithmetic & Logical Unit

Inf2C - Computer Systems Lecture 2 Data Representation

CS 64 Week 1 Lecture 1. Kyle Dewey

COMP2611: Computer Organization. Data Representation

Floating point. Today. IEEE Floating Point Standard Rounding Floating Point Operations Mathematical properties Next time.

±M R ±E, S M CHARACTERISTIC MANTISSA 1 k j

Computer Arithmetic Ch 8

Floating Point Representation in Computers

Implementation of IEEE754 Floating Point Multiplier

Data Representation Floating Point

Computer Arithmetic Ch 8

Floating Point January 24, 2008

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Digital Computer Arithmetic ECE 666

EE878 Special Topics in VLSI. Computer Arithmetic for Digital Signal Processing

Implementation of Double Precision Floating Point Multiplier in VHDL

A High-Speed FPGA Implementation of an RSD- Based ECC Processor

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Digital Computer Arithmetic ECE 666

UNIT IV: DATA PATH DESIGN

The Sign consists of a single bit. If this bit is '1', then the number is negative. If this bit is '0', then the number is positive.

Floating Point Arithmetic

Chapter 2 Float Point Arithmetic. Real Numbers in Decimal Notation. Real Numbers in Decimal Notation

Floating Point Puzzles. Lecture 3B Floating Point. IEEE Floating Point. Fractional Binary Numbers. Topics. IEEE Standard 754

Computer Architecture Chapter 3. Fall 2005 Department of Computer Science Kent State University

Number Systems and Binary Arithmetic. Quantitative Analysis II Professor Bob Orr

Systems I. Floating Point. Topics IEEE Floating Point Standard Rounding Floating Point Operations Mathematical properties

COMPUTER ARCHITECTURE AND ORGANIZATION. Operation Add Magnitudes Subtract Magnitudes (+A) + ( B) + (A B) (B A) + (A B)

COMP 122/L Lecture 2. Kyle Dewey

CS 261 Fall Floating-Point Numbers. Mike Lam, Professor.

3.5 Floating Point: Overview

Floating Point Puzzles. Lecture 3B Floating Point. IEEE Floating Point. Fractional Binary Numbers. Topics. IEEE Standard 754

Number Representations

COSC 243. Data Representation 3. Lecture 3 - Data Representation 3 1. COSC 243 (Computer Architecture)

Module 2: Computer Arithmetic

System Programming CISC 360. Floating Point September 16, 2008

CS Computer Architecture. 1. Explain Carry Look Ahead adders in detail

Floating Point Numbers

Floating Point Numbers

Module 12 Floating-Point Considerations

Operations On Data CHAPTER 4. (Solutions to Odd-Numbered Problems) Review Questions

Representing and Manipulating Floating Points

Some issues related to double roundings

Data Representation Floating Point

Chapter 03: Computer Arithmetic. Lesson 09: Arithmetic using floating point numbers

Computer Architecture and IC Design Lab. Chapter 3 Part 2 Arithmetic for Computers Floating Point

Floating Point Puzzles The course that gives CMU its Zip! Floating Point Jan 22, IEEE Floating Point. Fractional Binary Numbers.

1. NUMBER SYSTEMS USED IN COMPUTING: THE BINARY NUMBER SYSTEM

Bryant and O Hallaron, Computer Systems: A Programmer s Perspective, Third Edition. Carnegie Mellon

ECE331: Hardware Organization and Design

CS 261 Fall Floating-Point Numbers. Mike Lam, Professor.

Foundations of Computer Systems

Design and Optimized Implementation of Six-Operand Single- Precision Floating-Point Addition

Giving credit where credit is due

International Journal of Advanced Research in Electrical, Electronics and Instrumentation Engineering

Numeric Encodings Prof. James L. Frankel Harvard University

Giving credit where credit is due

Floating Point. CSE 238/2038/2138: Systems Programming. Instructor: Fatma CORUT ERGİN. Slides adapted from Bryant & O Hallaron s slides

CS 265. Computer Architecture. Wei Lu, Ph.D., P.Eng.

Floating-Point Matrix Product on FPGA

Chapter 3: Arithmetic for Computers

EE 109 Unit 20. IEEE 754 Floating Point Representation Floating Point Arithmetic

EC2303-COMPUTER ARCHITECTURE AND ORGANIZATION

MIPS Integer ALU Requirements

Number Systems. Both numbers are positive

EE 109 Unit 19. IEEE 754 Floating Point Representation Floating Point Arithmetic

Computer Architecture and Organization

Floating Point : Introduction to Computer Systems 4 th Lecture, May 25, Instructor: Brian Railing. Carnegie Mellon

Low Power Floating-Point Multiplier Based On Vedic Mathematics

A Binary Floating-Point Adder with the Signed-Digit Number Arithmetic

Design and Implementation of Effective Architecture for DCT with Reduced Multipliers

CS429: Computer Organization and Architecture

Computer Organisation CS303

fractional quantities are typically represented in computers using floating point format this approach is very much similar to scientific notation

Double Precision Floating-Point Multiplier using Coarse-Grain Units

Vendor Agnostic, High Performance, Double Precision Floating Point Division for FPGAs

Floating Point. EE 109 Unit 20. Floating Point Representation. Fixed Point

Floating-Point Arithmetic

Chapter Three. Arithmetic

What Every Computer Scientist Should Know About Floating-Point Arithmetic

2 General issues in multi-operand addition

The ALU consists of combinational logic. Processes all data in the CPU. ALL von Neuman machines have an ALU loop.

Written Homework 3. Floating-Point Example (1/2)

CO Computer Architecture and Programming Languages CAPL. Lecture 15

Transcription:

Measuring Improvement When Using HUB Formats to Implement Floating-Point Systems under Round-to- Nearest Abstract: This paper analyzes the benefits of using half-unitbiased (HUB) formats to implement floatingpoint (FP) arithmetic under a round-to-nearest mode from a quantitative point of view. Using the HUB formats to represent numbers allows the removal of the rounding logic of arithmetic units, including sticky-bit computation. This is shown for FP adders, multipliers, and converters. Experimental analysis demonstrates that the HUB formats and the corresponding arithmetic units maintain the same accuracy as the conventional ones. On the other hand, the implementation of these units, based on basic architectures, shows that the HUB formats simultaneously improve area, speed, and power consumption. In addition, based on the data obtained from the synthesis, an HUB single-precision adder is 14% faster but consumes 38% less area and 26% less power than the conventional adder. Similarly, an HUB single-precision multiplier is 17% faster, uses 22% less area, and consumes slightly less power than the conventional multiplier. At the same speed, the adder and the multiplier achieve area and power reductions of up to 50% and 40%, respectively. The proposed architecture of this paper analysis the logic size, area and power consumption using Xilinx 14.2. Enhancement of the project: Existing System: A totally different approach would be to use a new real-number encoding, in order to simplify the implementation of RN. Thus, the problem would change from optimizing the rounding operation to dealing with arithmetic operations under the new number representation. This proposal is found in with RN representations and with half-unit-biased (HUB) formats. Together with other advantages, these new formats allow performing RN simply by truncation. On the

other hand, these new formats are based on the simple modifications of the conventional formats and so could be applied to practically any conventional format. In this paper, we focus on the HUB FP formats. The efficiency of using the HUB formats for a fixedpoint representation has been demonstrated. By reducing the bitwidth while maintaining the same accuracy, the area cost and delay of finite impulse response filter implementations have been dramatically reduced, and similarly for the QR decomposition. HALF-UNIT-BIASED FP FORMATS An HUB FP format should include a significand that is represented using an HUB fixed-point number and an exponent that is represented in any conventional way [11]. An HUB fixed-point format is produced when the exactly represented numbers (ERNs) of a conventional representation are increased (or biased) by half unit in the last place (ULP). This shifting of the ERNs could be seen as an implicit least significant bit (ILSB) set to one. For example, the HUB version of the IEEE-754 single precision has 25 bits for the significand, where the first and last bits are implicit and equal to 1, but only 23 bits are stored, as in the conventional version. Fig. 1 shows an example for a binary FP format with 3-bit significand. Given a real value, its representation using either conventional or HUB formats will produce different rounding errors, although the accuracy of both the formats is the same.

Disadvantages: Fig. 1. Example of ERNs for a conventional FP format and its HUB version. RN is difficult Proposed System: The general procedure to operate with the HUB numbers involves the following steps: First, the ILSB is explicitly appended to the significand of input operands. Second, the operation is

performed in a classic way, such that all the bits of the significand result before rounding are obtained. Finally, the significand is rounded simply by truncation. FP Addition for HUB Numbers: A basic FP addition for conventional numbers requires several steps, which could be implemented using the significand datapath shown in Fig. 2(a). Fig. 2. Basic FP adder architectures for (a) standard and (b) HUB numbers. FP Multiplication for HUB Numbers: A classic FP multiplication is simpler than FP addition. The new exponent is computed by adding the exponents of the input operands, whereas the significand are multiplied, thus obtaining a value that is double the size.

Fig. 3. Basic FP multiplier for (a) standard and (b) HUB numbers. Conversion between FP Numbers of Different Sizes for HUB Numbers: Conversion between FP formats with different sizes is another operation that could benefit from using the HUB formats. We will focus on the conversion of the significand, since the conversion

of exponent is not affected when using the HUB formats (because the HUB FP formats use a conventional number to represent the exponent). Conversion Between Conventional and HUB Formats: The conversion between numbers under a conventional format and its corresponding HUB format (i.e., both the formats having the same number of explicit bits) may be required when data are exchanged between systems working in these different formats. Given that those formats do not share any ERN (except special cases), this conversion always implies a rounding and, consequently, a rounding error. In addition, each ERN of one format is always at the midpoint between two ERNs of the other format, i.e., it is a tie case. This fact simplifies the hardware, because the computation of the sticky bit is not required; however, the magnitude of rounding error is always 0.5 ULPs. Advantages: 2 s complement operation is simple by a bitwise inversion RN by truncation Software implementation: Modelsim Xilinx ISE