DEC64 is a number type. It can precisely represent
decimal fractions with 16 decimal places, which makes it very well suited
to all applications that are concerned with money. It can represent
values as gargantuan as `3.6028797018963967E+143` or as measly
as `1.0E-127`, which makes it well suited to most scientific
applications. It can provide very fast performance on integer values,
eliminating the need for a separate `int` type and avoiding the
terrible errors than can result from `int` truncation.

DEC64 is intended to be the only number type in the next generation of application programming languages.

DEC64 represents numbers as 64 bit values composed of 2 two’s complement components: a 56 bit `coefficient` and an 8 bit `exponent`. The `coefficient` is in the high order end, and the `exponent` is in the low order end. The `coefficient`’s decimal point is between bits 8 and 7.

63 | 8 | 7 | 0 |

coefficient |
exponent |

The `coefficient` is in the range -`36028797018963968` .. `36028797018963967. `The `exponent` is in the range `-127` .. `127`. Numbers may not use an `exponent` containing `-128``. The ``value` of a number is obtained from this formula:

`value`` = ``coefficient`` * 10`^{exponent}

Normalization is not required, and is usually not desired. Integers can have an `exponent` of `0` as long as the `coefficient` is less than 36 quadrillion. Addition of numbers with equal `exponents` could be performed in a single machine cycle.

There are 255 possible representations of zero. They are all considered to be equal.

There is a special value called `nan` that has a `coefficient` of `0` and an `exponent` of `-128`. The result of division by zero is `nan`. `nan` is also the result of operations that produce results that are too large to be represented. `nan` is equal to itself.

When an arithmetic operation has an input with an `exponent` of `-128`, the result will be `nan`. Applications are free to use the `coefficient` as they wish when the `exponent` is `-128`, since in that case the `coefficient` has no arithmetic significance. One possible use is to store object pointers in the `coefficient`.

DEC64 can be implemented efficiently in hardware or software.

Conversion to and from textual representations is simple and straightforward and free of the complexities that binary floating formats must wrestle with to minimize the inevitable errors caused by the fundamental incompatibility of the binary and decimal systems. DEC64 instead uses an internal representation that is very compatible with the `E` notation.

To convert an `int` to DEC64, shift it left 8 bits. To unpack a `coefficient`, shift it right 8 bits with sign extension. The `exponent` can be unpacked at no cost on x64 architecture because the least significant byte can be accessed directly.

There is a fast path for addition of integers that takes only 5 instructions:

; Add rdx to rax. mov cl,al ; load the exponent of rax into cl or cl,dl ; then or the two exponents together jnz slow_path ; if both are zero, take the fast path add rax,rdx ; add the coefficients together jo overflow ; if there was no overflow, we are done

The fast path for addition in hardware should take only 1 cycle when the two exponents are equal to each other and there is no overflow. The fast path for multiplication in hardware takes the time it takes to do a 56*56 signed multiply when there is no overflow.

A reference implementation is available at https://github.com/douglascrockford/DEC64. It produces an object file about 2K in size.

The idea of using powers of *ten* instead of powers of *two* is not new. For example,

Floating point subroutines and interpretive systems for early machines were coded by D. J. Wheeler and others, and the first publication of such routines was in

The Preparation of Programs for an Electronic Digital Computerby Wilkes, Wheeler, and Gill (Reading, Mass.: Addison-Wesley, 1951), subroutines A1-A11, pages 35-37 and 105-117. It is interesting to note that floatingdecimalsubroutines are described here, although a binary computer was being used; in other words, the numbers were represented as 10, not 2^{e}f, and therefore the scaling operations required multiplication or division by 10.^{e}f

The Art of Computer Programming, Volume 2: Seminumerical Algorithms, Third Editionby Donald Knuth (Addison-Wesley, 1998), page 226.

The book Knuth cited may have been the first software book. It described some of the libraries and conventions of Maurice Wilkes’s EDSAC, one of the first generation of Von Neumann machines. Some of its subroutines used a numeric format that was very similar to DEC64.

Floating point was so important that support for it was moved into hardware for better performance. This led to the development of binary floating point because a shift could be implemented much more easily than a divide by 10. It was discovered that by biasing the exponent and moving it to the position just after the sign bit that floating point numbers could be compared with integer opcodes, a nifty optimization. It was also discovered that because normalization always left a 1 bit in the most significant position of the significand, that that bit could be omitted, providing an additional bit of significance.

The Burroughs 5000 series had a floating point format in which an exponent of zero allowed the mantissa to be treated as an ordinary integer. DEC64 incorporates that brilliant idea.

Languages for scientific computing like FORTRAN provided multiple floating point types such as `REAL`

and `DOUBLE`

`PRECISION`

as well as `INTEGER`

, often also in multiple sizes. This was to allow programmers to reduce program size and running time. This convention was adopted by later languages like C and Java. In modern systems, this sort of memory saving is pointless. By giving programmers a choice of number types, programmers are required to waste their time making choices that don’t matter. Even worse, making a bad choice can lead to a loss of accuracy or destructive bugs. This is a bad practice that is very deeply ingrained.

Binary floating point trades away familiarity and decimal compatibility for performance. This made it unsuitable for business languages like COBOL. Decimal fractions cannot be represented accurately in binary floating point, which is a problem for programs that interact with humans, and is dangerous in programs that manipulate money. Exactness is required, so most business processing used BCD (Binary Coded Decimal) in which each digit is encoded in 4 bits. That created some inefficiency, but benefited from allowing a shift by 4 bits in place of the more complex divide by 10. For a time, mainframes could be ordered with optional floating point units for scientific computing, and optional BCD units for business computing.

The BASIC language eliminated much of the complexity of FORTRAN by having a single number type. This simplified the programming model and avoided a class of errors caused by selection of the wrong type. The efficiencies that could have gained from having numerous number types proved to be insignificant.

Business Basic was a dialect of BASIC that was developed by Basic/Four Corporation for its small business minicomputers. It used decimal floating point, much like the EDSAC, so the language could be used for both scientific and business applications. Business Basic could do everything that BASIC could do, and it could also handle money.

Intel undertook an ambitious architecture that was marketed as iAPX432. A decimal number type was considered for the 432 but rejected in favor of conventional binary floating point. The 432 failed, but its floating point unit was salvaged and repackaged as the 8087, a numeric coprocessor for the 8086. The 8087’s number types became the basis of the IEEE 754 floating point standard. The standard was so successful that it destroyed the market for decimal computing. A later revision of IEEE 754 attempted to remedy this, but the formats it recommended were so inefficient that it has not found much acceptance. DEC64 is a better alternative.