In mathematics, the repeating decimal 0.999... (which may also be written as 0.9, , 0.(9), or as "0.9" followed by any number of 9s in the repeating decimal) denotes a real number that can be shown to be the number one. In other words, the symbols 0.999... and 1 represent the same number. Proofs of this equality have been formulated with varying degrees of mathematical rigor, taking into account preferred development of the real numbers, background assumptions, historical context, and target audience.
In fact, every nonzero, terminating decimal has an equal twin representation with trailing 9s, such as 8.32 and 8.31999... The terminating decimal representation is almost always preferred, contributing to the misconception that it is the only representation. The same phenomenon occurs in all other bases or in any similar representation of the real numbers [1].
The equality of 0.999... and 1 is closely related to the absence of nonzero infinitesimals in the real number system, the most commonly used system in mathematical analysis. Some alternative number systems, such as the hyperreals, do contain nonzero infinitesimals. In most such number systems the standard interpretation of the expression 0.999... makes it equal to 1, but in some of these number systems, the symbol "0.999..." admits other interpretations that fall infinitesimally short of 1.
The equality 0.999... = 1 has long been accepted by mathematicians and taught in textbooks. Nonetheless, some students question or reject it. Some can be persuaded by an appeal to authority from textbooks and teachers, or by arithmetic reasoning, to accept that the two are equal.
Contents |
Algebraic proofs showing that 0.999... represents the number 1 use concepts such as fractions, long division, and digit manipulation to build transformations preserving equality from 0.999... to 1.
One reason that infinite decimals are a necessary extension of finite decimals is to represent fractions. Using long division, a simple division of integers like 1⁄9 becomes a recurring decimal, 0.111…, in which the digits repeat without end. This decimal yields a quick proof for 0.999… = 1. Multiplication of 9 times 1 produces 9 in each digit, so 9 × 0.111… equals 0.999… and 9 × 1⁄9 equals 1, so 0.999… = 1:
Another form of this proof multiplies ⅓ = 0.333… by 3.
One can give another proof using the a modified form of the usual division algorithm.[2]. For example, consider one can write out the long division algorithm as follows:
When a number in decimal notation is multiplied by 10, the digits do not change but each digit moves one place to the left. Thus 10 × 0.999... equals 9.999..., which is 9 greater than the original number. To see this, consider that in subtracting 0.999... from 9.999..., each of the digits after the decimal separator cancels, i.e. the result is 9 − 9 = 0 for each such digit. The final step uses algebra:
Although these proofs demonstrate that 0.999... = 1, the extent to which they explain the equation depends on the audience. In introductory arithmetic, such proofs help explain why 0.999... = 1 but 0.333... < 0.4. And in introductory algebra, the proofs help explain why the general method of converting between fractions and repeating decimals works. But the proofs shed little light on the fundamental relationship between decimals and the numbers they represent, which underlies the question of how two different decimals can be said to be equal at all.[3] William Byers argues that a student who agrees that 0.999... = 1 because of the above proofs, but hasn't resolved the ambiguity, doesn't really understand the equation.[4] Fred Richman argues that the first argument "gets its force from the fact that most people have been indoctrinated to accept the first equation without thinking".[5]
Once a representation scheme is defined, it can be used to justify the rules of decimal arithmetic used in the above proofs. Moreover, one can directly demonstrate that the decimals 0.999... and 1.000... both represent the same real number; it is built into the definition. This is done below.
Since the question of 0.999... does not affect the formal development of mathematics, it can be postponed until one proves the standard theorems of real analysis. One requirement is to characterize real numbers that can be written in decimal notation, consisting of an optional sign, a finite sequence of any number of digits forming an integer part, a decimal separator, and a sequence of digits forming a fractional part. For the purpose of discussing 0.999..., the integer part can be summarized as b0 and one can neglect negatives, so a decimal expansion has the form
It should be noted that the fraction part, unlike the integer part, is not limited to a finite number of digits. This is a positional notation, so for example the 5 in 500 contributes ten times as much as the 5 in 50, and the 5 in 0.05 contributes one tenth as much as the 5 in 0.5.
Perhaps the most common development of decimal expansions is to define them as sums of infinite series. In general:
For 0.999... one can apply the convergence theorem concerning geometric series:[6]
Since 0.999... is such a sum with a common ratio r = 1⁄10, the theorem makes short work of the question:
This proof (actually, that 10 equals 9.999...) appears as early as 1770 in Leonhard Euler's Elements of Algebra.[7]
The sum of a geometric series is itself a result even older than Euler. A typical 18th-century derivation used a term-by-term manipulation similar to the algebraic proof given above, and as late as 1811, Bonnycastle's textbook An Introduction to Algebra uses such an argument for geometric series to justify the same maneuver on 0.999...[8] A 19th-century reaction against such liberal summation methods resulted in the definition that still dominates today: the sum of a series is defined to be the limit of the sequence of its partial sums. A corresponding proof of the theorem explicitly computes that sequence; it can be found in any proof-based introduction to calculus or analysis.[9]
A sequence (x0, x1, x2, ...) has a limit x if the distance |x − xn| becomes arbitrarily small as n increases. The statement that 0.999... = 1 can itself be interpreted and proven as a limit:[10]
The last step, that 1⁄10n → 0 as n → ∞, is often justified by the Archimedean property of the real numbers. This limit-based attitude towards 0.999... is often put in more evocative but less precise terms. For example, the 1846 textbook The University Arithmetic explains, ".999 +, continued to infinity = 1, because every annexation of a 9 brings the value closer to 1"; the 1895 Arithmetic for Schools says, "...when a large number of 9s is taken, the difference between 1 and .99999... becomes inconceivably small".[11] Such heuristics are often interpreted by students as implying that 0.999... itself is less than 1.
The series definition above is a simple way to define the real number named by a decimal expansion. A complementary approach is tailored to the opposite process: for a given real number, define the decimal expansion(s) to name it.
If a real number x is known to lie in the closed interval [0, 10] (i.e., it is greater than or equal to 0 and less than or equal to 10), one can imagine dividing that interval into ten pieces that overlap only at their endpoints: [0, 1], [1, 2], [2, 3], and so on up to [9, 10]. The number x must belong to one of these; if it belongs to [2, 3] then one records the digit "2" and subdivides that interval into [2, 2.1], [2.1, 2.2], ..., [2.8, 2.9], [2.9, 3]. Continuing this process yields an infinite sequence of nested intervals, labeled by an infinite sequence of digits b0, b1, b2, b3, ..., and one writes
In this formalism, the identities 1 = 0.999... and 1 = 1.000... reflect, respectively, the fact that 1 lies in both [0, 1] and [1, 2], so one can choose either subinterval when finding its digits. To ensure that this notation does not abuse the "=" sign, one needs a way to reconstruct a unique real number for each decimal. This can be done with limits, but other constructions continue with the ordering theme.[12]
One straightforward choice is the nested intervals theorem, which guarantees that given a sequence of nested, closed intervals whose lengths become arbitrarily small, the intervals contain exactly one real number in their intersection. So b0.b1b2b3... is defined to be the unique number contained within all the intervals [b0, b0 + 1], [b0.b1, b0.b1 + 0.1], and so on. 0.999... is then the unique real number that lies in all of the intervals [0, 1], [0.9, 1], [0.99, 1], and [0.99...9, 1] for every finite string of 9s. Since 1 is an element of each of these intervals, 0.999... = 1.[13]
The Nested Intervals Theorem is usually founded upon a more fundamental characteristic of the real numbers: the existence of least upper bounds or suprema. To directly exploit these objects, one may define b0.b1b2b3... to be the least upper bound of the set of approximants {b0, b0.b1, b0.b1b2, ...}.[14] One can then show that this definition (or the nested intervals definition) is consistent with the subdivision procedure, implying 0.999... = 1 again. Tom Apostol concludes,
The fact that a real number might have two different decimal representations is merely a reflection of the fact that two different sets of real numbers can have the same supremum.[15]
Some approaches explicitly define real numbers to be certain structures built upon the rational numbers, using axiomatic set theory. The natural numbers – 0, 1, 2, 3, and so on – begin with 0 and continue upwards, so that every number has a successor. One can extend the natural numbers with their negatives to give all the integers, and to further extend to ratios, giving the rational numbers. These number systems are accompanied by the arithmetic of addition, subtraction, multiplication, and division. More subtly, they include ordering, so that one number can be compared to another and found to be less than, greater than, or equal to another number.
The step from rationals to reals is a major extension. There are at least two popular ways to achieve this step, both published in 1872: Dedekind cuts and Cauchy sequences. Proofs that 0.999... = 1 which directly use these constructions are not found in textbooks on real analysis, where the modern trend for the last few decades has been to use an axiomatic analysis. Even when a construction is offered, it is usually applied towards proving the axioms of the real numbers, which then support the above proofs. However, several authors express the idea that starting with a construction is more logically appropriate, and the resulting proofs are more self-contained.[16]
In the Dedekind cut approach, each real number x is defined as the infinite set of all rational numbers that are less than x.[17] In particular, the real number 1 is the set of all rational numbers that are less than 1.[18] Every positive decimal expansion easily determines a Dedekind cut: the set of rational numbers which are less than some stage of the expansion. So the real number 0.999... is the set of rational numbers r such that r < 0, or r < 0.9, or r < 0.99, or r is less than some other number of the form
Every element of 0.999... is less than 1, so it is an element of the real number 1. Conversely, an element of 1 is a rational number
which implies
Since 0.999... and 1 contain the same rational numbers, they are the same set: 0.999... = 1.
The definition of real numbers as Dedekind cuts was first published by Richard Dedekind in 1872.[20] The above approach to assigning a real number to each decimal expansion is due to an expository paper titled "Is 0.999 ... = 1?" by Fred Richman in Mathematics Magazine, which is targeted at teachers of collegiate mathematics, especially at the junior/senior level, and their students.[21] Richman notes that taking Dedekind cuts in any dense subset of the rational numbers yields the same results; in particular, he uses decimal fractions, for which the proof is more immediate. He also notes that typically the definitions allow { x : x < 1 } to be a cut but not { x : x ≤ 1 } (or vice versa) "Why do that? Precisely to rule out the existence of distinct numbers 0.9* and 1. [...] So we see that in the traditional definition of the real numbers, the equation 0.9* = 1 is built in at the beginning."[22] A further modification of the procedure leads to a different structure where the two are not equal. Although it is consistent, many of the common rules of decimal arithmetic no longer hold, for example the fraction 1/3 has no representation; see "Alternative number systems" below.
Another approach to constructing the real numbers uses the ordering of rationals less directly. First, the distance between x and y is defined as the absolute value |x − y|, where the absolute value |z| is defined as the maximum of z and −z, thus never negative. Then the reals are defined to be the sequences of rationals that have the Cauchy sequence property using this distance. That is, in the sequence (x0, x1, x2, ...), a mapping from natural numbers to rationals, for any positive rational δ there is an N such that |xm − xn| ≤ δ for all m, n > N. (The distance between terms becomes smaller than any positive rational.)[23]
If (xn) and (yn) are two Cauchy sequences, then they are defined to be equal as real numbers if the sequence (xn − yn) has the limit 0. Truncations of the decimal number b0.b1b2b3... generate a sequence of rationals which is Cauchy; this is taken to define the real value of the number.[24] Thus in this formalism the task is to show that the sequence of rational numbers
has the limit 0. Considering the nth term of the sequence, for n=0,1,2,..., it must therefore be shown that
This limit is plain;[25] one possible proof is that for ε = a/b > 0 one can take N = b in the definition of the limit of a sequence. So again 0.999... = 1.
The definition of real numbers as Cauchy sequences was first published separately by Eduard Heine and Georg Cantor, also in 1872.[20] The above approach to decimal expansions, including the proof that 0.999... = 1, closely follows Griffiths & Hilton's 1970 work A comprehensive textbook of classical mathematics: A contemporary interpretation. The book is written specifically to offer a second look at familiar concepts in a contemporary light.[26]
The result that 0.999... = 1 generalizes readily in two ways. First, every nonzero number with a finite decimal notation (equivalently, endless trailing 0s) has a counterpart with trailing 9s. For example, 0.24999... equals 0.25, exactly as in the special case considered. These numbers are exactly the decimal fractions, and they are dense.[27]
Second, a comparable theorem applies in each radix or base. For example, in base 2 (the binary numeral system) 0.111... equals 1, and in base 3 (the ternary numeral system) 0.222... equals 1. Textbooks of real analysis are likely to skip the example of 0.999... and present one or both of these generalizations from the start.[28]
Alternative representations of 1 also occur in non-integer bases. For example, in the golden ratio base, the two standard representations are 1.000... and 0.101010..., and there are infinitely many more representations that include adjacent 1s. Generally, for almost all q between 1 and 2, there are uncountably many base-q expansions of 1. On the other hand, there are still uncountably many q (including all natural numbers greater than 1) for which there is only one base-q expansion of 1, other than the trivial 1.000.... This result was first obtained by Paul Erdős, Miklos Horváth, and István Joó around 1990. In 1998 Vilmos Komornik and Paola Loreti determined the smallest such base, the Komornik–Loreti constant q = 1.787231650.... In this base, 1 = 0.11010011001011010010110011010011...; the digits are given by the Thue–Morse sequence, which does not repeat.[29]
A more far-reaching generalization addresses the most general positional numeral systems. They too have multiple representations, and in some sense the difficulties are even worse. For example:[30]
That all these different number systems suffer from multiple representations for some real numbers can be attributed to a fundamental difference between the real numbers as an ordered set and collections of infinite strings of symbols, ordered lexicographically. Indeed the following two properties account for the difficulty:
The first point follows from basic properties of the real numbers: L has a supremum and R has an infimum, which are easily seen to be equal; being a real number it either lies in R or in L, but not both since L and R are supposed to be disjoint. The second point generalizes the 0.999.../1.000... pair obtained for p1 = "0", p2 = "1". In fact one need not use the same alphabet for all positions (so that for instance mixed radix systems can be included) or consider the full collection of possible strings; the only important points are that at each position a finite set of symbols (which may even depend on the previous symbols) can be chosen from (this is needed to ensure maximal and minimal choices), and that making a valid choice for any position should result in a valid infinite string (so one should not allow "9" in each position while forbidding an infinite succession of "9"s). Under these assumptions, the above argument shows that an order preserving map from the collection of strings to an interval of the real numbers cannot be a bijection: either some numbers do not correspond to any string, or some of them correspond to more than one string.
Marko Petkovšek has proven that for any positional system that names all the real numbers, the set of reals with multiple representations is always dense. He calls the proof "an instructive exercise in elementary point-set topology"; it involves viewing sets of positional values as Stone spaces and noticing that their real representations are given by continuous functions.[31]
One application of 0.999... as a representation of 1 occurs in elementary number theory. In 1802, H. Goodwin published an observation on the appearance of 9s in the repeating-decimal representations of fractions whose denominators are certain prime numbers. Examples include:
E. Midy proved a general result about such fractions, now called Midy's theorem, in 1836. The publication was obscure, and it is unclear if his proof directly involved 0.999..., but at least one modern proof by W. G. Leavitt does. If one can prove that a decimal of the form 0.b1b2b3... is a positive integer, then it must be 0.999..., which is then the source of the 9s in the theorem.[32] Investigations in this direction can motivate such concepts as greatest common divisors, modular arithmetic, Fermat primes, order of group elements, and quadratic reciprocity.[33]
Returning to real analysis, the base-3 analogue 0.222... = 1 plays a key role in a characterization of one of the simplest fractals, the middle-thirds Cantor set:
The nth digit of the representation reflects the position of the point in the nth stage of the construction. For example, the point 2⁄3 is given the usual representation of 0.2 or 0.2000..., since it lies to the right of the first deletion and to the left of every deletion thereafter. The point 1⁄3 is represented not as 0.1 but as 0.0222..., since it lies to the left of the first deletion and to the right of every deletion thereafter.[34]
Repeating nines also turn up in yet another of Georg Cantor's works. They must be taken into account to construct a valid proof, applying his 1891 diagonal argument to decimal expansions, of the uncountability of the unit interval. Such a proof needs to be able to declare certain pairs of real numbers to be different based on their decimal expansions, so one needs to avoid pairs like 0.2 and 0.1999... A simple method represents all numbers with nonterminating expansions; the opposite method rules out repeating nines.[35] A variant that may be closer to Cantor's original argument actually uses base 2, and by turning base-3 expansions into base-2 expansions, one can prove the uncountability of the Cantor set as well.[36]
Students of mathematics often reject the equality of 0.999... and 1, for reasons ranging from their disparate appearance to deep misgivings over the limit concept and disagreements over the nature of infinitesimals. There are many common contributing factors to the confusion:
These ideas are mistaken in the context of the standard real numbers, although some may be valid in other number systems, either invented for their general mathematical utility or as instructive counterexamples to better understand 0.999...
Many of these explanations were found by David Tall, who has studied characteristics of teaching and cognition that lead to some of the misunderstandings he has encountered in his college students. Interviewing his students to determine why the vast majority initially rejected the equality, he found that "students continued to conceive of 0.999... as a sequence of numbers getting closer and closer to 1 and not a fixed value, because 'you haven’t specified how many places there are' or 'it is the nearest possible decimal below 1'".[40]
Of the elementary proofs, multiplying 0.333... = 1⁄3 by 3 is apparently a successful strategy for convincing reluctant students that 0.999... = 1. Still, when confronted with the conflict between their belief of the first equation and their disbelief of the second, some students either begin to disbelieve the first equation or simply become frustrated.[41] Nor are more sophisticated methods foolproof: students who are fully capable of applying rigorous definitions may still fall back on intuitive images when they are surprised by a result in advanced mathematics, including 0.999.... For example, one real analysis student was able to prove that 0.333... = 1⁄3 using a supremum definition, but then insisted that 0.999... < 1 based on her earlier understanding of long division.[42] Others still are able to prove that 1⁄3 = 0.333..., but, upon being confronted by the fractional proof, insist that "logic" supersedes the mathematical calculations.
Joseph Mazur tells the tale of an otherwise brilliant calculus student of his who "challenged almost everything I said in class but never questioned his calculator," and who had come to believe that nine digits are all one needs to do mathematics, including calculating the square root of 23. The student remained uncomfortable with a limiting argument that 9.99... = 10, calling it a "wildly imagined infinite growing process."[43]
As part of Ed Dubinsky's APOS theory of mathematical learning, Dubinsky and his collaborators (2005) propose that students who conceive of 0.999... as a finite, indeterminate string with an infinitely small distance from 1 have "not yet constructed a complete process conception of the infinite decimal". Other students who have a complete process conception of 0.999... may not yet be able to "encapsulate" that process into an "object conception", like the object conception they have of 1, and so they view the process 0.999... and the object 1 as incompatible. Dubinsky et al. also link this mental ability of encapsulation to viewing 1⁄3 as a number in its own right and to dealing with the set of natural numbers as a whole.[44]
With the rise of the Internet, debates about 0.999... have escaped the classroom and are commonplace on newsgroups and message boards, including many that nominally have little to do with mathematics. In the newsgroup sci.math, arguing over 0.999... is a "popular sport", and it is one of the questions answered in its FAQ.[45] The FAQ briefly covers 1⁄3, multiplication by 10, and limits, and it alludes to Cauchy sequences as well.
A 2003 edition of the general-interest newspaper column The Straight Dope discusses 0.999... via 1⁄3 and limits, saying of misconceptions,
The lower primate in us still resists, saying: .999~ doesn't really represent a number, then, but a process. To find a number we have to halt the process, at which point the .999~ = 1 thing falls apart. Nonsense.[46]
The Straight Dope cites a discussion on its own message board that grew out of an unidentified "other message board ... mostly about video games". In the same vein, the question of 0.999... proved such a popular topic in the first seven years of Blizzard Entertainment's Battle.net forums that the company issued a "press release" on April Fools' Day 2004 that it is 1:
We are very excited to close the book on this subject once and for all. We've witnessed the heartache and concern over whether .999~ does or does not equal 1, and we're proud that the following proof finally and conclusively addresses the issue for our customers.[47]
Two proofs are then offered, based on limits and multiplication by 10.
0.999... features also in mathematical folklore, specifically in the following joke:[48]
Q: How many mathematicians does it take to screw in a lightbulb?
A: 0.999999....
Although the real numbers form an extremely useful number system, the decision to interpret the notation "0.999..." as naming a real number is ultimately a convention, and Timothy Gowers argues in Mathematics: A Very Short Introduction that the resulting identity 0.999... = 1 is a convention as well:
However, it is by no means an arbitrary convention, because not adopting it forces one either to invent strange new objects or to abandon some of the familiar rules of arithmetic.[49]
One can define other number systems using different rules or new objects; in some such number systems, the above proofs would need to be reinterpreted and one might find that, in a given number system, 0.999... and 1 might not be identical. However, many number systems are extensions of —rather than independent alternatives to— the real number system, so 0.999... = 1 continues to hold. Even in such number systems, though, it is worthwhile to examine alternative number systems, not only for how 0.999... behaves (if, indeed, a number expressed as "0.999..." is both meaningful and unambiguous), but also for the behavior of related phenomena. If such phenomena differ from those in the real number system, then at least one of the assumptions built into the system must break down.
Some proofs that 0.999... = 1 rely on the Archimedean property of the real numbers: that there are no nonzero infinitesimals. Specifically, the difference 1 − 0.999... must be smaller than any positive rational number, so it must be an infinitesimal; but since the reals do not contain nonzero infinitesimals, the difference is therefore zero, and therefore the two values are the same.
However, there are mathematically coherent ordered algebraic structures, including various alternatives to the real numbers, which are non-Archimedean. For example, the dual numbers include a new infinitesimal element ε, analogous to the imaginary unit i in the complex numbers except that ε2 = 0. The resulting structure is useful in automatic differentiation. The dual numbers can be given a lexicographic order, in which case the multiples of ε become non-Archimedean elements.[50] Note however that, as an extension of the real numbers, the dual numbers still have 0.999... = 1. On a related note, while ε exists in dual numbers, so does ε/2, so ε is not "the smallest positive dual number," and, indeed, as in the reals, no such number exists.
Non-standard analysis provides a number system with a full array of infinitesimals (and their inverses).[51] A. H. Lightstone developed a decimal expansion for hyperreal numbers in (0, 1)∗.[52] Lightstone shows how to associate to each number a sequence of digits,
indexed by the hypernatural numbers. While he does not directly discuss 0.999..., he shows the real number 1/3 is represented by 0.333...;...333... which is a consequence of the transfer principle. As a consequence the number 0.999...;...999... = 1. With this type of decimal representation, not every expansion represents a number. In particular "0.333...;...000..." and "0.999...;...000..." do not correspond to any number.
The standard definition of the number 0.999... is the limit of the sequence 0.9, 0.99, 0.999, ... A different definition considers the equivalence class [(0.9, 0.99, 0.999, ...)] of this sequence in the ultrapower construction, which corresponds to a number that is infinitesimally smaller than 1. More generally, the hyperreal number uH=0.999...;...999000..., with last digit 9 at infinite hypernatural rank H, satisfies a strict inequality uH < 1. Accordingly, Karin Katz and Mikhail Katz have proposed an alternative interpretation of "0.999...":
All such interpretations of "0.999..." are infinitely close to 1. Ian Stewart characterizes this interpretation as an "entirely reasonable" way to rigorously justify the intuition that "there's a little bit missing" from 1 in 0.999....[54] Along with Katz & Katz, Robert Ely also questions the assumption that students' ideas about 0.999... < 1 are erroneous intuitions about the real numbers, interpreting them rather as nonstandard intuitions that could be valuable in the learning of calculus.[55][56] Jose Benardete in his book Infinity: An essay in metaphysics argues that some natural pre-mathematical intuitions cannot be expressed if one is limited to an overly restrictive number system:
Combinatorial game theory provides alternative reals as well, with infinite Blue-Red Hackenbush as one particularly relevant example. In 1974, Elwyn Berlekamp described a correspondence between Hackenbush strings and binary expansions of real numbers, motivated by the idea of data compression. For example, the value of the Hackenbush string LRRLRLRL... is 0.0101012... = 1/3. However, the value of LRLLL... (corresponding to 0.111...2) is infinitesimally less than 1. The difference between the two is the surreal number 1/ω, where ω is the first infinite ordinal; the relevant game is LRRRR... or 0.000...2.[58]
This is in fact true of the binary expansions of many rational numbers, where the values of the numbers are equal but the corresponding binary tree paths are different. For example, 0.10111...2 = 0.11000...2, which are both equal to 3⁄4, but the first representation corresponds to the binary tree path LRLRRR... while the second corresponds to the different path LRRLLL....
Another manner in which the proofs might be undermined is if 1 − 0.999... simply does not exist, because subtraction is not always possible. Mathematical structures with an addition operation but not a subtraction operation include commutative semigroups, commutative monoids and semirings. Richman considers two such systems, designed so that 0.999... < 1.
First, Richman defines a nonnegative decimal number to be a literal decimal expansion. He defines the lexicographical order and an addition operation, noting that 0.999... < 1 simply because 0 < 1 in the ones place, but for any nonterminating x, one has 0.999... + x = 1 + x. So one peculiarity of the decimal numbers is that addition cannot always be cancelled; another is that no decimal number corresponds to 1⁄3. After defining multiplication, the decimal numbers form a positive, totally ordered, commutative semiring.[59]
In the process of defining multiplication, Richman also defines another system he calls "cut D", which is the set of Dedekind cuts of decimal fractions. Ordinarily this definition leads to the real numbers, but for a decimal fraction d he allows both the cut (−∞, d ) and the "principal cut" (−∞, d ]. The result is that the real numbers are "living uneasily together with" the decimal fractions. Again 0.999... < 1. There are no positive infinitesimals in cut D, but there is "a sort of negative infinitesimal," 0−, which has no decimal expansion. He concludes that 0.999... = 1 + 0−, while the equation "0.999... + x = 1" has no solution.[60]
When asked about 0.999..., novices often believe there should be a "final 9," believing 1 − 0.999... to be a positive number which they write as "0.000...1". Whether or not that makes sense, the intuitive goal is clear: adding a 1 to the last 9 in 0.999... would carry all the 9s into 0s and leave a 1 in the ones place. Among other reasons, this idea fails because there is no "last 9" in 0.999....[61] However, there is a system that contains an infinite string of 9s including a last 9.
The p-adic numbers are an alternative number system of interest in number theory. Like the real numbers, the p-adic numbers can be built from the rational numbers via Cauchy sequences; the construction uses a different metric in which 0 is closer to p, and much closer to pn, than it is to 1. The p-adic numbers form a field for prime p and a ring for other p, including 10. So arithmetic can be performed in the p-adics, and there are no infinitesimals.
In the 10-adic numbers, the analogues of decimal expansions run to the left. The 10-adic expansion ...999 does have a last 9, and it does not have a first 9. One can add 1 to the ones place, and it leaves behind only 0s after carrying through: 1 + ...999 = ...000 = 0, and so ...999 = −1.[62] Another derivation uses a geometric series. The infinite series implied by "...999" does not converge in the real numbers, but it converges in the 10-adics, and so one can re-use the familiar formula:
(Compare with the series above.) A third derivation was invented by a seventh-grader who was doubtful over her teacher's limiting argument that 0.999... = 1 but was inspired to take the multiply-by-10 proof above in the opposite direction: if x = ...999 then 10x = ...990, so 10x = x − 9, hence x = −1 again.[62]
As a final extension, since 0.999... = 1 (in the reals) and ...999 = −1 (in the 10-adics), then by "blind faith and unabashed juggling of symbols"[64] one may add the two equations and arrive at ...999.999... = 0. This equation does not make sense either as a 10-adic expansion or an ordinary decimal expansion, but it turns out to be meaningful and true if one develops a theory of "double-decimals" with eventually repeating left ends to represent a familiar system: the real numbers.[65]