Sure, when you start decoupling the numbers from their actual values. The only thing this proves is that the fraction-to-decimal conversion is inaccurate. Your floating points (and for that matter, our mathematical model) don’t have enough precision to appropriately model what the value of 7/9 actually is. The variation is negligible though, and that’s the core of this, is the variation off what it actually is is so small as to be insignificant and, really undefinable to us - but that doesn’t actually matter in practice, so we just ignore it or convert it. But at the end of the day 0.999… does not equal 1. A number which is not 1 is not equal to 1. That would be absurd. We’re just bad at converting fractions in our current mathematical understanding.
Edit: wow, this has proven HIGHLY unpopular, probably because it’s apparently incorrect. See below for about a dozen people educating me on math I’ve never heard of. The “intuitive” explanation on the Wikipedia page for this makes zero sense to me largely because I don’t understand how and why a repeating decimal can be considered a real number. But I’ll leave that to the math nerds and shut my mouth on the subject.
Similarly, 1/3 = 0.3333…
So 3 times 1/3 = 0.9999… but also 3/3 = 1
Another nice one:
Let x = 0.9999… (multiply both sides by 10)
10x = 9.99999… (substitute 0.9999… = x)
10x = 9 + x (subtract x from both sides)
9x = 9 (divide both sides by 9)
x = 1
Not a proof, just wrong. In the “(substitute 0.9999… = x)” step, it was only done to one side, not both (the left side would’ve become 9.99999), therefore wrong.
Okay, but it equals one.
No, it equals 0.999…
2/9 = 0.222… 7/9 = 0.777…
0.222… + 0.777… = 0.999… 2/9 + 7/9 = 1
0.999… = 1
No, it equals 1.
Sure, when you start decoupling the numbers from their actual values. The only thing this proves is that the fraction-to-decimal conversion is inaccurate. Your floating points (and for that matter, our mathematical model) don’t have enough precision to appropriately model what the value of 7/9 actually is. The variation is negligible though, and that’s the core of this, is the variation off what it actually is is so small as to be insignificant and, really undefinable to us - but that doesn’t actually matter in practice, so we just ignore it or convert it. But at the end of the day 0.999… does not equal 1. A number which is not 1 is not equal to 1. That would be absurd. We’re just bad at converting fractions in our current mathematical understanding.
Edit: wow, this has proven HIGHLY unpopular, probably because it’s apparently incorrect. See below for about a dozen people educating me on math I’ve never heard of. The “intuitive” explanation on the Wikipedia page for this makes zero sense to me largely because I don’t understand how and why a repeating decimal can be considered a real number. But I’ll leave that to the math nerds and shut my mouth on the subject.
Similarly, 1/3 = 0.3333…
So 3 times 1/3 = 0.9999… but also 3/3 = 1
Another nice one:
Let x = 0.9999… (multiply both sides by 10)
10x = 9.99999… (substitute 0.9999… = x)
10x = 9 + x (subtract x from both sides)
9x = 9 (divide both sides by 9)
x = 1
My favorite thing about this argument is that not only are you right, but you can prove it with math.
Not a proof, just wrong. In the “(substitute 0.9999… = x)” step, it was only done to one side, not both (the left side would’ve become 9.99999), therefore wrong.
Except it doesn’t. The math is wrong. Do the exact same formula, but use .5555… instead of .9999…
Guess it turns out .5555… is also 1.
Lol you can’t do math apparently, take a logic course sometime
Let x=0.555…
10x=5.555…
10x=5+x
9x=5
x=5/9=0.555…