Volume 0, Issue 0
It took humanity at least 3000 years to reach its current level of mathematical
An essential part of that knowledge is that counting normally starts with
nothing. Zero cometh before one.
Is zero even?
Yes it is. Absolutely no mathematically literate person is questioning that fact.
There are several ways to look at it.
Either you define even numbers as
multiples of two and zero is 0 times 2; it's a muliple of 2, just like it is a
multiple of 3, 4, 5, etc. Or you may observe that every other number is even
and every other number is odd.
This extends without any problem to negative numbers as
well and zero must be considered even in order not to break the pattern:
-5, -3, -1, 1, 3, 5, 7, 9 are odd, whereas -6, -4, -2, 0, 2, 4, 6, 8 are even.
Now, a surprising number of people will somehow find this fact disturbing and hold
on to the backward notion that 0 is such a "special" number that it cannot be
treated like any other.
This includes, unfortunately, a few people who teach
or have been teaching mathematics.
When does a new century begin?
This is also a settled issue.
However, each time this happens in history, you will find the majority of people
taking the wrong side of the issue and "celebrate" the new century a year too early.
The 21st century and the 3rd millenium
(not the second millenium, please!) started at the beginning of January 1, 2001.
It took essentially a whole year to convince the majority of people of that fact.
Zero to the power of zero: 00=1.
Well, zero to the zeroth power
is actually equal to one. Like it or not.
This one is constantly used by professional mathematicians, even though
most will tell you that zero to the power of zero is undefined.
Zero has zero significant digits.
Leading zeros are ignored and cannot count as significant digits.
The strings 007, 07 and 7 denote the same integer (seven).
So do the strings 000, 00, 0 and the empty string.
You may not be able to write down the empty string "" without quotes
and that's why "0" is used instead to denote the integer zero.
However, when it comes to determining the length of the simplest decimal
representation of an integer, the empty string comes back into play:
The decimal length of zero is zero!
Zero is divisible into anthing but divides only itself.
Zero is divisible into anything
but it divides only itself!
The proper notion of divisibility makes interesting statements
meaningful and correct:
Two integers are distinct if and only if there is an integer which only one of them divides.
Zero is the only element which zero divides.
The Mathematical Universe:
Mathematics is first and foremost a language.
It's a human language. A set of conventions made over the centuries to allow things to be
recorded and communicated by human beings.
Like all languages it has a syntax (which
specifies how its symbols may be put together) and semantics. Semantics is the term used
by Computer Scientists to describe the meaning of certain sentences whose syntax
happens to be correct (not all syntactically correct sentences are meaningful).
Such a meaning is usually described in term of some "real" world.
What makes mathematics unique among all possible languages is that
the universe in terms of which
a meaning is assigned to mathematical statements is not "real" at all.
The mathematical discourse is about mathematics itself.
A Structure Built on Nothing
Anything Is True of Nothing
This one may be difficult to swallow at first, but logic itself just would not
be consistent without the convention that any statement which does not
apply to anything is "vacuously" true.
The negation of such a statement is also vacuously true.
In fact, if some statement and its negation are proven to be both true
of all things endowed with a given property, it is hereby shown that absolutely
nothing may have that property. Clear enough, isn't it?