Technically, 0^0 is an indeterminate form and has no specific solution. Accurate but unhelpful.
Practically, 0^0 highlights the issue that most of us don't have a good conceptual model for what exponents really do. How would you explain to a 10-year old why 3^0 = 1 beyond "it's necessary to make the algebra of powers work out".
to wrap my head around what exponents are really doing: some amount of growth (base) for some amount of time (power). This gives you a "multiplier effect". So, 3^0 means "3x growth for 0 seconds" which, being 0 seconds, changes nothing -- the multiplier is 1. "0x growth for 0 seconds" is also 1, since it was never applied. "0x growth for .00001 seconds" is 0, since a miniscule amount of obliteration still obliterates you.
Also good, but the last line doesn't work (in my opinion): to get from 0x1 to 1 you need to undo the x0, i.e. dividing by 0. Which is undefined ... Zero, confusing 10 year olds for centuries!
True. If you approach it from a standpoint of extrapolation from known values, you have a division by zero one way or the other.
What I'd intended was that the value of 1 was reached by applying the same algorithm that was applied to arrive at the other values: start with 1, multiply by the base once per instance of the exponent. No division involved.
It's still incorrect if we want to be strict, of course. That algorithm is not quite the definition of exponentiation, because that algorithm can't really be extended to work outside rational exponents. Exponentiation is defined across complex numbers (ignoring 0^0 for the moment). I think this is acceptable because I'm only shooting for an explanation, which doesn't need to be strict.
What is "indeterminate form"? What does it mean for expression to "have a specific solution"?
You see, 0^0 = 1, and it's obvious to a mathematician. The only problem is that the function f: [0, \infty) x R -> R, f(x, y) = x^y is discontinuous in (0, 0) and that's what causes problems -- for instance, this is the source of the whole "indeterminate form" notion. If a function f is continuous in (a, b), then for every two sequences a_n, b_n, such that lim a_n = a, lim b_n = b, we have lim f(a_n, b_n) = f(a, b). That's why lim (a_n)^(b_n) = a^b if (a, b) != (0, 0), and this is "determinate form". But if (a, b) = (0, 0), then no matter how we define 0^0, it does not follow that lim (a^n)^(b^n) = a^b = 0^0, because in this case, lim (a_n)^(b_n) can be every positive value, and so mathematicians used to call it "indeterminate form" (it's not common today, though). So, since this problem is unsolvable in a consistent (continuous) way, we define 0^0 = 1, to be consistent with exponentiation rules, at least.
I've never seen a need for an "intuitive" explanation of exponentiation -- the usual definition is as intuitive as one can get. The thing is, most people do not know, _why_ expressions like pi^e are supposed to make sense -- they just take exponentiation as given. Only then they need to make up some explanation why "exponentiation rules" are like this, and what exponentiation is about. Hell, people don't even know what real numbers are! How are they supposed to make sense of exponentiation with exponent other than natural number?
Mathematicians don't argue about what an expression "really is" (or at least, real mathematics doesn't involve this). They define functions and use axioms to prove theories about them.
"No really". Mathematics just isn't concerned with this stuff. Sometimes infinity it defined as single point making the real number compact, sometimes a "positive infinity" and a "negative infinity" are defined. Sometimes you add points to a given function to make it more tractable and sometimes you don't. But none of this "means" anything. The real number line can be embedded in a number of topological spaces. At least two division rings and various things (the complex numbers are most common). The way you extend a given function (say e^x) is going to vary depending on what space you're looking at as well as what topic you're interested in.
Math works with definition systems and get theorems out of them. If you want to know what something "really is", consult philosophy or something.
Math is a tool (and sometimes abused for pure pleasure, 200 years later applied to make hard crypto work). If your definition doesn't make sense for the application, fix your definition and get over it.
Another example I've recently often bitched about in discussions is modern measure theory and its application to probability calculations. People just don't get the concept of theorytically possible event, but probability 0, i.e. ignore this. But without Lebesgue integration L_p function spaces are not complete and an awful lot of stuff stops to work properly. Among them essentially all of modern physics.
The sane approach is to get over the "this doesn't make intuitive sense" bitchering and just use defintions to derive useful results. And after a few years of playing around with stuff and applying the un-intuitive definition, it's becoming intuitive ;-)
Interesting that you say that. I've skimmed but have been meaning to properly read Nelson's: Radically Elementary probability theory http://www.math.princeton.edu/~nelson/books/rept.pdf and http://www.stat.umn.edu/geyer/nsa/. They do away with that problem all together as well as infinite constructions (replaced by hyperfinite) by replacing measure theory with non standard analysis. The gain is at least increased intuitiveness.
It turns out to be so. If a reply is not a rebuttal, it is usually preceded with something like "To clarify, ..." or "I wanted to add, that ...". I just got confused without it.
I felt the need to state what I thought was the case. Maybe originally it was a rebuttal but you don't disagree, feel free to take it as a clarification.
You see, 0^0 = 1, and it's obvious to a mathematician . . . we define 0^0 = 1, to be consistent with exponentiation rules
Well, you're going to be inconsistent with them no matter how you define it, since, as you point out, x^y should be zero if you approach (0,0) along the x=0 axis, and it should be one if you approach along the y=0 axis.
0^0 is simply an expression that doesn't make sense. There isn't an answer, and there certainly isn't something we could agree to define it as. It is gibberish, nothing more, nothing less. One cannot assume just because there are mathematical symbols on paper that they make sense.
By "exponentiation rules" I mean algebraic equalities, like a^x * a^y = a^(x+y). Most of them work no matter if you define 0^0 = 1 or 0, but some of them are cleaner with 0^0 = 1. It's also consistent with cardinal and ordinal exponentiation (look it up). "Approaching along x axis" is not algebraic notion, it's analytic one.
0^0 makes no less sense than, say, -e^(i pi). They're both 1 because we define them like this. If you think that -e^(i pi) makes more sense than 0^0, please, explain me why.
Also, mathematicians agree in this, seriously. Go and ask one.
More precisely, as Arturo Magidin points out at http://math.stackexchange.com/questions/11150/zero-to-zero-p..., if we view exponentiation in the 'discrete setting', then $0^0$ must be $1$; whereas, if we view it in the continuous setting, there is simply no good answer—unlike $e^{i\pi}$, which also lives in the continuous setting, but has a perfectly good, unambiguous answer. (lotharbot gives a nice explanation below of the ways that this is consistent with existing mathematics; but it can also be derived from the definition of the exponential function, with no further arbitrary conventions needed.)
I agree with you and lotharbot that if you view things from a continuous perspective, 0^0 is indeterminate.
Thinking back on my math education, part of the difference in viewpoint may be the first time I rigorously met the continuous-domain exponential.
This was in real analysis. Exponentiation is defined first for positive integer exponents, and then for rational exponents. All elementary. Then it's extended to real-valued exponents by taking the limits of rational numbers, and appealing to continuity.
I just looked, this is exercise 6 in chapter 1 of baby Rudin.
So, because the notion of limit and continuity is embedded in this definition of the exponential function, it's natural to "approach" (groan) 0^0 as a special case, because the conditions of this definition (continuity) don't hold.
But my point is, if 0^0 = 1 is _the_ answer in discrete setting, and it's _an_ answer in continuous setting, why don't we just agree that 0^0 = 1 and stop creating confusing situation where sometimes it's defined and sometimes it's not. 0^0 = 1 does not make calculus theorems more complicated to state or prove with modern language. It was a problem in XIX century, when mathematicians did not have a solid foundation with concepts like limit or continuity, but it's over now.
Mathematician here; we do not.
It seems I was a little too bold with my claim. All the mathematicians I know (and I'm a mathematician as well) agree with 0^0 = 1. It's a folklore specific thing, I guess.
As am I, by training if not by profession. As is lotharbot. You're in a thread full of mathematicians. :)
Which is what I would expect on this site, actually. I'm always timid making technical claims here unless I'm sure I'm correct; it seems to be a place frequented by arbitrarily large fish.
To answer your question, if 0^0 = 1 is _the_ answer in discrete setting, and it's _an_ answer in continuous setting, why don't we just agree that 0^0 = 1 and stop creating confusing situation where sometimes it's defined and sometimes it's not.
I'm not persuaded it is always the answer. I think the fact that it is an indeterminate form in limits is a forceful enough demonstration of that. It all depends on context. If I came across a 0^0 in, say, an engineering context, my first instinct would be to check whether the formula was defined in that case, not to just assume that 1 would work.
I mean, it's like 1/0. If you're working in R, that's simply illegal. If you're working in R*, it's the infinite point. If you're taking a limit, it means "unbounded". If you're working in my favorite field, the hyperreals, it could be any number of flavors of infinity depending on the flavor of zero it was.
It would be foolhardy to try to define the symbol; without a context to supply some sort of sense, it is nonsense. And that is how I feel about 0^0 as well.
Please, treat exponentiation just like every other function out there. I don't get the whole limit argument at all. Given any function f: R x R -> R, if it happens that a_n -> a, b_n -> b, but lim f(a_n, b_n) != f(a, b), people just say that f is not continuous in (a, b), and the case is over. However, if f happens to be exponentiation function, people instead argue that f should not be defined in (a, b), forgetting about the fact that the theorem which lets you take a limit of an argument instead of a limit of a function values works only under assumption that f is continuous in a proper point. Instead of noting that there's no contradiction because the assumptions are not satisfied, people just run away from it, declaring 0^0 as undefined.
From this point of view, the whole notion of "indeterminate form" makes just as little sense as distinguishing some arbitrary class of functions and calling them "elementary". Why are some points of discontinuity of some functions more special than other points of discontinuity of other functions? Why sin is more elementary than gamma? Historical heritage of confusion, I guess.
Consider this related case: if you evaluate a limit and you get 0/0, you recognize that you need to do more work to find the actual limit. It could be 1, -1, 0, infinite, etc. depending on how you reached it (say, sin(x)/x versus sin(x)/x^2). The issue is not the continuity of x/x; the issue is whether setting a convention for 0/0 would give you the right value for a limit. Since it doesn't always, we call it "indeterminate".
Similarly, if you're evaluating a limit and you get 0^0 you need to do more work. You can't just stop and say "oh, that's 1". It depends on what function you used to get there -- x^x will give you a different answer from ( e^(-1/x) )^x. Again, it has nothing to do with the continuity of exponentiation. The issue is whether the convention of 0^0=1 is correct in the specific part of mathematics you're working in.
The same argument can be made if you're working in the hyperreals, or if you're working with field axioms -- the convention 0^0 doesn't work in that context.
Please, by all means, use the convention 0^0=1 when it's appropriate. But understand that it's not always appropriate. Not every mathematician works in the particular subschool that you do; not every mathematician is going to find your convention appropriate.
Because sometimes it's better not to. Sometimes it's inconsistent with our definitions.
Just like sometimes we agree that you can't divide by zero, and sometimes we agree that you can. Sometimes infinity is an actual value (say, in the extended reals), and sometimes it's just a symbol for "unbounded". Sometimes we agree that you can't take the square root of a negative number, and sometimes you can. Sometimes we use the axiom of choice, and sometimes we don't (and you can have an awful lot of fun either way!)
Mathematics is contextual. How various operations behave depends on which axioms and conventions are being used.
> Because sometimes it's better not to. Sometimes it's inconsistent with our definitions.
I'd love to see even one example of 0^0=1 being inconsistent with a definition. The closest I've ever seen is that it bothers people that for reasons of their own had their hearts set on (x,y) -> x^y having no discontinuities...
Perhaps it's more precise to say "Because sometimes it's better not to. Sometimes there is no canonical choice that follows from our definitions, and it doesn't help to assign an arbitrary value that doesn't help solve any related problems."
What is "x" equal to? In general, I mean, not in the context of any equation like "x+1=2". You could say "x=7 in the study of free variables over integers when no other constraints are given", and that is completely consistent with the rest of mathematics, and yet would not be particularly useful and introduces an ugly (philosophical weasel word, yes) asymmetry in the theory (I'd say it introduces a gauge invariance (https://secure.wikimedia.org/wikipedia/en/wiki/Gauge_theory), but I'm really not qualified to discuss that in a rigorous way.)
Someone like Scott Aaronson could put this claim on more solid footing, but I would state that, intuitively, "assigning a value to an indeterminate form leads to a more complex definition of a mathematical system" in some formal complexity-theory sense.
I haven't seen any reason why -e^(i pi)=1 could be considered incorrect. It's consistent with the Taylor Series expansion of e^x. It's consistent with the view of complex exponentiation as rotation. I don't know of any particular problems that arise from taking -e^(i pi)=1.
This is not the case with 0^0=1, which is inconsistent with many limits. That's why 0^0=1 is an agreed-upon convention sometimes. http://en.wikipedia.org/wiki/Exponentiation#Zero_to_the_zero... has a fairly nice summary of the issues involved in defining it.
This is not the case with 0^0=1, which is inconsistent with many limits.
So what? It's only a problem if you want the exponentation function to be continuous, so you escape the problem by leaving it undefined. You could place similar unbased requirements on complex exponentiation to make it seem incorrect. For instance, real exponentiation always gives a positive value for positive base, while complex does not, so e^(i pi) = -1 is wrong. I agree that this is ridiculous requirement, but leaving 0^0 undefined because the math is not as we want it to be (e.g. exponentation is not continuous) looks just as ridiculous and silly to me.
On the other hand, putting 0^0 = 1 makes it consistent many combinatoric formulas, and is also consistent with cardinal exponentation, where nobody objects to 0^0 = 1, when you look at 0 as the cardinal number.
It has nothing to do with wanting exponentiation to be continuous. It's simply a recognition that limits of the form 0^0 are indeterminate, which means a convention that 0^0=1 is not appropriate in the context of evaluating limits. This isn't an argument that it "seems" incorrect, like your bizarre argument about real vs complex exponentiation; it's an argument that it IS incorrect in that context. If you're evaluating limits, you have to treat 0^0 as indeterminate, not as 1.
Even the original article noted that we don't choose the 0^0 convention because it's "correct", but because it's "nice" -- which is why we define it that way in the contexts where it makes sense to define it that way. If you're working with combinatorics, 0^0=1. If you're working with cardinal exponentiation, 0^0=1. If you're taking limits, or working in the hyperreals, or in certain other contexts, the convention doesn't apply. In some circumstances, 0^0 isn't even a valid statement -- like if you're working directly with the field axioms of R.
Recognize what context you're working in, and what assumptions or conventions apply in that context. That's just good mathematics.
It's simply a recognition that limits of the form 0^0 are indeterminate, which means a convention that 0^0=1 is not appropriate in the context of evaluating limits.
But the whole point of distinguishing some "forms" as "indeterminate" is to work around the discontinuity of elementary functions! My absolutely first sentence in this thread is asking, what exactly the "indeterminate form" is. I'm asking this question, because this not a formal notion and you will not find any formal definition of it. Its existence is rooted in the fact that for no reason other than the tradition (and convenience) we use special notation for some functions. Instead of +: R x R -> R, +(2, 3) = 5, we write 2 + 3 = 5. The same goes for ^: [0, \infty] x R -> R. The only reason we have all those fancy limit evaluating laws is because these function are continuous most of the time. For instance, + is continuous everywhere, so lim +(a_n, b_n) = +(lim a_n, lim b_n), if both lim a_n and lim b_n make sense. Similarly, /: R x R - {0} -> R is also continuous everywhere, so lim /(a_n, b_n) = /(lim a_n, lim b_n), if the right hand expression makes sense. If it does not make sense, for instance when both lim a_n and lim b_n are equal to zero, we need cannot approach this problem in such a simple way. Now, some people would call /(0, 0) an "indeterminate form", which makes for me no more sense than calling f(0, 0) an indeterminate form, where f(x, y) = log_(1/x) (y) -- while f is continuous everywhere where defined, you cannot extend its domain to contain (0, 0) for it to stay continuous, just like you cannot do it with / function.
As I repeated many times, the whole affair is because ^ seem to be more familiar than beta function (we have a special notation for it, for instance), people want it to behave nicely, so that for instance it conforms to some arbitrary limit evaluating laws, missing the whole underlying concept of continuity.
The whole point of distinguishing some forms as "indeterminate" is to work around the fact that you're trying to conduct operations on the real numbers that are not defined under the field axioms of the real numbers (or the axioms of the extended reals [-inf,inf]). That's where its essence is rooted; that's why this whole affair exists -- the fact that 0/0, 0^0, 0xinf, inf-inf, etc. are not well defined by our axioms.
This actually relates to all three examples I've presented where the 0^0=1 convention fails. It should be treated as an indeterminate form in limits because it's not well-defined by the axioms of the real numbers; it's also not well-defined by the axioms of the hyperreals, but division of infinitesimals is well-defined in the hyperreals, which gives us an alternate method of computing limits that avoids the "indeterminate form" entirely.
Let me reiterate: 0^0 is not defined under the field axioms of the real numbers. The choice to define it as 1 is a convention which makes certain math easier, in certain areas of mathematics. It is by no means a universal convention; it is by no means the one and only correct definition of 0^0. You continue to argue for the convention, but miss the larger point that it is a convention which is chosen for convenience, and which is not always appropriate.
Exponential function is not defined by the axioms of real numbers, as opposed to addition and multiplication, so this point is irrelevant -- you can define it in any way you want. There's no inconvenience in defining 0^0 = 1, apart from misunderstanding the concept of limits by some people. Defining 0^0 = 1 is universal convention -- people either do it like this, or do not define 0^0 at all, which I'm fighting against.
The whole point of distinguishing some forms as "indeterminate" is to work around the fact that you're trying to conduct operations on the real numbers that are not defined
I am not. Are you? Let me reiterate: the whole concept of "indeterminate forms" (which, I repeat, is not formal at all) stems from misunderstanding the process of taking limits.
> "people either do it like this, or do not define 0^0 at all, which I'm fighting against."
I think it's silly to fight against it. There are circumstances in which leaving it undefined is good, and in which trying to define it as 1 would lead to either misunderstandings (in the case of beginners doing limits, a case you are too quick to dismiss) or actually incorrect (an equivalent problem in the hyperreals could violate the transfer principle).
It's a broad convention, but it is not universal, and it shouldn't be.
> x^y should be zero if you approach (0,0) along the x=0 axis, and it should be one if you approach along the y=0 axis.
Technically, 0^y is 0 only if you approach it from the right: y>0. To the left of 0 it is indeterminate or infinity depending on how you look at it.
x^0, however, makes sense for all x and is always 1
Yes. Tying to define f(0) as 0 or 1 won't make it continuous, as approaching from the lines x = 0 and y = 0 will make the limits differ (the definition of lack of a limit).
>most of us don't have a good conceptual model for what exponents really do
Instead of matching math to real world objects (1= one banana, 2 = two bananas, 1+2 = 3 bananas etc. ) and building up to exponentiation, multiplication etc. thereby introducing all sorts of paradoxes, group theory dodges all that and treats the whole thing as a very consistent rule-based system. Things fall into place quickly once the rules are laid out explicitly.
Consider:
finite abelian group with only 3 elements a,b,c.
Given a+b=c, a+c=a, what's b+b ?
Hmmm...okay, if a plus c is a, then c is acting like zero. So b+c must be b.
since addition is commutative (abelian gp), b+a must be a+b which you said was c.
So now we know b+a=c, b+c=b, so b+b better be a !
Students are easily convinced because you've laid out the rules very explicitly. In fact, they'll try to convince you that b plus b better be a because that's the only way to make the cayley table work out!
(http://en.wikipedia.org/wiki/Cayley_table)
There are several books that argue that the teaching of Abstract Algebra must precede Calculus for this very reason. With Calculus, the mapping of math to real-world objects leads to all sorts of messy realities. With group theory, you dodge that mess by simply stating rules upfront.
Group theory is formed by abstracting out the observed properties of number systems. If you want to show that some collection is a group, you will have to do the computations to show that they follow the rules, in which case, it helps to have an understanding of the mechanics of the computation.
> it helps to have an understanding of the mechanics
My claim is the exact opposite. I claim you don't need to understand the mechanics ( just blindly abide by the rules of the group or abelian group or finite simple group or whatever), which is why the approach is better. If you show a monkey red means stop and green means go and reinforce these rules by rewarding with a banana, eventually the monkey will stop when he sees the red. Not because he understands the mechanics of traffic management. Simply because he is abiding by the rules. Similarly, large portions of math can be approached by either the definitional route ( ie. rules ie. define propositions & theorems that logically follow if those props held ) or via trying to understand actual mechanics by mapping everything to real world phenomena ( x = distance, dx/dt = velocity, d/dt(dx/dt) = acceleration etc. ) which are problematic because the mapping breaks down due to the nature of physical reality ( like friction etc. )
How would one explain say Hilbert's 7th problem via the actual mechanics ?
If a is algebraic and b is irrational show a^b is transcendental.
You are probably right that such an approach is better at teaching students to be able to crank out solutions to problems, but I think it also reinforces the attitude a lot of students have that math is just a bunch of arbitrary rules that don't mean anything, and is therefore a waste of time. For students that aren't destined to be math majors, the most important aspect of math is being able to map real-world concepts to abstract rules. Being able to mechanically manipulate those rules is far secondary, especially with easy access to computers.
Only if n is an integer. Real exponentiation is something completely different, because it involves all aspects of real numbers - addition, multiplication, order and continuity, which are all interconnected, and the language of group theory is too weak to describe it. For instance, while 2^pi makes perfect sense in the realm of real numbers, it makes none in Z_3.
(Warning, ascii math is confusing and ambiguous to read. Sorr.)
Exponentiation of group "multiplication" does not immediately seem amenable to the reals, sure. But real exponentation does form a group, as shown here:
Define x_g(r) = the function that raises a Real/{0} (non-zero real) number r to the exponent x (in the sense of of some reasonable definition of exponentiaton of continuous functions). Define X = the set x_g() functions corresponding to all reals (including 0)
Define x_g y_g as composition: y_g(x_g(r)) = (r^x)^y = r^ (xy).
Then we have 0_g x_g = (r^0)^y = r ^ (0 y) = 1 = r ^ (y * 0) = (r^y)^0) = y_g 0_g -> identity
The group you described does not inherit any interesting structure from exponentiation -- indeed, one can easily see that it is isomorphic to the multiplicative group of reals. You could similarly construct a group isomorphic to an additive group of reals. This is an example of the fact that real exponentiation connects different aspects of real numbers, as well as the fact that just abstract algebra language is not enough to express properties of real numbers. You need to somehow relate the algebraic structure of reals to a topologic one, which stems from order imposed on reals and its continuity.
You write: "With Calculus, the mapping of math to real-world objects leads to all sorts of messy realities. With group theory, you dodge that mess by simply stating rules upfront."
(Prologue: I encountered group theory first by drawing pictures of pegboards and strings to illustrate permutations (before I knew the word!), not by reading the rules up front.)
You are describing a schism between pure/formal and applied mathematics (and between formalism and intuition, to some extent). It's completely cool for you to have an interest in pure math completely separated from applied math, real-world physics, programming etc. It is also cool for someone to pursue applied math,physics, etc without any pure math, but that is sad because a lot of beautiful symmetry and cross-disciplinary value would be last. (Goodbye, encryption!)
I personally strive to connect pure and applied mathematics. After getting burned (in an emotional/psychological way) by chasing pure math study beyond my ability to intuit and apply it, I now commit myself to learning theory and application in tandem. (I'll certainly appreciate the fact that pure theorists such as your ideal have gone several steps ahead and I can study their results without trying to discover them from scratch.)
In fact, my most recent flight of fancy / big dream is to write math/CS tutorials that provide such a tight integration of theory and application, abstract and concrete, general and specific. And I want to use modern web tools (hyperlink, animation, multi-dimensional page layout) to do so.
I'd love to talk to anyone interested in working with me on that :-)
Hi, I'm really interested in developing tutorials / explanations that help merge intuition and rigor, theory and application. You can reach me at kalid.azad@gmail.com.
> Technically, 0^0 is an indeterminate form and has no specific solution.
Precisely, as this is the true mathematician answer: "it depends where 0^0 comes from".
As a f(x,y): RxR->R function, come from the top of the R² plane and 0^0 is 0 but come from the right side and it's 1. Limits and extension by continuity give us this easily enough for fh:x->x^0 and fv:y->0^y.
Writing this I asked myself, what if we came from some funky other path, like the diagonal, or a curve?
h: R->RxR, x->(x, 0) defines "coming from the top", and foh = fh
v: R->RxR, x->(0, y) defines "coming from the top", and fov = fv
d: R->RxR, x->(x, x) defines coming along the diagonal, where things could get interesting.
s: R->RxR, t->(e^(at)sin(t), e^(at)cos(t)) defines coming along a log spiral whose tangent at t=0 is vertical, so fos looks like fun around t=0.
Now what happens if we build a path function p: RxR->RxR, (t, z)->? that endlessly approaches v when z->0? the log spiral with z=1/a as a parameter is a possible one. With such a p function, what does lim fop(x) when x->0 (which is a function of z) look like when subsequently z->0?
> How would you explain to a 10-year old why 3^0 = 1 beyond "it's necessary to make the algebra of powers work out".
Actually, that's exactly the reason 3^0=1: it was the definition that preserved the most identities. Agreed that this explanation doesn't really help intuition.
You add a bunch of stuff and you get a total. You add nothing and 0 is the total because it's the identity and starting point for addition. You multiply a bunch of stuff and you get a product. You multiply nothing and you get 1 because that's the identity for multiplication.
(That's the same as saying "the rules of algebra work out" but there's maybe something intuitive about multiplying nothing and getting back the thing that doesn't change the result of multiplication?)
Yeah -- that may have been the original motivation, but repeating it as an "explanation" reinforces the notion that math is a bunch of rules (vs. models you can construct and manipulate in your head).
0 probably started as a placeholder symbol for "naught", i.e. nothing to write, and the first scribes were taught "Just write a circle when you have nothing to report".
But, with greater understanding of numbers 0 evolved into its own entity and we saw numbers on a "line", a powerful mental model (why not 2d numbers? N-dimensional numbers? etc.)
Re-teaching that 3^0 = 1 "because the math is convenient" doesn't help us build a mental model of what exponents could be (I know you don't agree with this, just stating it again because the lack of intuitive explanations for math is a major pet peeve of mine).
the lack of intuitive explanations for math is a major pet peeve of mine
I'm a very visual thinker, and that is one reason I enjoy the new Art of Problem Solving textbook Prealgebra by Richard Rusczyk, David Patrick, and Ravi Boppana--
it is full of interesting visual "explanations" and substitute for proofs in a book intended for a young audience.
That said, I finally realized that I was limiting my mathematical development by insisting that every mathematical idea must appeal to my visual intuition. Some mathematical ideas are proven even if they don't appeal to visual intuition. In the words attributed to John von Neumann, "in mathematics you don't understand things. You just get used to them."
Thanks for the pointer! I found Needham's book awesome, I've barely made a dent in it but love the visualizations.
I don't think visualization is the only intuitive method -- you can have a general "sense", not sure how to put it more specifically -- I have a "sense" about growth of e without a specific diagram.
Agreed that not every concept can be understood... yet. There's a quote I love to rail on, in reference to Euler's formula:
"It is absolutely paradoxical; we cannot understand it, and we don't know what it means, but we have proved it, and therefore we know it must be the truth." (Benjamin Peirce, 19th-century Mathematician)
Really? Yes, it may be baffling at first, but we can _never_ understand it? Only if that's our attitude :).
It's an amusing way to put this, but yes, it's true. An example of such situation is the case of continuum hypothesis. Both it and its negation has been proven not to lead to contradiction, so while most mathematicians ignore it, effectively treating it as being true, many set and model theorists play with things like Martin's axiom, which makes sense only if continuum hypothesis is false.
The high school teacher in the link is a B.S. in math education. They're usually reflexive Platonists, believing that math is out there, and we merely discover it. This is a result of the teaching of undergrad math as essentially a series of completed works, with little history attached to it. This teacher probably hasn't thought critically about why, say, 1/x^2 = x^(-2).
By contrast, a mathematician has a Ph.D. in math, and has had to do original research and look at some history of topics. Thus, the mathematician knows that math is not a finished product, but is under constant refinement.
Now, let's talk about negative exponents. Students are taught that 1/x^2 = x^(-2). High school teachers often don't understand why. It is so ingrained that even asking "Why?" seems almost grammatically incorrect.
The reason why is that we know the following things about exponents:
1) x^n = x* x* ...* x for n a positive integer
2) As a consequence of 1, x^m* x^n = x^(m+n) for m,n positive integers
Now, the question is not "what is x^n if n is negative?" (which is what a high school teacher might ask). Rather, the question is "Can we define (!) x^n for negative n in a way consistent with the item two above?" (mathematician's framing). And, of course, we can. If x^n = 1/x^(-n) for n negative, then item two works.
So, a high school teacher most likely thinks that the negative exponent rule is simply a rule, handed down from the Gods of math. A mathematician recognizes that it is a convention, and such a smooth convention that there is simply no better choice.
Now, about 0^0: The HS teacher asks "What is 0^0?" and is therefore under the impression that 0^0 is undefined because according to certain reasonings it could be 0 or it could be 1. Textbooks (not written by mathematicians) wouldn't correct this. The TI-86 gives a domain error when 0^0 is input. The mathematician asks "What value of 0^0 makes my preferred formulae continue working?" and thereby defines 0^0=1.
> They're usually reflexive Platonists, believing that math is out there, and we merely discover it. This is a result of the teaching of undergrad math as essentially a series of completed works, with little history attached to it.
It's not entirely clear what you mean by this, but in the most obvious interpretation this idea is correct.
Here's what I mean by that: math deals with pure logic. All logical deductions (derivations from axioms to conclusions in formal systems) are in a sense "out there", waiting to be discovered. Certainly they have always been true, before people knew about them, and there are deductions that are true even though no one knows about them yet.
The other interpretation, which you probably meant, was the question of mathematical style and interest: which axioms do we study, and why do we care about them, and why are some theorems important and others not important? Certainly the definition that 1/x^2 = x^(-2) was made for consistency, which is an aspect of style. You could make some other definition and do formal reasoning just as well (although you'd have to change some other things - maybe the definition of exponentiation, maybe your interpretation of symbols on paper).
But either way, I think you are wrong that high school teachers don't think critically about why 1/x^2 = x^(-2). Some may be bad teachers, but I have known a lot of incredibly good high school math teachers, and I suspect they have thought about this. In fact, when I talk to mathematicians about high school education, they usually agree that people doing research know more parts of mathematics than high school teachers, but good high school teachers have a much deeper understanding of elementary math than people doing research, because they have had to approach it from many different angles in order to teach different students.
> Here's what I mean by that: math deals with pure logic. All logical deductions (derivations from axioms to conclusions in formal systems) are in a sense "out there", waiting to be discovered. Certainly they have always been true, before people knew about them, and there are deductions that are true even though no one knows about them yet.
From one viewpoint, math deals with pure logic. It's something of a poor viewpoint from my perspective: Newton couldn't back up his calculus with logic, Euler and Riemann made numerous unfounded assumptions when looking at the zeta function, Heaviside built a telegraph across the Atlantic despite lacking a proper logical foundation.
My personal viewpoint is that mathematics is a series of shortcuts for understanding and manipulating a wide variety of phenomena, and mathematical research is the development of further shortcuts. Often logic comes in, but it's usually after you get the result.
Nothing wrong with logic (I sure do like computers, for example!), it's just not the panacea for anything mathematical.
> But either way, I think you are wrong that high school teachers don't think critically about why 1/x^2 = x^(-2). Some may be bad teachers, but I have known a lot of incredibly good high school math teachers, and I suspect they have thought about this. In fact, when I talk to mathematicians about high school education, they usually agree that people doing research know more parts of mathematics than high school teachers, but good high school teachers have a much deeper understanding of elementary math than people doing research, because they have had to approach it from many different angles in order to teach different students.
You're absolutely right about this, and I was specifically thinking of two very bad high school teachers when I wrote my post. The fact is that I know very few HS math teachers, and so my opinion is quite clouded by these two.
Not entirely fair. While there is a lot of choice about how to define stuff around the edges of mathematics, after the axiomatic choices have been made, there's quite a lot to discover in their structure. Particularly if the relevant axioms relate to something outside of mathematics (the examples are too many to even scratch but start out thinking of the definitions of natural and rational numbers) then you really are discovering things in the same way as a physicist - in fact, this is how many physicists go about discovering things, for better or worse.
In most situations it makes sense to define exponentiation as repeated multiplication, and a^0 as the absence of multiplication by a, hence a^0 = 1 as the multiplicative identity. I wouldn't introduce the idea of anything else to a student unless they specifically asked me about one of the problems which can arise in choosing 0^0 = 1.
Math is out there. I don't understand how anyone can say that the Mandelbrot set was created or is formed from arbitrary axioms. It was discovered full stop.
Your parent never says anything about arbitrary axioms.
To your main point: are irrational numbers, say, "out there"? If so, where?
Until a few centuries ago it was mathematical standard practice to fudge 1/2 as 25/49 when taking it's square root. But then mathematicians invented (some would argue) the notion of an irrational, because it was, well... useful.
There are real metaphysical questions here; there have been since the greeks started reaching the limits of a purely geometric (read physical) understanding of math.
I don't think the word "invented" is much better an analogy for real numbers than "position" is for "where" the Mandelbrot Set is.
This goes beyond what is strictly math, but I don't think it's reasonable to say that properties of the real numbers (say, roots, pi, e, and so forth) are invented. In some sense they seem like the simplest thing that fits a few properties (and not that many). Similarly with Euclidean Geometry, and the natural numbers (with primes and their structure, etc.) I really actually think something akin to Occam's Razor applies in math!
You can of course tweak your starting points and get really interesting things too, say non-standard analysis or non-Euclidean geometries. And you can derive (I would say "discover") some rich and surprising properties and patterns in them.
But I don't think any of these things are accidental. The real numbers (and pi, e, etc) would surely be "invented" in the same (modulo shifts in convention) ways by other advanced civilizations, I think. Could you imagine this not being so?
This is also extra-mathematical, and I admit this may be hooey, but I believe (as a non-mathematician) that math gives hints at large patterns and tells you when things fit well or are funny. Like, I can argue whether Pi or 2*Pi is the more "natural" constant, and it's not a discussion completely devoid of content! Also, I think math tells me something is funny about 0^0 (because of the limit 0^x) but that 1 is the more natural fit.
I realize this puts me quite firmly into the category of people being belittled here!
> I realize this puts me quite firmly into the category of people being belittled here!
I hope I'm not participating in a discussion where people are being belittled! I think all the viewpoints here are fascinating.
Yours is certainly interesting and valid, I'm simply offering that the "full stop" at the end of your previous comment doesn't reflect the kind of deeper metaphysical questions that underly the whole progression of the history of math.
To your main point: are irrational numbers, say, "out there"? If so, where?
I believe this to be a bad counterpoint, because you could ask the same thing about natural numbers as well. I've personally never seen a natural number. Sure, I have seen and worked with lots of representations of natural numbers, but the numbers themselves are - as far as I understand it - not physical objects. There is no qualitative difference between natural numbers and any other mathematical objects in that respect. They are all on the same "plane of existence".
At least that's a valid world view or ontology. I know that not everybody thinks like that, but besides clarifying what the parent poster probably meant I think it's not a very meaningful discussion.
To your main point: are irrational numbers, say, "out there"? If so, where?
On the sheet of paper in front of me, as the hypotenuse of the isosceles right-angled triangle with unit length I drew a moment ago. Irrational numbers are probably a bad example of what you're talking about.
> Irrational numbers are probably a bad example of what you're talking about.
No, they're the perfect example and you're observation about the unit square is glib.
Trying to reason about the diagonal of the unit square basically destroyed the Pythagorean world view of integers as the fundamental building blocks of reality.
If mathematical entities are simply discovered and that-is-that, then why did it take nearly two millenia after this observation for western math to accept irrationals as numbers?
From M. Stifel, 1544:
"Now, that cannot be called a true number which is of such a nature that it lacks precision. Therefore, just as an infinite number is not a number, so an irrational number is not a true number, but lies hidden in a kind of cloud of infinity"
>as the hypotenuse of the isosceles right-angled triangle with unit length I drew a moment ago //
Doesn't the quantised nature of matter mean that 2^½ exists as a real measure of a material object only as much as a perfect circle actually is existent in our universe?
Although my mathematician friends would probably yell at me if they heard this, as far as I see things it's hardly any different from physics. Just a model we create to describe observed phenomenon.
Oh, math only works this way in intuitive fields, like elementary calculus, elementary probability and staticstis, graph theory or Euclidean geometry. However, there are fields in math that are different -- for instance, there are topological spaces that exhibit phenomenons unseen anywhere else and that are very hard to grasp intuitively (I had very hard time trying to imagine what Cech-Stone compactification construction based on ultrafilters look like, and eventually I gave up after I understood that this space is only supposed to satisfy an universal property, and not have any intuitive shape). Deep results in algebraic methods in topology say interesting things about the models, while being completely indescribable outside of them, apart from some trivial examples (for instance, what _exactly_ are cohomology classes?).
The bottom line is, if the only things you've seen are geometry and calculus, with intuitive concepts like speed of change, area, length, then yeah, it's only making our intuitions more formal. Otherwise, it's something completely different.
I'm exactly the intuitive thinker you describe, with only an undergrad degree in math. I have not a clue what a Cech-Stone compactification construction based on ultrafilters is.
But... I think that whatever it is, you decide on some fairly simple properties you want to satisfy, and then go off discovering what they lead to and what the consequences are. Sometimes (mostly all the time?) you get something trivial or that reduces to being isomorphic to something else, but sometimes you get out a lot more than you put in, in surprising ways. I call that a lot more like "discovery" than "invention" though both are strained as analogies.
Again, I would argue this is no different than physics. The strength and maturity of a model is not only based on its ability to describe what we can verify, but predict what we have never (and may never) seen, and sometimes what we have never even conceived of- and I think both physics and mathematics are doing that.
My point is, many concepts in math do not exist outside of the models describing them, while if we throw away physical models, the reality does not go away.
Bingo. Mathematics is a consistent model which can be applied to do useful things and make useful predictions, but like any logical system it springs from axioms that must simply be taken as given. Most of these axioms (Modus Ponens[1], for example) are so simple and obvious that it seems bizarre to construct a way of thinking without them, but that does not intrinsically make them "true".
The site isn't loading, but I find discussions about mathematical curiosities (though they're not that curious) come about because people are looking for a deeper meaning in mathematics. Math is playing with ontological objects in a system of definitions. There doesn't need to be an answer that "makes sense" for 0^0.
Depending on the context (are you working in set theory? are you making a new definitions for exponentiation?) you might have a different definition. But such operations are often defined recursively (e.g. in set theory, roughly, where S(x) = x+1 (or successor of x) Exp(x, 0) = 1, and Exp(x, S(y)) = x * Exp(x, y). Here you'll have 0^0 = 1, clearly.
For high school, 0^0 should be 1. It's necessary for problems high school students might encounter in calculous, and is the way it is defined in almost any field you'd be working in before graduate school.
High school teachers who insist that 0^0 != 1 likely don't understanding that it's a definition.
Clever student> hmmm...ohhh...aaahhh....WTF...I hate math I don't want this stupid stupid job lemme go back to coding monads in Haskell for my ubercool startup.
PITA interviewer >Don't let the door hit you on your way out.
The indeterminate form seems the most correct based on the analysis of the limit of f(x,y) = x^y as x approached zero from different paths.
I had always thought of it more of an algebraic identity thing; x^n * x^m = x^(n+m). Obviously x^(n) = x^(n+0) = x^n * x^0 which can only be satisfied if x^0 = 1. But this article (and really, the wikipedia treatment that beej71 linked to) made me think more about it.
This! The empty function is the function whose graph is the empty set. When you stop thinking of functions as symbolic expressions and make friends with the empty set, it's all clear as crystal. The empty set has perplexed a lot of people over the ages, so it's unsurprising that 0^0 prompts puzzlement. But with the benefit of modern hindsight, there's really no reason to stay confused. Set theory in the large is still a great mystery but we got the empty set well figured out by now.
The engineer in me says, "Why don't you guys pick something, and I'll go ahead and use it in my calculations (where it's appropriate)? I don't care much about the theory behind it."
This is completely off-topic, but did you mean to call it l'hospital as opposed to L'Hôpital? I remember even my calculus book had similar errors, and I always wondered if it was just because the two looked so similar (or if there was any more reasoning behind it). didn't mean to nitpick, your comment just triggered a repressed train of thought :)
His original name was actually Guillaume de l'Hospital; French spelling reforms later did away with a number of cases of silent 's' (which had been silent for a long time already), replacing it with a circumflex over the preceding vowel.
It's not an error. The french changed their spelling to replace a silent 's' after some letters with a circumflex over those letters. Using the silent s instead of a circumflex is considered correct when transcribing into English.
No. What would you apply it to in 0^0? The best you could do is come up with two functions, f(x) and g(x), with f(x) and g(x) going to 0 as x goes to 0, and try to use it to evaluate f(x)^g(x), but the result is going to depend on exactly what your choices are for f(x) and g(x).
A precondition of L'Hôpital's Rule is that the limit in question exists. So: Prove that the limit exists, and then you can bust out L'Hôpital's Rule to prove that it's 1.
And the cleverest student turns out to be right, given the convention asserted by the mathematician. If 0⁰ were 0, or indeterminate, then the limit wouldn't exist and, therefore, L'Hôpital's Rule wouldn't apply. But given that it's 1, the limit does exist, and all is well until Zermelo-Fraenkel is proven inconsistent.
Only after seeing how many comments on the article attempted to genuinely refute the article did I get a sense of how few people grasp the foundation of mathematics (that is, the composition of arbitrary assumptions to agreeable statements).
Doesn't the first argument have a logic error, in that x^(1-1) = x^1x^(-1) has a caveat for x not equal to 0, so it shouldn't prove anything for x = 0. In other words, it just says x^0 = 1 for any x =/= 0.
Using Abstract Algebra, I think 0^0 = 1 is completely accurate. The power function (y^x) could be defined to be the amount you times (x times) you apply the operation between the y on the identity element. In our usual numbers that looks like y(y(y...(y1)...)). When x is negative y becomes the multiplicative inverse of y and everything else remains the same.
The right convention is also "obvious" to practitioners of combinatorics. The exponentiation x^y, for integer x and y, is the number of possible strings of length y from a set of letters of cardinality x. (Hence 28 possible bytes.) How many ways are there to make a string of length 0, regardless of the alphabet size? Just one... you don't do anything.
We had this guy as a lecturer. Whether nullity exists or not (though James's argument is that it's as valid as j), and I have to admit his arithmetic does make some things simpler.
Yeah, I was just talking about the easier concept of x^0. Going the other way, 0^x it makes a lot less sense to me. I know that for all positive x, 0^x = 0. That it pops up to 1 at the origin means the thing is not smooth so drawing lines isn't going to help us.
Plotting z = y^x could be interesting but we're going to see asymptotic behaviour at x = y = 0 (when going in the x direction). So no easy wins there.
Practically, 0^0 highlights the issue that most of us don't have a good conceptual model for what exponents really do. How would you explain to a 10-year old why 3^0 = 1 beyond "it's necessary to make the algebra of powers work out".
I use an "expand-o-tron" analogy
http://betterexplained.com/articles/understanding-exponents-...
to wrap my head around what exponents are really doing: some amount of growth (base) for some amount of time (power). This gives you a "multiplier effect". So, 3^0 means "3x growth for 0 seconds" which, being 0 seconds, changes nothing -- the multiplier is 1. "0x growth for 0 seconds" is also 1, since it was never applied. "0x growth for .00001 seconds" is 0, since a miniscule amount of obliteration still obliterates you.
This can even be extended to understand, intuitively, why i^i is a real number (http://betterexplained.com/articles/intuitive-understanding-...).