Higher-Order Monotonicity
Remember in calculus class, when you learned that a function is “increasing” if , and that it’s “concave up” (or convex) if ? What about all the higher-order cases, where , and beyond? Apparently nobody cared enough to name those cases.
Well, let’s name them! If , then we will say that is -increasing (or positive -monotone). So, nonnegative functions are -increasing, while “increasing” functions are -increasing, and convex functions are -increasing.
Although we just defined these using derivatives, -increasing and -increasing (convex) functions can be defined more generally, without any requirement of differentiability.
- A function is -increasing if whenever .
- A function is -increasing (convex) if for any two points on the function’s graph, the line segment connecting these points lies above the graph.
In this post, we will see a different way of looking at these definitions which elegantly extends to any order of monotonicity.
Rethinking Convexity
When we turn the above definition of convexity into a formula, we get something like this:
Convexity (Definition 1):
A function is -increasing, or convex, if for every triple ,
The right-hand side of this formula is the height of the line segment connecting to , at the coordinate of , so this is a correct interpretation of the function lying under the line segment. Here’s an example of a convex function, where you can play around with the three points.
Intuitively, a function should be convex if it “curves up”, and never curves down. I think the above definition captures this in a clever way. If the function lies below the line segment connecting two points, it makes sense that it ultimately has to “curve up” as it goes from one end to another. You may object that it could squiggle up and down as much as it wants under the line segment, but that worry is quelled by the fact that the function has to lie under every line segment. If it ever curved down, you could find a line segment that the function would rise above, so it wouldn’t be convex as per the definition.
But despite this definition’s visual clarity, it feels a little… asymmetrical. The numbers and represent the endpoints of the line segment, and is the point at which we check that the function lies below the line segment. It just feels odd to me that the three points play different roles. It seems like there should be some symmetrical way to phrase this where , , and all play the same part, regardless of which is between the other two.
But more importantly, this definition seems too specialized. It is not obvious by this definition that -monotonicity is a “sequel” to -monotonicity, and it is not clear at all how this definition could be modified to capture -monotonicity. The whole “line segment” idea is specific to convexity and convexity alone. But fortunately, there’s an equivalent way to define convexity that avoids this issue.
Forget Lines. Use Parabolas!
The reason for the this definition’s asymmetry is that it involves three points, but a line segment is determined by two points. Thus one point must be the odd one out, playing a different role than the other two. But while a line is determined by two points, a parabola is determined by three points.
So let’s think about the unique parabola that passes through , , and . Here is an interactive demonstration:
No matter what you do, you can never get the parabola to curve down. This is why the function is convex: Every parabola passing through three of its points is convex.
You can convince yourself that this is indeed equivalent to the line-based definition. If the middle point lies under the line segment connecting the other two points, then the parabola must “curve up” to touch all three points. Conversely, if the parabola curves up, the middle point will always lie under the line connecting the other two points.
So this gives us a new definition of convexity:
Convexity (Definition 2):
A function is -increasing, or convex, if for every choice of three distinct numbers , , and , the leading coefficient of the unique parabola passing through , , and is nonnegative.
This definition is symmetrical! We did not have to specify the order of , , and ; all three numbers play the same role.
Extending the Definition to All Orders
Symmetry isn’t the only perk of this new definition. It’s also easy to extend to any order! We said that a function was -increasing if for any points, the degree polynomial passing through all those points has a nonnegative leading coefficient. Well, let’s adapt this to the other orders.
-Increasing (Nonnegative):
A function is -increasing (nonnegative) if for every number , the leading coefficient of the unique degree polynomial passing through is nonnegative.
This is trivially true. A degree polynomial is a constant, and the constant which passes through is obviously just . So this constant is nonnegative if and only if is nonnegative.
What about -increasing functions?
-Increasing:
A function is -increasing (or just increasing) if for every choice of two distinct numbers and , the leading coefficient of the unique degree polynomial passing through and is nonnegative.
What do you know, it works! A degree polynomial is just a line, and line is increasing whenever its leading coefficient is nonnegative. Also a line segment is increasing if and only if its left point is below its right point. Clearly this matches the condition that whenever .
So continuing this pattern, the general definition is clear!
n-Increasing:
A function is -increasing if for every choice of distinct numbers , the leading coefficient of the unique degree polynomial passing through for all is nonnegative.
Divided Differences
Alright, we have a derivative-free definition of all orders of monotonicity. But it relies on us being able to find the polynomial that passes through a collection of points. There are a couple ways of doing this, this simplest of which is Lagrange interpolation. But there’s another way discovered by Newton that relies on a concept called divided differences. I will denote divided differences by the name of a function, followed by one or several numbers in square brackets. They are defined recursively as follows:
For example,
I won’t go into the details here, but Newton found that is equal to the leading coefficient of the polynomial passing through at all the points . This gives us a nice notation for our definition:
n-Increasing (With Divided Differences):
A function is -increasing if for every choice of distinct numbers ,
Does This Match the Derivative Definition?
We started by defining a function as -increasing if . Then we found a more general definition using divided differences which does not require to be differentiable. But are these two definitions compatible?
Yes they are! If is -differentiable on an interval, then the two definitions are equivalent on that interval. Let’s prove it.
First, assume that is nonnegative. There is a generalization of the mean value theorem called the mean value theorem for divided differences, which states that there will always exist some number between the min and max values of such that
It follows that if is always nonnegative, then all forward differences of involving points will be nonnegative.
For the converse, assume that every divided difference of involving points is nonnegative. We can use the recursive definition of divided differences, along with their linearity, to show that
In , we build a telescoping sum by removing one “” at a time. In , we use the definition of divided differences to see that each difference quotient is a divided difference of points. Then, by assumption, every such divided difference is nonnegative, so the inequality holds.
We’ve shown that all forward differences of involving points are nonnegative. By repeating this process, we can eventually see that every forward difference of involving point is nonnegative. But forward differences involving point are defined as the function itself, so is always nonnegative. This completes the proof.
Final Notes
This definition of higher-order monotonicity was first used by Eberhard Hopf, who wrote about it here, and Tiberiu Popoviciu, whose paper can be found here. (I have not read these sources, but they were referenced in the article “A new look at Popoviciu’s concept of convexity for functions of two variables” by Sorin G. Gal and Constantin P. Niculescu.)
I would love to learn more about the relationship between -monotonicity and differentiability. It is known that, on open sets, -monotone functions have left-sided and right-sided limits everywhere, and -monotone functions are left and right differentiable everywhere (and therefore continuous). I wonder if these observations can be extended to higher orders. Are -monotone functions on open sets necessarily differentiable? Are they left and right 2-differentiable? Does the pattern continue? I haven’t been able to prove any of this, but I would like to look into it more. The answers may be in the papers I linked in the previous paragraph. Maybe one day I will find the time to pull out Google Translate and read them.