try ai
Popular Science
Edit
Share
Feedback
  • Newton Form of Interpolating Polynomials

Newton Form of Interpolating Polynomials

SciencePediaSciencePedia
Key Takeaways
  • The Newton form constructs an interpolating polynomial using divided differences, which represent successive, discrete rates of change within the data.
  • It is computationally efficient through Horner's method and highly adaptable, allowing new data points to be added without recalculating the entire polynomial.
  • The method is widely applied in robotics for path planning, finance for modeling yield curves, and as a core component of advanced numerical algorithms.
  • Deep mathematical connections reveal that the Newton form's recursive evaluation is fundamentally the same algorithm as the de Boor algorithm used for B-splines.

Introduction

The challenge of finding a smooth path that perfectly connects a set of data points is a classic problem in mathematics and science known as polynomial interpolation. While many are familiar with writing polynomials in a standard power-series form, this approach is often not the most efficient or insightful for practical applications. This raises the question: is there a more natural and powerful way to construct these functions, one that not only provides the answer but also reveals deeper relationships within the data?

This article explores such a method: the Newton form of the interpolating polynomial. We will uncover how this elegant approach, built on the genius of Isaac Newton, provides a superior framework for modeling, prediction, and computation.

First, in "Principles and Mechanisms," we will dissect the construction of the Newton form, introducing the intuitive concept of divided differences and highlighting its remarkable computational efficiency and adaptability. Following that, in "Applications and Interdisciplinary Connections," we will journey through its diverse uses, from guiding robotic arms and modeling financial markets to serving as the engine for complex numerical algorithms, revealing the unifying power of this mathematical tool.

Principles and Mechanisms

Imagine you have a handful of stars in the night sky and you want to trace a smooth path that passes through every single one. In mathematics, this is the classic problem of ​​polynomial interpolation​​. We seek a function—a polynomial—that perfectly connects a set of given data points. You might remember from algebra class that one way to write a polynomial is the familiar "power series" or "monomial" form: P(x)=anxn+an−1xn−1+⋯+a0P(x) = a_n x^n + a_{n-1} x^{n-1} + \dots + a_0P(x)=an​xn+an−1​xn−1+⋯+a0​. This form is built from simple blocks: 111, xxx, x2x^2x2, x3x^3x3, and so on. But what if there were a more clever, more natural set of building blocks for this task?

This is where the genius of Isaac Newton comes into play. The ​​Newton form​​ of an interpolating polynomial is a different, and in many ways superior, way of constructing this path through the stars. It's a journey of discovery that reveals not only the path itself but also the very nature of the relationships between the points.

A New Set of Building Blocks

Instead of using powers of xxx, the Newton form uses a basis that is custom-built from the locations of our data points, which we'll call ​​nodes​​ (x0,x1,x2,…x_0, x_1, x_2, \dotsx0​,x1​,x2​,…). The building blocks look like this:

  1. A constant term: 111
  2. A linear term: (x−x0)(x-x_0)(x−x0​)
  3. A quadratic term: (x−x0)(x−x1)(x-x_0)(x-x_1)(x−x0​)(x−x1​)
  4. A cubic term: (x−x0)(x−x1)(x−x2)(x-x_0)(x-x_1)(x-x_2)(x−x0​)(x−x1​)(x−x2​)
  5. ...and so on.

The full polynomial is a weighted sum of these blocks: P(x)=c0+c1(x−x0)+c2(x−x0)(x−x1)+c3(x−x0)(x−x1)(x−x2)+…P(x) = c_0 + c_1(x-x_0) + c_2(x-x_0)(x-x_1) + c_3(x-x_0)(x-x_1)(x-x_2) + \dotsP(x)=c0​+c1​(x−x0​)+c2​(x−x0​)(x−x1​)+c3​(x−x0​)(x−x1​)(x−x2​)+…

Notice something beautiful here. The second term, c1(x−x0)c_1(x-x_0)c1​(x−x0​), is zero when x=x0x = x_0x=x0​. The third term is zero when x=x0x = x_0x=x0​ or x=x1x = x_1x=x1​. Each new term we add is specifically designed to be zero at all the previous nodes! This property is the secret to the Newton form's power, as we shall see. The big question, of course, is: what are these mysterious coefficients, the ckc_kck​'s?

Divided Differences: The Blueprint for Construction

The coefficients of the Newton polynomial are not just arbitrary numbers; they have a profound and intuitive meaning. They are called ​​divided differences​​. Let's build them up step by step. Our data points are (x0,y0),(x1,y1),…(x_0, y_0), (x_1, y_1), \dots(x0​,y0​),(x1​,y1​),….

The first coefficient, c0c_0c0​, must ensure the polynomial passes through the first point (x0,y0)(x_0, y_0)(x0​,y0​). If we plug x=x0x=x_0x=x0​ into our Newton form, every term after the first one vanishes. So, we must have P(x0)=c0=y0P(x_0) = c_0 = y_0P(x0​)=c0​=y0​. The first coefficient is simply the value of our function at the starting point. We denote this as c0=f[x0]c_0 = f[x_0]c0​=f[x0​].

The second coefficient, c1c_1c1​, is determined by the second point, (x1,y1)(x_1, y_1)(x1​,y1​). We have P(x1)=c0+c1(x1−x0)=y1P(x_1) = c_0 + c_1(x_1-x_0) = y_1P(x1​)=c0​+c1​(x1​−x0​)=y1​. Solving for c1c_1c1​ gives: c1=y1−y0x1−x0c_1 = \frac{y_1 - y_0}{x_1 - x_0}c1​=x1​−x0​y1​−y0​​ This is nothing more than the slope of the line connecting the first two points! We call this the first-order divided difference, denoted f[x0,x1]f[x_0, x_1]f[x0​,x1​].

The third coefficient, c2c_2c2​, is found by looking at the third point, (x2,y2)(x_2, y_2)(x2​,y2​). It represents the "rate of change of the slope." It's a measure of curvature. The pattern continues, with each coefficient being defined recursively by the previous ones. The general kkk-th order divided difference is: f[xi,…,xi+k]=f[xi+1,…,xi+k]−f[xi,…,xi+k−1]xi+k−xif[x_i, \dots, x_{i+k}] = \frac{f[x_{i+1}, \dots, x_{i+k}] - f[x_i, \dots, x_{i+k-1}]}{x_{i+k} - x_i}f[xi​,…,xi+k​]=xi+k​−xi​f[xi+1​,…,xi+k​]−f[xi​,…,xi+k−1​]​ This formula might look intimidating, but it describes a wonderfully simple, recursive process. We can organize this entire calculation in a ​​divided difference table​​. We list our points, then calculate the first differences, then the second, and so on.

xix_ixi​f[xi]f[x_i]f[xi​]f[xi,xi+1]f[x_i, x_{i+1}]f[xi​,xi+1​]f[x0,x1,x2]f[x_0, x_1, x_2]f[x0​,x1​,x2​]
x0x_0x0​y0y_0y0​
f[x0,x1]f[x_0, x_1]f[x0​,x1​]
x1x_1x1​y1y_1y1​f[x0,x1,x2]f[x_0, x_1, x_2]f[x0​,x1​,x2​]
f[x1,x2]f[x_1, x_2]f[x1​,x2​]
x2x_2x2​y2y_2y2​

The magic is that the coefficients for our Newton polynomial—c0,c1,c2,…c_0, c_1, c_2, \dotsc0​,c1​,c2​,…—are precisely the entries along the top diagonal of this table. For example, if a calculation gives us the polynomial P(x)=7+2(x−0)−0.5(x−0)(x−3)P(x) = 7 + 2(x - 0) - 0.5(x - 0)(x - 3)P(x)=7+2(x−0)−0.5(x−0)(x−3), we can immediately identify the divided differences: f[x0]=7f[x_0] = 7f[x0​]=7, f[x0,x1]=2f[x_0, x_1] = 2f[x0​,x1​]=2, and f[x0,x1,x2]=−0.5f[x_0, x_1, x_2] = -0.5f[x0​,x1​,x2​]=−0.5. The structure of the polynomial directly reveals the coefficients.

The Elegance of Efficiency

So, we have this alternative way to write a polynomial. Why is it better? The first reason is stunning computational efficiency. Suppose we have our Newton polynomial and want to find its value at some new point, xxx. For instance, a sensor's behavior is modeled by a polynomial, and we want to predict the voltage at a time we didn't measure.

Look again at the Newton form, but let's write it in a nested way: P(x)=c0+(x−x0)(c1+(x−x1)(c2+(x−x2)[… ]))P(x) = c_0 + (x-x_0) \Big( c_1 + (x-x_1) \big( c_2 + (x-x_2) [ \dots ] \big) \Big)P(x)=c0​+(x−x0​)(c1​+(x−x1​)(c2​+(x−x2​)[…])) To evaluate this, we don't need to compute high powers of xxx. We can start from the inside and work our way out. For a fourth-degree polynomial, the calculation would be:

  1. Start with the innermost term: d3=c3+(x−x3)c4d_3 = c_3 + (x-x_3)c_4d3​=c3​+(x−x3​)c4​
  2. Use that result in the next layer: d2=c2+(x−x2)d3d_2 = c_2 + (x-x_2)d_3d2​=c2​+(x−x2​)d3​
  3. And again: d1=c1+(x−x1)d2d_1 = c_1 + (x-x_1)d_2d1​=c1​+(x−x1​)d2​
  4. Finally, the answer: P(x)=c0+(x−x0)d1P(x) = c_0 + (x-x_0)d_1P(x)=c0​+(x−x0​)d1​

This elegant procedure is a variation of ​​Horner's method​​. Each step involves just one multiplication and one addition. For a polynomial of degree nnn, this method requires only nnn multiplications and 2n2n2n additions/subtractions.

How does this compare to other methods, like the Lagrange form? While the Lagrange form is theoretically beautiful, evaluating it naively at a new point requires a number of operations that grows with the square of the degree, or O(n2)O(n^2)O(n2). The Newton-Horner method, by contrast, scales linearly, as O(n)O(n)O(n). For real-time systems like in robotics or path planning, where speed is critical, this difference is not just academic—it's the difference between a system that works and one that can't keep up.

The Power of Adaptability

The second, and perhaps most beautiful, advantage of the Newton form is its extensibility. Imagine you've done all the work to find the polynomial path through your first ten stars. Then, you discover an eleventh star you'd like to include. With the standard monomial or Lagrange forms, you'd have to throw away all your previous work and start the entire calculation from scratch.

Not so with the Newton form.

Let's say you have your polynomial Pn(x)P_n(x)Pn​(x) that passes through n+1n+1n+1 points. To find the new polynomial Pn+1(x)P_{n+1}(x)Pn+1​(x) that also includes the new point (xn+1,yn+1)(x_{n+1}, y_{n+1})(xn+1​,yn+1​), you simply add one more term: Pn+1(x)=Pn(x)+cn+1(x−x0)(x−x1)…(x−xn)P_{n+1}(x) = P_n(x) + c_{n+1} (x-x_0)(x-x_1)\dots(x-x_n)Pn+1​(x)=Pn​(x)+cn+1​(x−x0​)(x−x1​)…(x−xn​) This works because the new term we add, (x−x0)(x−x1)…(x−xn)(x-x_0)(x-x_1)\dots(x-x_n)(x−x0​)(x−x1​)…(x−xn​), is guaranteed to be zero at all of the original nodes x0,x1,…,xnx_0, x_1, \dots, x_nx0​,x1​,…,xn​. So, adding it doesn't disturb the fact that the polynomial already passes through the old points! All we need to do is calculate one new coefficient, the next divided difference cn+1=f[x0,…,xn+1]c_{n+1} = f[x_0, \dots, x_{n+1}]cn+1​=f[x0​,…,xn+1​], and tack on the new term. This is the mathematical equivalent of building a modular structure where you can always add another floor without having to rebuild the foundation.

Deeper Truths and Hidden Symmetries

The Newton form also reveals deeper truths about polynomials. A fundamental theorem states that for a given set of distinct points, there is only one unique polynomial of the lowest possible degree that passes through them all. But as we've seen, this unique polynomial can wear different "disguises." If you reorder your data points and construct a new Newton polynomial, the coefficients and the basis functions will look completely different. Yet, if you were to expand both forms into the standard power series, they would collapse into the exact same polynomial. This is a beautiful lesson in mathematics: the underlying reality is invariant, even when our description of it changes. It’s like describing an object from different points of view; it looks different, but it’s still the same object.

Interestingly, while most of the divided difference table changes when you reorder the points, the highest-order divided difference, f[x0,x1,…,xn]f[x_0, x_1, \dots, x_n]f[x0​,x1​,…,xn​], remains the same regardless of the permutation of the nodes. This is because this value has an invariant meaning: it is precisely the leading coefficient (ana_nan​) of the polynomial when written in the standard form P(x)=anxn+…P(x) = a_n x^n + \dotsP(x)=an​xn+…. It represents the overall, dominant "bend" of the curve across all the data points.

A crucial word of warning, however. The elegance of the Newton form doesn't make it a magic bullet. If you try to fit a very high-degree polynomial (say, degree 50) through many evenly-spaced points, you'll run into a monster known as ​​Runge's phenomenon​​, where the polynomial wiggles wildly between the nodes. The problem here is not the tool, but the task itself; high-degree interpolation on equispaced nodes is an ​​ill-conditioned​​ problem. While the Newton form is numerically more stable than trying to solve the problem with the monomial basis (which is notoriously unstable), it cannot overcome the intrinsic instability of the problem itself. The true expert knows not only how to use their tools, but also when not to use them, or how to change the problem (for instance, by using a different spacing of nodes, like Chebyshev nodes) to make it solvable.

The principle of divided differences is so powerful that it can even be extended. What if you wanted your curve not only to pass through a point, but to have a specific slope (derivative) there? This is called ​​Hermite interpolation​​. By treating a point with a specified derivative as two nodes that are infinitesimally close together, the divided difference framework can be generalized to solve this more complex problem, allowing us to design smooth trajectories for things like robotic arms.

The Newton form is more than just a formula; it's a way of thinking. It's about building a solution piece by piece, with each new piece respecting the work that has come before. It demonstrates the profound difference between a brute-force calculation and an elegant, structured approach—a theme that echoes throughout science and engineering.

Applications and Interdisciplinary Connections

Now that we have acquainted ourselves with the machinery of Newton's interpolating polynomial and the cleverness of divided differences, we might be tempted to view it as a neat mathematical curiosity. A clever trick, perhaps, for connecting dots. But to do so would be to miss the forest for the trees. The true power and beauty of this idea, like so many in science, lie not in its isolated elegance, but in its remarkable ability to describe, predict, and connect phenomena across a breathtaking range of disciplines. It is a universal key that unlocks problems in physics, finance, chemistry, and engineering. Let us now take a journey through some of these worlds and see what doors this key can open.

The Shape of Things: From Robot Arms to Light Beams

Perhaps the most intuitive application of interpolation is in defining shape and motion. Imagine you are designing the path for a robotic arm on an assembly line. You have a set of crucial "waypoints"—positions (x,y)(x, y)(x,y) that the arm must pass through at specific times ttt. How do you generate a smooth, continuous path connecting these points? The answer is to treat the problem as two separate, but synchronized, interpolation tasks. We construct one Newton polynomial, x(t)x(t)x(t), that takes the arm to the correct xxx-coordinate at each time tit_iti​, and a second polynomial, y(t)y(t)y(t), that does the same for the yyy-coordinate. The result is a smooth parametric curve, P(t)=(x(t),y(t))P(t) = (x(t), y(t))P(t)=(x(t),y(t)), that glides effortlessly through all the required waypoints.

But we can go further. We don't just want to know where the arm is; we want to know its velocity. A remarkable feature of the polynomial representation is that we can simply differentiate it. The structure of the Newton form provides a systematic way to find the derivative of our interpolating polynomial, giving us the velocity vector (x′(t),y′(t))(x'(t), y'(t))(x′(t),y′(t)) at any instant in time. This is not just an academic exercise; for a real robot, controlling velocity and acceleration is critical for precision and avoiding jerky, damaging movements.

The same principle of using interpolation to model a physical relationship applies in less visible, but equally fundamental, domains. Consider the phenomenon of dispersion, where a prism splits white light into a rainbow. This happens because the refractive index of the glass, nnn, changes with the wavelength of light, λ\lambdaλ. Physicists studying optics might have a few experimental measurements of nnn at specific wavelengths. Often, physical theory suggests that a relationship becomes simpler if we look at the right variables. For glass dispersion, a nearly straight-line relationship emerges if we plot nnn not against λ\lambdaλ, but against 1/λ21/\lambda^21/λ2. By applying Newton's method to this transformed data, we can build a highly accurate model, n(1/λ2)n(1/\lambda^2)n(1/λ2), that allows us to predict the refractive index—and thus how light will bend—at any wavelength in the visible spectrum. This is a beautiful example of a common theme in science: combining a physical insight (the variable transformation) with a powerful mathematical tool (interpolation) to create a predictive model from sparse data.

The Language of Change: Modeling Dynamic Systems

Many systems in science and finance are not static; they are constantly evolving. Here, the unique structure of the Newton form truly shines. Let's step into the world of computational finance, where analysts model the yield curve—a plot of the interest rate of bonds against their maturity time. We can create an initial model by interpolating the known yields for bonds with maturities of, say, 1, 5, and 10 years. Now, imagine a new bond with a 2-year maturity is issued, and its yield is announced. If we had used other interpolation methods, we might have to throw away our work and re-calculate the entire polynomial from scratch. But with the Newton form, the process is breathtakingly simple. Our existing polynomial remains the foundation; we simply calculate one new set of divided differences and add one more term to the end of our sum. It’s like adding a new Lego brick to a structure without having to rebuild it from the ground up. This "updatability" is of enormous practical value in fast-paced environments where new information arrives continuously.

Furthermore, the coefficients in the Newton form—the divided differences themselves—are not just abstract numbers; they have profound, intuitive meaning. Consider our yield curve, which relates maturity, ttt, to yield, f(t)f(t)f(t). The first-order divided difference, f[t1,t2]=(f(t2)−f(t1))/(t2−t1)f[t_1, t_2] = (f(t_2) - f(t_1)) / (t_2 - t_1)f[t1​,t2​]=(f(t2​)−f(t1​))/(t2​−t1​), is simply the slope of the line between two points. It's a discrete measure of the rate of change. The second-order divided difference, f[t1,t2,t3]f[t_1, t_2, t_3]f[t1​,t2​,t3​], measures how this slope itself is changing. It is, in essence, a discrete approximation of the second derivative. In finance, the second derivative of a price function is known as its convexity, a critical measure of risk. So, the sign of our second divided difference tells us immediately whether the yield curve is locally "convex up" (bending upwards) or "convex down" (bending downwards). The mathematical tool gives us direct insight into a key economic concept.

This same pattern of transforming data to reveal a simpler structure, and then interpolating, appears in physical chemistry when studying reaction rates. The famous Arrhenius equation predicts that the natural logarithm of a reaction rate constant, ln⁡(k)\ln(k)ln(k), should be a linear function of the inverse temperature, 1/T1/T1/T. By plotting experimental data in this way (an "Arrhenius plot"), chemists can interpolate a handful of measurements to predict the reaction rate at any temperature within the range, which is essential for designing and controlling chemical processes.

The Engine of Computation: Building Better Algorithms

So far, we have used interpolation to model the world. But one of its most powerful applications is in building the very engines of scientific computation itself. Consider one of the central problems in science: solving ordinary differential equations (ODEs), which describe everything from planetary orbits to the spread of a disease. A common family of numerical methods, the Adams-Bashforth methods, are built on a beautiful idea. To find the value of our solution y(t)y(t)y(t) at the next time step, tn+1t_{n+1}tn+1​, we look at the recent history of its derivative, f(t,y)f(t,y)f(t,y), at times tn,tn−1,tn−2,…t_n, t_{n-1}, t_{n-2}, \dotstn​,tn−1​,tn−2​,…. We then fit an interpolating polynomial through these past derivative values and integrate it from tnt_ntn​ to tn+1t_{n+1}tn+1​. This integral gives us our estimate for the change in yyy.

Why is the Newton form so crucial here? Because sophisticated ODE solvers don't use a fixed step size; they take large steps when the solution is smooth and small steps when it changes rapidly. This means the time points tn,tn−1,…t_n, t_{n-1}, \dotstn​,tn−1​,… are not equally spaced. The Newton form, built on divided differences, handles non-uniform spacing with perfect ease, making it the ideal foundation for these powerful and efficient "adaptive" algorithms.

Interpolation also serves as a core component in other numerical algorithms, like root-finding. Newton's method for finding roots approximates a function with a tangent line. The Secant method uses a line passing through two points. Müller's method takes this one step further. It takes the last three approximations to a root, fits a parabola through them using a quadratic Newton polynomial, and finds where that parabola intersects the axis. This is often more robust and can find complex roots. Once again, polynomial interpolation serves as a fundamental building block for a more sophisticated computational tool.

Beyond the Line: Painting on a Grid and Unifying Theories

Our journey so far has been along a one-dimensional line. But the world has more dimensions. What if a quantity, like the pressure of a gas, depends on two variables, say volume VVV and temperature TTT? Can we interpolate on a 2D grid of data points? The principle of divided differences extends with stunning elegance. We can think of it as interpolation in stages. First, for each fixed temperature in our data, we perform a 1D Newton interpolation across all the volume measurements. This gives us a set of interpolating polynomials, one for each temperature. The coefficients of these polynomials, however, vary with temperature. So, in the second stage, we interpolate the coefficients themselves as functions of temperature! This iterated process produces a single, smooth surface P(V,T)P(V, T)P(V,T) that passes through every point in our data grid. It’s a powerful testament to the recursive and generalizable nature of the underlying mathematics.

We end our tour with a glimpse into an even deeper connection, a moment of profound unity that would have made Feynman smile. In the world of computer-aided design (CAD) and computer graphics, shapes are often defined by B-spline curves. These curves are not typically defined by points they must pass through, but by a set of "control points" that guide their shape. They are evaluated using a clever and stable recursive recipe known as the de Boor algorithm. On the surface, this seems a world away from our interpolating polynomials.

Yet, it is not. The deep truth, revealed by a branch of mathematics known as blossoming or polar forms, is that Newton's divided difference evaluation and the de Boor algorithm are, at their core, the very same algorithm. They are two different dialects of the same fundamental geometric language. Both are recursive schemes that build up the final value through a cascade of simple weighted averages. By performing a change of basis—translating the interpolation points and divided differences of our problem into an equivalent set of B-spline knots and control points—the two computational tables become identical. What appeared to be two separate, ingenious inventions are revealed to be two faces of a single, more fundamental mathematical structure. It is in these moments of unexpected unification that we see the true beauty and power of mathematical ideas, echoing across the landscape of science and engineering.