try ai
Popular Science
Edit
Share
Feedback
  • Second-Order Optimization

Second-Order Optimization

SciencePediaSciencePedia
Key Takeaways
  • Second-order optimization methods use the Hessian matrix to understand a function's curvature, enabling faster and more direct convergence to a minimum compared to first-order methods.
  • The classic Newton's method offers quadratic convergence but is computationally infeasible for large problems due to the O(n³) cost of solving its core linear system.
  • Quasi-Newton methods, especially L-BFGS, provide a practical compromise by approximating curvature information with linear memory and computational scaling, making them suitable for large-scale optimization.
  • Problem-specific approximations, like the Gauss-Newton method for least-squares, exploit the function's structure to create efficient second-order-like updates.
  • Second-order concepts are crucial across diverse fields, from finding molecular energy minima in chemistry to ensuring structural stability in engineering via Second-Order Cone Programming (SOCP).

Introduction

In the vast landscape of mathematical optimization, the simplest approach is often to follow the steepest path downhill. This strategy, known as gradient descent, serves as the foundation for solving countless problems. However, it suffers from a critical flaw: it is shortsighted, relying only on local slope. This can lead it to become hopelessly stuck in flat regions or at saddle points, mistaking a mountain pass for a valley floor. To navigate these complex terrains effectively, we need a more sophisticated map—one that reveals not just the slope, but the very curvature of the landscape.

This article delves into the powerful world of second-order optimization, a class of methods that leverage second-derivative information to achieve faster and more robust convergence. By understanding a function's local curvature, these techniques can make intelligent leaps toward a solution, overcoming the pitfalls that plague simpler first-order methods. We will first explore the core principles and mechanisms, uncovering the mathematics of the Hessian matrix, the elegance of Newton's method, and the practical compromises like Quasi-Newton methods that make these ideas viable for large-scale applications. Following this, we will journey through a diverse range of interdisciplinary connections, witnessing how these same mathematical concepts are applied to solve fundamental problems in quantum chemistry, engineering control, statistical analysis, and solid mechanics.

Principles and Mechanisms

Beyond Going Downhill: The Need for a Better Map

Imagine you're a hiker lost in a dense fog, trying to find the lowest point in a vast, rolling landscape. All you have is a special device that tells you which direction is steepest downhill right where you're standing. This is the essence of the most fundamental optimization algorithm, ​​gradient descent​​. You check the direction of the gradient, take a small step, and repeat. It's a simple, robust strategy.

But what happens when you reach a spot that seems perfectly flat? Your device reads zero slope in all directions, so you stop. Have you found the bottom of a valley? Perhaps. But you might also be on the perfectly flat peak of a hill, or, more perplexingly, at a ​​saddle point​​—a mountain pass that slopes down in front and behind you, but up to your left and right. A first-order method like gradient descent, which only looks at the slope, can't tell the difference. Any point where the gradient is zero, known as a ​​critical point​​, is a potential stopping point, but not necessarily the minimum we seek.

To navigate effectively, you need more than just the slope. You need a sense of the curvature of the land. Is this flat spot shaped like a bowl (a minimum), a dome (a maximum), or a saddle? To gain this insight, we must turn to second-order information.

The Hessian: A Mathematical Lens for Curvature

In the language of calculus, the tool that describes the local curvature of a function is the ​​Hessian matrix​​. For a function fff that depends on a vector of variables x=(x1,x2,…,xn)\mathbf{x} = (x_1, x_2, \dots, x_n)x=(x1​,x2​,…,xn​), the Hessian, denoted ∇2f(x)\nabla^2 f(\mathbf{x})∇2f(x), is the n×nn \times nn×n matrix of all its second-order partial derivatives.

∇2f(x)=(∂2f∂x12∂2f∂x1∂x2⋯∂2f∂x1∂xn∂2f∂x2∂x1∂2f∂x22⋯∂2f∂x2∂xn⋮⋮⋱⋮∂2f∂xn∂x1∂2f∂xn∂x2⋯∂2f∂xn2)\nabla^2 f(\mathbf{x}) = \begin{pmatrix} \frac{\partial^2 f}{\partial x_1^2} & \frac{\partial^2 f}{\partial x_1 \partial x_2} & \cdots & \frac{\partial^2 f}{\partial x_1 \partial x_n} \\ \frac{\partial^2 f}{\partial x_2 \partial x_1} & \frac{\partial^2 f}{\partial x_2^2} & \cdots & \frac{\partial^2 f}{\partial x_2 \partial x_n} \\ \vdots & \vdots & \ddots & \vdots \\ \frac{\partial^2 f}{\partial x_n \partial x_1} & \frac{\partial^2 f}{\partial x_n \partial x_2} & \cdots & \frac{\partial^2 f}{\partial x_n^2} \end{pmatrix}∇2f(x)=​∂x12​∂2f​∂x2​∂x1​∂2f​⋮∂xn​∂x1​∂2f​​∂x1​∂x2​∂2f​∂x22​∂2f​⋮∂xn​∂x2​∂2f​​⋯⋯⋱⋯​∂x1​∂xn​∂2f​∂x2​∂xn​∂2f​⋮∂xn2​∂2f​​​

If the gradient ∇f\nabla f∇f tells us how the function's value changes as we move, the Hessian ∇2f\nabla^2 f∇2f tells us how the function's gradient changes. It's the "rate of change of the rate of change." Calculating it is a straightforward, if sometimes tedious, application of differentiation. For instance, given the gradient of a potential energy function, we can find all the second derivatives to assemble the Hessian and analyze its properties at any point.

For many functions encountered in science and engineering, particularly those that are quadratic in nature, the Hessian reveals a deep connection to the function's underlying structure. For a general quadratic cost function, the Hessian matrix is constant and directly related to the matrices defining the quadratic form, giving us a complete picture of the function's global curvature.

A particularly beautiful example comes from statistics. The log-probability density of the ubiquitous multivariate normal distribution is a perfect quadratic function. Its Hessian is a constant matrix, equal to the negative inverse of the covariance matrix, −Σ−1-\Sigma^{-1}−Σ−1. This profound result connects the statistical notion of covariance directly to the geometric concept of curvature in optimization.

What the Hessian Tells Us: Valleys, Peaks, and Passes

The true power of the Hessian emerges at a critical point where the gradient is zero. Here, the properties of the Hessian matrix tell us everything about the local geometry. We analyze the Hessian by looking at its ​​eigenvalues​​.

  • If all eigenvalues are strictly positive, the Hessian is ​​positive definite​​. The surface curves upwards in every direction, like a bowl. We have found a ​​local minimum​​.

  • If all eigenvalues are strictly negative, the Hessian is ​​negative definite​​. The surface curves downwards in every direction, like a dome. We have found a ​​local maximum​​.

  • If some eigenvalues are positive and some are negative, the Hessian is ​​indefinite​​. The surface curves up in some directions and down in others. This is the signature of a ​​saddle point​​.

  • If some eigenvalues are zero and the rest have the same sign (all positive or all negative), the Hessian is ​​positive or negative semidefinite​​. The situation is ambiguous; it could be a minimum, a maximum, or a more complex flat region.

The ​​second-order necessary condition​​ for a point to be a local minimum is that its Hessian must be positive semidefinite. If we calculate the Hessian at a critical point and find it has a negative eigenvalue, we can say with certainty that it is not a local minimum. The Hessian gives us the tool to avoid getting trapped on hilltops or saddle points.

The Master Stroke: Newton's Method

Knowing the curvature doesn't just help us classify critical points; it allows us to devise a much more powerful way to find them. This is the genius of ​​Newton's method​​.

The idea is breathtakingly simple and elegant. At our current position xk\mathbf{x}_kxk​, we don't just look at the slope. We use both the gradient ∇f(xk)\nabla f(\mathbf{x}_k)∇f(xk​) and the Hessian ∇2f(xk)\nabla^2 f(\mathbf{x}_k)∇2f(xk​) to build a perfect quadratic approximation of our function—a local paraboloid that matches the value, slope, and curvature of the true landscape at that exact spot. Then, instead of taking a small step downhill, we make one giant leap directly to the minimum of that quadratic model.

This leads to the famous Newton update rule:

xk+1=xk−[∇2f(xk)]−1∇f(xk)\mathbf{x}_{k+1} = \mathbf{x}_k - [\nabla^2 f(\mathbf{x}_k)]^{-1} \nabla f(\mathbf{x}_k)xk+1​=xk​−[∇2f(xk​)]−1∇f(xk​)

In practice, we don't actually compute the inverse of the Hessian. Instead, we solve the linear system of equations ∇2f(xk)pk=−∇f(xk)\nabla^2 f(\mathbf{x}_k) \mathbf{p}_k = -\nabla f(\mathbf{x}_k)∇2f(xk​)pk​=−∇f(xk​) for the Newton step pk\mathbf{p}_kpk​, and then update xk+1=xk+pk\mathbf{x}_{k+1} = \mathbf{x}_k + \mathbf{p}_kxk+1​=xk​+pk​.

When this method works, it works like magic. For a truly quadratic function, its quadratic model is perfect, and Newton's method finds the exact minimum in a single step!. Near a minimum, its convergence is ​​quadratically fast​​, meaning the number of correct digits in the solution roughly doubles with every iteration. It leaves simple gradient descent in the dust.

The Price of Power: The Curse of the Hessian

So, if Newton's method is so wonderful, why don't we use it for everything? The answer lies in its staggering computational cost, a "curse" that renders it unusable for the large-scale problems that dominate modern science and technology, like training massive neural networks.

Let's consider a problem with nnn variables. The costs break down as follows:

  1. ​​Forming the Hessian:​​ We need to compute roughly n(n+1)2\frac{n(n+1)}{2}2n(n+1)​ second derivatives, an operation with a complexity of O(n2)\mathcal{O}(n^2)O(n2).
  2. ​​Storing the Hessian:​​ We need to hold an n×nn \times nn×n matrix in memory, requiring O(n2)\mathcal{O}(n^2)O(n2) storage.
  3. ​​Solving for the Newton step:​​ Solving the n×nn \times nn×n linear system is the killer. For a dense Hessian, this costs O(n3)\mathcal{O}(n^3)O(n3) floating-point operations.

While a gradient descent step costs a mere O(n)\mathcal{O}(n)O(n), the Newton step costs O(n3)\mathcal{O}(n^3)O(n3). Now, imagine training a large language model with n=50n=50n=50 million parameters. The O(n2)\mathcal{O}(n^2)O(n2) memory requirement to store the Hessian would be on the order of petabytes (101510^{15}1015 bytes), far beyond any computer's RAM. The O(n3)\mathcal{O}(n^3)O(n3) computation for a single step would take longer than the age of the universe. For large problems, the classical Newton's method is not just impractical; it's a fantasy.

The Art of Approximation: Living Without the True Hessian

Does this mean we must abandon the power of curvature and retreat to the slow crawl of gradient descent? Not at all! The failure of Newton's method on large problems sparked a revolution in numerical optimization, leading to a beautiful array of techniques that capture the spirit of Newton's method without paying its exorbitant price. The secret is to approximate.

Path 1: Quasi-Newton and the Art of the Update

The most successful family of such methods is ​​Quasi-Newton​​. The core idea is this: if the Hessian is too expensive, let's not compute it. Instead, let's build an approximation of it (or, more cleverly, its inverse) iteratively.

The most celebrated of these is the ​​BFGS​​ (Broyden–Fletcher–Goldfarb–Shanno) algorithm. It maintains an approximation to the inverse Hessian, let's call it BkB_kBk​. At each step, after moving from xk\mathbf{x}_kxk​ to xk+1\mathbf{x}_{k+1}xk+1​, it uses the change in position (sk=xk+1−xks_k = \mathbf{x}_{k+1} - \mathbf{x}_ksk​=xk+1​−xk​) and the change in gradient (yk=∇f(xk+1)−∇f(xk)y_k = \nabla f(\mathbf{x}_{k+1}) - \nabla f(\mathbf{x}_k)yk​=∇f(xk+1​)−∇f(xk​)) to "update" its approximation from BkB_kBk​ to Bk+1B_{k+1}Bk+1​. This update is a simple, computationally cheap formula that nudges the approximation to be more accurate, particularly along the direction of the last step. It's like learning about the landscape's curvature by observing how the slope changes as you walk. This approach replaces the O(n3)\mathcal{O}(n^3)O(n3) system solve with an O(n2)\mathcal{O}(n^2)O(n2) matrix-vector product.

But for our neural network, O(n2)\mathcal{O}(n^2)O(n2) is still too much. This leads to an even more ingenious refinement: ​​Limited-memory BFGS (L-BFGS)​​. L-BFGS realizes that we don't even need to store the approximate Hessian matrix BkB_kBk​. Instead, the search direction can be calculated using only the last mmm pairs of position and gradient changes (where mmm is a small number, like 10 or 20). By storing only this short history, the memory requirement plummets from O(n2)\mathcal{O}(n^2)O(n2) to a manageable O(mn)\mathcal{O}(mn)O(mn), and the computational cost per step drops to O(mn)\mathcal{O}(mn)O(mn). Since mmm is a small constant, the costs scale linearly with the number of parameters, nnn. L-BFGS is the beautiful compromise that has become a workhorse of large-scale optimization: it's much faster than gradient descent because it incorporates curvature information, yet it remains feasible for problems with millions or even billions of variables.

Path 2: Exploiting Problem Structure with Gauss-Newton

Another way to approximate the Hessian is to exploit the specific structure of the problem. A classic example is ​​nonlinear least squares​​, where we aim to minimize a sum of squared errors: f(x)=12∑i=1m[ri(x)]2f(x) = \frac{1}{2} \sum_{i=1}^{m} [r_i(x)]^2f(x)=21​∑i=1m​[ri​(x)]2. This structure is fundamental to data fitting.

If we write out the true Hessian of this function, it splits into two parts:

∇2f(x)=J(x)TJ(x)⏟Part 1+∑i=1mri(x)∇2ri(x)⏟Part 2\nabla^2 f(x) = \underbrace{J(x)^T J(x)}_{\text{Part 1}} + \underbrace{\sum_{i=1}^{m} r_i(x) \nabla^2 r_i(x)}_{\text{Part 2}}∇2f(x)=Part 1J(x)TJ(x)​​+Part 2i=1∑m​ri​(x)∇2ri​(x)​​

where J(x)J(x)J(x) is the Jacobian of the residual functions ri(x)r_i(x)ri​(x). The ​​Gauss-Newton method​​ makes a wonderfully pragmatic approximation: it simply ignores Part 2!. This is brilliant because Part 1 only involves first derivatives (the Jacobian), which are much cheaper to compute than the second derivatives in Part 2.

This approximation is excellent if the model fits the data well (so the residuals ri(x)r_i(x)ri​(x) are small near the solution) or if the model components are nearly linear (so their Hessians ∇2ri(x)\nabla^2 r_i(x)∇2ri​(x) are small). It's another path to an efficient, second-order-like method, tailored to the problem's innate structure.

A Final Word of Caution: The Treacherous Canyons of Ill-Conditioning

Even with these powerful tools, some landscapes are simply treacherous. Imagine a valley that is not a round bowl but a long, narrow canyon. The function is extremely steep across the canyon walls but almost flat along its floor.

This corresponds to an ​​ill-conditioned​​ Hessian, where the ratio of its largest to smallest eigenvalue—the ​​condition number​​ κ\kappaκ—is very large. A large condition number poses serious practical problems for second-order methods:

  1. ​​Numerical Instability:​​ The linear system for the Newton step becomes numerically fragile. Tiny floating-point errors during the calculation get amplified by the condition number, resulting in a computed step that can be wildly inaccurate. This can cause the algorithm to stall, unable to make further progress.

  2. ​​Shrinking Convergence Basin:​​ The magical quadratic convergence of Newton's method is only guaranteed very close to the solution. A large condition number can shrink this region of guaranteed convergence to be infinitesimally small. Far from the solution, the pure Newton step can be a disastrously poor move, sending the iterate flying away from the minimum.

This is why practical implementations always include globalization strategies like ​​line searches​​ (damping the step) or ​​trust regions​​ (restricting the step size) to ensure stability. The condition number, in essence, is a measure of the intrinsic difficulty of the optimization problem. It's a reminder that even with our most sophisticated maps and tools, navigating the vast landscapes of optimization requires care, cleverness, and a healthy respect for the terrain.

Applications and Interdisciplinary Connections

Having grasped the principles of second-order optimization—the art of navigating a complex landscape by not only knowing the direction of steepest descent but also feeling the very curvature of the ground beneath our feet—we can now embark on a journey to see these ideas in action. You might think this is merely a mathematician's game, a set of abstract tools for abstract problems. But nothing could be further from the truth. The principle of using second-derivative information, the Hessian, to find a goal more efficiently is a universal theme, a thread of profound insight that weaves through an astonishing tapestry of scientific and engineering disciplines. We will see how this single idea helps us design molecules, control rockets, decode signals, quantify our own uncertainty, and even ensure that bridges stand firm.

The Quest for the Minimum: Sculpting Reality at the Atomic Scale

At its heart, nature is lazy. A water droplet minimizes its surface area to form a sphere. A soap bubble does the same. A molecule, in its most stable state, will twist and fold itself into a shape that minimizes its total energy. For the quantum chemist, the holy grail is to predict this final, lowest-energy configuration. This is a minimization problem of staggering complexity, set in a high-dimensional landscape of electronic orbitals and nuclear positions.

A first-order method is like being a blindfolded hiker on a mountain, only able to feel the slope at your feet (the gradient) to take a small step downhill. You'll get there eventually, but it's a slow and meandering path. A second-order, Newton-type method gives the hiker a special sense—the ability to feel the curvature of the entire valley. Instead of just taking a step, you can make a calculated leap toward the very bottom. In the language of quantum chemistry, this means solving the famous Newton-Raphson equation, HΔκ=−g\mathbf{H} \Delta\boldsymbol{\kappa} = -\mathbf{g}HΔκ=−g, where g\mathbf{g}g is the energy gradient and H\mathbf{H}H is the Hessian. The solution, Δκ\Delta\boldsymbol{\kappa}Δκ, isn't just a direction; it's a precise, calculated update to the molecular orbitals that promises the fastest convergence to the minimum energy state.

This power, however, comes at a tremendous price. As any good engineer knows, there is no free lunch. Calculating the full "curvature map" of the energy landscape—the Hessian matrix—is computationally brutal. It requires evaluating how every parameter change affects every other parameter, a task that can be far more demanding than simply finding the slope. This is the great trade-off of second-order methods in chemistry: they offer the promise of quadratic convergence, a gloriously fast "homing in" on the solution, but at the cost of building and solving a massive linear system that involves coupling between all the moving parts of the wavefunction.

The story doesn't end in the valleys. Chemical reactions happen when molecules traverse "mountain passes"—saddle points on the energy landscape known as transition states. A transition state is a point of maximum energy along the reaction path but minimum energy in all other directions. How could we possibly find such a specific point? The gradient alone is useless, as it is zero at the top of the pass just as it is at the bottom of a valley. Here, the Hessian is not just helpful; it is essential. A transition state is defined as a point where the Hessian has exactly one negative eigenvalue. The second-order method is our guide, uniquely capable of leading us not just to stable minima, but to the crucial, fleeting geometries that govern the rates of all chemical reactions. The practical application of this idea requires further cleverness, as the choice of coordinate system—whether simple Cartesian or chemically-intuitive internal coordinates—can dramatically affect the stability and efficiency of the search, forcing us to switch between representations to avoid mathematical singularities.

Engineering Control and Taming Uncertainty

Let's leave the world of molecules and enter the realm of engineering and information. Imagine you are tasked with designing a control system for a satellite. At every moment, you must decide which thrusters to fire to keep it on course, minimizing fuel consumption while staying pointed at a target. This is the world of Model Predictive Control (MPC), a strategy of "predicting the future to choose the best action now."

For many physical systems, the "cost" associated with a sequence of control actions—how far you'll be from the target, how much fuel you'll use—can be described by a perfect, beautiful quadratic function, a bowl-shaped valley in the space of all possible control inputs. In this ideal scenario, the Hessian is a constant matrix, and the quadratic model is not an approximation but the exact reality. The Newton equation, HU∗=−f\mathbf{H} \mathbf{U}^{*} = -\mathbf{f}HU∗=−f, doesn't just give a good step; it gives the single perfect step to the exact optimal control sequence U∗\mathbf{U}^{*}U∗. The curvature of the cost function gives us the complete answer in one fell swoop.

This same mathematical structure appears when we try to extract information from noisy data. Consider a radar station trying to determine the direction of an incoming signal. This is a problem of Direction of Arrival (DOA) estimation. We can construct a "likelihood" function, which measures how probable a given direction is, based on the signals received by an array of antennas. The peak of this function corresponds to the most likely direction. How do we find this peak? Once again, we turn to Newton's method. By calculating the gradient and Hessian of the (logarithm of the) likelihood function, we can efficiently climb to its summit. The problem of finding an energy minimum and the problem of finding the most probable explanation are, from a mathematical perspective, two sides of the same coin.

Perhaps the most profound connection is in the field of statistics and uncertainty quantification. Finding the "best" parameters for a model is only half the battle; the other half is knowing how certain we are of those parameters. Imagine you've fit a model of a complex biochemical reaction network to experimental data. The Laplace approximation, a method rooted in second-order information, provides a brilliant answer. The best-fit parameters correspond to the peak of a posterior probability distribution. The Hessian at that peak tells us about the shape of the peak. If the curvature is sharp (large Hessian eigenvalues), the peak is narrow, meaning our parameters are tightly constrained and our uncertainty is low. If the curvature is flat (small Hessian eigenvalues), the peak is broad and smeared out, meaning the data provides little information and our uncertainty is high. The inverse of the Hessian gives us an estimate of the covariance matrix—a direct quantification of our uncertainty.

This reveals a beautiful trade-off. A full-scale simulation like Markov Chain Monte Carlo (MCMC) might take days to meticulously map out the entire probability landscape, capturing every nuance of a "sloppy" model with complex, banana-shaped ridges of uncertainty. The Laplace approximation, using just the local curvature at the peak, can give us an answer in minutes. It might miss some of the global structure, but it provides an incredibly efficient estimate of our uncertainty. The Hessian, once again, is the key that unlocks not just the solution, but our confidence in it.

A New Language for Optimization: The Power of the Cone

So far, we have viewed second-order methods as algorithms—dynamic processes for finding a point. But there is another, more modern perspective: using the geometry of second-order constraints to describe a problem. This leads us to the powerful framework of Second-Order Cone Programming (SOCP).

The fundamental building block is deceptively simple: the inequality ∥x∥2≤t\|\mathbf{x}\|_2 \le t∥x∥2​≤t, where x\mathbf{x}x is a vector and ttt is a scalar. This describes a set of points that form an "ice cream cone" in (n+1)(n+1)(n+1)-dimensional space. It turns out that a vast array of practical constraints can be expressed using this conic form. For instance, requiring a point to be within a certain distance of another, or to lie in the intersection of two spheres, can be written as a set of simple second-order cone constraints.

The real magic happens when we use this framework to reformulate problems that don't initially look like they fit. A classic problem in signal processing is denoising: given a corrupted measurement b\mathbf{b}b and a model matrix A\mathbf{A}A, find the signal x\mathbf{x}x that best explains the measurement. A natural goal is to find the x\mathbf{x}x that minimizes the residual error, ∥Ax−b∥2\|\mathbf{A}\mathbf{x} - \mathbf{b}\|_2∥Ax−b∥2​. This is a non-linear objective function. However, by using a simple but profound trick known as the epigraph formulation, we can turn it into an SOCP. We introduce a new variable ttt and rephrase the problem as: "minimize ttt subject to the constraint ∥Ax−b∥2≤t\|\mathbf{A}\mathbf{x} - \mathbf{b}\|_2 \le t∥Ax−b∥2​≤t." The objective is now linear, and the constraint is precisely the standard second-order cone form. We have transformed a difficult problem into a standard one that can be solved with breathtaking efficiency by modern optimization software.

This power finds a stunning application in the field of solid mechanics. When will a material like soil or concrete fail under load? Engineers use "yield criteria" to predict this. One of the most effective, the Drucker-Prager criterion, can be expressed as an inequality relating the shear stress and the pressure within the material. Miraculously, this physical law can be written exactly as a second-order cone constraint. This means that complex problems in geotechnical and structural engineering—like assessing the stability of a foundation under a building—can be formulated as SOCPs and solved with guaranteed reliability. This stands in stark contrast to older models like the Mohr-Coulomb criterion, whose yield surface contains "corners." These non-smooth points, where the curvature is undefined, break the elegant machinery of many optimization algorithms. The smooth, convex geometry of the second-order cone provides not just a computational convenience, but a powerful and robust language for ensuring the safety of the world we build.

From the quantum dance of electrons to the stability of the ground beneath our feet, the principle of second-order optimization proves itself to be a unifying concept of immense power and beauty. It is a testament to how a single mathematical insight—understanding curvature—can provide a clearer and more direct path to solving the most challenging problems across science and engineering.