Skip to main content
  • Submit
  • About
    • Editorial Board
    • PNAS Staff
    • FAQ
    • Accessibility Statement
    • Rights and Permissions
    • Site Map
  • Contact
  • Journal Club
  • Subscribe
    • Subscription Rates
    • Subscriptions FAQ
    • Open Access
    • Recommend PNAS to Your Librarian
  • Log in
  • My Cart

Main menu

  • Home
  • Articles
    • Current
    • Latest Articles
    • Special Features
    • Colloquia
    • Collected Articles
    • PNAS Classics
    • Archive
  • Front Matter
  • News
    • For the Press
    • Highlights from Latest Articles
    • PNAS in the News
  • Podcasts
  • Authors
    • Information for Authors
    • Editorial and Journal Policies
    • Submission Procedures
    • Fees and Licenses
  • Submit
  • About
    • Editorial Board
    • PNAS Staff
    • FAQ
    • Accessibility Statement
    • Rights and Permissions
    • Site Map
  • Contact
  • Journal Club
  • Subscribe
    • Subscription Rates
    • Subscriptions FAQ
    • Open Access
    • Recommend PNAS to Your Librarian

User menu

  • Log in
  • My Cart

Search

  • Advanced search
Home
Home

Advanced Search

  • Home
  • Articles
    • Current
    • Latest Articles
    • Special Features
    • Colloquia
    • Collected Articles
    • PNAS Classics
    • Archive
  • Front Matter
  • News
    • For the Press
    • Highlights from Latest Articles
    • PNAS in the News
  • Podcasts
  • Authors
    • Information for Authors
    • Editorial and Journal Policies
    • Submission Procedures
    • Fees and Licenses

New Research In

Physical Sciences

Featured Portals

  • Physics
  • Chemistry
  • Sustainability Science

Articles by Topic

  • Applied Mathematics
  • Applied Physical Sciences
  • Astronomy
  • Computer Sciences
  • Earth, Atmospheric, and Planetary Sciences
  • Engineering
  • Environmental Sciences
  • Mathematics
  • Statistics

Social Sciences

Featured Portals

  • Anthropology
  • Sustainability Science

Articles by Topic

  • Economic Sciences
  • Environmental Sciences
  • Political Sciences
  • Psychological and Cognitive Sciences
  • Social Sciences

Biological Sciences

Featured Portals

  • Sustainability Science

Articles by Topic

  • Agricultural Sciences
  • Anthropology
  • Applied Biological Sciences
  • Biochemistry
  • Biophysics and Computational Biology
  • Cell Biology
  • Developmental Biology
  • Ecology
  • Environmental Sciences
  • Evolution
  • Genetics
  • Immunology and Inflammation
  • Medical Sciences
  • Microbiology
  • Neuroscience
  • Pharmacology
  • Physiology
  • Plant Biology
  • Population Biology
  • Psychological and Cognitive Sciences
  • Sustainability Science
  • Systems Biology
Research Article

Discovering governing equations from data by sparse identification of nonlinear dynamical systems

Steven L. Brunton, Joshua L. Proctor, and J. Nathan Kutz
PNAS April 12, 2016 113 (15) 3932-3937; first published March 28, 2016; https://doi.org/10.1073/pnas.1517384113
Steven L. Brunton
aDepartment of Mechanical Engineering, University of Washington, Seattle, WA 98195;
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • For correspondence: sbrunton@uw.edu
Joshua L. Proctor
bInstitute for Disease Modeling, Bellevue, WA 98005;
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
J. Nathan Kutz
cDepartment of Applied Mathematics, University of Washington, Seattle, WA 98195
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  1. Edited by William Bialek, Princeton University, Princeton, NJ, and approved March 1, 2016 (received for review August 31, 2015)

  • Article
  • Figures & SI
  • Info & Metrics
  • PDF
Loading

Significance

Understanding dynamic constraints and balances in nature has facilitated rapid development of knowledge and enabled technology, including aircraft, combustion engines, satellites, and electrical power. This work develops a novel framework to discover governing equations underlying a dynamical system simply from data measurements, leveraging advances in sparsity techniques and machine learning. The resulting models are parsimonious, balancing model complexity with descriptive ability while avoiding overfitting. There are many critical data-driven problems, such as understanding cognition from neural recordings, inferring climate patterns, determining stability of financial markets, predicting and suppressing the spread of disease, and controlling turbulence for greener transportation and energy. With abundant data and elusive laws, data-driven discovery of dynamics will continue to play an important role in these efforts.

Abstract

Extracting governing equations from data is a central challenge in many diverse areas of science and engineering. Data are abundant whereas models often remain elusive, as in climate science, neuroscience, ecology, finance, and epidemiology, to name only a few examples. In this work, we combine sparsity-promoting techniques and machine learning with nonlinear dynamical systems to discover governing equations from noisy measurement data. The only assumption about the structure of the model is that there are only a few important terms that govern the dynamics, so that the equations are sparse in the space of possible functions; this assumption holds for many physical systems in an appropriate basis. In particular, we use sparse regression to determine the fewest terms in the dynamic governing equations required to accurately represent the data. This results in parsimonious models that balance accuracy with model complexity to avoid overfitting. We demonstrate the algorithm on a wide range of problems, from simple canonical systems, including linear and nonlinear oscillators and the chaotic Lorenz system, to the fluid vortex shedding behind an obstacle. The fluid example illustrates the ability of this method to discover the underlying dynamics of a system that took experts in the community nearly 30 years to resolve. We also show that this method generalizes to parameterized systems and systems that are time-varying or have external forcing.

  • dynamical systems
  • machine learning
  • sparse regression
  • system identification
  • optimization

Advances in machine learning (1) and data science (2) have promised a renaissance in the analysis and understanding of complex data, extracting patterns in vast multimodal data that are beyond the ability of humans to grasp. However, despite the rapid development of tools to understand static data based on statistical relationships, there has been slow progress in distilling physical models of dynamic processes from big data. This has limited the ability of data science models to extrapolate the dynamics beyond the attractor where they were sampled and constructed.

An analogy may be drawn with the discoveries of Kepler and Newton. Kepler, equipped with the most extensive and accurate planetary data of the era, developed a data-driven model for planetary motion, resulting in his famous elliptic orbits. However, this was an attractor-based view of the world, and it did not explain the fundamental dynamic relationships that give rise to planetary orbits, or provide a model for how these bodies react when perturbed. Newton, in contrast, discovered a dynamic relationship between momentum and energy that described the underlying processes responsible for these elliptic orbits. This dynamic model may be generalized to predict behavior in regimes where no data were collected. Newton’s model has proven remarkably robust for engineering design, making it possible to land a spacecraft on the moon, which would not have been possible using Kepler’s model alone.

A seminal breakthrough by Bongard and Lipson (3) and Schmidt and Lipson (4) has resulted in a new approach to determine the underlying structure of a nonlinear dynamical system from data. This method uses symbolic regression [i.e., genetic programming (5)] to find nonlinear differential equations, and it balances complexity of the model, measured in the number of terms, with model accuracy. The resulting model identification realizes a long-sought goal of the physics and engineering communities to discover dynamical systems from data. However, symbolic regression is expensive, does not scale well to large systems of interest, and may be prone to overfitting unless care is taken to explicitly balance model complexity with predictive power. In ref. 4, the Pareto front is used to find parsimonious models. There are other techniques that address various aspects of the dynamical system discovery problem. These include methods to discover governing equations from time-series data (6), equation-free modeling (7), empirical dynamic modeling (8, 9), modeling emergent behavior (10), and automated inference of dynamics (11⇓–13); ref. 12 provides an excellent review.

Sparse Identification of Nonlinear Dynamics (SINDy)

In this work, we reenvision the dynamical system discovery problem from the perspective of sparse regression (14⇓–16) and compressed sensing (17⇓⇓⇓⇓–22). In particular, we leverage the fact that most physical systems have only a few relevant terms that define the dynamics, making the governing equations sparse in a high-dimensional nonlinear function space. The combination of sparsity methods in dynamical systems is quite recent (23⇓⇓⇓⇓⇓⇓–30). Here, we consider dynamical systems (31) of the formddtx(t)=f(x(t)).[1]The vector x(t)∈ℝn denotes the state of a system at time t, and the function f(x(t)) represents the dynamic constraints that define the equations of motion of the system, such as Newton’s second law. Later, the dynamics will be generalized to include parameterization, time dependence, and forcing.

The key observation is that for many systems of interest, the function f consists of only a few terms, making it sparse in the space of possible functions. Recent advances in compressed sensing and sparse regression make this viewpoint of sparsity favorable, because it is now possible to determine which right-hand-side terms are nonzero without performing a combinatorially intractable brute-force search. This guarantees that the sparse solution is found with high probability using convex methods that scale to large problems favorably with Moore’s law. The resulting sparse model identification inherently balances model complexity (i.e., sparsity of the right-hand-side dynamics) with accuracy, avoiding overfitting the model to the data. Wang et al. (23) have used compressed sensing to identify nonlinear dynamics and predict catastrophes; here, we advocate using sparse regression to mitigate noise.

To determine the function f from data, we collect a time history of the state x(t) and either measure the derivative x˙(t) or approximate it numerically from x(t). The data are sampled at several times t1,t2,⋯,tm and arranged into two matrices:X=[xT(t1)xT(t2)⋮xT(tm)]=[x1(t1)x2(t1)⋯xn(t1)x1(t2)x2(t2)⋯xn(t2)⋮⋮⋱⋮x1(tm)x2(tm)⋯xn(tm)]→state↓timeX˙=[x˙T(t1)x˙T(t2)⋮x˙T(tm)]=[x˙1(t1)x˙2(t1)⋯x˙n(t1)x˙1(t2)x˙2(t2)⋯x˙n(t2)⋮⋮⋱⋮x˙1(tm)x˙2(tm)⋯x˙n(tm)].Next, we construct a library Θ(X) consisting of candidate nonlinear functions of the columns of X. For example, Θ(X) may consist of constant, polynomial, and trigonometric terms:Θ(X)=[1||X||XP2||XP3||⋯sin(X)||cos(X)||⋯].[2]Here, higher polynomials are denoted as XP2,XP3, etc., where XP2 denotes the quadratic nonlinearities in the state x:XP2=[x12(t1)x1(t1)x2(t1)⋯x22(t1)⋯xn2(t1)x12(t2)x1(t2)x2(t2)⋯x22(t2)⋯xn2(t2)⋮⋮⋱⋮⋱⋮x12(tm)x1(tm)x2(tm)⋯x22(tm)⋯xn2(tm)].Each column of Θ(X) represents a candidate function for the right-hand side of Eq. 1. There is tremendous freedom in choosing the entries in this matrix of nonlinearities. Because we believe that only a few of these nonlinearities are active in each row of f, we may set up a sparse regression problem to determine the sparse vectors of coefficients Ξ=[ξ1ξ2⋯ξn] that determine which nonlinearities are active:X˙=Θ(X)Ξ.[3]This is illustrated in Fig. 1. Each column ξk of Ξ is a sparse vector of coefficients determining which terms are active in the right-hand side for one of the row equations x˙k=fk(x) in Eq. 1. Once Ξ has been determined, a model of each row of the governing equations may be constructed as follows:x˙k=fk(x)=Θ(xT)ξk.[4]Note that Θ(xT) is a vector of symbolic functions of elements of x, as opposed to Θ(X), which is a data matrix. Thus,x˙=f(x)=ΞT(Θ(xT))T.[5]

Fig. 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Fig. 1.

Schematic of the SINDy algorithm, demonstrated on the Lorenz equations. Data are collected from the system, including a time history of the states X and derivatives X˙; the assumption of having X˙ is relaxed later. Next, a library of nonlinear functions of the states, Θ(X), is constructed. This nonlinear feature library is used to find the fewest terms needed to satisfy X˙=Θ(X)Ξ. The few entries in the vectors of Ξ, solved for by sparse regression, denote the relevant terms in the right-hand side of the dynamics. Parameter values are σ=10,β=8/3,ρ=28, (x0,y0,z0)T=(−8,7,27)T. The trajectory on the Lorenz attractor is colored by the adaptive time step required, with red indicating a smaller time step.

Each column of Eq. 3 requires a distinct optimization to find the sparse vector of coefficients ξk for the kth row equation. We may also normalize the columns of Θ(X), especially when entries of X are small, as discussed in the SI Appendix.

For examples in this paper, the matrix Θ(X) has size m×p, where p is the number of candidate functions, and m≫p because there are more data samples than functions; this is possible in a restricted basis, such as the polynomial basis in Eq. 2. In practice, it may be helpful to test many different function bases and use the sparsity and accuracy of the resulting model as a diagnostic tool to determine the correct basis to represent the dynamics in. In SI Appendix, Appendix B, two examples are explored where the sparse identification algorithm fails because the dynamics are not sparse in the chosen basis.

Realistically, often only X is available, and X˙ must be approximated numerically, as in all of the continuous-time examples below. Thus, X and X˙ are contaminated with noise so Eq. 3 does not hold exactly. Instead,X˙=Θ(X)Ξ+ηZ,[6]where Z is modeled as a matrix of independent identically distributed Gaussian entries with zero mean, and noise magnitude η. Thus, we seek a sparse solution to an overdetermined system with noise. The least absolute shrinkage and selection operator (LASSO) (14, 15) is an ℓ1-regularized regression that promotes sparsity and works well with this type of data. However, it may be computationally expensive for very large data sets. An alternative based on sequential thresholded least-squares is presented in Code 1 in the SI Appendix.

Depending on the noise, it may be necessary to filter X and X˙ before solving for Ξ. In many of the examples below, only the data X are available, and X˙ are obtained by differentiation. To counteract differentiation error, we use the total variation regularization (32) to denoise the derivative (33). This works quite well when only state data X are available, as illustrated on the Lorenz system (SI Appendix, Fig. S7). Alternatively, the data X and X˙ may be filtered, for example using the optimal hard threshold for singular values described in ref. 34. Insensitivity to noise is a critical feature of an algorithm that identifies dynamics from data (11⇓–13).

Often, the physical system of interest may be naturally represented by a partial differential equation (PDE) in a few spatial variables. If data are collected from a numerical discretization or from experimental measurements on a spatial grid, then the state dimension n may be prohibitively large. For example, in fluid dynamics, even simple 2D and 3D flows may require tens of thousands up to billions of variables to represent the discretized system. The proposed method is ill-suited for a large state dimension n, because of the factorial growth of Θ in n and because each of the n row equations in Eq. 4 requires a separate optimization. Fortunately, many high-dimensional systems of interest evolve on a low-dimensional manifold or attractor that is well-approximated using a low-rank basis Ψ (35, 36). For example, if data X are collected for a high-dimensional system as in Eq. 2, it is possible to obtain a low-rank approximation using dimensionality reduction techniques, such as the proper orthogonal decomposition (POD) (35, 37).

The proposed sparse identification of nonlinear dynamics (SINDy) method depends on the choice of measurement variables, data quality, and the sparsifying function basis. There is no single method that will solve all problems in nonlinear system identification, but this method highlights the importance of these underlying choices and can help guide the analysis. The challenges of choosing measurement variables and a sparsifying function basis are explored in SI Appendix, section 4.5 and Appendixes A and B.

Simply put, we need the right coordinates and function basis to yield sparse dynamics; the feasibility and flexibility of these requirements is discussed in Discussion and SI Appendix section 4.5 and Appendixes A and B. However, it may be difficult to know the correct variables a priori. Fortunately, time-delay coordinates may provide useful variables from a time series (9, 12, 38). The ability to reconstruct sparse attractor dynamics using time-delay coordinates is demonstrated in SI Appendix, section 4.5 using a single variable of the Lorenz system.

The choice of coordinates and the sparsifying basis are intimately related, and the best choice is not always clear. However, basic knowledge of the physics (e.g., Navier–Stokes equations have quadratic nonlinearities, and the Schrödinger equation has |x|2x terms) may provide a reasonable choice of nonlinear functions and measurement coordinates. In fact, the sparsity and accuracy of the proposed sparse identified model may provide valuable diagnostic information about the correct measurement coordinates and basis in which to represent the dynamics. Choosing the right coordinates to simplify dynamics has always been important, as exemplified by Lagrangian and Hamiltonian mechanics (39). There is still a need for experts to find and exploit symmetry in the system, and the proposed methods should be complemented by advanced algorithms in machine learning to extract useful features.

Results

We demonstrate the algorithm on canonical systems*, ranging from linear and nonlinear oscillators (SI Appendix, section 4.1), to noisy measurements of the chaotic Lorenz system, to the unsteady fluid wake behind a cylinder, extending this method to nonlinear PDEs and high-dimensional data. Finally, we show that bifurcation parameters may be included in the models, recovering the parameterized logistic map and the Hopf normal form from noisy measurements. In each example, we explore the ability to identify the dynamics from state measurements alone, without access to derivatives.

It is important to reiterate that the sparse identification method relies on a fortunate choice of coordinates and function basis that facilitate sparse representation of the dynamics. In SI Appendix, Appendix B, we explore the limitations of the method for examples where these assumptions break down: the Lorenz system transformed into nonlinear coordinates and the glycolytic oscillator model (11⇓–13).

Chaotic Lorenz System.

As a first example, consider a canonical model for chaotic dynamics, the Lorenz system (40):x˙=σ(y−x),[7a]y˙=x(ρ−z)−y,[7b]z˙=xy−βz.[7c]Although these equations give rise to rich and chaotic dynamics that evolve on an attractor, there are only a few terms in the right-hand side of the equations. Fig. 1 shows a schematic of how data are collected for this example, and how sparse dynamics are identified in a space of possible right-hand-side functions using convex ℓ1 minimization.

For this example, data are collected for the Lorenz system, and stacked into two large data matrices X and X˙, where each row of X is a snapshot of the state x in time, and each row of X˙ is a snapshot of the time derivative of the state x˙ in time. Here, the right-hand-side dynamics are identified in the space of polynomials Θ(X) in (x,y,z) up to fifth order, although other functions such as sin,cos,exp, or higher-order polynomials may be included:Θ(X)=[x(t)||y(t)||z(t)||x(t)2||x(t)y(t)||⋯z(t)5||].Each column of Θ(X) represents a candidate function for the right-hand side of Eq. 1. Because only a few of these terms are active in each row of f, we solve the sparse regression problem in Eq. 3 to determine the sparse vectors of coefficients Ξ=[ξ1ξ2⋯ξn] that determine which terms are active in the dynamics. This is illustrated schematically in Fig. 1, where sparse vectors ξk are found to represent the derivative x˙k as a linear combination of the fewest terms in Θ(X).

In the Lorenz example, the ability to capture dynamics on the attractor is more important than the ability to predict an individual trajectory, because chaos will quickly cause any small variations in initial conditions or model coefficients to diverge exponentially. As shown in Fig. 1, the sparse model accurately reproduces the attractor dynamics from chaotic trajectory measurements. The algorithm not only identifies the correct terms in the dynamics, but it accurately determines the coefficients to within .03% of the true values. We also explore the identification of the dynamics when only noisy state measurements are available (SI Appendix, Fig. S7). The correct dynamics are identified, and the attractor is preserved for surprisingly large noise values. In SI Appendix, section 4.5, we reconstruct the attractor dynamics in the Lorenz system using time-delay coordinates from a single measurement x(t).

PDE for Vortex Shedding Behind an Obstacle.

The Lorenz system is a low-dimensional model of more realistic high-dimensional PDE models for fluid convection in the atmosphere. Many systems of interest are governed by PDEs (24), such as weather and climate, epidemiology, and the power grid, to name a few. Each of these examples is characterized by big data, consisting of large spatially resolved measurements consisting of millions or billions of states and spanning orders of magnitude of scale in both space and time. However, many high-dimensional, real-world systems evolve on a low-dimensional attractor, making the effective dimension much smaller (35).

Here we generalize the SINDy method to an example in fluid dynamics that typifies many of the challenges outlined above. In the context of data from a PDE, our algorithm shares some connections to the dynamic mode decomposition, which is a linear dynamic regression (41⇓–43). Data are collected for the fluid flow past a cylinder at Reynolds number 100 using direct numerical simulations of the 2D Navier–Stokes equations (44, 45). The nonlinear dynamic relationship between the dominant coherent structures is identified from these flow-field measurements with no knowledge of the governing equations.

The flow past a cylinder is a particularly interesting example because of its rich history in fluid mechanics and dynamical systems. It has long been theorized that turbulence is the result of a series of Hopf bifurcations that occur as the flow velocity increases (46), giving rise to more rich and intricate structures in the fluid. After 15 years, the first Hopf bifurcation was discovered in a fluid system, in the transition from a steady laminar wake to laminar periodic vortex shedding at Reynolds number 47 (47, 48). This discovery led to a long-standing debate about how a Hopf bifurcation, with cubic nonlinearity, can be exhibited in a Navier–Stokes fluid with quadratic nonlinearities. After 15 more years, this was resolved using a separation of timescales and a mean-field model (49), shown in Eq. 8. It was shown that coupling between oscillatory modes and the base flow gives rise to a slow manifold (Fig. 2, Left), which results in algebraic terms that approximate cubic nonlinearities on slow timescales.

Fig. 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Fig. 2.

Example of high-dimensional dynamical system from fluid dynamics. The vortex shedding past a cylinder is a prototypical example that is used for flow control, with relevance to many applications, including drag reduction behind vehicles. The vortex shedding is the result of a Hopf bifurcation. However, because the Navier–Stokes equations have quadratic nonlinearity, it is necessary to use a mean-field model with a separation of timescales, where a fast mean-field deformation is slave to the slow vortex shedding dynamics. The parabolic slow manifold is shown (Left), with the unstable fixed point (C), mean flow (B), and vortex shedding (A). A POD basis and shift mode are used to reduce the dimension of the problem (Middle Right). The identified dynamics closely match the true trajectory in POD coordinates, and most importantly, they capture the quadratic nonlinearity and timescales associated with the mean-field model.

This example provides a compelling test case for the proposed algorithm, because the underlying form of the dynamics took nearly three decades for experts in the community to uncover. Because the state dimension is large, consisting of the fluid state at hundreds of thousands of grid points, it is first necessary to reduce the dimension of the system. The POD (35, 37), provides a low-rank basis resulting in a hierarchy of orthonormal modes that, when truncated, capture the most energy of the original system for the given rank truncation. The first two most energetic POD modes capture a significant portion of the energy, and steady-state vortex shedding is a limit cycle in these coordinates. An additional mode, called the shift mode, is included to capture the transient dynamics connecting the unstable steady state (“C” in Fig. 2) with the mean of the limit cycle (49) (“B” in Fig. 2). These modes define the x,y,z coordinates in Fig. 2.

In the coordinate system described above, the mean-field model for the cylinder dynamics is given by (49)x˙=μx−ωy+Axz,[8a]y˙=ωx+μy+Ayz,[8b]z˙=−λ(z−x2−y2).[8c]If λ is large, so that the z dynamics are fast, then the mean flow rapidly corrects to be on the (slow) manifold z=x2+y2 given by the amplitude of vortex shedding. When substituting this algebraic relationship into Eqs. 8a and 8b, we recover the Hopf normal form on the slow manifold.

With a time history of these three coordinates, the proposed algorithm correctly identifies quadratic nonlinearities and reproduces a parabolic slow manifold. Note that derivative measurements are not available, but are computed from the state variables. Interestingly, when the training data do not include trajectories that originate off of the slow manifold, the algorithm incorrectly identifies cubic nonlinearities, and fails to identify the slow manifold.

Normal Forms, Bifurcations, and Parameterized Systems.

In practice, many real-world systems depend on parameters, and dramatic changes, or bifurcations, may occur when the parameter is varied. The SINDy algorithm is readily extended to encompass these important parameterized systems, allowing for the discovery of normal forms (31, 50) associated with a bifurcation parameter μ. First, we append μ to the dynamics:x˙=f(x;μ),[9a]μ˙=0.[9b]It is then possible to identify f(x;μ) as a sparse combination of functions of x as well as the bifurcation parameter μ.

Identifying parameterized dynamics is shown in two examples: the 1D logistic map with stochastic forcing,xk+1=μxk(1−xk)+ηk,and the 2D Hopf normal form (51),x˙=μx+ωy−Ax(x2+y2)y˙=−ωx+μy−Ay(x2+y2).The logistic map is a classical model for population dynamics, and the Hopf normal form models spontaneous oscillations in chemical reactions, electrical circuits, and fluid instability.

The noisy measurements and the sparse dynamic reconstructions for both examples are shown in Fig. 3. In the logistic map example, the stochastically forced trajectory is sampled at 10 discrete parameter values, shown in red. From these measurements, the correct parameterized dynamics are identified. The parameterization is accurate enough to capture the cascade of bifurcations as μ is increased, resulting in the detailed bifurcation diagram in Fig. 3. Parameters are identified to within .1% of true values (SI Appendix, Appendix C).

Fig. 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Fig. 3.

SINDy algorithm is able to identify normal forms and capture bifurcations, as demonstrated on the logistic map (Left) and the Hopf normal form (Right). Noisy data from both systems are used to train models. For the logistic map, a handful of parameter values μ (red lines), are used for the training data, and the correct normal form and bifurcation sequence is captured (below). Noisy data for the Hopf normal form are collected at a few values of μ, and the total variation derivative (33) is used to compute time derivatives. The accurate Hopf normal form is reproduced (below). The nonlinear terms identified by the algorithm are in SI Appendix, section 4.4 and Appendix C.

In the Hopf normal-form example, noisy state measurements from eight parameter values are sampled, with data collected on the blue and red trajectories in Fig. 3 (Top Right). Noise is added to the position measurements to simulate sensor noise, and the total variation regularized derivative (33) is used. In this example, the normal form is correctly identified, resulting in accurate limit cycle amplitudes and growth rates (Bottom Right). The correct identification of a normal form depends critically on the choice of variables and the nonlinear basis functions used for Θ(x). In practice, these choices may be informed by machine learning and data mining, by partial knowledge of the physics, and by expert human intuition.

Similarly, time dependence and external forcing or feedback control u(t) may be added to the vector field:x˙=f(x,u(t),t),t˙=1.Generalizing the SINDy algorithm makes it possible to analyze systems that are externally forced or controlled. For example, the climate is both parameterized (50) and has external forcing, including carbon dioxide and solar radiation. The financial market is another important example with forcing and active feedback control.

Discussion

In summary, we have demonstrated a powerful technique to identify nonlinear dynamical systems from data without assumptions on the form of the governing equations. This builds on prior work in symbolic regression but with innovations related to sparse regression, which allow our algorithms to scale to high-dimensional systems. We demonstrate this method on a number of example systems exhibiting chaos, high-dimensional data with low-rank coherence, and parameterized dynamics. As shown in the Lorenz example, the ability to predict a specific trajectory may be less important than the ability to capture the attractor dynamics. The example from fluid dynamics highlights the remarkable ability of this method to extract dynamics in a fluid system that took three decades for experts in the community to explain. There are numerous fields where this method may be applied, where there are ample data and the absence of governing equations, including neuroscience, climate science, epidemiology, and financial markets. Finally, normal forms may be discovered by including parameters in the optimization, as shown in two examples. The identification of sparse governing equations and parameterizations marks a significant step toward the long-held goal of intelligent, unassisted identification of dynamical systems.

We have demonstrated the robustness of the sparse dynamics algorithm to measurement noise and unavailability of derivative measurements in the Lorenz system (SI Appendix, Figs. S6 and S7), logistic map (SI Appendix, section 4.4.1), and Hopf normal form (SI Appendix, section 4.4.2) examples. In each case, the sparse regression framework appears well-suited to measurement and process noise, especially when derivatives are smoothed using the total-variation regularized derivative.

A significant outstanding issue in the above approach is the correct choice of measurement coordinates and the choice of sparsifying function basis for the dynamics. As shown in SI Appendix, Appendix B, the algorithm fails to identify an accurate sparse model when the measurement coordinates and function basis are not amenable to sparse representation. In the successful examples, the coordinates and function spaces were somehow fortunate in that they enabled sparse representation. There is no simple solution to this challenge, and there must be a coordinated effort to incorporate expert knowledge, feature extraction, and other advanced methods. However, in practice, there may be some hope of obtaining the correct coordinate system and function basis without knowing the solution ahead of time, because we often know something about the physics that guide the choice of function space. The failure to identify an accurate sparse model also provides valuable diagnostic information about the coordinates and basis. If we have few measurements, these may be augmented using time-delay coordinates, as demonstrated on the Lorenz system (SI Appendix, section 4.5). When there are too many measurements, we may extract coherent structures using dimensionality reduction. We also demonstrate the use of polynomial bases to approximate Taylor series of nonlinear dynamics (SI Appendix, Appendix A). The connection between sparse optimization and dynamical systems will hopefully spur developments to automate and improve these choices.

Data science is not a panacea for all problems in science and engineering, but used in the right way, it provides a principled approach to maximally leverage the data that we have and inform what new data to collect. Big data are happening all across the sciences, where the data are inherently dynamic, and where traditional approaches are prone to overfitting. Data discovery algorithms that produce parsimonious models are both rare and desirable. Data science will only become more critical to efforts in science in engineering, such as understanding the neural basis of cognition, extracting and predicting coherent changes in the climate, stabilizing financial markets, managing the spread of disease, and controlling turbulence, where data are abundant, but physical laws remain elusive.

Acknowledgments

We are grateful for discussions with Bing Brunton and Bernd Noack, and for insightful comments from the referees. We also thank Tony Roberts and Jean-Christophe Loiseau. S.L.B. acknowledges support from the University of Washington. J.L.P. thanks Bill and Melinda Gates for their support of the Institute for Disease Modeling and their sponsorship through the Global Good Fund. J.N.K. acknowledges support from the US Air Force Office of Scientific Research (FA9550-09-0174).

Footnotes

  • ↵1To whom correspondence should be addressed. Email: sbrunton{at}uw.edu.
  • Author contributions: S.L.B., J.L.P., and J.N.K. designed research; S.L.B. performed research; S.L.B., J.L.P., and J.N.K. analyzed data; and S.L.B. wrote the paper.

  • The authors declare no conflict of interest.

  • ↵*Code is available at faculty.washington.edu/sbrunton/sparsedynamics.zip.

  • This article is a PNAS Direct Submission.

  • This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1517384113/-/DCSupplemental.

Freely available online through the PNAS open access option.

View Abstract

References

  1. ↵
    1. Jordan MI,
    2. Mitchell TM
    (2015) Machine learning: Trends, perspectives, and prospects. Science 349(6245):255–260
    .
    OpenUrlAbstract/FREE Full Text
  2. ↵
    1. Marx V
    (2013) Biology: The big challenges of big data. Nature 498(7453):255–260
    .
    OpenUrlCrossRefPubMed
  3. ↵
    1. Bongard J,
    2. Lipson H
    (2007) Automated reverse engineering of nonlinear dynamical systems. Proc Natl Acad Sci USA 104(24):9943–9948
    .
    OpenUrlAbstract/FREE Full Text
  4. ↵
    1. Schmidt M,
    2. Lipson H
    (2009) Distilling free-form natural laws from experimental data. Science 324(5923):81–85
    .
    OpenUrlAbstract/FREE Full Text
  5. ↵
    1. Koza JR
    (1992) Genetic Programming: On the Programming of Computers by Means of Natural Selection (MIT Press, Cambridge, MA), Vol 1
    .
  6. ↵
    1. Crutchfield JP,
    2. McNamara BS
    (1987) Equations of motion from a data series. Complex Syst 1(3):417–452
    .
    OpenUrl
  7. ↵
    1. Kevrekidis IG, et al.
    (2003) Equation-free, coarse-grained multiscale computation: Enabling microscopic simulators to perform system-level analysis. Commun Math Sci 1(4):715–762
    .
    OpenUrlCrossRef
  8. ↵
    1. Sugihara G, et al.
    (2012) Detecting causality in complex ecosystems. Science 338(6106):496–500
    .
    OpenUrlAbstract/FREE Full Text
  9. ↵
    1. Ye H, et al.
    (2015) Equation-free mechanistic ecosystem forecasting using empirical dynamic modeling. Proc Natl Acad Sci USA 112(13):E1569–E1576
    .
    OpenUrlAbstract/FREE Full Text
  10. ↵
    1. Roberts AJ
    (2014) Model Emergent Dynamics in Complex Systems (SIAM, Philadelphia)
    .
  11. ↵
    1. Schmidt MD, et al.
    (2011) Automated refinement and inference of analytical models for metabolic networks. Phys Biol 8(5):055011
    .
    OpenUrlCrossRefPubMed
  12. ↵
    1. Daniels BC,
    2. Nemenman I
    (2015) Automated adaptive inference of phenomenological dynamical models. Nat Commun 6:8133
    .
    OpenUrlCrossRefPubMed
  13. ↵
    1. Daniels BC,
    2. Nemenman I
    (2015) Efficient inference of parsimonious phenomenological models of cellular dynamics using S-systems and alternating regression. PLoS One 10(3):e0119821
    .
    OpenUrlCrossRefPubMed
  14. ↵
    1. Tibshirani R
    (1996) Regression shrinkage and selection via the lasso. J R Stat Soc, B 58(1):267–288
    .
    OpenUrl
  15. ↵
    1. Hastie T,
    2. Tibshirani R,
    3. Friedman J
    (2009) The Elements of Statistical Learning (Springer, New York), Vol 2
    .
  16. ↵
    1. James G,
    2. Witten D,
    3. Hastie T,
    4. Tibshirani R
    (2013) An Introduction to Statistical Learning (Springer, New York)
    .
  17. ↵
    1. Donoho DL
    (2006) Compressed sensing. IEEE Trans Inf Theory 52(4):1289–1306
    .
    OpenUrlCrossRef
  18. ↵
    1. Candès EJ,
    2. Romberg J,
    3. Tao T
    (2006) Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans Inf Theory 52(2):489–509
    .
    OpenUrlCrossRef
  19. ↵
    1. Candès EJ,
    2. Romberg J,
    3. Tao T
    (2006) Stable signal recovery from incomplete and inaccurate measurements. Commun Pure Appl Math 59(8):1207–1223
    .
    OpenUrlCrossRef
  20. ↵
    1. Candès EJ,
    2. Wakin MB
    (2008) An introduction to compressive sampling. IEEE Signal Processing Magazine 25(2):21–30
    .
    OpenUrlCrossRef
  21. ↵
    1. Baraniuk RG
    (2007) Compressive sensing. IEEE Signal Process Mag 24(4):118–120
    .
    OpenUrl
  22. ↵
    1. Tropp JA,
    2. Gilbert AC
    (2007) Signal recovery from random measurements via orthogonal matching pursuit. IEEE Trans Inf Theory 53(12):4655–4666
    .
    OpenUrlCrossRef
  23. ↵
    1. Wang WX,
    2. Yang R,
    3. Lai YC,
    4. Kovanis V,
    5. Grebogi C
    (2011) Predicting catastrophes in nonlinear dynamical systems by compressive sensing. Phys Rev Lett 106(15):154101
    .
    OpenUrlCrossRefPubMed
  24. ↵
    1. Schaeffer H,
    2. Caflisch R,
    3. Hauck CD,
    4. Osher S
    (2013) Sparse dynamics for partial differential equations. Proc Natl Acad Sci USA 110(17):6634–6639
    .
    OpenUrlAbstract/FREE Full Text
  25. ↵
    1. Ozoliņš V,
    2. Lai R,
    3. Caflisch R,
    4. Osher S
    (2013) Compressed modes for variational problems in mathematics and physics. Proc Natl Acad Sci USA 110(46):18368–18373
    .
    OpenUrlAbstract/FREE Full Text
  26. ↵
    1. Mackey A,
    2. Schaeffer H,
    3. Osher S
    (2014) On the compressive spectral method. Multiscale Model Simul 12(4):1800–1827
    .
    OpenUrlCrossRef
  27. ↵
    1. Brunton SL,
    2. Tu JH,
    3. Bright I,
    4. Kutz JN
    (2014) Compressive sensing and low-rank libraries for classification of bifurcation regimes in nonlinear dynamical systems. SIAM J Appl Dyn Syst 13(4):1716–1732
    .
    OpenUrlCrossRef
  28. ↵
    1. Proctor JL,
    2. Brunton SL,
    3. Brunton BW,
    4. Kutz JN
    (2014) Exploiting sparsity and equation-free architectures in complex systems (invited review). Eur Phys J Spec Top 223:2665–2684
    .
    OpenUrlCrossRef
  29. ↵
    1. Bai Z, et al.
    (2014) Low-dimensional approach for reconstruction of airfoil data via compressive sensing. AIAA J 53(4):920–930
    .
    OpenUrl
  30. ↵
    1. Arnaldo I,
    2. O’Reilly UM,
    3. Veeramachaneni K
    (2015) Building predictive models via feature synthesis. Proceedings of the 2015 Annual Conference on Genetic and Evolutionary Computation (ACM, New York), pp 983–990
    .
  31. ↵
    1. Holmes P,
    2. Guckenheimer J
    (1983) Nonlinear oscillations, dynamical systems, and bifurcations of vector fields. Applied Mathematical Sciences (Springer, Berlin), Vol 42
    .
  32. ↵
    1. Rudin LI,
    2. Osher S,
    3. Fatemi E
    (1992) Nonlinear total variation based noise removal algorithms. Physica D 60(1):259–268
    .
    OpenUrlCrossRef
  33. ↵
    1. Chartrand R
    (2011) Numerical differentiation of noisy, nonsmooth data. ISRN Applied Mathematics 2011(2011):164564
    .
    OpenUrl
  34. ↵
    1. Gavish M,
    2. Donoho DL
    (2014) The optimal hard threshold for singular values is 4/3. IEEE Trans Inf Theory 60(8):5040–5053
    .
    OpenUrlCrossRef
  35. ↵
    1. Holmes PJ,
    2. Lumley JL,
    3. Berkooz G,
    4. Rowley CW
    (2012) Turbulence, Coherent Structures, Dynamical Systems and Symmetry, Cambridge Monographs in Mechanics (Cambridge Univ Press, Cambridge, UK), 2nd Ed
    .
  36. ↵
    1. Majda AJ,
    2. Harlim J
    (2007) Information flow between subspaces of complex dynamical systems. Proc Natl Acad Sci USA 104(23):9558–9563
    .
    OpenUrlAbstract/FREE Full Text
  37. ↵
    1. Berkooz G,
    2. Holmes P,
    3. Lumley JL
    (1993) The proper orthogonal decomposition in the analysis of turbulent flows. Annu Rev Fluid Mech 23:539–575
    .
    OpenUrl
  38. ↵
    1. Takens F
    (1981) Detecting strange attractors in turbulence. Lect Notes Math 898:366–381
    .
    OpenUrlCrossRef
  39. ↵
    1. Marsden JE,
    2. Ratiu TS
    (1999) Introduction to Mechanics and Symmetry (Springer, New York), 2nd Ed
    .
  40. ↵
    1. Lorenz EN
    (1963) Deterministic nonperiodic flow. J Atmos Sci 20:130–141
    .
    OpenUrlCrossRef
  41. ↵
    1. Rowley CW,
    2. Mezić I,
    3. Bagheri S,
    4. Schlatter P,
    5. Henningson D
    (2009) Spectral analysis of nonlinear flows. J Fluid Mech 645:115–127
    .
    OpenUrl
  42. ↵
    1. Schmid PJ
    (2010) Dynamic mode decomposition of numerical and experimental data. J Fluid Mech 656:5–28
    .
    OpenUrlCrossRef
  43. ↵
    1. Mezic I
    (2013) Analysis of fluid flows via spectral properties of the Koopman operator. Annu Rev Fluid Mech 45:357–378
    .
    OpenUrlCrossRef
  44. ↵
    1. Taira K,
    2. Colonius T
    (2007) The immersed boundary method: A projection approach. J Comput Phys 225(2):2118–2137
    .
    OpenUrlCrossRef
  45. ↵
    1. Colonius T,
    2. Taira K
    (2008) A fast immersed boundary method using a nullspace approach and multi-domain far-field boundary conditions. Comput Methods Appl Mech Eng 197:2131–2146
    .
    OpenUrlCrossRef
  46. ↵
    1. Ruelle D,
    2. Takens F
    (1971) On the nature of turbulence. Commun Math Phys 20:167–192
    .
    OpenUrlCrossRef
  47. ↵
    1. Jackson CP
    (1987) A finite-element study of the onset of vortex shedding in flow past variously shaped bodies. J Fluid Mech 182:23–45
    .
    OpenUrlCrossRef
  48. ↵
    1. Zebib Z
    (1987) Stability of viscous flow past a circular cylinder. J Eng Math 21:155–165
    .
    OpenUrlCrossRef
  49. ↵
    1. Noack BR,
    2. Afanasiev K,
    3. Morzynski M,
    4. Tadmor G,
    5. Thiele F
    (2003) A hierarchy of low-dimensional models for the transient and post-transient cylinder wake. J Fluid Mech 497:335–363
    .
    OpenUrlCrossRef
  50. ↵
    1. Majda AJ,
    2. Franzke C,
    3. Crommelin D
    (2009) Normal forms for reduced stochastic climate models. Proc Natl Acad Sci USA 106(10):3649–3653
    .
    OpenUrlAbstract/FREE Full Text
  51. ↵
    1. Marsden JE,
    2. McCracken M
    (1976) The Hopf Bifurcation and Its Applications (Springer, New York), Vol 19
    .
PreviousNext
Back to top
Article Alerts
Email Article

Thank you for your interest in spreading the word on PNAS.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Discovering governing equations from data by sparse identification of nonlinear dynamical systems
(Your Name) has sent you a message from PNAS
(Your Name) thought you would like to see the PNAS web site.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Citation Tools
Sparse identification of nonlinear dynamics
Steven L. Brunton, Joshua L. Proctor, J. Nathan Kutz
Proceedings of the National Academy of Sciences Apr 2016, 113 (15) 3932-3937; DOI: 10.1073/pnas.1517384113

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Request Permissions
Share
Sparse identification of nonlinear dynamics
Steven L. Brunton, Joshua L. Proctor, J. Nathan Kutz
Proceedings of the National Academy of Sciences Apr 2016, 113 (15) 3932-3937; DOI: 10.1073/pnas.1517384113
del.icio.us logo Digg logo Reddit logo Twitter logo CiteULike logo Facebook logo Google logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Mendeley logo Mendeley
Proceedings of the National Academy of Sciences: 113 (15)
Table of Contents

Submit

Sign up for Article Alerts

Article Classifications

  • Physical Sciences
  • Applied Mathematics

Jump to section

  • Article
    • Abstract
    • Sparse Identification of Nonlinear Dynamics (SINDy)
    • Results
    • Discussion
    • Acknowledgments
    • Footnotes
    • References
  • Figures & SI
  • Info & Metrics
  • PDF

You May Also be Interested in

Penguin swimming
Origin and diversification of penguins
Juliana Vianna and Rauri Bowie explain the origin and diversification of penguins.
Listen
Past PodcastsSubscribe
Opinion: Cultural and linguistic diversities are crucial pillars of biodiversity
To best manage natural systems, modern societies must consider alternative views and interpretations of the natural world.
Inner Workings: Sub buoys prospects for 3D map of marine microbial communities
Implications range from elucidating metabolic pathways that help facilitate greenhouse gas release, to revealing compounds for medicine or pollution remediation.
Image credit: Mak Saito (Woods Hole Oceanographic Institution, Woods Hole, MA).
Ancient genomes reveal demographic history of France
A large genomic dataset reveals ancient demographic events that accompanied the transition to agriculture and changes in metallurgic practices in France.
Image credit: Pixabay/DavidRockDesign.
Satellite in orbit
Orbital-use fees in satellite industry
A study finds that imposing a tax on orbiting satellites could increase the value of the satellite industry from $600 billion to $3 trillion by 2040 by decreasing collision risks and space debris.
Image credit: NASA.

Similar Articles

Site Logo
Powered by HighWire
  • Submit Manuscript
  • Twitter
  • Facebook
  • RSS Feeds
  • Email Alerts

Articles

  • Current Issue
  • Latest Articles
  • Archive

PNAS Portals

  • Anthropology
  • Chemistry
  • Classics
  • Front Matter
  • Physics
  • Sustainability Science
  • Teaching Resources

Information

  • Authors
  • Editorial Board
  • Reviewers
  • Librarians
  • Press
  • Site Map
  • PNAS Updates

Feedback    Privacy/Legal

Copyright © 2020 National Academy of Sciences. Online ISSN 1091-6490