Abstract
This paper is a first attempt to marry constructive nonlinear control theory techniques with active inference. Specifically, we are interested in the relationship between differential flatness and the design of generative models for use in control settings. We place specific emphasis on the pathwise properties of differentially flat systems that inherit from their definition in terms of successive temporal derivatives and relate this to the use of generalised coordinates of motion in formulating continuous-time generative models in active inference. To illustrate the basic concepts, we appeal to the example of oculomotor control.
Keywords: differential flatness, active inference, periodic smooth random functions, pathwise formulations
1. Introduction
Active inference is one, if not the most, promising formal framework for computational neuroscience, with many applications in a number of areas (see, e.g., [1] and the references therein). The recent developments foregrounding pathwise formulations and Bayesian mechanics—developed in [2,3] among others—furnish a principled and natural setting to address many aspects of perception, planning, and control. The central idea that we will appeal to in this paper is that (natural or artificial) creatures make use of implicit generative world models both to draw inferences about the world and to control their sensed world. It is the latter that is more salient for our current purposes. In fact, one could regard the inferential (i.e., perceptual) role of generative models as merely a method of attaining tight bounds on marginal likelihoods—the quantity of interest when optimising sensory data through action.
We here consider some conditions whose fulfilment may help to construct generative models well suited for controlling one’s environment. Crucially, conditions are in direct agreement with constructive frameworks developed in nonlinear control (see, e.g., [4,5,6]). We thus try to show how this framework, and in particular the differential flatness structural property, can be leveraged in applications of active inference to control problems.
Many insightful works have already been published in this general area, both on the links and differences between active inference and classical control schemes and on deploying active inference for control (see, e.g., [7,8,9,10,11,12,13,14] just to name a few).
At first sight, differential flatness and active inference seem quite distant frameworks. The first aims to reduce a trajectory tracking error to zero, while the second minimises surprise or variational free energy; the first is inherently deterministic, and the second naturally deals with stochastic fluctuations. We shall see that the trajectory tracking error is indeed a form of surprise and that the very definition of differential flatness can be adapted to smooth random fluctuations, particularly apt for neuroscientific applications. There are two main approaches to minimise surprise or discrepancy:
-
○
The first is to envision the problem as an optimisation procedure, as in active inference (but also in optimal control or model predictive control); in this respect, the goal is to fulfil an optimisation criterion, leading to the minimisation of target discrepancies.
-
○
The second is to derive a deterministic control (action) scheme for the goal and estimate the fluctuations online to actively compensate for them.
The first route is that taken by active inference, while the second underwrites the differential flatness approach. We shall nevertheless see that the two frameworks are not as different as one may think. Indeed, the mode dynamics enforced by free energy minimisation—when those dynamics enjoy the differential flatness property—yield a functional parametrisation linking the two schemes intimately. Moreover, this parametrisation appears to be a most fruitful and promising tool to study nonlinear dynamics in neuroscience and biology.
More broadly, here, we try to foreground some potentially useful features of our framework, which draws from dynamical control system structural properties in a typical pathwise and physically preserving formulation. In particular, we shall see how the most striking feature of differential flatness—namely, differential parametrisation—can be of use within the active inference framework. More specifically, this differential parametrisation induces, among other relations, an invertible mapping from the sensation to action.
The problem we will be interested in in Section 5 is one in which we have a generative model that describes how (we believe) some system will evolve over time as a function of some control variables. Given a desired state or path for this system, our interest is in understanding how the control variables would have to be set in order to realise that desired configuration. More specifically, we are interested in the properties a generative model should possess in order to infer the control variables that realise some goal.
This paper is structured as follows. We first provide a brief introduction to the idea of a generative model for control and introduce some of the key definitions we will appeal to subsequently. Several of these definitions turn out to rely upon differentiable fluctuations, which leads us to a consideration of the interpretation of stochastic fluctuations that we will commit to (i.e., the use of random periodic functions). Having outlined the basic structure of the sorts of generative models we are interested in, we consider the way in which the principles that underwrite active inference constrain choices we might make when designing these models and equipping them with goals. We use the forms of the key objective functions from active inference (variational and expected free energy) to consider some of the generic properties of good generative models and consider whether these properties can be motivated by appealing to the notion of differential flatness. Following these theoretical considerations, we demonstrate the way these ideas work in practice through a worked example based upon oculomotor control. We conclude with a discussion of the relationship between some of these ideas and active inference, with a particular focus on the notion of generalised coordinates of motion, which inherit from similar ideas to that of differential flatness but turn out to play a very different role.
2. Preliminary Notions: Generative Model, Action, State, and Fluctuation Choice
2.1. Generative Models
As outlined above, we are interested in control problems that are formulated in terms of generative models. As an example to make this clear, which will be unpacked in much greater detail later on, consider how we might go about controlling the positions of our eyes such that we achieve a specific fixation point on a surface or track a moving point. To decide how to move our eyes, our brains might employ a model in which action variables (), like the contraction of extra-ocular muscles, might influence the dynamics of states (), such as the angle of our gaze or fixation point, and in which these states return some observable (e.g., visual) data (). We now precisely outline the notion of a generative model.
Definition 1 (Generative model).
A generative model is a set of stochastic differential equations of the following form:
(1a)
(1b) where is a vector representing the state of the system (e.g., oculomotor apparatus) to be controlled (see Definition for a precise definition), is the action vector (an external control input), the signal or instruction we might send to that system, θ is a vector of the agent’s model parameters, is the output vector, the agent’s sampling of the sensor signal, and and are the state and output fluctuations, considered, in the Stratonovich sense, as smooth random functions (see Section 2.3 below). The functions f and h are here supposed to be meromorphic in their arguments, and we suppose that belong to a functional space such that , with being the space of continuous functions in time.
It is often convenient, in applications of active inference, to write a generative model in terms of generalised coordinates of motion. These coordinates are the coefficients of a Taylor series expansion of variables around the current time. In doing so, we replace a single nonlinear stochastic differential equation (for x) with a set of locally linearised equations for each order of motion: , where the superscript in brackets indicates the order of motion (i.e., the term in the Taylor series to which the coefficient belongs or the order of the temporal derivative one must take to get to these variables from our initial variable).
The motivation for this formulation is twofold. First, it allows one to express noise processes of varying smoothness in terms of the covariance between the fluctuations assigned to each order of motion. Second, through applications of the chain rule, it allows one to determine the gradients of generalised coordinates of y with respect to u via the gradients of the generalised coordinates of x. This will be important later, when we look to the way in which descent of variational free energy gradients—by changing u—can be understood as analogous to the enaction of spinal reflex arcs that bring proprioceptive data in line with anticipated setpoints.
The need to find mappings from sensation to action is one of the key ideas common to active inference—using generalised coordinates of motion—and differential flatness as applied to control. It is interesting that both approach this problem by appealing to successive derivatives of the variables in a generative model.
Definition 2 (Mean generative model).
Given a generative model as above, the mean, or deterministic, generative model is given by the following set of differential equations:
(2a)
(2b) It thus corresponds to the deterministic part of the corresponding stochastic differential equation system.
2.2. Action, Output, and State
We here give precise definitions for the notions of action, or control input, output, state, and realisation, which will be useful in Section 6.2. These were given in a differential algebraic setting in [4]; see also [15] for an elementary treatment, analogous to the one below (although stated in a deterministic setting). Let us note that the subsequent definitions, especially the differential flatness definitions, are made in the presence of fluctuations. This is not to say that we lose the deterministic character of this notion, since these perturbations may well be deterministic functions. They may also be stochastic, if they are sufficiently smooth, i.e., differentiable up a certain order.
Note that the action , or the control input, are functions enabling us to act on the system in order to fulfil a specified goal. The dynamics equations thus form an undetermined system of differential equations, since the control functions are not a priori determined. Once the control variables are fixed (i.e., substituted with known functions of time), the system (1) becomes determined (i.e., can be solved or integrated).
An output is generated by the model. These outputs or observations may represent signals coming from the senses, in case the agent is a living being, or from the sensors in the case of an artificial system. More precisely, we have the following definition.
Definition 3 (action (control input) and output).
Consider a model with variables , fluctuations , and equations
(3) A control input, or action, of the model is an m-tuple (with ) of variables with the following properties.
Endogenous character.
The components of can be (locally and generically) expressed as a function of and its derivatives:
(4) Differential independence.
There does not exist any nontrivial differential relation of the form
(5) Any system variable is influenced by the input.
Every , satisfies
(6) An output (with ) is a p-tuple, where the s are functions of the model’s variables:
(7)
Remark 1 (Local genericity).
Generally, in (4), we have to deal with implicit functions:
(8) And, locally, in the neighbourhood of generic regular points (i.e., points where the Jacobian of Ψ with respect to is regular), we can solve (8) for using the implicit function theorem. This will be termed locally and generically here and in the forthcoming definitions.
The state variables represent the instantaneous memory of the system: once the control (action) variables have been determined, the knowledge of the state variables (at time t) enables the prediction of the future state (at time ). A complementary formulation is the following: the state of a dynamical system is a set of physical quantities, the specification of which (in the absence of external excitation) completely determines the evolution of the system. More precisely, we have the following definition.
Definition 4 (state).
Consider a model with variables , input , fluctuations , and equations
(9) A state of the model is an n-tuple (with ) of variables with the following properties.
Endogenous character.
The components of can be (locally and generically) expressed as a function of and its derivatives:
(10) Independence with respect to the input.
There does not exist any nontrivial differential relation of the form
(11) Representation property.
Every variable that is a function of and of its time derivatives can be expressed through , and its derivatives. In other words, there exists (locally and generically) a representation of the form
(12)
By virtue of the above representation property, for all , the time derivatives are expressed through , and its time derivatives. This can be otherwise stated as follows.
Definition 5 (State representation).
Consider a model with variables , input , fluctuations , equations
(13) and state . Then, there exists a so-called state representation of the form
(14) When the right-hand side of (14) does not depend on the derivatives of the input (i.e., when for all ), the state representation is called classical.
A realisation of a model consists of a state and a state representation for this model, as the following definition states.
Definition 6 (Realisation).
Consider a model with variables , fluctuations , and equations
(15) where comprises an input and output : . A realisation of this model consists in the existence of a state and a state representation of the form (14) for the model.
2.3. Fluctuation Choice
An interesting point about the definitions above is that they depend upon there being differentiable (i.e., smooth or analytic) fluctuations. This means we need to think carefully about what we mean by fluctuations—an important topic generally in the study of stochastic dynamical systems. Several choices can be made for the fluctuations , appearing in the above generative models. These include the following:
-
○
Stochastic processes in the Itô sense.
-
○
Stochastic processes in the Stratonovich sense.
- ○
-
○
Stochastic processes with Hölder continuous sample paths, yielding radom ODEs (RODEs) (see, e.g., [19]);
-
○
Rough paths (see, e.g., [20]);
-
○
Random Fourier series (RFS) (see, e.g., [21,22] for metric and convergence properties, with extensions on Riemannian manifolds [23] and locally compact groups [24]; see also [25] for an engineer’s view).
Let us choose the latter because they may furnish a convenient form of fluctuations, and the solution of random ODEs can be shown to converge to solutions of a Stratonovich stochastic differential equation (see, e.g., [25], Theorem 5.1). We may thus consider so-called periodic smooth random functions:
| (16) |
where , each and is an independent sample from (where denotes the real normal distribution of mean and variance V), and is the floor function. This type of function is L-periodic, entire, and -band-limited.
Then, multiplying (16) by and taking the variance to be instead of , we obtain the notion of a big periodic smooth random function. Note that since , we have . Thus, in the big normalisation, the random coefficients of the sum have variances essentially independent of as . Then, as , indefinite integrals of big smooth random functions converge with probability 1 to standard Brownian paths (see, e.g., Theorem 4.3 of [25] or Theorem 2, p. 236 of [21]). Regarding smooth random functions, the reader may also consult the very interesting works of R.J. Adler and J.E. Taylor, [26,27,28], with insightful chapters on geometry and smooth random manifolds, among others. Interestingly, this technology underwrites generative models in classical brain imaging analysis, namely, statistical parametric mapping, based upon random field theory and topological inference [29,30]. The previous convergence can also be related to the Wong–Zakai theorems (see, e.g., [31]) used in the insightful paper [32], Subsection 1.3.2.
Smooth random functions may not be an apt choice at atomic scales, where a particle’s movements are highly erratic. However, they become particularly appropriate at the cell and mesoscopic scales and most probably appropriate at macroscopic scales at which many fluctuations are generated by dynamical systems that evolve over a timescale faster than that considered for a given control problem (see [1] and the seminal observations of Stratonovich: “a certain care must be taken in replacing an actual process by Markov process, since Markov processes have many special features, and, in particular, differ from the processes encountered in radio engineering by their lack of smoothness… Any random process actually encountered in radio engineering is analytic, and all its derivatives are finite with probability one” ([33], pp. 122–124)).
Remark 2 (Wavelet random series).
One may be tempted, in the spirit of the above, to consider wavelet random series, since wavelet expansions are better behaved than their Fourier counterparts. However, recent works by C. Esser, S. Jaffard, and B. Vede [34] (see also [35]) suggest that caution is in order; in contrast to Fourier series, the randomisation of almost every continuous function gives an almost surely nowhere locally bounded function.
We sill refer to the notion of a tube around a nominal function, i.e., for , where f is a deterministic function and a periodic smooth random function. We shall be concerned with estimating the probability of leaving a tube of width (for ). For instance, one has the following estimate (see [36]):
with , . The reader may consult [36] for generalisations of this bound inequality. The notable paper by [37] also contains results for the volume of tubes using Gaussian–Minkowski functionals.
Example 1 (Smooth random functions).
The following two plots are examples of smooth random functions with varying wavelength λ. More precisely, the left plot is a smooth random function (a sum of sines and cosines of the form (16)) with parameters , ; recall that the sum has terms, with . Thus, when , , we obtain terms in the sum. The right plot is an analogous sum with terms (See Figure 1).
Figure 1.
Examples of smooth random functions with parameters , (left plot) and , (right plot).
In order to offer a more concrete intuition as to the notions of state, control input, and output, let us consider a simple, although generic, example.
Example 2 (Simple generic example; model).
Consider the following simple example with scalar action (control) and sensor output:
(17a)
(17b)
(17c) where f is a smooth function: . The fluctuations are taken as smooth random Gaussian functions:
(18a)
(18b)
(18c)
(18d) with . Here (2) are the equations of a generative model, in the sense of Definition 1, is the state of the generative model, in the sense of Definition 4, u is the input, and y is an output, in the sense of Definition 3.
This type of model includes all models predicated upon Newton’s law of motion, such as
(19) where M is a mass, F is a model for internal forces depending on the position y and velocity , and v is an external force.
A more general case would be the following:
(20a)
(20b)
(20c)
(20d) with s being smooth functions of their arguments, invertible with respect to the final function. All the computations made for the simple example (17) can be readily extended to the above.
3. Free Energy, Flatness, and Conceptual Similarities
3.1. Free and Expected Free Energy
One can understand both active inference- and differential flatness-informed control schemes as identifying mappings from sensation to action from models that detail the influence of action on sensation. Flatness rests upon there being an invertible mapping between action and sensation such that a desired sensory trajectory uniquely determines the actions that generate it. Active inference involves the selection of actions that bring sensations in line with the mode of a marginal density of sensory data implied by a generative model. This is mediated by reflexive actions determined by sensory data. The marginal density that identifies desired sensory trajectories is often specified in terms of an expected free energy—whose role is to determine prior plausibilities of alternative action sequences based upon their capacity to minimise the Kullback–Lieber divergence (also known as risk) between desired and anticipated sensory trajectories.
Consistent with control-theoretic formulations of the sort outlined above, active inference can be formulated as optimising a functional of a model that relates controllable variables to some observable outcomes. Specifically, it depends upon optimisation (minimisation) of a variational free energy that acts as an upper bound on the surprise or negative log marginal likelihood of those observations. Variational free energy can be formulated in several ways to quantify the performance of a system engaging in active inference in terms of energies (i.e., surprise) and divergences (i.e., relative entropies):
| (21a) |
| (21b) |
| (21c) |
The above presents variational free energy as a functional of two probability distributions (for discrete states) or densities (for continuous states). The density labelled p is that associated with our generative model, while q represents a density variously referred to as a recognition density, an approximate posterior density, or a variational density. For the purposes of this paper, the variational density is assumed to have already been optimised such that . Each formulation of the free energy depends upon different factorisations of the generative model. When expressed as a joint density, the minimisation of free energy can be seen as a constrained maximum entropy problem. On factorising into conditionals and marginal likelihoods, the free energy is seen to be an upper bound on surprise —the negative log marginal likelihood, Bayesian model evidence, or improbability of an observation under a given generative model. Finally, factorising a generative model into priors and likelihoods gives us a balance between complexity—how far we must move from prior beliefs to explain observations — and the accuracy with which we can account for sensory inputs.
Further to this, one can formulate expected free energies that highlight other forms of discrepancy that are especially relevant for the optimisation of control. When formulated explicitly in a pathwise setting, we have
| (22a) |
| (22b) |
The expected free energy is typically used for planning, where we might integrate this quantity along future paths and assign paths of control states higher probabilities for lower expected free energies. Of particular relevance for the discussion that follows is the idea that, by placing priors over the path of observations we wish to obtain (here indicated by conditioning upon a goal ), optimisation of expected free energy involves determining the set of control paths that would realise these outcomes.
We can now identify at least three different aspects of optimality, namely surprise, inadequacy, and discrepancy:
The discrepancy in observation (inaccuracy) in gathering information from the world. This discrepancy is implicit in the (negative) accuracy term of the variational free energy, which under Gaussian assumptions will reduce to the square of a prediction error: , where is the predicted observation under beliefs about . This is the (sensory) prediction error that features in predictive coding or Bayesian filtering, linear quadratic control, and model predictive control.
The discrepancy in action in the way the agent acts in the world. This discrepancy is between the goal , a predefined trajectory to be followed, and the actual position : . This is implicit in the risk term of the expected free energy, which quantifies the divergence between the distribution anticipated under a set of control states and the distribution anticipated given the goal.
The discrepancy in modelling in the way the agent represents the world internally. This discrepancy would normally be quantified using the marginal likelihood of sensory observations under the model and is reflected in the surprise term of the variational free energy. When free energy is minimal with respect to q, it becomes a tight bound on this discrepancy. For this reason, variational free energy is often used as a tractable method of approximating Bayes factors to compare alternative models in statistical inference.
The free energies above ( and ) can be seen as so-called global Lyapunov functions (see, e.g., [38]) that capture these aspects seen in optimal control (see, e.g., [39,40]). In what follows, we shall mostly be interested in , and specifically the risk term, leaving the other aspects to future works. More precisely, we are interested in whether the notion of differential flatness coheres with the selection of generative models that optimise and . The derivation of trajectory tracking action laws on differentially flat models will be seen to minimise the above risk in (see Section 4.6).
Table 1 summarises the notations used so far, where bold symbols are vectors.
Table 1.
Notation table.
| Variable | Meaning |
|---|---|
| Hidden state vector | |
| Output vector (e.g., sensor data) | |
| Action vector | |
| Fluctuations in hidden state dynamics | |
| Fluctuations in sensor data | |
| Variational free energy | |
| Expected free energy |
3.2. Differential Flatness
3.2.1. Controllability
A ubiquitous notion—when one wishes to steer a system—is the global controllability, as stated in the following definition (see, e.g., [41]).
Definition 7.
A system such as (1), but idealised such that we assume the fluctuations are infinitely precise,
(23a)
(23b) is said to be globally controllable if, for any time instants and and initial and final states and , there exists an action (a control law) in a space of admissible controls steering the system from to .
The reader may note that this definition is purely descriptive: it does not contain any constructive procedure for steering the system; it is not possible to infer, from reading the definition alone, the form of the control law to be applied to go from to . This is a definition of the existence-of-a-solution kind, not of a solution construction to the given problem.
The reader may note that this definition is purely descriptive, in that it does not contain any constructive procedure for steering the system. Moreover it is pointwise in spirit, rather than pathwise. Indeed, this definition does not say anything about the path that links the initial to the final state. This path can be implausible and still fulfil the controllability requirements. We shall see in the next subsection a stronger and, in our sense, more useful property, namely, differential flatness.
3.2.2. Motivation Through Observation and Action
The premise of active inference is that an agent seeks to minimise the surprise or divergence between its beliefs or expectations about the surrounding environment and the actual state it experiences. This minimisation can, in principle, be enforced through effective information gathering and action or be engrained in the very structure of the agent’s model (such as being refined through learning and evolution).
We will examine a case where the requisite efficiency is encoded in the agent’s model structure itself. More precisely the imperatives for perception and action are directly fulfilled through the following:
-
(Odsf)
Observation discrepancy structural fulfilment. The state can be recovered through what the agent is able to know directly (i.e., without any inference, reflection, or computation), that is , , and their time derivatives. This amounts to the system being constructively observable.
-
(Adsf)
Action discrepancy structural fulfilment. The link between the goal and the action —required to reach that goal—is direct, in that the action is given as a function of the goal and its time derivatives. This amounts to the system being left-invertible.
A system such as (23) is said to be left-invertible with respect to if the action is a function of and its derivatives (see Property 5 of [42]). Thus, for a dynamical system to fulfil both (Odsf) and (Adsf), one needs to have a function such that both the state and the action can be expressed in terms of and its time derivatives. This corresponds to differential flatness [5,43,44], a property shared by a great number of practical dynamical systems (see, e.g., [45] and the references therein).
3.2.3. Motivation Through Direct and Inverse Views
The fundamental property of flat systems is that all their solutions can be functionally parameterised by a finite number of functions and their time derivatives. Although flatness is a relatively recent notion, introduced in control theory in the 1990s [5,43,44], it actually has a long history: a similar notion— of systems of undetermined differential equations integrable without integration—dates back to Hilbert [46] and Cartan [47]. Indeed, the control system can be seen as an underdetermined differential system consisting of n equations , for , and variables (n states and m action variables). The difference between the number of variables and the number of equations gives the degrees of freedom number of the system. It follows that m functions can be chosen freely. In the context of control systems, one usually chooses freely the input and then integrates in order to compute the state . But is this the only way to achieve this?
In order to answer this question, consider the following simple single-input system:
| (24a) |
| (24b) |
We can freely choose , then integrate it once to compute , and then integrate it a second time to compute . Let us now choose freely , differentiate it once to get , and then differentiate to obtain . It follows that the system with Equation (24) admits two functional parameterisations: one via the input, for which we have to integrate twice (which is, in the general case of , difficult and, sometimes, impossible analytically), and one via the state , for which we have to differentiate twice (which is always possible in a straightforward way). Hence, there are underdetermined differential systems (also known as control systems) that are solvable without integration, namely, differentially flat systems.
3.2.4. Formal Definition
Let us consider the central definition of this subsection, stated for systems with specified parameters, for ease of reading (see, e.g., [15,42,45]; see also the Supplementary Material, File AIandFP-FlatnessAndSRR-HMTPKF-2026-SupplMaterial-v1.pdf, Section D and [48] for a Python library, version 0.10.2).
Definition 8 (Differential flatness).
Consider a model with variables , fluctuations , and equations
(25) The model is called differentially flat if there exists an m-tuple of variables, , named flat outputs, with the following three properties.
Endogenous character. We have (locally and generically)
(26) In other words, the components of the flat output are combinations of the system’s variables.
Functional parameterisation.
The system’s variables can be (locally and generically) expressed through the flat output and a finite number of its derivatives:with integers, such that the system’s equations
(27) are identically verified.
Differential independence. The components of the flat output are differentially independent; i.e., any differential relationis necessarily trivial: .
Remark 3 (Differential flatness–state-space form case).
In case the model has the following form:
(28) with state , action , and fluctuation , the functional parametrisation (27) becomes
(29a)
(29b)
Remark 4.
Note that the conventional definition of differential flatness is made for determinsitic (mean) models:
(30) with state and action , in which case the functional parametrisation (27) becomes
(31a)
(31b) The fluctuations considered above in (29a) and (29b) are not necessarily supposed to be known functions of time, but rather are seen as generic, sufficiently differentiable functions. The relations (29) are valid whatever the fluctuations ζ may be at this structural level. See Section 3.2.5 below for a discussion on this functional parametrisation.
Remark 5 (Local genericity).
As in Remark 1, the relations of endogenous character in (26) and of functional parameterisation in (27), naturally yield implicit functional relations:
(32a)
(32b) And, locally, in the neighbourhood of generic regular points (i.e., points where the Jacobian of H with respect to ω and with respect to is regular), we can solve (32a) for ω and (32b) for using the implicit function theorem.
Let us note that the differential flatness definition is made in the presence of fluctuations. The latter may be deterministic or smooth stochastic functions (i.e., differentiable up a certain order). We will consider relations furnishing, in particular, the action as a functional of the sensory and fluctuation paths. This enables one to study precisely and quantitatively the influence that each perturbation function may exert on the action. Moreover the relation yielding the action—respectively, the state—as a function of sensor and fluctuation paths is generic, in the sense that it is valid for any (sufficiently smooth) sensory and fluctuation paths. In this sense, the notion of differential flatness is agnostic with respect to the stochasticity of the generative model. The latter may be deterministic (as is the case for mean generative models), subject to deterministic, but unkown, perturbations or subject to stochastic (and sufficiently smooth) fluctuations. The previous definition readily affords the following characterisation:
Proposition 1.
A differentially flat system in state-space form is a system like (28), i.e., , which is observable and left invertible with respect to ω.
To summarise what has been said so far, in order to express what we think we know about the system (our beliefs) and what we wish for the system (our expectations), we require a model structure that expresses first the agent system’s constructive observability (what we know or believe) and second the agent system’s left invertibility (what we expect and where the action is expressed in terms of the flat output goal, ). The latter amounts to differential flatness, which is no more than the generator character of the flat output, i.e., the functional parameterisation and the differential polynomial independence of the flat output components. The last item enables one to choose the various components of independently of each other. Both properties are reminiscent of the basis notion in vector spaces (i.e., the minimally generating and maximally independent characters); indeed, in a differential algebraic setting, the notion of flat output corresponds to a differential transcendence basis (see [5]).
An interesting point of contact with active inference is that optimisation of expected free energy functionals implies a high degree of mutual information between different components of a model (specifically, between states and observations). This implies precise mappings from actions, via states, to observations. Crucially, that same mutual information means that we would expect observations to be highly informative about states and, possibly, actions. The potential to recover subsets of variables from others in such models is heuristically compatible with the differential flatness concept. Furthermore, the need for differential independence has an interesting link with the interpretation of variational free energy as an objective function for constrained maximum entropy inference—where in the absence of the ‘energy’ constraints, the best configuration is that with maximum entropy without mutual constraints between the components of a system.
Remark 6 (Tracking an arbitrary output).
Suppose that the agent’s model is differentially flat. The agent’s goal may coincide with the flat output goal: . If this is not the case, consider the expression of the agent’s output as a function of the flat output ω and its derivatives:
(33) and view this expression as a differential equation in ω:
(34) Substituting by the goal yields the flat output goal as a solution of
(35)
3.2.5. Functional Parameterisation
The functional parameterisation property is an essential, if not by far the most essential feature, of differential flatness. Indeed, the original model
| (36a) |
| (36b) |
is totally equivalent to its functional parametric form:
| (37a) |
| (37b) |
| (37c) |
And, crucially, the functional parametric form (37) is quasi-static, whereas the original model (36) has a dynamical form. Hence, when one wishes to deal with (36), one is naturally tempted to seek the solution of the ODE system (36a), and most often, an analytical solution cannot be found. But when dealing with the functional parametric form (37), all system variables, i.e., and here, are parametrised by the function , since once we know the function , then we readily know the functions and through (37) (assuming, in a structural step such as this one, that the fluctuations are known or measured).
Note that this parametrisation is a pathwise one and that it carries over the whole fluctuation; it can thus be envisioned either as a tube around the mean (see, e.g., [27,28,36]) or as a function sheaf (see, e.g., [49,50]; see also [51] for an elementary introduction and [52] for more complete while still accessible references).
Thus, the action, the control , and the state are envisioned as functionals of the flat output and of the fluctuations, i.e., as functions over spaces of functions:
| (38a) |
| (38b) |
And it is the study of these functionals that may be of great interest to the active inference community.
3.3. Conceptual Similarities
We can now see—with a simple concrete example—that both differential flatness and free energy minimisation enforce an inverse mapping from the sensed output to the action, namely, the control input.
Example 3 (Simple generic example; similarities).
Recall the following simple example with scalar action (control) and sensor output:
(39a)
(39b)
(39c) This model is differentially flat with flat output y. Indeed, the functional parameterisation is obtained as follows. Equation (39c) yields
(40) Equation (39a) yields
(41) Then, using Equation (39b),
(42) Thus, the functional parametrisation associated with the differential flatness property is
(43a)
(43b)
(43c) We see in the last equation the inverse mapping from the sensory output to the control input . Note that this relation embeds all the influences of fluctuations on the action, since it is a functional in y , , and .
In the case of a deterministic system,
(44a)
(44b)
(44c) The parametrisation becomes
(45a)
(45b)
(45c) Note that (44) is dynamically equivalent to (45).
Then, minimising the free energy enforces the dynamical mode (assuming it is possible to set all terms of this free energy to zero, at least approximately), i.e., the dynamical model obtained without fluctuations:
(46a)
(46b)
(46c) where , , is the mode of ∗ under the recognition density. By virtue of the previous equivalence, (46) is dynamically equivalent to
(47a)
(47b)
(47c) We now see that the minimisation of free energy enforces an (at least approximate) inverse mapping from the sensed output to the action. There is a further relationship that we will return to later, predicated upon generalised coordinates of motion, which we have not addressed here and which relates to the final equation above. As the action influences the rate of change of the states, which then influence the sensory data, the gradient of current sensory data with respect to the action is zero. This can be seen explicitly by noting that there is no term in the free energy in which sensory data and the mode of the action jointly appear. However, as highlighted here, there can be a non-zero gradient associated with temporal derivatives of the sensory data (i.e., higher orders of generalised coordinates of motion). This is essential for the reflexive formulation of action under active inference.
4. References, Flatness-Based Trajectory Tracking, and Perceptual and Active Inferences
4.1. Equivalence to Linearity
4.1.1. Differential Flatness Characterisation
The class of differentially flat systems is—despite the fact that it occurs quite frequently in practice—the simplest nonlinear class with respect to the feedback equivalence classes. Indeed, we have the following.
Proposition 2.
A system is flat if, and only if, it is linearisable by endogenous feedback and a change of coordinates.
A dynamic feedback is called endogenous if it does not include any external dynamics. More precisely, the following holds.
Definition 9.
Consider the dynamics . The feedback
(48a)
(48b) (where is the new input) is called a dynamic endogenous feedback if the original dynamics is equivalent to the transformed one
(49a)
(49b) Two systems are called equivalent if there exists an invertible transformation that exchanges their trajectories.
Crucially the preceding linearisation is not local but global, and the category of flat systems is not very far from the one of linear systems, since they are equivalent through endogenous feedback and coordinate change.
4.1.2. Dynamical Extension Algorithm
This procedure enables one to determine if an m-tuple is a flat output or not and to obtain a linearizing feedback.
Phase I—Weak Brunovský Index Gathering
-
(1)Differentiate until a combination of control inputs appears; denote by the number of successive differentiations:
with and . -
(2)Differentiate until a combination of control inputs (independent of the previous ones) appears; denote by the number of successive differentiations:
with and .⋮
-
(m)Differentiate until a combination of control inputs (independent of the previous ones) appears; denote by the number of successive differentiations:
with and .
Phase II—Flatness Character Determination
Then, if (n being the state dimension), the system admits as a flat output. If not, is not a flat output.
Phase III—Linearizing Feedback
In case is a flat output, the linearizing feedback is given by
| (50a) |
| (50b) |
| (50c) |
where are the new action (new control input) variables. Since in Phase I, the s () are functionally independent, the latter Equations (50) is invertible in :
| (51a) |
| (51b) |
| (51c) |
with .
The original dynamics is then transformed, via the linearising endogenous feedback (50), to a linear dynamics of the form
| (52a) |
| (52b) |
with the new input .
Thus, the agent’s model (36) is equivalent, through the feedback (50), to the so-called flat output dynamics:
| (53a) |
| (53b) |
Thus, the original model (36) with n equations has been reduced, exactly, i.e., without any approximation, to the flat output dynamics (53) with m equations, where m is, in most practical cases, significantly smaller than n.
4.2. Differential Flatness and Controllability
A natural question a reader may ask is as follows: when is the differential flatness property verifiable? In other words, what are the checkable conditions ensuring flatness for a given system? The answer, for general nonlinear systems, is still an open problem. There are some conditions for restricted system classes or feedback equivalence classes. For instance, conditions are known for single-input systems and for static state feedback equivalence (see the Supplementary Material, File AIandFP-FlatnessAndSRR-HMTPKF-2026-SupplMaterial-v1.pdf, Section C).
There are some simple classes that are trivially flat. In the case of single-input systems, the class of systems that are, up to a feedback equivalence, in a cascade form like the following,
| (54a) |
| (54b) |
| (54c) |
with the functions being invertible in their last argument, are differentially flat, with as a flat output. The systems that are, up to a feedback equivalence, in a form with a finite number m of blocks like (54) are also differentially flat. We shall nevertheless see that differential flatness is a very strong property, as the following proposition suggests.
Proposition 3.
A differentially flat system is globally controllable.
Proof.
The global controllability property (see Definition 7) amounts to the following: for any initial and final states and of , and any initial and final times and , there exists a control law , in the so-called space of admissible controls , driving the model from the initial state to the final state . Consider now a flat system with equations
with flat output . We thus have, by virtue of the flatness property, the following functional parametrisation:
Consider two arbitrary states and in , and . Let be a polynomial such that
The control steering the system from to is
Thus the system is globally controllable. □
Example 4 (Oculomotor dynamics and flatness).
As a simple example, consider a model of oculomotor control, a problem that has been addressed extensively through active inference: see, e.g., [53,54,55,56,57]. Here, we will consider a simplified kinematics and dynamics of eye movement. Consider a human agent whose eyes are located at a fixed distance d from a vertical visual scene. Let us concentrate of the movement of a single eye and denote the pupil’s center by E. The eye tracks a point P with coordinates in the visual scene’s fixed reference frame. The agent’s head is supposed to be fixed. Let , with coordinates , be the orthogonal projection of E on the visual scene plane. We shall first describe the generative model and then demonstrate that this is differentially flat. We then consider how it might be linearised as outlined above.
According to Listing’s law [58], two angles characterise the movement of the pupil: ψ, the orientation, or yaw angle, and ϕ, the elevation, or pitch angle. Let us note that this model is extremely similar to the one of a two-degree-of-freedom gimbal system (see, e.g., [59]). We then have
(55a)
(55b) Let us set
The differentiation of the expressions of X and X in (55) with respect to time yields (See Figure 2)
(56a)
(56b) Then, using the expressions of X and Y in (55), we get expressions of and as functions of X and Y:
Thus, the relations expressing and in (56) become
(57a)
(57b) The relationship between and the torques exerted by the eye muscles can be written as follows:
(58a)
(58b) with being the rotational inertia of the eye, supposed to be spherical (see the Supplementary Material, Section A, File AIandFP-FlatnessAndSRR-HMTPKF-2026-SupplMaterial-v1.pdf, for the reason why the inertia is doubled in the yaw equation above). Let us set
The full generative model will then be written as follows:
(59a)
(59b)
(59c)
(59d) where , , are fluctuations, which may be due to various alterations in human vision and the eye inertia , with and being the eye’s mass and radius.
This model is differentially flat, with as a flat output. Indeed, one has the following:
From the first two lines of (59), we get the expressions of and :
(60a)
(60b) From the final two lines of (59), we get the expressions of the action, the control inputs and :
(61a)
(61b) Note that the previous parametrisation is a pathwise one and that it carries over all of the fluctuations. Let us consider the reference trajectories and for x and y, depicted in Figure 3. More precisely, the plotted trajectory has the following form:
(62) where the simulation is performed on , is the stiffness of the trajectory, and is the time where has reached half of the maximum. The minimum is ; the maximum is .
The open-loop components of the states and , corresponding to Equation (60), are shown in Figure 4, together with the influence of the fluctuations , and their first derivatives. These fluctuations were chosen to be
(63) with ζ being of the form (16), , , and . Recall the form (16):
(64) where , each and is an independent sample from (where denotes the real normal distribution of mean and variance V), and is the floor function. Since and , we have .
The corresponding fluctuations and are depicted in Figure 5.
The open-loop action components and , corresponding to Equation (61), are described in Figure 6. Although the fluctuations in Figure 5 are extremely small, their influence is already noticeable, although quite modest, in the plot.
If one considers , which forces and to be still very small, the corresponding plots of and change radically, as depicted in Figure 7.
Consider again Example 4 and its generative model:
(65a)
(65b)
(65c)
(65d) To exactly linearise the model (65), one first differentiates (65c) and (65d), so that the action (control input) , can appear:
(66a)
(66b) or, in matrix form,
(67)
(68) To exactly linearise the dynamics (65), we shall equate to , where , are new action functions. Hence, we obtain
(69) Then, the nonlinear dynamics (65) is transformed into
Figure 2.
This graphic is intended to illustrate the set-up of our oculomotion example. On the left, we see the set-up from above, with an eye looking towards a screen at distance d. The pair of equations in the lower left panel give the (linear) flows for the yaw and pitch of the eyeball. The graphic on the right shows the screen with the fixation point P relative to the orthogonal projection . The equations in the lower right panel show the corresponding (stochastic) motion of the fixation point. The expression in terms of fixation points provides a useful nonlinear system in which we can unpack the differential flatness concept. It also provides a clear example of a control problem, as we might expect the target of an eye movement to be the fixation location that helps us resolve uncertainty about something in visual space, as opposed to a desired oculomotor angle.
Figure 3.
Simple reference trajectory examples .
Figure 4.
Orientation and elevation speeds and for very small fluctuations .
Figure 5.
Fluctuations from Equation (63), with an extremely small scale (, , ). Red line is the mean; grey lines are the signals with fluctuations.
Figure 6.
Orientation and elevation torques for extremely small fluctuations (, , ). Red lines are the means, i.e., without fluctuations; grey lines are the signals with fluctuations.
Figure 7.
Orientation and elevation torques for very small fluctuations (, , ). Red lines are the means, i.e., without fluctuations; grey lines are the signals with fluctuations.
4.3. Trajectory Design and Planning
In order to perform tracking of a predefined trajectory, one has first to design this trajectory, the goal of the subsequent action. In some cases, this trajectory will be obvious to design, e.g., for an eye to move along a line, or in a circular motion, or for an arm to grab an object when no obstacle is present. In other circumstances, the task of planning a trajectory may be highly complex, especially in cases where both the agent and some obstacles are moving. The literature on planning is huge, and it has been studied extensively in fields like robotics, where it plays a crucial role (see, e.g., [60]).
4.4. Synthesis Law Computations: Tracking Controller
4.4.1. General Action Tracking Law
There are numerous ways to achieve trajectory tracking with stability, i.e., ensure that the discrepancy in action tends to zero asymptotically, and the corresponding literature is huge (see, e.g., [41] for a classic reference). For our purposes, i.e., the fulfilment of the inference guidelines, all of the laws described in [41] are inappropriate. This is because they do not rely on pathwise properties, and more importantly they do not rely on a basis-like property. In contrast, the flatness-based framework is inherently pathwise while embedding the physics of the agent’s model.
Considering a differentially flat model , we here wish to derive an action law (a controller) able to follow any reference trajectory . In order to compensate for model mismatch and poorly known initial conditions, one has to complement the open loop (obtained through flatness) with a closed-loop corrective term depending on the error .
Knowing that the dynamics is flat, with flat output , it can be transformed via the linearising endogenous feedback (50) to a linear dynamics of the form
| (70a) |
| (70b) |
with the new input . Then, the elementary tracking feedback law
| (71) |
with appropriately chosen gains renders the error dynamics
| (72) |
asymptotically stable. In vector form, the preceding law (71) is expressed as follows:
| (73) |
where and . Finally, the tracking action (tracking feedback control) law is expressed as follows:
| (74a) |
| (74b) |
| (74c) |
| (74d) |
| (74e) |
which ensures the tracking of to , with stability through driving the discrepancy to zero.
Remark 7 (Open-loop and model-free).
The preceding action law is named a linearising feedback controller due to the fact that the flat output dynamics is exactly linearised in a first step. Another possibility—most probably more fruitful—is to use one of the following control laws. A first possible choice is an open-loop controller, i.e., an action law obtained through the use of (29b), p. 13, in which ω is substituted with , a reference trajectory, with this open-loop law being supplemented by a model-free controller, in the spirit of [61]. Other possible choices include (74) supplemented by a model-free controller, an ADRC (active disturbance rejection control) one [62,63], or a sliding mode control one [64,65].
Example 5 (Simple generic example; tracking).
Consider again the following simple example with scalar action (control) and sensor output:
(75a)
(75b)
(75c) In order to elaborate on the tracking action laws, we have to obtain the sensor output y dynamics, which is obtained by differentiating (75a), thus obtaining :
(76) and using (75c):
(77) Setting , we have . Then, the y-dynamics (77) becomes
(78) The trajectory tracking goal is
(79) When expressing this goal in a dynamical setting, this renders the solution of a target differential equation:
(80) with such that all solutions of (80) are exponentially decreasing functions of time.
In order to transform (78) into (80), we set
(81) Hence, the control law ensuring the tracking of with stability is given by
(82) This control law assumes the full knowledge of the fluctuations , , and . When this knowledge is not available, we have to derive a control law based on a deterministic generative model (which is all that the agent knows), estimate the fluctuations, and compensate for them (see, e.g., [66] for the so-called model-free control or [62,65] for other schemes).
4.4.2. Oculomotor Example
Example 6 (Oculomotor tracking).
Consider again Example 4 and its generative model:
(83a)
(83b)
(83c)
(83d) and suppose that one wishes to track a reference trajectory for . We have seen in Example 4 that the following action feedback:
(84)
(85) transforms the nonlinear dynamics (83) into the following linear dynamics:
The trajectory tracking can then be enforced through the following tracking feedback law in , :
(86a)
(86b)
(86c) where , , , and are suitable constants (the action or feedback control gains) ensuring the stability of the closed-loop error dynamics:
Here, it is sufficient to take all , . Note that the preceding laws and in (86) can be replaced by other ones, such as, for instance, model-free control ones (see, e.g., [66]). Finally, to express the original control law , we must invert (84), yielding and as functions of and :
(87) Then substituting (86a) into (87) yields the final action tracking controller:
(88) which ensures the tracking of to with stability, i.e., the fulfilment of (IG-Perfo) through driving the discrepancy to zero. The quantity sensed by the agent is the position of the point P in the visual scene.
Remark 8 (A more realistic example).
Note that we could have considered a more realistic example than (83) by including, as in [53], a delay in sensing and an advance in acting. In order to achieve predictions, the agent uses a form of open-loop control. A possible choice could be the use of reference signals, using the functional parametrisation offered by the flatness property. In other words, using (61) after substitution of with , we get the open-loop action law:
(89a)
(89b) Another (heavier) choice would be to simulate the generative model to get a prediction for .
4.4.3. Simulations
The following plots refer to the simulation of the occulomotor model in (83), with the following parameters:
| (90) |
The fluctuations are all smooth random functions (i.e., sums of cosine and sine with coefficients that are random variables with a normal distribution). More precisely, they are of the form depicted in Equation (16):
| (91) |
| (92) |
where , each and is an independent sample from (where denotes the real normal distribution of mean and variance V), and is the floor function. Here, ; hence . Let us consider a smooth eye tracking of a point along a regular curve, here a quatrefoil, shown in Figure 8 with the reference curve in red and the actual tracking (i.e., the eye movement) in blue. The trajectory type has been chosen because of its smoothness. Its precise time parametrisation is
| (93a) |
| (93b) |
| (93c) |
Figure 8.
Reference quatrefoil trajectory in red and actual movement from the generative model in blue.
The corresponding references , and the actual values , are plotted in Figure 9.
Figure 9.
Plots of , in red and of , in blue for the quatrefoil.
The corresponding action tracking laws and are shown in Figure 10. As a last example, let us consider the hypocycloid in Figure 11. We chose this type of trajectory because it may be seen as typical of saccadic eye movements. Its precise time parametrisation is
| (94a) |
| (94b) |
| (94c) |
Figure 10.
Plots of the actions and for the quatrefoil.
Figure 11.
Reference hypocycloid trajectory in red and actual movement from the generative model in blue.
4.5. Active Inference
The principle of active inference can then be stated as extremisation of free energy via both action and perception, under a prior belief that action will extremise expected free energy (see, e.g., Figure 7 of [2]):
| (95a) |
| (95b) |
| (95c) |
| (95d) |
| (95e) |
with being the agent’s estimate of the hidden state , and as in (21c) (complexity/accuracy) and (22b) (risk/ambiguity), written in a compact form.
The above extremisations can be formulated as the solutions to a gradient descent.
4.6. Link with Flatness-Based Tracking
The imperative for active inference is the minimisation of surprise, namely, the discrepancy between expectations, or beliefs, and actual values. In the light of the previous decomposition of variational free energy in Section 3.1, we conclude by foregrounding the links between active inference and nonlinear control. Consider the agent’s dynamics . The dynamic feedback (see (48))
| (96a) |
| (96b) |
| (96c) |
(where is the new input) transforms the agent’s dynamics to
| (97a) |
| (97b) |
| (97c) |
The above Equation (97) then yield the following action surprise dynamics, or action discrepancy dynamics:
| (98) |
This surprise dynamics will not, in general (i.e., for any feedback law (96)), imply the action discrepancy’s limit to be zero, i.e., a minimisation of the free energy. Let us now consider the tracking action (tracking feedback control) law stemming from (74):
| (99a) |
| (99b) |
| (99c) |
| (99d) |
| (99e) |
which yields a special form of (98):
| (100) |
where is a linear function of its arguments such that (100) admits solely exponentially stable solutions. The tracking action (tracking feedback control) law (99) thus ensures the tracking of to , with stability through driving the discrepancy to zero, minimising the risk in the expected free energy .
Example 7 (Oculomotor tracking control law).
Consider again the oculomotor Example 4. The tracking action (tracking feedback control) law (99) takes the form of (88):
(101)
5. Prediction as a Link Between Active Inference and Differential Flatness
5.1. Delays and -Flatness
In real systems stemming from neuroscience or physiology, delays are present, both in sensing and acting. This has not yet been taken into account in the previous sections, although it remains of utmost importance. Indeed, there are fundamental differences between a delay-free model and one including delays; the most striking difference is the infinite-dimensional character of delay systems. To be more precise, consider a delay-differential equation of the following form:
| (102) |
If we want to integrate such an equation, we do not need to know a pointwise initial condition such as but have to know a whole function on a time interval . Another view on this is that there is not just one operator acting on the variables—time differentiation—but two: time differentiation and the time delay. Therefore, the structural properties (controllability and observability) of such systems become more complex, even for linear delay systems (see, e.g., [67]). The notion of flatness has been extended to the case of delay systems (see [68,69]). We shall informally recall the definition of the so-called -flatness (a special case of the -flatness where the delays do not only occur on the sensor outputs and the action variables but also on the hidden state) and then expose two different kinds of predictors, which are needed for tracking. We denote a delay operator of amplitude by :
| (103) |
In informal terms, a -flat system is a system of delay differential equations that is differentially flat when one allows for delays. Let us see how this notion unfolds in a concrete example.
Example 8.
Consider Example 2 with a delay in the action:
(104a)
(104b)
(104c) with . This model is easily seen to be δ-flat, with y as a δ-flat output. Indeed, we have
(105a)
(105b)
(105c)
5.2. Trajectory Tracking and Predictors
When one wishes to perform trajectory tracking, as in (99), the presence of delays in the sensor output and/or in the action control will induce the necessity to predict some or all of the hidden states or sensor output. Consider this under the previous example.
Example 9.
The tracking control scheme (82) is now transformed into
(106a)
(106b)
(106c)
We now review two kinds of predictors, which may be used in tracking control schemes. Consider the following type of mean generative model with delayed action and a linear relation between the sensory output and hidden state:
| (107a) |
| (107b) |
with . This model can be transformed into the advanced generative model
| (108a) |
| (108b) |
| (108c) |
The first scheme is a so-called delayed observer (see [70]):
| (109) |
with being the predictor gain matrix. The simulation of Equation (109) yields an estimate of . It is proven [70] that the so-called prediction observer error tends to zero as t tends to infinity.
The second scheme is an integral form one (see [71,72,73]):
| (110a) |
| (110b) |
where is such that is globally asymptotically stable at ; i.e., for every trajectory , we have as . Equation (110b) comes from the following simple observation:
| (111) |
The initial condition for the integral Equation (110b) for is defined by
| (112) |
The predictor state is given by the implicit relation (110b), which can be solved using various approximation strategies for the integral on the right-hand side. Then, the predictive control law of the simple example shall be implemented as follows.
Example 10.
The tracking control scheme (106a) is now transformed into
(113a)
(113b)
(113c)
5.3. Generalised Coordinates
In the active inference framework, there is a crucial distinction between what is called motion of the mean compared to mean of the motion. To unpack this distinction we first need the notion of generalised coordinates, which is obtained through a linearised differentiation. Generally a differentiation operation ∂ in algebra is defined as one such that, for any variables (here time functions) and , the chain rule is fulfilled:
| (114) |
A linearised (or first-order) differentiation, , here denoted as ′, is an operator following the above chain rule such that, for any time functions and ,
| (115) |
whenever , and denoting by the jth iterated application of the ′ operator to . This type of differentiation is linked to first-order stochastic realisation problems, as briefly discussed below.
Higher-order differentiations can also be considered. A k-th-order differentiation is an operator following the above chain rule such that, for any time functions and ,
| (116) |
whenever , . This type of differentiation is linked to higher-order stochastic realisation problems, although these may induce quite involved computations.
In other words, suppose the solution of
| (117) |
can be expressed through a series expansion:
| (118) |
e.g., it could be an analytic function of time or a Gevrey series (i.e., belonging to a class of functions in between the analytic and the full one—see, e.g., [74,75]; see also [76] for extensions). Then, the various derivatives at the time origin can be recovered as follows:
| (119a) |
| (119b) |
| (119c) |
assuming the fluctuations are sufficiently smooth and
| (120a) |
| (120b) |
The so-called local linear approximation, in terms of [32], amounts to neglecting all higher-order terms in Equation (120), i.e., considering
| (121) |
Such an approximation, when valid, enables one to solve the so-called stochastic realisation problem quite easily (recall that a realisation is a differential equation involving a hidden state obtained from an input/action–output/sensor differential equation—see, e.g., Definition 6; see, e.g., [77,78] on the stochastic realisation problem). This approximation is also justified when studying generalised Bayesian filtering under the Laplace approximation (see, e.g., [32], Subsections 3.3.4 and 3.3.5).
The generalised coordinates may be seen as a coordinate frame that moves with the current point. In this view, it is linked with one-form transformations one can apply to nonlinear systems (see, e.g., [79,80]); the latter are more general than the endogenous dynamical feedbacks used here. It is also linked with the Cartan moving frame method (see, e.g., [81]).
6. Conclusions, Limitations, and Future Directions
We have considered the utility of differential flatness through the lens of active inference (see, e.g., [1]). This utility has been detailed in terms of control as inference. Specifically, one might conclude that if the generative models that underwrite active inference or control as inference can be limited to the class of differentially flat models, we obtain an extremely efficient control-theoretic scheme. This work therefore enables a control-theoretic perspective on active inference as control as inference by focusing on action trajectories and the minimisation of various discrepancies (afforded by variational and expected free energy). From the perspective of active inference, this work is a primer on differential flatness and its particular relevance to the kinds of generative models one might consider and be committed to. Crucially, this paper is the first (provisional) attempt to consider expected free energy in the setting of continuous state-space models.
In addition to their roles in developing systems for control, one could consider other applications of the frameworks outlined in this paper. One area is the field of computational psychiatry (see, e.g., [82,83,84]), where one can develop generative models of decision-making tasks—solved with active inference—and use these to understand the computational mechanisms of psychopathology. Such models are often formulated in terms of the selection among discrete alternatives. However, as noted by our reviewers, differential flatness pertains to continuous, differentiable state spaces. While it is true that many applications of computational psychiatry focus on discrete probabilities, there are important areas of psychiatry that depend upon continuous variables of the sort addressed in differential flatness accounts. These include the altered motor dynamics associated with catatonia in psychotic disorders (see, e.g., [85]) and altered smooth pursuit eye movements in schizophrenia (see, e.g., [86,87]).
There are several related notions that could not be addressed here due to the lack of space. These include Liouvillian aspects (see [88,89,90] related to automatic control and [91,92] for the examination of this property in the setting of hypothalamic–pituitary–adrenal axis models and Wilson–Cowan population networks), and robust tracking with model-free control (see [66]). Liouvillian aspects, in particular, offer an opportunity to extend the notion of flatness when a model is not differentially flat. Future work will consider these and other issues related to energy transmission and controllability (see, e.g., [93,94] for a formulation of the corresponding problem) and variational function transmission. These deal with characterising a model in terms of salient features and understanding how reparameterisations might transform such features.
We discuss the following points, emphasising limitations of the current treatment and sketching some future directions.
Generalised coordinates, where some limitations—due to its approximate character—may be avoided through smooth random realisations.
Extensions of flatness: Liouvillian characters to deal with cases where the model is not differentially flat.
Observers and algebraic estimators to estimate the hidden state from sensor output.
Robust control law synthesis to cope with uncertainty under the generative model, including fluctuations.
Constraint fulfilment, where constraints are imposed on the hidden state, the action, and their time derivatives.
Feature transmission: How a feature of interest; e.g., the energy ( norm), the slope or curvature, etc., is transformed through functional parametrisation.
6.1. Generalised Coordinate Limitations
The generalised coordinates are not appropriate for deriving tracking feedback action in the sense being considered here. Indeed, the dynamical extension algorithm needs to make the action appear through time differentiation, which will not be the case unless the action was already present in the state-space dynamics. More precisely, for the components of the flat output whose Brunovksỳ index is strictly superior to 2, the generalised coordinates are not appropriate, since successive application of will not make the action appear. For the components whose index or 2, the generalised coordinates will be sufficient.
Let us illustrate this through three simple examples.
-
First consider the longitudinal motion of a car, used, for instance, in ACCs (Automatic Cruise Controllers):
with M being the car’s mass, the longitudinal speed of the vehicle’s centre of gravity, the force due to the wind and to friction of the tyres on the ground, r the mean wheel radius, and u the traction engine propulsion torque, taken as the action. This system is trivially flat, with flat output , and the preceding equation is rewritten as(122) The action is already present in the dynamics equation, , and there is no need to differentiate more. The generalised coordinate dynamics will lead to the same action tracking controller as the original one.
- Second, consider the oculomotor example again. Then one supplementary differentiation of (resp. ) will be sufficient. Both of these will involve and its partial derivatives with respect to and in the flat output dynamics (see (69)):
(123a) (123b) -
Third, consider one of the most popular model for diabetes, namely the Bergman minimal model (see, e.g., [95,96]):
(124a) (124b) (124c)
where is the concentration of blood glucose; is the concentration of insulin in the tissue fluid; is the concentration of insulin in the blood; and are the basal concentrations of glucose and insulin; and is the glycaemic influence of a meal, seen as a fluctuation. In addition, , , and are positive-valued parameters that control the rates of appearance and disappearance of glucose and insulin; is the meal intake time (bolus intake).(124d) We wish to control the glucose concentration in order to track a desired trajectory known in advance. To do so, we apply the dynamical extension algorithm, and we have to differentiate Equation (124a) two times in order for the action u to appear. We get
and
and the control will appear in , present in . The generalised coordinates unfold as follows. The preceding model is rewritten, within the differentiation , as follows:(125a) (125b) (125c) (125d) The successive application of yields
andThus, the term is missing when compared to its counterpart with time differentiation. And, if one attempted to use this in order to produce an action tracking feedback, the resulting error dynamics would be of the form
with and being a reference glucose trajectory. Although the are such that the solutions of(126)
are exponentially decreasing, the solutions of (126) may not tend to zero, since it is steered by .
6.2. Smooth Random Realisation
The use of this type of differentiation is also linked to the stochastic realisation problem (see, e.g., [10], 4.(c).(i), p. 15), which may be quite complex in the general case. In contrast to this, the differential flatness property yields a weak Brunovský canonical form through the flat output (see [97], Subsection 4.1 and Definition 4.3). This canonical form yields the so-called flat output dynamics, which readily gives a smooth random realisation, as the following proposition states.
Proposition 4 (Smooth random realisation).
Consider a differentially flat model, with flat output . Then, there exists integers such that is a state. This state, through the flat output dynamics, yields a smooth random realisation of the model.
Proof.
The existence of the integers such that is a state is ensured by the dynamical extension algorithm. To see this, and to obtain the associated realisation, recall the so-called flat output dynamics (Equation (53) above), which can be written as the following state-space representation:
(127a)
(127b)
(127c)
(127d)
(127e)
(127f)
(127g)
(127h) with , since , and
Then is a state that yields a smooth random realisation of the original system, based on the differential flatness property. □
The present framework can be seen as an adequate proposal for the future direction in Subsection 4.1,, paragraph “Stochastic control via generalized coordinates” of [32]. It also embeds, in a rather simple fashion, non-stationary smooth random signals (see Remark 4.1.1 in [32]).
6.3. Other Future Directions
We now briefly consider future directions that could not be unpacked in this paper due to lack of space.
6.3.1. Extensions of Flatness: Liouvillian Characters
Although many practical system models are differentially flat, some, among specific classes, are not; this is especially the case for many biological and neuroscience population models. Fortunately, another analogous property is shared by a much wider class, the one of so-called Liouvillian systems. Liouvillian systems can be seen as an extension of flat systems [88,89,98]. The most striking property of the latter is that all state and control variables of the system can be directly expressed—without integration of any differential equation—in terms of the flat output and a finite number of its time derivatives. So-called Liouvillian systems share a similar property, but in order to derive the trajectories of a Liouvillian system, we also need integration of a few differential equations whose solutions are known analytically. It follows that flatness-based control approaches can be extended up to solving a finite number of differential equations.
6.3.2. Observers and Algebraic Estimators
We have here dealt with control law synthesis, but we did not touch on the equally important subject of observer or estimator synthesis, i.e., the procedures aimed at estimating the hidden state from sensor measurements. Two main paths are available: The first—of so-called observers—amounts to a simulation of the mean generative model, driven by the error between the sensor output and its observation model, a function of the state (see, e.g., [99,100]). The second, more direct, approach is to directly make use of the so-called constructive observability (see, e.g., [4,101]), where the state is a function of sensor output, action, and their derivatives: . This last procedure requires the estimation of sensor output derivatives (see [6,102]).
6.3.3. Robust Control Law Synthesis
When the mean generative model is a crude approximation to the real system, the cumulative effect of the fluctuations needs to be compensated for in the control scheme. So-called robust control laws aim to fulfil a control goal, such as trajectory tracking, despite the fluctuations, or perturbations, and unmodelled dynamics (i.e., model mismatch). These can be achieved via the so-called model-free control scheme, where the cumulative effects of fluctuations are estimated online and compensated for on the fly (see, e.g., [66]). The recent HEOL scheme (see [103]) achieves the same goal through slightly different techniques. While model-free control is often associated with flatness-based feedback tracking control laws synthesised on a nominal, or mean generative, model, HEOL one uses an open-loop flatness-based scheme and the tangent, or variational system, associated with the simplified flat system, i.e., the linearised system around a reference trajectory of the simplified flat system. Other instances include, among others, sliding mode control and active disturbance rejection control (see, e.g., [62,64,65]).
6.3.4. Constraint Fulfilment
In real-world applications, dynamical systems are always subject to constraints: on the state (for example the configuration space of a robot is not the whole space) and on the action (e.g., muscles have finite power). These can be handled in the present framework through optimisation-based planning of the flat output trajectories (see, e.g., [104,105,106]). A promising framework is the one of model-free predictive control (MFPC, see [107]), mixing the popular predictive control (see, e.g., [108,109]) and model-free control, cited above.
6.3.5. Feature Transmission
Another interesting characteristic—of the functional parametrisation associated with differential flatness—is one of feature transmission. This includes the transmission of geometric features: how the curvature of the flat output trajectory is related to the curvature of the action; in other words, deducing the action’s curvature as a function of the flat output curvature and its time derivatives from the relation yielding the action as a function of the flat output and its derivatives.
Supplementary Materials
The following supporting information can be downloaded at https://github.com/hugues-mounier/AI-FP-differential-flatness-and-smooth-random-realisation/.
Author Contributions
Methodology, H.M., T.P. and K.F.; Software, H.M.; Writing—original draft, H.M.; Writing—review & editing, T.P. and K.F. All authors have read and agreed to the published version of the manuscript.
Data Availability Statement
The data and code presented in this study are openly available in https://github.com/hugues-mounier/AI-FP-differential-flatness-and-smooth-random-realisation/ (accessed on 29 October 2025).
Conflicts of Interest
The authors declare no conflicts of interest.
Funding Statement
This research was funded by Wellcome Trust grant number 226793/Z/22/Z. The Third author (TR) is supported by an NIHR Academic Clinical Fellowship [ref: ACF-2023-13-013].
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Friston K. A free energy principle for a particular physics. arXiv. 2019 doi: 10.48550/arXiv.1906.10184.1906.10184 [DOI] [Google Scholar]
- 2.Friston K., Da Costa L., Sajid N., Heins C., Ueltzhöffer K., Pavliotis G.A., Parr T. The free energy principle made simpler but not too simple. Phys. Rep. 2023;1024:1–29. doi: 10.1016/j.physrep.2023.07.001. [DOI] [Google Scholar]
- 3.Ramstead M.J.D., Sakthivadivel D.A.R., Heins C., Koudahl M., Millidge B., Da Costa L., Klein B., Friston K.J. On Bayesian mechanics: A physics of and by beliefs. Interface Focus. 2023;13:20220029. doi: 10.1098/rsfs.2022.0029. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Fliess M., Glad S.T. Essays on Control. Birkhäuser; Boston, TX, USA: 1993. An Algebraic Approach to Linear and Nonlinear Control; pp. 223–267. [DOI] [Google Scholar]
- 5.Fliess M., Lévine J., Martin P., Rouchon P. Flatness and defect of non-linear systems: Introductory theory and examples. Int. J. Control. 1995;61:1327–1361. doi: 10.1080/00207179508921959. [DOI] [Google Scholar]
- 6.Fliess M., Join C., Ramirez H.S. Non-linear estimation is easy. Int. J. Model. Identif. Control. 2008;4:12. doi: 10.1504/IJMIC.2008.020996. [DOI] [Google Scholar]
- 7.Baltieri M. PhD thesis. University of Sussex; Brighton, UK: 2019. Active Inference: Building a New Bridge Between Control Theory and Embodied Cognitive Science. [Google Scholar]
- 8.Baltieri M., Buckley C. PID Control as a Process of Active Inference with Linear Generative Models. Entropy. 2019;21:257. doi: 10.3390/e21030257. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Lanillos P., Meo C., Pezzato C., Meera A.A., Baioumy M., Ohata W., Tschantz A., Millidge B., Wisse M., Buckley C.L., et al. Active Inference in Robotics and Artificial Agents: Survey and Challenges. arXiv. 2021 doi: 10.48550/arXiv.2112.01871.2112.01871 [DOI] [Google Scholar]
- 10.Da Costa L., Friston K., Heins C., Pavliotis G.A. Bayesian mechanics for stationary processes. Proc. R. Soc. A Math. Phys. Eng. Sci. 2021;477:20210518. doi: 10.1098/rspa.2021.0518. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Meera A.A., Wisse M. A Brain Inspired Learning Algorithm for the Perception of a Quadrotor in Wind. arXiv. 2021 doi: 10.48550/arXiv.2109.11971.2109.11971 [DOI] [Google Scholar]
- 12.Tschantz A., Barca L., Maisto D., Buckley C.L., Seth A.K., Pezzulo G. Simulating homeostatic, allostatic and goal-directed forms of interoceptive control using Active Inference. Biol. Psychol. 2022;169:108266. doi: 10.1016/j.biopsycho.2022.108266. [DOI] [PubMed] [Google Scholar]
- 13.Baioumy M., Pezzato C., Ferrari R., Hawes N. Unbiased Active Inference for Classical Control. arXiv. 2022 doi: 10.48550/arXiv.2207.13409.2207.13409 [DOI] [Google Scholar]
- 14.Bos F., Meera A.A., Benders D., Wisse M. Free Energy Principle for State and Input Estimation of a Quadcopter Flying in Wind; Proceedings of the 2022 International Conference on Robotics and Automation (ICRA); Philadelphia, PA, USA. 23–27 May 2022; Piscataway, NJ, USA: IEEE; 2022. [DOI] [Google Scholar]
- 15.Rudolph J. Flatness-Based Control—An Introduction. 1st ed. Shaker GmbH; Düren, Germany: 2021. [DOI] [Google Scholar]
- 16.Francine Diener M.D., editor. Nonstandard Analysis in Practice. Springer; Berlin/Heidelberg, Germany: 1995. [DOI] [Google Scholar]
- 17.Fliess M. Probabilités et fluctuations quantiques. Comptes Rendus. MathéMatique. 2007;344:663–668. doi: 10.1016/j.crma.2007.04.001. [DOI] [Google Scholar]
- 18.Fliess M., Join C. Towards a new viewpoint on causality for time series. ESAIM Proc. Surv. 2015;49:37–52. doi: 10.1051/proc/201549004. [DOI] [Google Scholar]
- 19.Han X., Kloeden P.E. Random Ordinary Differential Equations and Their Numerical Solution. Springer; Singapore: 2017. [DOI] [Google Scholar]
- 20.Friz P.K., Hairer M. A Course on Rough Paths: With an Introduction to Regularity Structures. Springer International Publishing; Cham, Switzerland: 2020. [DOI] [Google Scholar]
- 21.Kahane J.P. Some Random Series of Functions Series. 2nd ed. Cambridge University Press; Cambridge, UK: 1993. Number 5 in Cambridge Studies in Advanced Mathematics. [Google Scholar]
- 22.Hill M. Convergence of random Fourier series; Proceedings of the University of Chicago 2012 Summer Program for Undergraduate; 2012. [(accessed on 29 October 2025)]. Available online: https://math.uchicago.edu/~may/REU2012/REUPapers/Hill.pdf. [Google Scholar]
- 23.Tzvetkov N. Riemannian analogue of a Paley-Zygmund theorem. Séminaire Équations aux Dérivées Partielles (2008–2009) [(accessed on 29 October 2025)]. pp. 1–14. Available online: https://eudml.org/doc/11189.
- 24.Marcus M.B., Pisier G. Random Fourier Series with Applications to Harmonic Analysis (AM-101) Princeton University Press; Princeton, NJ, USA: 1981. [Google Scholar]
- 25.Filip S., Javeed A., Trefethen L.N. Smooth Random Functions, Random ODEs, and Gaussian Processes. SIAM Rev. 2019;61:185–205. doi: 10.1137/17M1161853. [DOI] [Google Scholar]
- 26.Robert J., Adler J.E.T. Random Fields and Geometry. Springer; New York, NY, USA: 2007. [DOI] [Google Scholar]
- 27.Adler R.J., Taylor J.E. Topological Complexity of Smooth Random Functions: École d’Été de Probabilités de Saint-Flour XXXIX-2009. Springer; Berlin/Heidelberg, Germany: 2011. [DOI] [Google Scholar]
- 28.Adler R.J., Taylor J.E., Worsley K.J. Applications of Random Fields and Geometry: Foundations and Case Studies. Book in Progress. 2016. [(accessed on 29 October 2025)]. Available online: https://cris.technion.ac.il/en/publications/applications-of-random-fields-and-geometry-foundations-and-case-s/
- 29.Friston K.J., Holmes A.P., Worsley K.J., Poline J., Frith C.D., Frackowiak R.S.J. Statistical parametric maps in functional imaging: A general linear approach. Hum. Brain Mapp. 1994;2:189–210. doi: 10.1002/hbm.460020402. [DOI] [Google Scholar]
- 30.Worsley K.J., Marrett S., Neelin P., Vandal A.C., Friston K.J., Evans A.C. A unified statistical approach for determining significant signals in images of cerebral activation. Hum. Brain Mapp. 1996;4:58–73. doi: 10.1002/(SICI)1097-0193(1996)4:1<58::AID-HBM4>3.0.CO;2-O. [DOI] [PubMed] [Google Scholar]
- 31.Gyöngy I., Michaletzky G. On Wong–Zakai approximations with δ–martingales. Proc. R. Soc. London. Ser. A Math. Phys. Eng. Sci. 2004;460:309–324. doi: 10.1098/rspa.2003.1244. [DOI] [Google Scholar]
- 32.Da Costa L., Da Costa N., Heins C., Medrano J., Pavliotis G.A., Parr T., Meera A.A., Friston K. A Theory of Generalized Coordinates for Stochastic Differential Equations. Stud. Appl. Math. 2025;154:e70062. doi: 10.1111/sapm.70062. [DOI] [Google Scholar]
- 33.Stratonovich R.L. Topics in the Theory of Random Noise. Volume 2 Macmillan Education; South Yarra, Australia: 1967. [Google Scholar]
- 34.Esser C., Jaffard S., Vedel B. Regularity properties of random wavelet series. arXiv. 2023 doi: 10.1090/tpms/1205.2304.00811 [DOI] [Google Scholar]
- 35.Aubry J.M., Jaffard S. Emergent Nature. World Scientific; Singapore: 2002. Random wavelet series: Theory and applications. [DOI] [Google Scholar]
- 36.Adler R.J. On excursion sets, tube formulas and maxima of random fields. Ann. Appl. Probab. 2000;10:1–74. doi: 10.1214/aoap/1019737664. [DOI] [Google Scholar]
- 37.Krishnan S.R., Taylor J.E., Adler R.J. The Intrinsic geometry of some random manifolds. Electron. Commun. Probab. 2017;22:1–12. doi: 10.1214/16-ecp4763. [DOI] [Google Scholar]
- 38.Yuan R.S., Ma Y.A., Yuan B., Ao P. Lyapunov function as potential function: A dynamical equivalence. Chin. Phys. B. 2014;23:010505. doi: 10.1088/1674-1056/23/1/010505. [DOI] [Google Scholar]
- 39.Parr T., Da Costa L., Friston K. Markov blankets, information geometry and stochastic thermodynamics. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2019;378:20190159. doi: 10.1098/rsta.2019.0159. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Wang H., Hu W., Gan X., Ao P. The generalized Lyapunov function as Ao’s potential function: Existence in dimensions 1 and 2. J. Appl. Anal. Comput. 2023;13:359–375. doi: 10.11948/20220149. [DOI] [Google Scholar]
- 41.Khalil H.K. Nonlinear Control. Pearson; Upper Saddle River, NJ, USA: 2014. [Google Scholar]
- 42.Fliess M. A note on the invertibility of nonlinear input-output differential systems. Syst. Control Lett. 1986;8:147–151. doi: 10.1016/0167-6911(86)90073-3. [DOI] [Google Scholar]
- 43.Fliess M., Lévine J., Martin P., Rouchon P. On Differentially Flat Nonlinear Systems. IFAC Proc. Vol. 1992;25:159–163. doi: 10.1016/s1474-6670(17)52275-2. [DOI] [Google Scholar]
- 44.Fliess M., Lévine J., Martin P., Rouchon P. A Lie-Bäcklund approach equivalence and flatness of nonlinear systems. IEEE Trans. Automat. Control. 1999;44:922–937. doi: 10.1109/9.763209. [DOI] [Google Scholar]
- 45.Sira-Ramírez H. Differentially Flat Systems. CRC Press; Boca Raton, FL, USA: 2004. [Google Scholar]
- 46.Hilbert D. Über den Begriff der Klasse von Differentialgleichungen. Math. Ann. 1912;73:95–108. doi: 10.1007/BF01456663. [DOI] [Google Scholar]
- 47.Cartan M. Sur l’équivalence absolue de certains systèmes d’équations différentielles et sur certaines familles de courbes. Bull. Soc. Math. Fr. 1914;2:12–48. doi: 10.24033/bsmf.938. [DOI] [Google Scholar]
- 48.Fuller S., Greiner B., Moore J., Murray R., van Paassen R., Yorke R. The Python Control Systems Library (python-control); Proceedings of the 60th IEEE Conference on Decision and Control (CDC); Austin, TX, USA. 14–17 December 2021; Piscataway, NJ, USA: IEEE; 2021. pp. 4875–4881. [Google Scholar]
- 49.Perrin D. Algebraic Geometry. Springer; London, UK: 2008. [DOI] [Google Scholar]
- 50.Eisenbud D., Harris J. The Geometry of Schemes. 1st ed. Springer; New York, NY, USA: 2000. Graduate Texts in Mathematics. [Google Scholar]
- 51.Agrios M. A Very Elementary Introduction to Sheaves. arXiv. 2022 doi: 10.48550/arXiv.2202.01379.2202.01379 [DOI] [Google Scholar]
- 52.Curry J. Sheaves, Cosheaves and Applications. arXiv. 20131303.3255 [Google Scholar]
- 53.Perrinet L.U., Adams R.A., Friston K.J. Active inference, eye movements and oculomotor delays. Biol. Cybern. 2014;108:777–801. doi: 10.1007/s00422-014-0620-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Adams R.A., Aponte E., Marshall L., Friston K.J. Active inference and oculomotor pursuit: The dynamic causal modelling of eye movements. J. Neurosci. Methods. 2015;242:1–14. doi: 10.1016/j.jneumeth.2015.01.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Parr T., Friston K.J. Active inference and the anatomy of oculomotion. Neuropsychologia. 2018;111:334–343. doi: 10.1016/j.neuropsychologia.2018.01.041. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Parr T., Friston K.J. The computational pharmacology of oculomotion. Psychopharmacology. 2019;236:2473–2484. doi: 10.1007/s00213-019-05240-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Harris D., Vine S., Wilson M., Arthur T. The Relationship Between Environmental Statistics and Predictive Gaze Behaviour During a Manual Interception Task: Eye Movements as Active Inference. Comput. Brain Behav. 2023;7:225–241. doi: 10.1007/s42113-023-00190-5. [DOI] [Google Scholar]
- 58.Wong A. Listing’s law: Clinical significance and implications for neural control. Surv. Ophthalmol. 2004;49:563–575. doi: 10.1016/S0039-6257(04)00134-1. [DOI] [PubMed] [Google Scholar]
- 59.Ekstrand B. Equations of motion for a two-axes gimbal system. IEEE Trans. Aerosp. Electron. Syst. 2001;37:1083–1091. doi: 10.1109/7.953259. [DOI] [Google Scholar]
- 60.Siciliano B., Khatib O., editors. Springer Handbook of Robotics. Springer International Publishing; Cham, Switzerland: 2016. [DOI] [Google Scholar]
- 61.Fliess M., Join C., Moussa K., Djouadi S.M., Alsager M.W. Toward simple in silico experiments for drugs administration in some cancer treatments. IFAC-PapersOnLine. 2021;54:245–250. doi: 10.1016/j.ifacol.2021.10.263. [DOI] [Google Scholar]
- 62.Sira-Ramirez H., Luviano-Juárez A., Ramírez-Neria M., Zurita-Bustamante E.W. Active Disturbance Rejection Control of Dynamic Systems: A Flatness Based Approach. Butterworth-Heinemann; Oxford, UK: 2017. [Google Scholar]
- 63.Sira-Ramírez H., Gómez-León B.C., Aguilar-Orduña M.A. A two stage Active Disturbance Rejection Control design for under-actuated nonlinear systems. IFAC-PapersOnLine. 2023;56:4539–4544. doi: 10.1016/j.ifacol.2023.10.949. [DOI] [Google Scholar]
- 64.Utkin V.I. Sliding Modes in Control and Optimization. Springer; Berlin/Heidelberg, Germany: 1992. [DOI] [Google Scholar]
- 65.Shtessel Y., Edwards C., Fridman L., Levant A. Sliding Mode Control and Observation. Springer; New York, NY, USA: 2014. [DOI] [Google Scholar]
- 66.Fliess M., Join C. Model-free control. Int. J. Control. 2013;86:2228–2252. doi: 10.1080/00207179.2013.810345. [DOI] [Google Scholar]
- 67.Fliess M., Mounier H. Controllability and observability of linear delay systems: An algebraic approach. ESAIM Control. Optim. Calc. Var. 1998;3:301–314. doi: 10.1051/cocv:1998111. [DOI] [Google Scholar]
- 68.Mounier H., Rudolph J. Flatness-based control of nonlinear delay systems: A chemical reactor example. Int. J. Control. 1998;71:871–890. doi: 10.1080/002071798221614. [DOI] [Google Scholar]
- 69.Mounier H., Rudolph J. Flatness and quasi-static state feedback in non-linear delay systems. Int. J. Control. 2008;81:445–456. doi: 10.1080/00207170701579437. [DOI] [Google Scholar]
- 70.Estrada-Sánchez I., Velasco-Villa M., Rodríguez-Cortés H. Prediction-Based Control for Nonlinear Systems with Input Delay. Math. Probl. Eng. 2017;2017:7415418. doi: 10.1155/2017/7415418. [DOI] [Google Scholar]
- 71.Krstic M. Input Delay Compensation for Forward Complete and Strict-Feedforward Nonlinear Systems. IEEE Trans. Autom. Control. 2010;55:287–303. doi: 10.1109/TAC.2009.2034923. [DOI] [Google Scholar]
- 72.Bresch-Pietri D., Petit N., Krstic M. Prediction-based control for nonlinear state- and input-delay systems with the aim of delay-robustness analysis; Proceedings of the 2015 54th IEEE Conference on Decision and Control (CDC); Osaka, Japan. 15–18 December 2015; Piscataway, NJ, USA: IEEE; 2015. pp. 6403–6409. [DOI] [Google Scholar]
- 73.Karafyllis I., Krstic M. Predictor Feedback for Delay Systems: Implementations and Approximations. Springer International Publishing; Cham, Switzerland: 2017. [DOI] [Google Scholar]
- 74.Rodino . Linear Partial Differential Operators in Gevrey Spaces. World Scientific; Singapore: 1993. pp. 5–59. [DOI] [Google Scholar]
- 75.Balser W. Formal Power Series and Linear Systems of Meromorphic Ordinary Differential Equations. Springer; New York, NY, USA: 2000. [DOI] [Google Scholar]
- 76.Teofanov N., Tomić F., Žigić M. An introduction to extended Gevrey regularity. arXiv. 2024 doi: 10.3390/axioms13060352.2404.17366 [DOI] [Google Scholar]
- 77.Lindquist A., Picci G. On the Stochastic Realization Problem. SIAM J. Control Optim. 1979;17:365–389. doi: 10.1137/0317028. [DOI] [Google Scholar]
- 78.Veeravalli T., Raginsky M. Revisiting Stochastic Realization Theory using Functional Itô Calculus. arXiv. 2024 doi: 10.1016/j.ifacol.2024.10.190.2402.10157 [DOI] [Google Scholar]
- 79.van Nieuwstadt M., Rathinam M., Murray R. Differential flatness and absolute equivalence; Proceedings of the 1994 33rd IEEE Conference on Decision and Control; Orlando FL, USA. 14–16 December, 1994; Piscataway, NJ, USA: IEEE; 1994. CDC-94. [DOI] [Google Scholar]
- 80.Olver P.J. Equivalence, Invariants and Symmetry. Cambridge University Press; Cambridge, UK: 1995. [DOI] [Google Scholar]
- 81.Ivey T.A.t.A. Cartan for Beginners. American Mathematical Society; Providence, RI, USA: 2003. Graduate Studies in Mathematics. [Google Scholar]
- 82.Pio-Lopez L., Kuchling F., Tung A., Pezzulo G., Levin M. Active inference, morphogenesis, and computational psychiatry. Front. Comput. Neurosci. 2022;16:988977. doi: 10.3389/fncom.2022.988977. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Badcock P.B., Davey C.G. Active Inference in Psychology and Psychiatry: Progress to Date? Entropy. 2024;26:833. doi: 10.3390/e26100833. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84.Harris H.W. Active inference and psychodynamics: A novel integration with applications to depression and stress disorders. Front. Psychiatry. 2025;16:1630858. doi: 10.3389/fpsyt.2025.1630858. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Adams R.A., Stephan K.E., Brown H.R., Frith C.D., Friston K.J. The Computational Anatomy of Psychosis. Front. Psychiatry. 2013;4:47. doi: 10.3389/fpsyt.2013.00047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Adams R.A., Perrinet L.U., Friston K. Smooth Pursuit and Visual Occlusion: Active Inference and Oculomotor Control in Schizophrenia. PLoS ONE. 2012;7:e47502. doi: 10.1371/journal.pone.0047502. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Perrinet L.U., Adams R.A., Friston K. Active inference, eye movements and oculomotor delays. BMC Neurosci. 2013;14:P133. doi: 10.1186/1471-2202-14-S1-P133. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Chelouah A. Extensions of differential flat fields and Liouvillian systems; Proceedings of the 36th IEEE Conference on Decision and Control; San Diego, CA, USA. 12 December 1997; Piscataway, NJ, USA: IEEE; 1997. pp. 4268–4273. CDC-97. [DOI] [Google Scholar]
- 89.Chelouah A. Diffieties and Liouvillian Systems. arXiv. 2010 doi: 10.48550/arXiv.1010.3909.1010.3909 [DOI] [Google Scholar]
- 90.Chetverikov V.N. Liouville systems and symmetries. Differ. Equ. 2012;48:1639–1651. doi: 10.1134/S0012266112120099. [DOI] [Google Scholar]
- 91.Nicolau F., Mounier H., Androulakis I.P. HPA axis differential flatness and Liouvillian study for higher resiliency investigations. IMA J. Math. Control Inf. 2023;40:746–788. doi: 10.1093/imamci/dnad030. [DOI] [Google Scholar]
- 92.Nicolau F., Mounier H. Flatness of Networks of Synaptically Coupled Excitatory-Inhibitory Neural Modules. ESAIM Control. Optim. Calc. Var. 2023;29:89. doi: 10.1051/cocv/2023082. [DOI] [Google Scholar]
- 93.Karrer T.M., Kim J.Z., Stiso J., Kahn A.E., Pasqualetti F., Habel U., Bassett D.S. A practical guide to methodological considerations in the controllability of structural brain networks. J. Neural Eng. 2020;17:026031. doi: 10.1088/1741-2552/ab6e8b. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Baggio G., Pasqualetti F., Zampieri S. Energy-Aware Controllability of Complex Networks. Annu. Rev. Control. Robot. Auton. Syst. 2022;5:465–489. doi: 10.1146/annurev-control-042920-014957. [DOI] [Google Scholar]
- 95.Nandi S., Singh T. Global Sensitivity Analysis on the Bergman Minimal Model. IFAC-PapersOnLine. 2020;53:16112–16118. doi: 10.1016/j.ifacol.2020.12.431. [DOI] [Google Scholar]
- 96.Bergman R.N. Origins and History of the Minimal Model of Glucose Regulation. Front. Endocrinol. 2021;11:583016. doi: 10.3389/fendo.2020.583016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Rudolph J. Well-formed dynamics under quasi-static state feedback. Banach Cent. Publ. 1995;32:349–360. doi: 10.4064/-32-1-349-360. [DOI] [Google Scholar]
- 98.Crespo T., Hajto Z., Mohseni R. Real Liouvillian extensions of partial differential fields. Symmetry Integr. Geom. Methods Appl. 2021;17:095. doi: 10.3842/SIGMA.2021.095. [DOI] [Google Scholar]
- 99.Besançon G. Nonlinear Observers and Applications. Springer; Berlin/Heidelberg, Germany: 2007. [DOI] [Google Scholar]
- 100.Isidori A. Nonlinear Control Theory for Automation. In: Nof S.Y., editor. Springer Handbook of Automation. Springer International Publishing; Cham, Switzerland: 2023. pp. 163–187. [DOI] [Google Scholar]
- 101.Diop S. Some Control Observation Problems and Their Differential Algebraic Partial Solutions. In: Quadrat A., Zerz E., editors. Algebraic and Symbolic Computation Methods in Dynamical Systems. Springer International Publishing; Cham, Switzerland: 2020. pp. 147–160. [DOI] [Google Scholar]
- 102.Sira-Ramirez H.J., Garcia Rodriguez C., Cortes Romero J.A., Luviano Juarez A. Algebraic Identification and Estimation Methods in Feedback Control Systems. John Wiley & Sons; Nashville, TN, USA: 2014. Wiley Series in Dynamics and Control of Electromechanical Systems. [Google Scholar]
- 103.Join C., Delaleau E., Fliess M. Flatness-based control revisited: The HEOL setting. Comptes Rendus. Mathématique. 2024;362:1693–1706. doi: 10.5802/crmath.674. [DOI] [Google Scholar]
- 104.Greco L., Mounier H., Bekcheva M. An approximate characterisation of the set of feasible trajectories for constrained flat systems. Automatica. 2022;144:110484. doi: 10.1016/j.automatica.2022.110484. [DOI] [Google Scholar]
- 105.Beaver L.E., Malikopoulos A.A. Optimal control of differentially flat systems is surprisingly easy. Automatica. 2024;159:111404. doi: 10.1016/j.automatica.2023.111404. [DOI] [Google Scholar]
- 106.Join C., Delaleau E., Fliess M. The Euler-Lagrange Equation and Optimal Control: Preliminary Results; Proceedings of the 2024 12th International Conference on Systems and Control (ICSC); Batna, Algeria. 3–5 November 2024; Piscataway, NJ, USA: IEEE; 2024. pp. 155–160. [DOI] [Google Scholar]
- 107.Join C., Delaleau E., Fliess M. Model-Free Predictive Control: Introductory Algebraic Calculations, and a Comparison with HEOL and ANNs. IFAC-PapersOnLine. 2025;59:255–260. doi: 10.1016/j.ifacol.2025.10.044. [DOI] [Google Scholar]
- 108.Richalet J., Rault A., Testud J., Papon J. Model predictive heuristic control: Applications to industrial processes. Automatica. 1978;14:413–428. doi: 10.1016/0005-1098(78)90001-8. [DOI] [Google Scholar]
- 109.Mayne D.Q. Model predictive control: Recent developments and future promise. Automatica. 2014;50:2967–2986. doi: 10.1016/j.automatica.2014.10.128. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data and code presented in this study are openly available in https://github.com/hugues-mounier/AI-FP-differential-flatness-and-smooth-random-realisation/ (accessed on 29 October 2025).











