Skip to main content
bioRxiv
  • Home
  • About
  • Submit
  • ALERTS / RSS
Advanced Search
New Results

Reduction of a Stochastic Model of Gene Expression: Lagrangian Dynamics Gives Access to Basins of Attraction as Cell Types and Metastabilty

View ORCID ProfileElias Ventre, View ORCID ProfileThibault Espinasse, View ORCID ProfileCharles-Edouard Bréhier, View ORCID ProfileVincent Calvez, View ORCID ProfileThomas Lepoutre, View ORCID ProfileOlivier Gandrillon
doi: https://doi.org/10.1101/2020.09.04.283176
Elias Ventre
1Univ Lyon, ENS de Lyon, Univ Claude Bernard, CNRS UMR 5239, INSERM U1210, Laboratory of Biology and Modelling of the Cell, 46 allée d’Italie, Site Jacques Monod, F-69007 Lyon, France
2Inria Team Dracula, Inria Center Grenoble Rhone-Alpes, Grenoble, France
3Univ Lyon, Université Claude Bernard Lyon 1, CNRS UMR 5208, Institut Camille Jordan, 43 blvd du 11 novembre 1918, F-6962 Villeurbanne, France
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Elias Ventre
  • For correspondence: elias.ventre@ens-lyon.fr
Thibault Espinasse
2Inria Team Dracula, Inria Center Grenoble Rhone-Alpes, Grenoble, France
3Univ Lyon, Université Claude Bernard Lyon 1, CNRS UMR 5208, Institut Camille Jordan, 43 blvd du 11 novembre 1918, F-6962 Villeurbanne, France
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Thibault Espinasse
Charles-Edouard Bréhier
3Univ Lyon, Université Claude Bernard Lyon 1, CNRS UMR 5208, Institut Camille Jordan, 43 blvd du 11 novembre 1918, F-6962 Villeurbanne, France
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Charles-Edouard Bréhier
Vincent Calvez
2Inria Team Dracula, Inria Center Grenoble Rhone-Alpes, Grenoble, France
3Univ Lyon, Université Claude Bernard Lyon 1, CNRS UMR 5208, Institut Camille Jordan, 43 blvd du 11 novembre 1918, F-6962 Villeurbanne, France
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Vincent Calvez
Thomas Lepoutre
2Inria Team Dracula, Inria Center Grenoble Rhone-Alpes, Grenoble, France
3Univ Lyon, Université Claude Bernard Lyon 1, CNRS UMR 5208, Institut Camille Jordan, 43 blvd du 11 novembre 1918, F-6962 Villeurbanne, France
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Thomas Lepoutre
Olivier Gandrillon
1Univ Lyon, ENS de Lyon, Univ Claude Bernard, CNRS UMR 5239, INSERM U1210, Laboratory of Biology and Modelling of the Cell, 46 allée d’Italie, Site Jacques Monod, F-69007 Lyon, France
2Inria Team Dracula, Inria Center Grenoble Rhone-Alpes, Grenoble, France
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Olivier Gandrillon
  • Abstract
  • Full Text
  • Info/History
  • Metrics
  • Preview PDF
Loading

Abstract

Differentiation is the process whereby a cell acquires a specific phenotype, by differential gene expression as a function of time. This is thought to result from the dynamical functioning of an underlying Gene Regulatory Network (GRN). The precise path from the stochastic GRN behavior to the resulting cell state is still an open question. In this work we propose to reduce a stochastic model of gene expression, where a cell is represented by a vector in a continuous space of gene expression, to a discrete coarse-grained model on a limited number of cell types. We develop analytical results and numerical tools to perform this reduction for a specific model characterizing the evolution of a cell by a system of piecewise deterministic Markov processes (PDMP). Solving a spectral problem, we find the explicit variational form of the rate function associated to a large deviations principle, for any number of genes. The resulting Lagrangian dynamics allows us to define a deterministic limit of which the basins of attraction can be identified to cellular types. In this context the quasipotential, describing the transitions between these basins in the weak noise limit, can be defined as the unique solution of an Hamilton-Jacobi equation under a particular constraint. We develop a numerical method for approximating the coarse-grained model parameters, and show its accuracy for a symmetric toggle-switch network. We deduce from the reduced model an approximation of the stationary distribution of the PDMP system, which appears as a Beta mixture. Altogether those results establish a rigorous frame for connecting GRN behavior to the resulting cellular behavior, including the calculation of the probability of jumps between cell types.

Introduction

Differentiation is the process whereby a cell acquires a specific phenotype, by differential gene expression as a function of time. Measuring how gene expression changes as differentiation proceeds is therefore of essence to understand this process. Advances in measurement technologies now allow to obtain gene expression levels at the single cell level. It offers a much more accurate view than population-based measurements, that has been obscured by mean population-based averaging [1], [2]. It has been established that there is a high cell-to-cell variability in gene expression, and that this variability has to be taken into account when investigating a differentiation process at the single-cell level [3], [4], [5], [6], [7], [8], [9], [10], [11].

A popular vision of the cellular evolution during differentiation, introduced by Waddington in [12], is to compare cells to marbles following probabilistic trajectories, as they roll through a developmental landscape of ridges and valleys. These trajectories are represented in the gene expression space: a cell can be described by a vector, each coordinate of which represents the expression of a gene [13], [14]. Thus, the state of a cell is characterized by its position in the gene expression space, i.e its specific level for all of its expressed genes. This landscape is often considered to be shaped by the underlying gene regulatory network (GRN), the behavior of which can be influenced by many factors, such as proliferation or cell-to-cell communication. Theoretically, the number of states a cell can take is equal to the number of possible combination of protein quantities associated to each gene. This number is potentially huge [15]. But metastability seems inherent to cell differentiation processes, as evidenced by limited number of existing cellular phenotypes [16], [17], providing a rationale for dimension reduction approaches [18]. Indeed, since [19] and [20], many authors have identified cell types with the basins of attraction of a dynamical system modeling the differentiation process, although the very concept of “cell type” has to be interrogated in the era of single-cell omics [21].

Adapting this identification for characterizing metastability in the case of stochastic models of gene expression has been studied mostly in the context of stochastic diffusion processes [22], [23], [24], but also for stochastic hybrid systems [25]. In the weak noise limit, a natural development of this analysis consists in describing the transitions between different macrostates within the large deviations framework [26], [27].

We are going to apply this strategy for a piecewise-deterministic Markov process (PDMP) describing GRN dynamics within a single cell, introduced in [28], which corresponds accurately to the non-Gaussian distribution of single–cell gene expression data. Using the work of [29], the novelty of this article is to provide analytical results for characterizing the metastable behavior of the model for any number of genes, and to combine them with a numerical analysis for performing the reduction of the model in a coarse-grained discrete process on cell types. We detail the model in Section 1, and we present in Section 2 how the reduction of this model in a continuous-time Markov chain on cell types allows to characterize the notion of metastability. For an arbitrary network, we provide in Section 3.1 a numerical method for approximating each transition rate of this coarse-grained model, depending on the probability of a rare event. In Section 3.2, we show that this probability is linked to a large deviations principle. The main contribution of this article is to derive in Section 4.1 the explicit variational form of the rate function associated to a Large deviations principle (LDP) for this model. We discuss in Sections 4.2 and 4.3 the conditions for which a unique quasipotential exists and allows to describe transitions between basins. We replace in Section 4.4 these results in the context of studying metastability. Finally, we apply in Section 5 the general results to a toggle-switch network. We also discuss in Section 6.1 some notions of energy associated to the LDP and we propose in Section 6.2 a non-Gaussian mixture model for approximating proteins distribution.

1 Model description

The model which is used throughout this article is based on a hybrid version of the well-established two-state model of gene expression [30], [31]. A gene is described by the state of the promoter, which can be {on, off}. If the state of the promoter is on, mRNAs are transcripted and translated into proteins, which are considered to be produced at a rate s. If the state of the promoter is off, only degradation of proteins occurs at a rate d (see Figure 1). kon and koff denote the exponential rates of transition between the states on and off. This model is a reduction of a mechanistic model including both mRNA and proteins, which is described in Appendix A.

Figure 1:
  • Download figure
  • Open in new tab
Figure 1:

Simplified two-states model of gene expression [28], [31].

Neglecting the molecular noise associated to proteins quantity, we obtain the hybrid model: Embedded Image where E(t) denotes the promoter state, P (t) denotes the protein concentration at time t, and we identify the state off with 0, the state on with 1.

The key idea for studying a GRN is to embed this two-states model into a network. Denoting the number of genes by n, the vector (E, P) describing the process is then of dimension 2n. The jump rates for each gene i are expressed in terms of two specific functions kon,i and koff,i. To take into account the interactions between the genes, we consider that for all i = 1, · · ·, n, kon,i is a function which depends on the full vector P via the GRN, represented by a matrix Θ of size n. We assume that kon,i is upper and lower bounded by a positive constant for all i. The function is chosen such that if gene i activates gene j, then Embedded Image. For the sake of simplicity, we consider that koff,i does not depend on the protein level.

We introduce a typical time scale Embedded Image for the rates of promoters activation kon,i, and a typical time scale Embedded Image for the rates of proteins degradation. Then, we define the scaling factor Embedded Image which characterizes the difference in dynamics between two processes: 1. gene bursting dynamics and 2. protein dynamics. It is generally considered that promoter switches are fast with respect to protein dynamics, i.e that ε ≪ 1, at least for eukaryotes [32]. Driven by biological considerations, we will consider values of ε smaller than 1/5 (see Appendix A).

We then rescale the time of the process by Embedded Image. We also rescale the quantities kon,i and koff,i by Embedded Image, and di by Embedded Image, for any gene i, in order to simplify the notations. Finally, the parameters si can be removed by a simple rescaling of the protein concentration Pi for every gene by its equilibrium value when Ei = 1 (see [28] for more details). We obtain a reduced dimensionless PDMP system modeling the expression of n genes in a single cell: Embedded Image

Here, X describes the protein vector in the renormalized gene expression space Ω := (0, 1)n and E describes the promoters state, in PE := {0, 1}n. We will refer to this model, that we will use throughout this article, as the PDMP system.

As card(PE) = 2n, we can write the joint probability density u(t, e, x) of (Et, Xt) as a 2n-dimensional vector Embedded Image. The master equation on u can be written: Embedded Image

For all i = 1, · · ·, n, for all x ∈ Ω, Fi(x) and Ki(x) are matrices of size 2n. Each Fi is diagonal, and the term on a line associated to a promoter state e corresponds to the drift of gene i: di(ei − xi). Ki is not diagonal: each state e is coupled with every state e’ such that only the coordinate ei changes in e, from 1 to 0 or conversely. Each of these matrices can be expressed as a tensorial product of (n − 1) two-dimensional identity matrices with a two-dimensional matrix corresponding to the operator associated to an isolated gene: Embedded Image

We detail in Appendix B the case of n = 2 for a better understanding of this tensorial expression.

2 Model reduction in the small noise limit

2.1 Deterministic approximation

The model (1) describes the promoter state of every gene i at every time as a Bernoulli random variable. We use the biological fact that promoter switches are frequent compared to protein dynamic, i.e ε < 1 with the previous notations. When ε ≪ 1, we can approximate the conditional distribution of the promoters knowing proteins, ρ, by its quasistationary approximation Embedded Image: Embedded Image which is derived from the stationary distribution of the Markov chain on the promoters states, defined for a given value of the protein vector X = x by the matrix Embedded Image (see [33], [34]). Thus, the PDMP model (1) can be coarsely approximated by a system of ordinary differential equations: Embedded Image

Intuitively, these trajectories correspond to the mean behaviour of a cell in the weak noise limit, i.e when promoters jump much faster than proteins concentration changes. More precisely, a random path Embedded Image converges in probability to a trajectory ϕt solution of the system (4), when ε → 0 [35]. The diffusion limit, which keeps a residual noise scaled by Embedded Image, can also be rigorously derived from the PDMP system [36], which is detailed in Appendix C.1.

In the sequel, we assume that every limit set of a trajectory solution of the system (4) as t → +∞ is reduced to a single equilibrium point, described by one of the solutions of: Embedded Image

Note that the condition above strongly depends on the interaction functions {kon,i}i=1,···,n. Alternatively speaking, in this work we rule out the existence of attractive limit cycles or more complicated orbits. We also assume that the closure of the basins of attraction which are associated to the stable equilibria of the system (5) covers the gene expression space Ω.

Without noise, the fate of a cell trajectory is fully characterized by its initial state x0. Generically, it converges to the attractor of the basin of attraction it belongs to, which is a single point by assumption. However, noise can modify the deterministic trajectories in at least two ways. First, in short times, a stochastic trajectory can deviate significantly from the deterministic one. In the case of a single, global, attractor, the deterministic system generally allows to retrieve the global dynamics of the process, i.e the equilibrium and the order of convergence between the different genes, for realistic ε (see Figure 2).

Figure 2:
  • Download figure
  • Open in new tab
Figure 2:

Comparison between the average on 100 simulated trajectories with ε = 1/7 (2a), ε = 1/30 (2b) and the trajectories generated by the deterministic system (2c) for a single pathway network: gene 1 → gene 2 → gene 3.

Second, in long times, stochastic dynamics can even push the trajectory out of the basin of attraction of one equilibrium state to another one, changing radically the fate of the cell. These transitions cannot be catched by the deterministic limit, and happen on a time scale which is expected to be of the order of Embedded Image (owing to a Large deviations principle studied below), where C is an unknown constant depending on the basins. In Figure 3a, we illustrate this situation for a toggle-switch network of two genes. We observe possible transitions between three basins of attraction. Two examples of random paths, the stochastic evolution of promoters and proteins along time, are represented in Figures 3b and 3c for different values of ε. All the details on the interaction functions and the parameters used for this network can be found respectively in the Appendices D and E.

Figure 3:
  • Download figure
  • Open in new tab
Figure 3:

3a: Phase portrait of the deterministic approximation for a symmetric toggle-switch with strong inhibition: two genes which activate themselves and inhibit each other. 3b: Example of a stochastic trajectory generated by the toggle-switch, for ε = 1/7. 3c: Example of a stochastic trajectory generated by the toggle-switch, for ε = 1/30.

2.2 Metastability

When the parameter ε is small, transitions from one basin of attraction to another are rare events: in fact the mean escape time from each basin is much larger than the time required to reach a local equilibrium (quasi-stationary state) in the basin.

Adopting the paradigm of metastability mentioned in the introduction, we identify each cell type to a basin of attraction associated to a stable equilibrium of the deterministic system (4). In this point of view, a cell type corresponds to a metastable sub-region of the gene expression space. It also corresponds to the notion of macrostate used in the theory of Markov State Models which has been recently applied to a discrete cell differentiation model in [37]. Provided we can describe accurately the rates of transition between the basins on the long run, the process can then be coarsely reduced to a new discrete process on the cell types.

More precisely, let m be the number of stable equilibrium points of the system (4), called attractors. We denote Z the set of the m basins of attraction associated to these m attractors, that we order arbitrarily: Z = {Z1, · · ·, Zm}. The attractors are denoted by Embedded Image. Each attractor is associated to a unique basin of attraction. By assumption, the closure of these m basins, Embedded Image, covers the gene expression space Ω. To obtain an explicit characterization of the metastable behavior, we are going to build a discrete process Embedded Image, with values in Z. From a random path Embedded Image of the PDMP system such that Embedded Image, we define a discrete process Embedded Image describing the cell types: Embedded Image where Embedded Image is a sequence of stopping times defined by: Embedded Image. Note that Embedded Image are the successive metastable states, and that Embedded Image are the successive times of transition between them. From the convergence of any random path to a solution of the deterministic system (4), that we mentioned in Section 2.1, we know that for every basin Zi such that Embedded Image, whatever is the point on the boundary of Zi which has been first attained, Embedded Image reaches any small neighborhood of the attractor Embedded Image of Zi before leaving the basin, with probability converging to 1 as ε → 0. In addition, for any basin Zj, the probability Embedded Image is asymptotically independent of the value of Embedded Image, and is then asymptotically determined by the value of Embedded Image. In other words, Embedded Image converges to a Markov chain when ε → 0. We refer to [38] to go further in the analysis of the coupling between the processes Embedded Image and Xε for general Markov processes.

For small ε, it is natural to approximate the distribution of the exit time from a basin by an exponential distribution. The in silico distribution represented in Figure 4 suggests that this assumption seems accurate for the toggle-switch network, even for a realistic value of ε. Note, however, that the exponential approximation slightly overestimates the probability that the exit times are small.

Figure 4:
  • Download figure
  • Open in new tab
Figure 4:

Comparison between the distribution of the exit time from a basin, obtained with a Monte-Carlo method, and the exponential law with appropriate expected value, for ε = 1/7. We represent the two densities in normal scale (on the left-hand side) and in logarithmic scale (on the right-hand side) to observe that the exponential law overestimates the probability that the exit times are small.

To completely characterize the coarse-grained resulting process, it remains to compute the transition rates Embedded Image of the time-continuous Markov chain on the basins, that we define for all pair of basins (Zi, Zj) ∈ Z2, i ≠ j, by: Embedded Image where Embedded Image is called the Mean First Exit Time of exit from Zi. This Markov process with discrete state space Z represents accurately, when ε is small enough, the main dynamics of the metastable system in the weak noise limit [39]. This reduced model is fully described by m2 transition rates: when the number of genes n is large, it is significantly smaller than the n2 parameters characterizing the GRN model (see Appendix D).

This collection of transition rates are characterized by rare events: when ε ≪ 1 or when the number of genes is large, it would be too expensive to compute them with a crude Monte-Carlo method. We are then going to present a method for approximating these transition rates from probabilities of some rare events. We will detail afterwards how these probabilities can be computed either by an efficient numerical method or an analytical approximation.

3 Computing the transition rates

3.1 Transition rates from probabilities of rare events

In this Section, we approximate each transition rate between any pair of basins (Zi,Zj), j ≠ i in terms of the probability that a random path realizes a certain rare event in the weak noise limit.

Let us consider two small parameters r, R such that 0 < r < R. We denote Embedded Image the r-neighborhood of Embedded Image, and Embedded Image its R-neighborhood. For a random path Embedded Image of the PDMP system starting in Embedded Image, we denote the probability of reaching a basin Zj, j ≠ i before any other basin Zk, k ≠ i, j, and before entering in Embedded Image: Embedded Image where Embedded Image is the hitting time of a set A ⊂ Ω.

The method developed in [40] aims to show how it is possible, from the knowledge of the probability (7), to approximate the transition rate Embedded Image presented in (6). Briefly, it consists in cutting a transition path into two pieces, a piece going from Embedded Image to Embedded Image and another reaching Zj from Embedded Image: the transition rates aij can be then approximated by the reverse of the mean number of attempts of reaching Zj from Embedded Image before entering in Embedded Image, which is close to the inverse of the rare event probability given by (7) when Embedded Image, multiplied by the average time of each excursion, that we denote Embedded Image. We obtain: Embedded Image

It is worth noticing that to be rigorous, this method need to redefine the neighborhoods Embedded Image and Embedded Image by substituting to the squared euclidean distance a new function based on the probability of reaching the (unknown) boundary: Embedded Image. The details are provided in Appendix F.

We observe that the average time Embedded Image can be easily computed by a crude Monte-Carlo method: indeed, the trajectories entering in Embedded Image are not rare. It thus only remains to explain the last ingredient for the approximation of the transition rates, which is how to estimate the probabilities of the form (7).

3.2 Computing probabilities of the form (7)

A powerful method for computing probabilities of rare events like (7), is given by splitting algorithms. We decide to adapt the Adaptative Multilevel Splitting Algorithm (AMS) described in [41] to the PDMP system: all the details concerning this algorithm can be found in Appendix G. In Section 5, we will verify that the probabilities given by the AMS algorithm are consistent with the ones obtained by a crude Monte-Carlo method for the toggle-switch network.

However, estimating the probability (7) becomes hard when both the number of genes of interest increases and ε decreases. Indeed, the AMS algorithm allows to compute probabilities much smaller than the ones we expect for biologically relevant parameters (ε ≈ 0.1), but the needed number of samples grows at least with a polynomial factor in ε−1. If the number of genes considered is large, these simulations can make the algorithm impossible to run in a reasonable time. A precise analysis of the scalability of this method for the PDMP system is beyond the scope of this article, but we have been able to get consistent results on a laptop in time less than one hour for a network of 5 genes, with ε > 1/15. The resulting probabilities were of order 5.10−3.

In order to overcome this problem, we are now going to develop an analytical approach for approximating these probabilities, by means of optimal trajectories exiting each basin of attraction. The later can be computed within the context of Large deviations. As we will see in Section 5, this approach is complementary to the AMS algorithm, and consistent with it.

3.2.1 Large deviations setting

In this section, we derive a variational principle for approximating the transition probability (7) introduced in Section 3.1. A powerful methodology for rigorously deriving a variational principle for optimal paths is Large deviations theory. It has been developed extensively within the context of Stochastic Differential Equations (SDE) [39][42]. For the sake of simplicity, we present here only an heuristic version. There exists a Large deviations principle (LDP) for a stochastic process with value in Ω if for all x0 ∈ Ω there exists a lower semi-continuous function defined on the set of continuous trajectories from [0, T] to Ω, JT : C0T (ℝn) → [0, ∞], such that for all set of trajectories A ⊂ C0T (ℝn): Embedded Image

The function JT is called the rate function of the process in [0, T], and the quantity JT (ϕ) is called the cost of the trajectory ϕ over [0, T].

The particular application of this theory to stochastic hybrid systems has been developed in detail in [43] and [35]. We now present consequences of results developed in [29].

Definition 1.

The Hamiltonian is the function H: Ω × ℝn ↦ ℝ, such that for all (x, p) ∈ Ω × ℝn, H(x, p) is the unique eigenvalue associated to a nonnegative right-eigenvector Embedded Image, (which is unique up to a normalization), of the following spectral problem: Embedded Image where the matrix Embedded Image is defined by: Embedded Image

We remark that the matrix M (x, p) has off-diagonal nonnegative coefficients. Moreover, the positivity of the functions kon,i makes M irreducible (the matrix allows transition between any pair Embedded Image after at most n steps). Thereby, the Perron Frobenius Theorem may be applied, and it justifies the existence and uniqueness of H(x, p). Moreover, from a general property of Perron eigenvalues when the variable p appears only on the diagonal of the matrix, H is known to be convex [44].

The following result is a direct consequence of theoretical results of [29] applied to the PDMP system (1).

Theorem 1.

Let us denote Embedded Image and Embedded Image its interior.

The Fenchel-Legendre transform of the Hamiltonian H is well-defined and satisfies: ∀x ∈ Ω, Embedded Image, Embedded Image

Moreover, the PDMP system (1) satisfies a LDP, and the associated rate function has the form of a classical action: the cost of any piecewise differentiable trajectory ϕt in C0T (ℝn) satisfying for all t ∈ [0, T), Embedded Image, can be expressed as Embedded Image

The function L is called the Lagrangian of the PDMP system.

We are now going to show how in certain cases, trajectories which minimize the quantity (12) between two sets in Ω can be defined with the help of solutions of an Hamilton-Jacobi equation with Hamiltonian H.

3.2.2 WKB approximation and Hamilton-Jacobi equation

The Hamiltonian defined in (10) also appears in the WKB (Wentzell, Kramer, Brillouin) approximation of the master equation [34], [29]. This approximation consists in injecting in the master equation (2) of the PDMP system, a solution of the form: Embedded Image where Embedded Image then denotes the marginal distribution on proteins of the distribution u at time t, and πe(x, t) is a probability vector denoting the conditional distribution of the promoters knowing that proteins are fixed to X = x at t. The expression (13) is justified under the assumption that the density ue is positive at all times.

Under the regularity assumptions S ∈ C1(Ω × ℝ+, ℝ) and Embedded Image, we can perform a Taylor expansion in ε of the functions S and πe, for any e ∈ PE, and keeping in the resulting master equation only the leading order terms in ε, that we denote S0 and Embedded Image, we obtain: Embedded Image

Identifying the vectors π0 and ∇xS0 with the variables ζ and p in equation (10), we obtain that S0 is solution of an Hamilton-Jacobi equation: Embedded Image

More precisely, if at any time t, the marginal distribution on proteins of the PDMP process, denoted u(·, t), follows a LDP and if its rate function, denoted V (·, t), is differentiable on Ω, then the function H(·, ∇xV (·, t)) appears as the time derivative of V (·, t) at t. Moreover, the WKB method presented above shows that the rate function V (·, t) is identified for any time t with the leading order approximation in ε of the function S(·, t) = −ε log(u(·, t)). Note that (13) is also reminiscent of the Gibbs distribution associated with a potential S. Some details about the interpretation of this equation and its link with the quasistationary approximation can be found in [34].

Next, we consider a function V, solution of the Hamilton-Jacobi equation (14), that we assume being of class C1(Ω × ℝ+, ℝ) for the sake of simplicity. Then, for any piecewise differentiable trajectory ϕt ∈ C0T (Ω) such that Embedded Image for all t ∈ [0, T), one has, by definition of the Fenchel-Legendre transform: Embedded Image

Moreover, when H is strictly convex in p, we have: Embedded Image

Then, the equality in (15) is exactly reached at any time for trajectories ϕt such that for all t ∈ [0, T), i = 1, · · ·, n: Embedded Image

3.2.3 General method for computing probabilities of the form (7)

We now detail the link existing between the regular solutions V of the Hamilton-Jacobi equation (14) and the probabilities of the form (7). For this, we introduce the notion of quasipotential.

Definition 2.

Denoting Embedded Image the set of piecewise differentiable trajectories in C0T (Ω), we define the quasipotential as follows: for two sets A, B ⊂ Ω and a set R ⊂ Ω \ (A ∪ B), Embedded Image

We call a trajectory Embedded Image an optimal trajectory between the two subsets A, B ⊂ Ω in Ω \ R, if it reaches the previous infimum.

For the sake of simplicity, if R = ∅, we will write QR(A, B) = Q(A, B).

With these notations, the LDP principle allows to approximate for any basin Zj, i ≠ j, the probability Embedded Image defined in (7), which is the probability of reaching Zj, from a point Embedded Image Embedded Image, before Embedded Image, by the expression: Embedded Image

A direct consequence of the inequality (15), in the case where equality is reached, is that a regular solution V of the Hamilton-Jacobi equation (14) defines trajectories for which the cost (12) is minimal between any pair of its points. Moreover, if V is a stationary solution of (14), the cost of such trajectories does not depend on time: these trajectories are then optimal between any pair of its points among every trajectory in any time. We immediately deduce the following lemma:

Lemma 1.

For a stationary solution V ∈ C1(Ω, ℝ) of (14) and for all T > 0, any trajectory Embedded Image satisfying the system (16) associated to V is optimal in Ω between ϕ(0) and ϕ(T), and we have: Embedded Image

Thus, for approximating the probability of interest (7), between any pair of basin (Zi, Zj), we are going to build a trajectory Embedded Image, which verifies the system (16) associated to a stationary solution V of (14), with Embedded Image, and which reaches in a time T a point x ∈ ∂Zi ∩ ∂Zj such that Embedded Image. For such trajectory, from Lemma 1, we could then approximate the probability (7) by the formula Embedded Image where Cij is an appropriate prefactor. Unfortunately, if there exists an explicit expression of Cij in the one-dimensional case [34], and that an approximation has been built for multi-dimensional SDE model [45], they are intractable or not applicable in our case. In general, the prefactor does not depend on ε [46]. In that case − ln Embedded Image is asymptotically an affine function of ε−1, the slope of which is Embedded Image and the initial value − ln(Cij). Then, the strategy we propose simply consists in approximating the prefactor by comparison between the probabilities given by the AMS algorithm and the Large deviations approximation (17) for a fixed ε (large enough to be numerically computed.)

To conclude, for every pair of basins (Zi, Zj), i ≠ j, one of the most efficient methods for computing the probability (7) is to use the AMS algorithm. When the dimension is large, and for values of ε which are too small for this algorithm to be efficiently run, we can use the LDP approximation (17) if the corresponding optimal trajectories Embedded Image can be explicitly found. The latter condition is studied in the next sections. The AMS algorithm is then still employed to approximate the prefactors, which is done using intermediate values of ε by the regression procedure mentioned above.

4 Analytical approximation of probabilities of the form (7) for the PDMP system

4.1 Expressions of the Hamiltonian and the Lagrangian

In this section, we identify the Perron eigenvalue H(x, p) of the spectral problem (10), and prove that its Fenchel-Legendre transform L with respect to the variable p is well defined on ℝn. We then obtain the explicit form of the Hamiltonian and the Lagrangian associated to the LDP for the PDMP system (1).

Theorem 2.

For all n in ℕ*, the Hamiltonian is expressed as follows: for all (x, p) ∈ Ω × ℝn, the unique solution of the spectral problem (10) (with nonnegative eigenvector) is: Embedded Image Moreover, the function H is strictly convex with respect to p.

Theorem 3.

The Lagrangian is expressed as follows: for all (x, v) ∈ Ω × ℝn, one has: Embedded Image In addition, for all x ∈ Ω, L(x, v) = 0 if and only if for all i = 1, · · ·, n: Embedded Image

As detailed in Appendix C.2, we remark that the Lagrangian of the PDMP process defined in (19) is not equal to the Lagrangian of the diffusion approximation defined in Appendix C.1, which is: Embedded Image

More precisely, the Lagrangian of the diffusion approximation is a second order approximation of the Taylor expansion of the Lagrangian of the PDMP system around the velocity field associated to the deterministic limit system (4). Observe that the Lagrangian of the diffusion approximation is a quadratic mapping in v, which is expected since the diffusion approximation is described by a Gaussian process. On the contrary, the Lagrangian L given by (19) is not quadratic in v. As it had been shown in [47] for Fast-Slow systems, this highlights the fact that the way rare events arise for the PDMP system is fundamentally different from the way they would arise if the dynamics of proteins was approximated by an SDE.

Proof of Theorem 2. Defining the 2 × 2 matrix Embedded Image the Perron eigenproblem associated to M (i) Embedded Image implies immediately that Embedded Image

If we impose the constraint Embedded Image, i.e that there exists for all x, pi, αp,i(x) ∈ (0, 1) such that Embedded Image, we obtain the following equation: Embedded Image

Since Tx(0) = −kon,i(x) and Tx(1) = koff,i, Tx has one and only one root in (0, 1). After a quick computation, one gets for all x, p ∈ Ω × ℝn: Embedded Image

Considering the tensorial structure of the problem, denoting Mi(x, p) = piFi(x) + Ki(x) (the tensorial version, see Section 1), we have by definition of M (11): Embedded Image

For Embedded Image, we obtain: Embedded Image

Since ζ > 0, one obtains the expression (18) for the Hamiltonian: Embedded Image

We verify that H is strongly convex with respect to p, which follows from the following computation: for all i, j = 1, · · ·, n, Embedded Image and the cross-derivatives are clearly 0. This concludes the proof of Theorem 2.

Proof of Theorem 3. The objective is to compute the Fenchel-Legendre transform of the Hamil-tonian H given by (18) in Theorem 2.

For all x ∈ Ω and for all vi ∈ ℝ, the function g: Pi ↦ pivi − Hi(x, pi) is concave. An asymptotic expansion (when pi → ±∞) gives: Embedded Image

Let us study three cases. If vi ∈ (−dixi, di(1 − xi)), g goes to −∞ when pi → ±∞: thus g reaches a unique maximum in ℝ. At the boundary vi = −dixi (resp. vi = di(1 − xi)), g goes to −∞ as Pi goes to +∞ (resp. −∞) and converges to kon,i(x) (resp. koff,i) as Pi goes to −∞ (resp. +∞): then g is upper bounded and the sup is well defined. If vi ∉ [−dixi, di(1 − xi)], g(Pi) goes to +∞ when either Pi → −∞ of Pi → +∞, thus g is not bounded from above. As a consequence, Embedded Image is finite if and only if vi ∈ [−dixi,di(1 − xi)].

The Fenchel-Legendre transform of H is then given as follows: for all x ∈ Ω and v ∈ ℝn Embedded Image and L(x, v) is finite for every v ∈ Ωv(x). To find an expression for L(x, v), we have to find for all i = 1, · · ·, n the unique solution pv,i(x) of the invertible equation: Embedded Image. Developing the term on the right-hand side, we obtain: Embedded Image where ui = 2(vi + dixi) − di, ci = 4kon,i(x)koff,i > 0, zi = dipi + kon,i(x) − koff,i.

When vi ∈ (−dixi, di(1 − xi)), we have ui ∈ (−di, di). Thus, we obtain Embedded Image and as Zi and ui must have the same sign, we can conclude for every vi ∈ (−dixi, di(1 − xi)): Embedded Image

Injecting this formula in the expression of the Fenchel-Legendre transform, we obtain after straightforward computations: Embedded Image

We finally obtain the expression when v ∈ Ωv(x): Embedded Image

Finally, if v ∉ Ωv(x), i.e. if there exists i such that vi ∉ [−dixi, di(1 − xi)], then L(x, v) = Li(xi, vi) = ∞. As expected, the Lagrangian is always nonnegative. In addition, it is immediate to check that L(x, v) = 0 if and only if the velocity field v is the drift of the deterministic trajectories, defined by the system (4).

4.2 Stationary Hamilton-Jacobi equation

We justified in Section 3.2.3 that the stationary solutions of the Hamilton-Jacobi equation (14) are central for finding an analytical approximation of the transition rates described in Section 3.1. Thus, we are going to study the existence and uniqueness (under some conditions) of functions V ∈ C1(Ω, ℝ) such that for all x ∈ Ω: Embedded Image

Recalling that from (21), Embedded Image, we construct two classes of solutions V, such that for all Embedded Image or Embedded Image.

The first class of solutions contains all the constant functions on Ω. From the expression (20). The second class contains all functions V such that for all x ∈ Ω: Embedded Image

In particular, we show in Appendix D that the condition (26) holds for the toggle-switch network described in Appendix E and studied in Section 5.

We will see in the next section that the class of constant solutions are associated to the deterministic system (4), which are the trajectories of convergence within the basins. We describe in Section 4.3.2 a more general class of solutions than (26), which defines the optimal trajectories of exit from the basins of attraction of the deterministic system.

4.3 Optimal trajectories

In the sequel we study some properties of the optimal trajectories associated to the two classes of solutions of the stationary Hamilton-Jacobi equation (25) introduced above.

4.3.1 Deterministic limit and relaxation trajectories

From Lemma 1, for every constant function V (·) = C on Ω, the associated collection of paths ϕt satisfying the system (16) is optimal in Ω between any pair of its points. Replacing p = ∇xV = 0 in (16), we find that these trajectories verify at any time t > 0 the deterministic limit system (4): Embedded Image

Moreover, for every trajectory ϕt solution of this system, we have for any T > 0: Embedded Image

We call such trajectories the relaxation trajectories, as they characterize the optimal path of convergence within every basin. From Theorem 3, these relaxation trajectories are the only zero-cost admissible trajectories.

4.3.2 Quasipotential and fluctuation trajectories

We now characterize the optimal trajectories of exit from the basins. We are going to show that the condition (C) defined below is sufficient for a solution V ∈ C1(Ω, ℝ) of the equation (25) to define optimal trajectories realizing the rare events described by the probabilities (7).

Definition 3.

We define the following condition on a function V ∈ C1(Ω, ℝ): (C) The set {x ∈ Ω | ∇xV (x) = 0} is reduced to isolated points.

The results presented below in Theorem 4 are mainly an adaptation of Theorem 3.1, Chapter 4, in [39]. In this first Theorem, we state some properties of solutions V ∈ C1(Ω, ℝ) of (25) satisfying the condition (C):

Theorem 4.

Let V ∈ C1(Ω, ℝ) be a solution of (25).

  1. For any optimal trajectory ϕt satisfying the system (16) associated to V, for any time t we have the equivalence: Embedded Image

  2. The condition (C) implies that the gradient of V vanishes only on the stationary points of the system (4).

  3. If V satisfies (C), then V is strictly increasing on any trajectory which solves the system (16), such that the initial condition is not an equilibrium point of the system (4). Moreover, for any basin of attraction Zi associated to an attractor Embedded Image, we have: Embedded Image

  4. If V satisfies the condition (C), under the assumption that Embedded Image, we have the formula: Embedded Image

  5. Let us consider V, Embedded Image two solutions of (25) satisfying the condition (C). The stable equilibria of the system defined for every time t by Embedded Image are exactly the attractor of the deterministic system (4) Embedded Image. We denote Embedded Image the basins of attraction which are associated to these equilibria: at least on Embedded Image, the relation Embedded Image is satisfied.

Moreover, under the assumptions 1. that Embedded Image, and 2. that between any pair of basins Embedded Image, we can build a serie of basins Embedded Image such that u0 = 1, um = j and for all k < m, Embedded Image, then V and Embedded Image are equal in Ω up to a constant.

Note that the point (iii) makes these solutions consistent with the interpretation of the function V as the rate function associated to the stationary distribution of the PDMP system, presented in Section 3.2.2. Indeed, as every random path converges in probability when ε → 0 to the solutions of the deterministic system (4) [35], the rate function has to be minimal on the attractors of this system, which then corresponds to the points of maximum likelihood at the steady state. It should also converge to +∞ on ∂Ω, as the cost of reaching any point of the boundary is infinite (see Corollary 2, in the proof of Theorem 5). However, we see in (v) that the uniqueness, up to a constant, needs an additional condition on the connection between basins which remains not clear for us at this stage, and which will be the subject of future works.

If V ∈ C1(Ω, ℝ) is a solution of (25) saitsfying (C), we call a trajectory solution of the system (16) associated to V a fluctuation trajectory.

We observe that any function satisfying the relation (26) belongs to this class of solutions of (25), and then that in particular, such C1 function exists for the toggle-switch network. In that special case, we can explicitly describe all the fluctuation trajectories: for any time t, replacing Embedded Image in the system (16), we obtain Embedded Image

In the second theorem (Theorem 5), we justify that the fluctuation trajectories are the optimal trajectories of exit:

Theorem 5.

Let us assume that there exists a function V ∈ C1(Ω, ℝ) which is solution of (25) and satisfies the condition (C). For any basin Zi ∈ Z, there exists at least one basin Zj, j ≠ i, such that there exists a couple (x0, ϕt), where Embedded Image and ϕt is a fluctuation trajectory, and such that ϕ(0) = x0, Embedded Image.

Let us denote Embedded Image and Embedded Image Embedded Image. Under the following assumption y∈Xij

(A) any relaxation trajectory starting in ∂Zi ∩ ∂Zj stays in ∂Zi ∩ ∂Zj, we have Embedded Image and: Embedded Image

In particular, if there exists a fluctuation trajectory between any attractor Embedded Image and every saddle points of the deterministic system (16) on the boundary ∂Zi, and if the assumption (A) of Theorem 5 is verified for every basin Zj, j ≠ i, the function V allows to quantify all the optimal costs of transition between the basins. This is generally expected because the attractors are the only stable equilibria for the reverse fluctuations (see the proof of Theorem 4.(v)). The proofs of Theorems 4 and 5 use classical tools from Hamiltonian system theory and are postponed to Appendix H.

When a solution V ∈ C1(Ω, ℝ) satisfying (C) exists, the saddle points of the deterministic system (4) are then generally the bottlenecks of transitions between basins and the function V characterizes the energetic barrier between them. The function Embedded Image depends on the basin Zi, which is a local property: it explains why the function V is generally called the global quasipotential, and Embedded Image the local quasipotential of the process [48].

The precise analysis of the existence of a regular solution V satisfying (C) for a given network is beyond the scope of this article. When it is impossible to find a regular solution, more general arguments developed within the context of Weak KAM Theory can allow to link the viscosity solutions of the Hamilton-Jacobi equation to the optimal trajectories in the gene expression space [49].

4.4 Partial conclusion

We have obtained in Theorem 3 the form of the Lagrangian in the variational expression (12) for the rate function JT associated to the LDP for the PDMP system (1). We have also highlighted the existence and interpretation of two types of optimal trajectories.

The first class consists in relaxation trajectories, which characterize the convergence within the basins. The fact that they are the only trajectories which have zero cost justifies that any random path Embedded Image converges in probability to a relaxation trajectory.

When there exists a function V ∈ C1(Ω, ℝ) satisfying (26), the system (28) defines the second class of optimal trajectories, called the fluctuation trajectories. From Theorem 5, for every basin Zi, there exists at least one basin Zj, j ≠ i, and a trajectory Embedded Image which verifies this system, starts on Embedded Image and reaches a point of x ∈ ∂Zi ∩ ∂Zj such that Q(x0, x) = Q(x0, ∂Zi ∩ ∂Zj). This trajectory then realizes the rare event of probability pij(x0). Injecting the velocity field defining (28) in the Lagrangian (19), we deduce: Embedded Image

If the assumption (A) of Theorem 5 is verified, this minimum is necessarily reached on a saddle point of V on ∂Zi ∩ ∂Zj and in that case, the time T must be taken infinite. Then, the formula (29) can be injected in the approximation (17), and the method described in Section 3.2.2 allows to compute the probability of the form (7) for the pair (Zi, Zj).

Moreover, for every basin Zk, k ≠ i, if the assumption (A) of Theorem 5 is verified and if there exists, for any saddle point Embedded Image, a trajectory satisfying the system (28) which starts at x0 and reaches Embedded Image (at T → ∞), the formula (29) can also be injected in the approximation (17) for the pair (Zi, Zk), and the method described in Section 3.2.3 allows then to compute the probabilities of the form (7) for any pair of basins (Zi, Zk)k=1,···,m.

5 Application to the toggle-switch network

In this section, we consider the class of interaction functions defined in Appendix D for a network with two genes (n = 2). This function comes from a chromatin model developed in [28] and is consistent with the classical Hill function characterizing promoters switches. Using results and methods described in the previous sections, we are going to reduce the PDMP system when the GRN is the toggle-switch network described in Appendix E. After defining the attractors of the deterministic system (4), building the optimal fluctuation trajectories between these attractors and the common boundaries of the basins, we will compute the cost of the trajectories and deduce, from the approximation (17), the transition probabilities of the form (7) as a function of ε, up to the prefactor. We will compute these probabilities for some ε with the AMS algorithm described in Appendix G for obtaining the prefactor. We will then approximate the transition rates characterizing the discrete Markov chain on the cellular types, given by the formula (8), for many values of ε. We will finally compare these results to the ones given by a crude Monte-Carlo method.

5.1 Computation of the attractors, saddle points and optimal trajectories

First, we compute the stable equilibrium points of the PDMP system (1). The system (5) has no explicit solution. We present a simple method to find them, which consists in sampling a collection of random paths in Ω: the distribution of their final position after a long time approximates the marginal on proteins of the stationary distribution. We use these final positions as starting points for simulating the relaxation trajectories, described by (4), with an ODE solver: each of these relaxation trajectories converges to one of the stable equilibrium points. This method allows to obtain all the stable equilibrium corresponding to sufficiently deep potential wells (see Figure 5). Possible other potential wells can be omitted because they correspond to basins where the process has very low probability of going, and which do not impact significantly the coarse-grained Markov model.

Figure 5:
  • Download figure
  • Open in new tab
Figure 5:

5a: 100 cells are plotted under the stationary distribution. The relaxation trajec-tories allow to link every cell to its associated attractor. 5b: 1000 cells are plotted under the stationary distribution. They are then classified depending on their attractor, and this figure sketches the kernel density estimation of proteins within each basin. 5c: The ratio of cells that are found within each basin gives an estimation of the stationary distribution on the basins.

Second, we need to characterize the fluctuation trajectories. In Appendix D, we introduced the interaction function and proved that for any symmetric two-dimensional network defined by this function, i.e such that for any pair of genes (i, j), θij = θji (where θ is the matrix characterizing the interactions between genes), there exists a function V such that the relation (26) is verified. This is then the case for the toggle-switch network, which is symmetric. We have proved in Section 4.2 that such function V solves the Hamilton-Jacobi equation (25), and verifies the condition (C). Thus, the system (28) defines the fluctuation trajectories.

Third, we need to find the saddle points of the system (4). As we know that for any attractor, there exists at least one fluctuation trajectory which starts on the attractor and reaches a saddle point (in an infinite time), a naive approach would consist in simulating many trajectories with different initial positions around every attractors, until reaching many saddle points of the system. This method is called a shooting method and may be very efficient in certain cases. But for the toggle-switch, we observe that the fluctuation trajectories are very unstable: this method does not allow to obtain the saddle points.

We develop a simple algorithm which uses the nonnegative function l(·) = L(·, νv(·)), which corresponds to the Lagrangian evaluated on the drift νv of the fluctuation trajectories defined by the system (28). We have: Embedded Image

As expected, since νv cannot be equal to the drift of a relaxation trajectory except on the stationary points of the relaxation trajectories and since the Lagrangian L(x, v) is equal to 0 if and only if v corresponds to the drift of a relaxation trajectory, the function l vanishes only on these stationary points. If there exists a saddle point connecting two attractors, this function will then vanish there. The algorithm is described in Appendix I. For the toggle-switch network, it allows to recover all the saddle points of the system (4).

Fourth, we want to compute the optimal trajectories between every attractors and the saddle points on the boundary of its associated basin. Using the reverse of the fluctuation trajectories, for which the attractors of the system (4) are asymptotically stable (see the proof of Theorem 4.(v)), we can successively apply a shooting method around every saddle points. We observe that for the toggle-switch network, for any saddle point at the boundary of two basins, there exists a reverse fluctuation trajectory which converges to the attractors of both basins. For any pair of basins (Zi, Zj), we then obtain the optimal trajectories connecting the attractor Embedded Image and the saddle points belonging to the common boundary ∂Zi ∩ ∂Zj (see Figure 6).

Figure 6:
  • Download figure
  • Open in new tab
Figure 6:

The optimal fluctuation trajectories from a first attractor continued by the relaxation trajectories reaching the second attractor, for the pair (Z+−, Z−+). We omit the other pairs of attractors (Z+−, Z−−) and (Z−−, Z−+), because their optimal trajectories are simply straight lines.

Finally, we want to compute the optimal transition cost between any pair of basins (Zi, Zj). We observe that every relaxation trajectories starting on the common boundary of two basins stay on this boundary and converge to a unique saddle point inside: the assumption (A) of Theorem 5 is then verified. It follows from this theorem that the optimal trajectory between any basin Zi and Zj necessarily reaches ∂Zi ∩ ∂Zj on a saddle point, and then that the optimal transition cost is given by the trajectory which minimizes the cost among all those found previously between the attractor and the saddle points. We denote this optimal trajectory Embedded Image. Its cost is explicitly described by the formula (29) (with T → ∞), which is then the optimal cost of transition between Zi and Zj.

The LDP ensures that for all δ, η > 0, there exists ε′ such that for all ε ∈ (0, ε′), a random path Embedded Image reaching Zj from Embedded Image before Embedded Image, verifies: Embedded Image with probability larger than 1 − η. In other words, given a level of resolution δ, we could then theoretically find ε such that any trajectory of exit from Zi to Zj would be indistinguishable from trajectory Embedded Image at this level of resolution. But in practice, the event Embedded Image is too rare to be simulated directly for such ε.

We plot in Figure 7 two sets of random exit paths, simulated for two different ε, illustrating the fact that the probability of an exit path to be far from the optimal fluctuation trajectory decreases with ε.

Figure 7:
  • Download figure
  • Open in new tab
Figure 7:

Comparison between the optimal trajectory of the figure 6 and 30 random paths conditioned on reaching, from a point of the boundary of the R-neighborhood of an attractor Embedded Image, the r-neighborhood of a new attractor Embedded Image before the r neighborhood of the first attractor Embedded Image, with r < R. We represent this comparison for 7a: ε = 1/7 and 7b: ε = 1/21. For each figure, one of these random paths is colored, separating the fluctuation and the relaxation parts.

5.2 Comparison between predictions and simulations

For each pair of basins (Zi, Zj), the expression (29) provides an approximation of the probability of the rare event Embedded Image, up to a prefactor, and the approximation (17) allows to deduce the associated transition rate. We plot in Figure 8 the evolution of these two estimations, as ε decreases, comparing respectively to the probabilities given by the AMS algorithm and the transition rates computed with a Monte-Carlo method. As in [50], we decide to plot these quantities in logarithmic scale. We observe that, knowing the prefactor, the Large deviations approximation is accurate even for ε > 0.1, and that induced transition rates are close to the ones observed with a Monte-Carlo method too. We represent in Figure 10b the variance of the estimator of the transition rates given by the AMS method.

Figure 8:
  • Download figure
  • Open in new tab
Figure 8:

8a: Comparison between the probabilities (7) between the basins Z+− and Z−−, in logarithmic scale, given by the Large deviations approximation (in red) and the AMS algorithm (in green). The prefactor is computed for ε = 1/8 and the red curve is then adjusted to fit the numerical results. The blue curve corresponds to the probabilities obtained with a Monte-Carlo method. 8b: Comparison between the transition rates between the basins Z+− and Z−−, in logarithmic scale, given by the formula (8), where the probability (7) is given by the Large deviations approximation (in red) and the AMS algorithm (in green). The blue curve corresponds to the transition rates obtained with a Monte-Carlo method, by the formula (6). The quantities obtained by a Monte-Carlo method, in blue, are not represented after ε = 1/8 because the transition rates become too small to be efficiently computed.

We also remark that our analysis provides two ways of estimating the stationary measure of the discrete coarse-grained model. On the one hand, we can obtain a long-time proteins distribution of thousands of cells by simulating the PDMP system (1) from random initial conditions: by identifying each cell with a basin, as shown in Figure 9a, we can find a vector μb describing the ratio of cells belonging to each basin. When the number and length of the simulations are large enough, this vector μb should be a good approximation of the stationary measure on the basins. On the other hand, the transition rates allows to build the transition matrix M of the discrete Markov process on the basins, Embedded Image, defined in Section 2.2. If the exponential approximation of the first passage time from every basin is accurate, then the stationary distribution on the basins should be well approximate by the unique probability vector such that μzM = 0 (see Figure 9b).

Figure 9:
  • Download figure
  • Open in new tab
Figure 9:

Comparison between the two methods for obtaining estimators of the stationary distributions on the basins: μb (9a) and μz (9b).

Figure 10:
  • Download figure
  • Open in new tab
Figure 10:

10a: The total variation of the difference between μb and μz as a function of ε−1. 10b: Boxplots representing the variation of the transition rates for 10 iterations of the method used in 10a, between each pair of basins for ε = 1/7.

Monte-Carlo methods for approximating the transition rates have a very high computational cost when ε is small. Thus, comparing these two stationary distributions appears as a good alternative for verifying the accuracy of the transition rates approximations. We plot in Figure 10a the evolution of the total variation distance between these two stationary distributions as ε decreases. We observe that the total variation is small even for realistic values of ε. The variance of the estimator μb is very small (given it is estimated after a time long enough) but the estimator μz accumulates all numerical errors coming from the estimators needed to compute the transition rates: this is likely to explain the unexpected small increases observed in this curve for ε = 1/6. We represent in Figure 10b the variance of the transition rates estimators between every pair of attractors used for estimating the distribution μz in Figure 10a, for ε = 1/7: as expected, this variance increases with the transition rates.

The similarity between the two distributions μz and μb seems to justify the Markovian approximation of the reduced process Embedded Image for small but realistic ε: at least for the toggle-switch network, the coarse-grained model, evolving on the basins of attractions seen as cellular types, describes accurately the complex behaviour of a cell in the gene expression space.

5.3 Applicability for more complex networks

It is in general very complex to find a solution V ∈ C1(Ω, ℝ) to the stationary Hamilton-Jacobi equation (25) which satisfies the condition (C) for general networks, when the number of genes is greater than 2. In order to apply the strategy developed in Section 4, for computing the cost of the optimal trajectories of transition between two basins, it would be then necessary to build a computational method for approximating such solution. Although the most common approach in this case consists in finding optimal trajectories without computing a potential (see [51] or [52] for more recent works), some methods have been recently built for SDEs model, like Langevin dynamics [53]. Such computational method for the PDMP system is beyond the scope of the article. However, we remark that even if there are no reasons for the trajectories satisfying the system (28) to be optimal when no function satisfying the relation (26) can be found, our computational method still allows to compute these trajectories, and we observe that they generally still bound the attractors and the saddle points of the deterministic system (4). Their costs can then be used as a proxy for the probabilities of the form (7): we observe in Figures 16b and 17b in Appendix J that for two non-symmetric networks of respectively 3 and 4 genes, our method still provides good results.

6 Discussion

Using the WKB approximation presented in Section 3.2.2 and the explicit formulas for the Hamiltonian and the Lagrangian detailed in Section 4.1, we are going now to analyze more precisely how the LDP for the proteins can be interpreted in regards to the dynamics of promoters, and we will see how two classical notions of energies can be interpreted in light of this analysis.

6.1 Correspondences between velocities and promoters frequency lead to energetic interpretations

The main idea behind the LDP principle for the PDMP system is that a slow dynamics on proteins coupled to the fast Markov chain on promoters rapidly samples the different states of PE according to some probability measure Embedded Image. The value Embedded Image corresponds then to the parameter of the Bernoulli describing the random variable Ei, and can be interpreted as the frequency of the promoter of gene i.

The point of view of [35] consisted in stating a LDP for the PDMP system by studying the deviations of π from the quasistationary distribution (3). The work of [29] consists in averaging the flux associated to the transport of each protein over the measure π, in order to build a new expression of this LDP which depends only on the protein dynamics. Its coupling with an Hamiltonian function through a Fenchel-Legendre transform allows to apply a wide variety of analytical tools to gain insight on the most probable behaviour of the process, conditioned on rare events. In this Section, we see how correspondences between these different points of view on the LDP shed light on the meaning of the Hamiltonian and Lagrangian functions and lead to some energetic interpretations.

6.1.1 Correspondence between velocity and promoter frequency

Let us fix the time. The velocity field of the PDMP system, that we denote Φ, is a n-dimensional vector field, function of the random vectors E, X, which can be written for any i = 1, · · ·, n: Embedded Image with the functions voff,i: xi ↦ −dixi and von,i: xi ↦ di(1 − xi) for any x ∈ Ω.

For all i = 1, · · ·, n, let Embedded Image denote the conditional expectation of the promoter Ei knowing a protein vector X. As presented in Section 2.1, the quasistationary approximation identifies the vector field ρ to the invariant measure of the Markov chain on the promoter states.

For a given conditional expectation of promoters ρ, the vector field Embedded Image is defined for all x ∈ Ω by: Embedded Image

Denoting Ωv the set of vector fields v continuous on Ω, such that for all x ∈ Ω, v(x) ∈ Ωv(x), we see that vρ ∈ Ωv. Conversely, the formula (31) can be inverted for associating to every velocity field v ∈ Ωv, characterizing the protein dynamics, a unique conditional expectation of promoters states knowing proteins, ρv, which is the unique solution to the reverse problem Embedded Image, and which is defined by: Embedded Image

6.1.2 Dynamics associated to a protein field

We detailed above the correspondence between any admissible velocity field v ∈ Ωv and a unique vector field ρv describing a conditional expectation of promoters states knowing proteins. Moreover, the proof of Theorem 2 reveals that for any vector field p: Ω ↦ ℝn, we can define a unique vector field αp: Ω ↦ (0, 1)n by the expression (20).

As presented in Section 3.2.2, we denote V the leading order term of the Taylor expansion in ε of the function S defined in (13), such that the distribution of the PDMP system is defined at a fixed time t, and for all e ∈ PE, by Embedded Image, where π(x) is a probability vector in SE for all x ∈ Ω.

On the one hand, we have seen in Section 3.2.2 that for all x ∈ Ω, the eigenvector ζ(x, ∇xV (x)) of the spectral problem (10) (for p = ∇xV (x)) corresponds to the leading order term of the Taylor expansion in ε of π(x). For all i = 1, · · ·, n, the quantity Embedded Image then represents the leading order approximation of the conditional expectation Embedded Image. On the other hand, if we denote the gradient field p = ∇xV defined on Ω, we recall that for all Embedded Image Embedded Image. We then obtain: Embedded Image

This interpretation of the vector αp, combined with the relation (32), allows us to state that the velocity field defined for all x ∈ Ω by Embedded Image characterizes, in the weak noise limit, the protein dynamics associated to the proteins distribution Embedded Image. We see that the velocity field Embedded Image corresponds to the drift of the deterministic system (4) if and only if Embedded Image, and then if and only if p = 0 (see Section 4.2). The gradient field p can be understood as a deformation of the deterministic drift, in the weak noise limit.

We recall that for all p ∈ ℝn, we have from (21): Embedded Image

With the previous notations, the Lagrangian associated to a velocity field v can then be written on every x ∈ Ω as a function of αp and ρv: Embedded Image where p(x) = pv(x) is defined by the expression (24). Thus, we see that the duality between the Lagrangian and the Hamiltonian, that we intensively used in this article for analyzing the optimal trajectories of the PDMP system, and which is expressed through the relation (24) between the variables v and p, also corresponds to a duality between two promoters frequencies ρv and αp associated to the velocity fields v and Embedded Image.

The situation is then the following: for a given proteins distribution Embedded Image such that the first order approximation of S in ε, V, is differentiable on Ω, the velocity field v associated by duality to the gradient field p = ∇xV, and which characterizes a collection of optimal trajectories of the PDMP system (satisfying the system (16) associated to V) when u is the stationary distribution, does not correspond to the protein velocity Embedded Image associated to the distribution u in the weak noise limit, except when the Lagrangian vanishes on (x, v). Alternatively speaking, the optimal trajectories associated to a distribution in the sense of Large deviations, characterized by the velocity field v, do not correspond to the trajectories expected in the weak noise limit, characterized by the velocity field Embedded Image. This is an important limit for developing a physical interpretation of the Hamiltonian system in analogy with Newtonian mechanics. However, the correspondence between promoters states distributions and velocity fields developed above leads us to draw a parallel with some notions of energy.

6.1.3 Energetic interpretation

Following a classical interpretation in Hamiltonian system theory, we introduce a notion of energy associated to a velocity field:

Definition 4.

Let us consider x ∈ Ω and v ∈ Ωv. The quantity Ev(x) = H(x, pv(x)) is called the energy of the velocity field v on x, where pv(x) is defined by the expression (24).

Interestingly, combining the expression of the Hamiltonian given in Theorem 2 with the expressions (24) and (32), the energy of a velocity v on every x ∈ Ω can be rewritten: Embedded Image where for all i = 1, · · ·, n, Φi is the random variable defined by the expression (30), which follows, conditionally to proteins, a Bernoulli distribution of parameter ρv,i, and Embedded Image Embedded Image denotes its standard deviation.

Finally, we have Embedded Image, and Embedded Image denotes the quasistationary distribution described in (3).

Formally, the energy of a promoter distribution can then be decomposed in two terms: a first term describing its velocity in absolute terms, scaled by its standard deviation, and a second term depending on the network. A high energy distribution on a point x is characterized by a fast and deterministic protein dynamics in regards respectively to the velocity of the quasistationary approximation on x and the standard deviation of its associated promoter distribution. We remark that this notion of energy does not depend on the proteins distribution, but only on the promoters frequency ρv around a certain location x. Depending on x only through the vector field ρv (and the functions kon,i), it is likely to be interpreted as the kinetic energy of a cell.

The potential Embedded Image, where Embedded Image is the marginal on proteins of the stationary distribution of the stochastic process, is classically interpreted as a notion of potential energy, not depending on the effective promoter frequency. Apparently, this notion of energy is not related to the one described previously. Once again, the difficulty for linking these two notions of energy comes from the fact that the dynamics associated to the “momentum” p = ∇xV, which is characterized by the velocity field v defined by the formula (23), is not the same that the protein dynamics associated in the weak noise limit to the marginal distribution on proteins Embedded Image, which is defined by the promoters frequency Embedded Image.

6.2 Mixture model

The results of Section 5 lead us to consider the coarse-grained model as promising for capturing the dynamics of the metastable system, even for realistic ε. We are now going to introduce a mixture model which provides an heuristic link between the protein dynamics and the coarse-grained model, and appears then promising for combining both simplicity and ability to describe the main ingredients of cell differentiation process.

When ε is small, a cell within a basin Zj ∈ Z is supposed to be most of the time close to its attractor: a rough approximation consists in identifying the activation rate of a promoter ei in each basin by the dominant rate within the basin, corresponding to the value of kon,i on the attractor. For any gene i = 1, · · ·, n and any basin Zj ∈ Z, we can then consider: Embedded Image

Combining this approximation of the functions kon,i by their main mode within each basin with the description of metastability provided in Section 2.2, we build another process described by the 2n + 1-dimensional vector of variables (Z(t), E(t), X(t)), representing respectively the cell type, the promoter state and the protein concentration of all the genes (see Figure 11).

Figure 11:
  • Download figure
  • Open in new tab
Figure 11:

Weak noise approximate model. The Markov chain on the set of basins Z is here illustrated by the one corresponding to the toggle-switch network of Figure 3a.

Considering that the PDMP system spends in each basin a time long enough to equilibrate inside, we decide to approximate the distribution of the vector (E(t), X(t)) in a basin Zj by its quasistationary distribution. It is then equivalent to the stationary distribution of a simple two states model with constant activation function, which is a product of Beta distributions [28]. Thus, the marginal on proteins of the stationary distribution of this new model, that we denote u, can be approximated by a mixture of Beta distributions: Embedded Image where μz is the stationary distribution of the Markov chain characterizing the coarse-grained model.

In that point of view, the marginal distribution on proteins of a single cell X is characterized by a hidden Markov model: in each basin Zj, which corresponds to the hidden variable, the vector X is randomly chosen under the quasistationary distribution Embedded Image of the reduced process (E, X | Zj). This simplified model provides a useful analytical link between the proteins distribution of the PDMP system (depending on the whole GRN) and the coarse-grained model parameters.

This mixture also provides an approximation for the potential of the system on Ω: Embedded Image

We remark that this new phenomenological model is a generalization of the local approximations of both the potential and the distribution within each basin that we have used for building the isocomittor surfaces and the score function of the AMS algorithm in Appendices F.3 and G.

6.3 One application for the mixture model

An interesting application for the mixture approximation presented in Section 6.2 is the computation of the potential energy of the system, as defined in the previous section. The potential energy of a population of cells C located on (xc)c∈C can be approximated by the sum Embedded Image, where V is defined by (34)

We represent in Figure 12 the evolution of the potential energy of a population of cells during the differentiation process, simulated from the PDMP system associated to the toggle-switch network presented in Appendix E. The population is initially centered on the attractor of the undifferentiated state Z_ _. We observe that the potential energy reaches a peak before decreasing.

Figure 12:
  • Download figure
  • Open in new tab
Figure 12:

Evolution of the potential energy V of a population of 500 cells along the differentiation process.

We remark that in [54], the authors have revealed the universality of such feature during cell differentiation, for what they called the transcriptional uncertainty landscape, for many available single-cell gene expression data sets. This transcriptional uncertainty actually corresponds to the stationary potential V of our model, approximated for each cell from the exact stationary distribution of an uncoupled system of PDMPs (i.e with a diagonal interaction matrix). Although it cannot be formally linked to intracellular energetic spending yet, we can note that one of the authors recently described a peak in energy consumption during the erythroid differentiation sequence [55].

The mixture model also paves the way for interpreting non-stationary behaviours. Indeed, let us denote μz,t the distribution of the basins at any time t. The mixture distribution can be used as a proxy for non stationary distributions of a PDMP system: Embedded Image

In that case, the only time-dependent parameters are the coordinates of the vector μz,t ∈ [0, 1]m where m is the number of basins, and μz,t = μz if t is such that the stationary distribution is reached. The parameters Embedded Image could be inferred from omics data at any time t, for example with an EM algorithm [56], [57].

Conclusion

Reducing a model of gene expression to a discrete coarse-grained model is not a new challenge, ([26], [25]), and it is often hard to perform when the dimension is large. This reduction is closely linked to the notion of landscape through the quasipotential, the analysis of which has been often performed for non mechanistic models, where the random effects are considered as simple noise ([53], [22]), or for a number of genes limited to 2.

In this work, we propose a numerical method for approximating the transition rates of a multidimensional PDMP system modeling genes expression in a single cell. This method allows to compute these transition rates from the probabilities of some rare events, for which we have adapted an AMS algorithm. Although this method theoretically works for any GRN, the computation cost of the AMS algorithm may explode when both the number of genes increases and the scaling factor ε decreases.

In order to approximate these probabilities within the Large deviations context, we provided an explicit expression for the Hamiltonian and Lagrangian of a multidimensional PDMP system, we defined the Hamilton-Jacobi equation which characterizes the quasipotential, for any number of genes, and we provided the explicit expression of the associated variational problem which characterizes the landscape. We have deduced for some networks an analytical expression of the energetic costs of switching between the cell types, from which the transition rates can be computed. These approximations are accurate for a two-dimensional toggle-switch. We also verified that these analytical approximations seem accurate even for networks of 3 or 4 genes for which the energetic cost provided by the method is not proved to be optimal. However, testing the accuracy of this method for describing more complex networks would imply to build an approximate solution to the stationary Hamilton-Jacobi equation (25), which would be the subject of future works.

Finally, we have derived from the coarse-grained model a Beta-mixture model able to approximate the stationary behavior of a cell in the gene expression space. As far as we know, this is the first time that such an explicit link between a PDMP system describing cell differentiation and a non-Gaussian mixture model is proposed.

Altogether this work establishes a formal basis for the definition of a genetic/epigenetic landscape, given a GRN. It is tempting to now use the same formalism to assess the inverse problem of inferring the most likely GRN, given an (experimentally-determined) cell distribution in the gene expression space, a notoriously difficult task [58], [28].

Such random transitions between cell states have been recently proposed as the basis for facilitating the concomitant maintenance of transcriptional plasticity and stem cell robustness [59]. In this case, the authors have proposed a phenomenological view of the transition dynamics between states. Our work lays the foundation for formally connecting this cellular plasticity to the underlying GRN dynamics.

Finally our work provides the formal basis for the quantitative modelling of stochastic state transitions underlying the generation of diversity in cancer cells ([60], [61]), including the generation of cancer stem cells [62].

Code availability

The code for reproducing the main figures of the article is available at https://gitbio.ens-lyon.fr/eventr01/jomb_reduction. It also contains the functions for the AMS algorithm, which is detailed in the appendix.

Acknowledgments

This work was supported by funding from French agency ANR (SingleStatOmics; ANR-18-CE45-0023-03). We thank Ulysse Herbach for having highlighted the notions of main modes for the stochastic hybrid model of gene expression, and for critical reading of the manuscript. We would like to thank the referees and the associated editor for carefully reading our manuscript and for their constructive comments which helped improving the quality of the paper. We also thank all members of the SBDM and Dracula teams, and of the SingleStatOmics project, for enlightening discussions. We also thank the BioSyL Federation and the LabEx Ecofect (ANR-11-LABX-0048) of the University of Lyon for inspiring scientific events.

A Mechanistic model and fast transcription reduction

We recall briefly the full PDMP model, which is described in details in [28], based on a hybrid version of the well-established two-state model of gene expression [30], [31] including both mRNA and protein production [63] and illustrated in Figure 13.

Figure 13:
  • Download figure
  • Open in new tab
Figure 13:

The two-states model of gene expression [28], [31].

A gene is described by the state of a promoter, which can be {on, off}. If the promoter is on, mRNAs will be transcripted with a rate sm and degraded with a rate dm. If it is off, only mRNA degradation occurs. Translation of mRNAs into proteins happens regardless of the promoter state at a rate sp, and protein degradation at a rate dm. Neglecting the molecular noise of proteins and mRNAs, we obtain the hybrid model: Embedded Image where (E(t), M (t), P (t)) denote respectively the promoter, mRNA and protein concentration at time t. As detailed in Section 1, the key idea is then to put this two-states model into a network by characterizing the jump rates of each gene by two specific functions kon,i and koff,i, depending at any time on the protein vector X(t).

In order to obtain the PDMP system (1) that we use throughout this article, we exploit the two modifications that are performed in [28] to this mechanistic model. First, the parameters sm and sp can be removed to obtain a dimensionless model, from which physical trajectories can be retrieved with a simple rescaling.

Second, a scaling analysis leads to simplify the model. Indeed, degradation rates play a crucial role in the dynamics of the system. The ratio Embedded Image controls the buffering of promoter noise by mRNAs and, since koff,i » kon,i, the ratio Embedded Image controls the buffering of mRNA noise by proteins. In line with several experiments [64] [65], we consider that mRNA bursts are fast in regard to protein dynamics, i.e Embedded Image with Embedded Image fixed. The correlation between mRNAs and proteins produced by the gene is then very small, and the model can be reduced by removing mRNA and making proteins directly depend on the promoters. We then obtain the PDMP system (1).

Denoting Embedded Image the mean value of the function kon,i, i.e its value where there is no interaction between gene i and the other genes, the value of a scaling factor Embedded Image can then be decomposed in two factors: one describing the ratio between the degradation rates of mRNA and proteins, Embedded Image, which is evaluated around 1/5 in [66], and one characterizing the ratio between promoter jumps frequency and the degradation rates of mRNA, Embedded Image. This last ratio is very difficult to estimate in practice. Assuming that it is smaller than 1, i.e that the mean exponential decay of mRNA when the promoter Ei is off is smaller than the mean activation rate, we can consider that εi is smaller than 1/5. Finally, for obtaining the model (1), we consider two typical timescales Embedded Image and Embedded Image, for the rates of proteins degradation and promoters activation respectively, such that for all genes i, Embedded Image and Embedded Image are of order 1 (when the disparity between genes is not too important). We then define Embedded Image.

B Tensorial expression of the master equation of the PDMP system

We detail the tensorial expression of the master equation (2) for a two-dimensional network. We fix ε = 1 for the sake of simplicity.

The general form for the infinitesimal operator can be written: Embedded Image where F is the vectorial flow associated to the PDMP and Q the matrix associated to the jump operator.

A jump between two promoters states e, e′ is possible only if there is exactly one gene for which the promoter has a different state in e than in e′: in this case, we denote e ~ e′.

We have, for any x: F (e, x) = (d0(e0 − x0), · · ·, dn(en − xn))T. Then, for all e ∈ PE, the infinitesimal operator can be written: Embedded Image

For a two-dimensional process (n = 2), there are four possible configurations for the promoter state: e00 = (0, 0), e01 = (0, 1), e10 = (1, 0), e11 = (1, 1). It is impossible to jump between the states e00 and e11. If we denote u(t, x) the four-dimensional vector: Embedded Image, we can write the infinitesimal operator in a matrix form: Embedded Image

We remark that each of these matrices can be written as a tensorial product of the corresponding two-dimensional operator with the identity matrix: Embedded Image

The master equation (2) is obtained by taking the adjoint operator of L: Embedded Image where K(x) = QT (x) is the transpose matrix of Q.

C Diffusion approximation

C.1 Definition of the SDE

In this section, we apply a key result of [36] to build the diffusion limit of the PDMP system (1). Let us denote Embedded Image a trajectory satisfying the ODE system: Embedded Image where Embedded Image characterizes the deterministic system (4). We consider the process Embedded Image defined by: Embedded Image where Xtε verifies the PDMP system. Then, from the theorem 2.3 of [36] the sequence of processes Embedded Image converges in law when ε → 0 to a diffusion process which verifies the system: Embedded Image where Bt denotes the Brownian motion. The diffusion matrix Σ(x) = σ(x)σT (x) is defined by: Embedded Image where ∀e ∈ PE, Embedded Image, and ϕ is solution of a Poisson equation: Embedded Image

Let ζ be a probability vector in SE representing the stationary measure of the jump process on promoters knowing proteins: ∀x ∈ Ω, ζ(x, ·)Q(x) = 0. We have: Embedded Image.

It is straightforward to see that for all Embedded Image. Then, let us define ϕ such that: Embedded Image

We verify that this vector ϕ is solution to the Poisson equation (36) for all x. The matrix Σ(x) is then a diagonal matrix defined by: Embedded Image

For all x ∈ Ω, the matrix σ(x) is then also diagonal and defined by: Embedded Image and we have defined all the terms of the diffusion limit (35).

C.2 The Lagrangian of the diffusion approximation is a second-order approximation of the Lagrangian of the PDMP system

It is well known that the diffusion approximation satisfies a LDP of the form (12) [39]. The formula (37) allows to define the Lagrangian associated to this LDP, that we denote Ld. From the theorem 2.1 of [39], we have: Embedded Image

Note that for any fixed x ∈ Ω, Ld(x, ·) is a quadratic function.

We recall that the Lagrangian associated to the LDP for the PDMP system, that we found in Theorem 3, is defined for all x, v ∈ Ω × Ωv(x) by: Embedded Image

Expanding this Lagrangian with respect to v around Embedded Image (the drift of the relaxation trajectories), we obtain: Embedded Image

Thus, we proved that the Lagrangian of the diffusion approximation of the PDMP process corresponds to the two first order terms in Embedded Image of the Taylor expansion of the real Lagrangian.

D Example of interaction function

We recall that we assume that the vector koff does not depend on the protein vector.

The specific interaction function chosen comes from a model of the molecular interactions at the promoter level, described in [28]: Embedded Image with:

  • k0,i the basal rate of expression of gene i,

  • k1,i the maximal rate of expression of gene i,

  • mi,j an interaction exponent, representing the power of the interaction between genes i and j,

  • σi is the rescaling factor depending on the parameters of the full model including mRNAs,

  • θ a matrix defining the interactions between genes, corresponding to a matrix with diagonal terms defining external stimuli, and

  • Embedded Image

For a two symmetric two-dimensional network, we have for any x = (x1, x2) ∈ Ω: Embedded Image

When m11 = m22 = m12 = m21 and θ12 = θ21, we have then for every x ∈ Ω: Embedded Image

Thus, for all x ∈ Ω, when d1 = d2 we have:

Embedded Image

As a consequence, owing to the Poincaré lemma, there exists a function V ∈ C1(Ω, ℝ) such that the condition (26) is satisfied: one has Embedded Image

E Description of the toggle-switch network

This table describes the parameters of the symmetric two-dimensional toggle-switch used all along the article. These values correspond to the parameters used for the simulations. The rescaling in time by the parameter scale Embedded Image, for the model presented in Section 1, corresponds to divide every k0,i, k1,i, di by Embedded Image. The mean values Embedded Image and di are then, as expected, of order 1 for every gene i.

View this table:
  • View inline
  • View popup
  • Download powerpoint

F Details on the approximation of the transition rate as a function of probability (7)

In this section, we adapt the method developed in [40] to justify the formula (8) provided in Section 3.1, which approximate for every pair of basins (Zi, Zj) the transition rate aij as a function of the probability (7).

F.1 General setting

Let us consider r, R such that 0 < r < R, we recall that Embedded Image and Embedded Image denote respectively the r-neighborhood and the R-neighborhood of the attractor Embedded Image. Let us consider a random path Embedded Image of the PDMP system, with initial condition Embedded Image. We define the series of stopping times Embedded Image such that Embedded Image and for all l ∈ ℕ*:

  • Embedded Image

  • Embedded Image

We then define Embedded Image. If Embedded Image, we set Embedded Image and the chain Embedded Image stops.

Figure 14:
  • Download figure
  • Open in new tab
Figure 14:

Illustration of the stopping times σl and μl describing respectively the lth entrance of a random path Xt in a r-neighborhood γ of an attractor Xeq, and its lth exit from a R-neighborhood Γ.

From the formula (6) characterizing the transition rates, we can write: Embedded Image where we define the random variable: Embedded Image.

Let us denote Embedded Image. We can make the following approximation: Embedded Image

Indeed, the quantity on the left hand side is close to the mean number of attempts for reaching, from Embedded Image, a basin Zk, k ≠ i, before Embedded Image, which is equal to Embedded Image, multiplied by the mean time of each attempt (knowing that at each step l, Embedded Image), which is exactly Embedded Image. We should add the mean time for reaching ∂Zj from Embedded Image at the last step, but it is negligible when the number of attempts is large, which is the case in the small noise limit.

F.2 Method when Embedded Image is reduced to a single point

We consider the case when Embedded Image is reduced to a single point. It can happen for example when we consider only one gene (Ω = (0, 1)) and when the attractor Embedded Image is located at a distance smaller than r from one of the boundaries of the gene expression space Embedded Image or Embedded Image. In such situation, a random path crosses necessarily the same point x0 to both exit Embedded Image and come back to Embedded Image (if it does not reach a basin Zj before): the Markov property of the PDMP process then justifies that the quantities Embedded Image and Embedded Image do not depend of l. Then Embedded Image behaves like a discrete homogeneous Markov chain with two states, 1 being absorbing.

Let us define a second random variable Embedded Image. The homogeneity of the Markov chain Embedded Image ensures that Embedded Image follows a geometric distribution. Its expected value is then the reverse of the parameter of the geometric law, i.e: Embedded Image.

Moreover, it is straightforward to see that from the same reasoning applied to any Zk, k ≠ i: Embedded Image

Thus, from (40) we can approximate the transition rate by the formula (8).

F.3 Method in the general case

The difficulty for generalizing the approach described above, when Embedded Image is not reduced to a single point, is to keep the Markov property, which has been used to cut the trajectories into pieces. Heuristically, the same argument which led us to approximate the PDMP system by a Markov jump process can be used to justify the asymptotic independence on l of the quantity Embedded Image: for ε ≪ 1, any trajectory starting on Embedded Image will rapidly loose the memory of its starting point after a mixing time within Embedded Image. But it is more complicated to conclude on the independence from l of the quantity Embedded Image, which may depend on the position of Embedded Image when the gene expression space is multidimensional.

We introduce two hypersurfaces Embedded Image and Embedded Image Embedded Image, where c1 < c2 are two small constants. We substitute to the squared euclidean distance, used for characterizing the neighborhood Embedded Image and Embedded Image, a new function based on the probability of reaching the (unknown) boundary: ∀x, y ∈ Zi, Embedded Image. The function Embedded Image is generally called committor, and the hypersurfaces Embedded Image and Embedded Image isocommittor surfaces. The committor function is not known in general; if it was, employing a Monte-Carlo method would not be necessary for obtaining the probabilities (7). However, it can be approximated from the potential of the PDMP system within each basin, defined in the equilibrium case by the well-known Boltzman law: Embedded Image being the marginal on proteins of the stationary distribution of the process. Indeed, for reasons that are precisely the subject of Section 3.2 (studied within the context of Large deviations), the probability Embedded Image is generally linked in the weak noise limit to the function V by the relation: Embedded Image where Cij is a constant specific to each pair of basins (Zi, Zj). We remark that when ε is small, a cell within a basin Zj ∈ Z is supposed to be most of the time close to its attractor: a rough approximation could lead to identify the activation rate of a promoter ei in each basin by the dominant rate within the basin, corresponding to the value of kon,i on the attractor. For any gene i = 1, · · ·, n and any basin Zj ∈ Z, we can then approximate: Embedded Image

Under this assumption, the stationary distribution of the process is close to the stationary distribution of a simple two states model with constant activation function, which is a product of Beta distributions [28]. We then obtain an approximation of the marginal on proteins of the stationary distribution within each basin Zj: Embedded Image

By construction, this approximation is going to be better in a small neighborhood of the attractor Embedded Image. Thus, this expression provides an approximation of the potential V around each attractor Embedded Image: Embedded Image

In every basin Zi, and for all x ∈ Zi close to the attractor, the hypersurfaces where Embedded Image is constant will be then well approximated by the hypersurfaces where the explicitly known function Embedded Image is constant.

For each attractor Embedded Image, we can then approximate the two isocommittor surfaces described previously: Embedded Image where c′1 and c′2 are two constants such that Embedded Image.

We then replace Embedded Image and Embedded Image by, respectively, Embedded Image and Embedded Image in the definitions of the stopping times Embedded Image and Embedded Image provided in Section F.1. From the proposition 1. of [40], we obtain that, as in the simple case described in Section F.2, Embedded Image is independent of l. Defining Embedded Image, the definition of Embedded Image allows to ensure that Embedded Image does not depend on the point x0 of Embedded Image which is crossed at each step Embedded Image. This random variable follows then a geometric distribution, with expected value Embedded Image, and we can derive an expression of the form (8).

G AMS algorithm

We use an Adaptive Multilevel Splitting algorithm (AMS) described in [41]. The algorithm provides for every Borel sets (A, B) an unbiased estimator of the probability: Embedded Image

It is supposed that the random process attains easily A from x, more often than B, called the target set.

The crucial ingredient we need to introduce is a score function ξ(·) to quantify the adaptive levels describing how close we are from the target set B from any point x. The variance of the algorithm strongly depends on the choice of this function.

The optimal score function is the function Embedded Image itself, called the committor which is unknown. It is proved, at least for multilevel splitting algorithms applied to stochastic differential equations in [67], [68], that if a certain scalar multiplied by the score function is solution of the associated stationary Hamilton-Jacobi equation, where the Hamiltonian comes from the Large deviations setting, the number of iterations by the algorithm to estimate the probability in a fixed interval confidence grows sub-exponentially in ε.

For the problem studied in this article, for every basin Zj ∈ Z, we want to estimates probabilities substituting A to γj and B to another basin Zk, k ≠ j. Using the approximation of V given by the expression (34), we obtain the following score function, up to a specific constant specific to each basin: Embedded Image

We remark that this last approximation allows to retrieve the definition of the local potential (41) defined on Appendix F.3, when the boundary of the basins are approximated by the leading term in the Beta mixture. The approximation is justified by the fact that for small ε, the Beta distributions are very concentrated around their centers, meaning that for every basin Zk ∈ Z, k ≠ j: Embedded Image

We supposed that ∀Zj ∈ Z, μz(Zj) > 0, where μz denotes the distributions on the basins. This is a consequence of the more general assumption that the stationary distribution of the PDMP system is positive on the whole gene expression space, which is necessary for rigorously deriving an analogy of the Gibbs distribution for the PDMP system (see Section 3.2.2).

We modify the score function to be adapted for the study of the transitions from each basin Zj to Zk, k ≠ j: Embedded Image

This function is specific to each transition to a basin Zk but defined in the whole gene expression space. We verify: ξk(x) ≤ 0 if x ∈ Ω \ Zk and ξk(x) = 0 if x ∈ Zk. We use ξk as the score function for the AMS algorithm.

In order to estimate probabilities of the type Embedded Image for x ∈ Zj, we need to approximate the boundaries of the basins of attraction, which are unknown. For this sake, we use again the approximate potential function ξ ≈ V to approximate the basins only from the knowledge of their attractor: Embedded Image

We use the Adaptative Multilevel Splitting algorithm described in Section 4 of [41], with two slight modifications in order to take into account the differences due to the underlying model and objectives:

  • First, a random path associated to the PDMP system does not depend only on the protein state but is characterized at each time t by the 2n-dimensional vector: (Xt, Et). For any simulated random path, we then need to associate an initial promoter state. However, we know that in the weak noise limit, for a protein state close to the attractor of a basin, the promoter states are rapidly going to be sampled by the quasistationary distribution: heuristically, this initial promoter state will not affect the algorithm. We decide to initially choose it randomly under the quasistationary distribution. For every x0 ∈ Γmin,j beginning a random path in a basin Zj, we choose for the promoter state of any gene i, Embedded Image, following a Bernoulli distribution: Embedded Image

  • Compared with [41], an advanced algorithm is used to improve the sampling of the entrance time in a set γmin,j. In practice timestepping is required to approximate the protein dynamics, and it may happen that the exact solution enters γmin,j between two time steps, whereas the discrete-time approximation remains outside γmin,j. We propose a variant of the algorithm studied in [69] for diffusion processes, where a Brownian Bridge approximation gives a more accurate way to test entrance in the set γmin,j.

In the case of the PDMP system, we replace the Brownian Bridge approximation, by the solution of the ODE describing the protein dynamics: considering that the promoter state e remains constant between two timepoints, the protein concentration of every gene i, xi(t) is a solution of the ODE: Embedded Image, which implies: Embedded Image

We show that for one gene, the problem can be easily solved. Indeed, let us denote Embedded Image the ith component of the vector Embedded Image. The function: Embedded Image is differentiable and its derivative Embedded Image vanishes if and only if Embedded Image, i.e when Embedded Image

Then, if ci ≤ 0 or ci ≥ Δt, the minimum of the squared euclidean distance of the i-th coordinate of the path to the attractor is reached at one of the points xi(0) or xi(Δt). If 0 ≤ ci ≤ Δt, the extremum is reached at xi(ci). This value, if it is a minimum, allows us to determine if the process has reached any neighborhood of an attractor Embedded Image between two timepoints.

For more than one gene, the minimum of the sum: Embedded Image is more complicated to find. If for all i = 1, · · ·, n, di = d, which is the case of the two-dimensional toggle-switch studied in Section 5, the extremum can be explicitly computed: Embedded Image

But we recall that for more than one gene, the set of interest is the isocommittor surface γmin,j and not a neighborhood γj. An approximation consists in identifying γmin,j to the r-neighborhood of Embedded Image, where r is the mean value of Embedded Image for x ∈ γmin,j.

If the parameters di are not all similar, we have to make the hypothesis that the minimum is close to the minimum for each gene. In this case, we just verify that for any gene i, the value of the minimum xi(ci) for every gene is not in the set Embedded Image: if it is the case for one gene, we consider that the process has reached the neighborhood Embedded Image of the basins Zj between the two timepoints.

H Proofs of Theorems 4 and 5

First, we recall the theorem of characteristics applied to Hamilton-Jacobi equation [70], which states that for every solution V ∈ C1(Ω, ℝ) of (25), the system (16) associated to V Embedded Image is equivalent to the following system of ODEs on (x, p) ∈ Ω × ℝn, for x(0) = ϕ(0) and p(0) = ΔxV (x(0)): Embedded Image

A direct consequence of this equivalence with an ODE system is that two optimal trajectories associated to two solutions of the stationary Hamilton-Jacobi equation cannot cross each other with the same velocity. We then have the following lemma:

Lemma 2.

Let V1 and V2 be two solutions of (25) in C1(Ω, ℝ).

For any trajectories Embedded Image solutions of the system (16) associated respectively to V1 and V2, if there exists t ∈ [0, T] such that ϕ1(t) = ϕ2(t) and Embedded Image, then one has ϕ1(t) = ϕ2(t) for all t ∈ [0, T].

This corollary is important for the two first items of the proof of Theorem 4:

Corollary 1.

For any solution V ∈ C1(Ω, ℝ) of (25) and any trajectory Embedded Image satisfying the system (16) associated to V, we have the equivalence: Embedded Image

Proof. We recall that the relaxation trajectories correspond to trajectories satisfying the system (16) associated to a constant function V, i.e such that ∇xV = 0 on the whole trajectory. At any time t, the correspondence between any velocity field v of Ωv and a unique vector field p, proved in Theorem 3 with the relation (24), allows to ensure that: Embedded Image

The lemma 2 ensures that any trajectory which verifies the same velocity field than a relaxation trajectory at a given time t is a relaxation trajectory: we can then conclude.

Finally, the following lemma is important for the first item of the proof of Theorem 4:

Lemma 3.

∀i ∈ {1, · · ·, n}, ∀x ∈ Ω we have: Embedded Image

Proof. We have seen in the proof of Theorem 2 that for all i = 1, · · ·, n and for all x ∈ Ω, Hi(x, ·) is strictly convex, and that Hi(x, pi) → ∞ as Pi → ±∞. Moreover, Hi(x, pi) vanishes on two points Embedded Image and Embedded Image inside ℝ.

Then, the min on pi is reached on the unique critical point Embedded Image, and we have: Embedded Image

Finally: Embedded Image We now prove the theorem 4.

Proof of Theorem 4.(i). We consider a trajectory ϕt satisfying the system (16) associated to V. We recall that the Fenchel-Legendre expression of the Lagrangian allows to state that the vector field p associated to the velocity field Embedded Image by the relation (24) is precisely p = ∇xV (ϕ(t)). When V is such that H(·, ∇xV (·)) = 0 on Ω, we have then for any time t: Embedded Image

We recall that from Theorem 3: Embedded Image

From this and (43), we deduce that for such an optimal trajectory: Embedded Image

The velocity field then vanishes only at the equilibrium points of the deterministic system. Conversely, we recall that for such trajectory we have for any t: Embedded Image

Assume that for all Embedded Image. Then, by Lemma 3, we have: Embedded Image for all i.

Thereby, H(ϕ(t), ∇ xV (ϕ(t))) = 0 if and only if for all i Hi(ϕ(t), Embedded Image Embedded Image, which implies: Embedded Image. The lemma is proved.

Proof of Theorem 4.(ii). From the Corollary 1, if ∇xV (ϕ(t)) = 0, the trajectory is a relaxation trajectory, along which the gradient is uniformly equal to zero. The condition (C) implies that it is reduced to a single point: Embedded Image. Conversely, with the same reasoning that for the proof of (i): Embedded Image

We recognize the equation of a relaxation trajectory, which implies: ∇xV (ϕ(t)) = 0. Thus, for any optimal trajectory satisfying the system (16) associated to a solution V of the equation (25) satisfying the condition (C), we have for any time t: Embedded Image

From the equivalence proved in (i), the condition (C) then implies that the gradient of V vanishes only on the stationary points of the system (4).

Proof of Theorem 4.(iii). As a consequence of (ii), for any optimal trajectory ϕt associated to a solution V of (25) which satisfies the condition (C), we have for all t > 0: Embedded Image

Then, if there exists t > 0 such that Embedded Image, it cannot be equal to the drift of a relaxation trajectory, defined by the deterministic system (4), which is known to be the unique velocity field for which the Lagrangian vanishes (from Theorem (3)). Then it implies: Embedded Image

The relation (43), combined to the fact that the Lagrangian is always nonnegative allow to conclude: Embedded Image

Thus, the function V strictly increases on these trajectories.

Furthermore, on any relaxation trajectory ϕr(t), from the inequality (15) we have for any times T 1 < T 2: Embedded Image

The equality holds between T 1 and T 2 if and only if for any t ∈ [T 1, T 2]: L(ϕ(t), Embedded Image. In that case, from Theorem 3the drift of the trajectory is necessarily the drift of a relaxation trajectory between the two timepoints and then, from Corollary 1, ∇xV = 0 on the set of points {ϕ(t), tℝ+}, which is excluded by the condition (C) (when this set is not reduced to a single point). Thus, if ϕ(T1) /= ϕ(T2), we have: V (ϕ(T1)) > V (ϕ(T2)).

By definition, for any basin Zi and for all x ∈ Zi there exists a relaxation trajectory connecting x to the associated attractor Embedded Image. So ∀x ∈ Zi, Embedded Image.

Proof of Theorem 4.(iv). Let V be a solution of (25) satisfying the condition (C). We consider trajectories solutions of the system defined by the drift Embedded Image. We recall that from (iii), the condition (C) ensures that V decreases on these trajectories, and that for all t1 < t2: V (ϕ(t1)) − V (ϕ(t2)) = Q(ϕ(t2), ϕ(t1)) > 0. Then, the hypothesis Embedded Image ensures that such trajectories cannot reach the boundary ∂Ω: if it was the case, we would have a singularity inside Ω, which is excluded by the condition V ∈ C1(Ω, ℝ). The same reasoning also ensures that there is no limit cycle or more complicated orbits for this system.

Recalling that from (i) and (ii), the fixed point of this system are reduced to the points where ∇x,V = 0 on Ω, we conclude that for all x ∈ Ω, there exists a fixed point a ∈ Ω, satisfying ∇xV (a) = 0, such that a trajectory solution of this system converges to a, i.e: V (x) − V (a) = Q(a, x).

As from the inequality (15), we have for every point a the relation V (x) − V (a) ≤ Q(a, x), the previous equality corresponds to a minimum and we obtain the formula: Embedded Image

Proof of Theorem 4.(v). Let V be a solution of (25) satisfying the condition (C). We denote by νV the drift of the optimal trajectories ϕt on [0, T] satisfying the system (16) associated to V: ∀t ∈ [0, T], Embedded Image, ∇xV (ϕ(t))) = νV (ϕ(t)). We call trajectories solution of this system reverse fluctuations trajectories.

For any basin Zi associated to the stable equilibrium of the deterministic system Embedded Image, we have:

  • From (i), Embedded Image and Embedded Image.

  • From (iii), we know that V increases on these trajectories: Embedded Image Embedded Image.

  • From (iii), we also have: Embedded Image.

Without loss of generality (since we only use ∇xV), we can assume Embedded Image. We have then: Embedded Image, V (x) > 0. Moreover, since we have assumed that Embedded Image is isolated, there exists δV > 0 such that Zi contains a ball Embedded Image. Therefore, V reaches a local minimum at Embedded Image. Conversely if V reaches a local minimum at a point Embedded Image, then Embedded Image is necessarily an equilibrium (from (ii)), and the fact that V strictly decreases on the relaxation trajectories ensures that it is a Lyapunov function for the deterministic system, and then that Embedded Image is a stable equilibrium. The stable equilibria of the deterministic system are thereby exactly the local minima of V, and for any attractor Embedded Image, V is also a Lyapunov function for the system defined by the drift −νV, for which Embedded Image is then a locally asymptotically stable equilibrium. Thereby, stable equilibria of the deterministic system are also stable equilibria of the system defined by the drift −νV.

It remains to prove that no unstable equilibria the deterministic system is stable for the system defined by the drift −νV. Let Embedded Image be an unstable equilibrium of the relaxation system, then V does not reach a local minimum at Embedded Image. Therefore, as close as we want of Embedded Image there exists x such that Embedded Image. We recall that reverse fluctuations trajectories ϕt starting from such a point and remaining in Ω will have V (ϕ(t)) striclty decreasing: by Lyapunov La Salle principle, they shall be attracted towards the set {y, ∇〈V (y), νV (y)〉 = 0}, which contains from (iii) only the critical points of V, which are from (i) the equilibria of both deterministic and reverse fluctuation systems. In particular, either ϕt leaves Ω (and the equilibrium is unstable) or ϕt converges to another equilibrium (since they are isolated) and this contradicts the stability. So we have proved that stable equilibria of both systems are the same.

We then obtain that for any Zi, there exists δV such that: Embedded Image

Reverting time, any point of Embedded Image can then be reached from any small neighborhood of Embedded Image. We deduce from Lemma 1 that: Embedded Image

Applying exactly the same reasoning to another function Embedded Image solution of (25) and satisfying (C), this ensures that Embedded Image, at least for Embedded Image. We recall that that from Lemma 2, two optimal trajectories ϕt, Embedded Image solutions of the system (16), associated respectively to two solutions V and Embedded Image of the equation (25), cannot cross each other without satisfying Embedded Image along the whole trajectories. Thereby, we can extend the equality Embedded Image on the basins of attraction associated to the stable equilibrium Embedded Image for both systems defined by the drifts −νV or Embedded Image. Thus, we have proved that the basins associated to the attractors are the same for both systems. We denote Embedded Image these common basins.

Under the assumption 2. of the theorem, we obtain by continuity of V that for every pair of basin Embedded Image. It follows that under this assumption, there exists a constant c ∈ ℝ such that for every attractor Embedded Image: Embedded Image

Moreover, the assumption 1. ensures that from Theorem 4.(iv), there exists a fixed point a1 ∈ Ω (with ∇xV (a1) = 0), such that a trajectory solution of the system defined by the drift −νV converges to a1, i.e: V (x) − V (a1) = Q(a1, x). On one side, if a1 is unstable, it necessarily exists on any neighborhood of a1 a point x2 such that V (x2) < V (a1). As for all y ∈ Ω, Q(·, y) is positive definite, we have then another fixed point a2 ≠ a1 such that V (x2) = Q(a2, x2) + V (a2). We obtain: V (x) > h(x, a1) + Q(a2, x2) + V (a2). On the other side, by continuity of the function Q(a2, ·), for every δ1 > 0, x2 can be chosen close enough to a1 such that: Q(a2, x2) ≥ Q(a2, a1) − δ1. We obtain: Embedded Image

Repeating this procedure until reaching a stable equilibrium at a step N, which is necessarily finite because we have by assumption a finite number of fixed points, we obtain the inequality Embedded Image where every ak denotes a fixed point and aN is an attractor. Using the triangular inequality satisfied by Q, and passing to the limit δ → 0, we find that V (x)−V (aN) ≥ Q(aN, x). Moreover, from the inequality (15), we have necessarily Embedded Image. It then follows from (44) that Embedded Image.

Applying exactly the same reasoning for building a serie of fixed point Embedded Image such that Embedded Image is an attractor and Embedded Image, we obtain Embedded Image. We can conclude: Embedded Image

Proof of Theorem 5. First, we prove the following lemma:

Lemma 4.

∀i ∈ {1, · · · , n}, we have:

  1. ∃δl > 0, ∃ηl > 0, such that ∀x, y ∈ Ω, if yi < xi ≤ δl, then we have: Embedded Image

  2. ∃δr < 1, ∃ηr > 0, such that ∀x, y ∈ Ω, if yi > xi ≥ δr, then we have: Embedded Image

Proof. (i) We denote Embedded Image. We have mi > 0 by assumption. We choose a real number δ which satisfies these two conditions:

  1. Embedded Image,

  2. Embedded Image.

On the one hand, we recall that the function Embedded Image is convex and vanishes only on Embedded Image. Then, Li(x, ·) is decreasing on Embedded Image.

On the other hand, for all x ∈ Ω, if xi ≤ δl, we have necessarily: Embedded Image Embedded Image from the condition 1. Then, we obtain that for all x ∈ Ω, if xi ≤ δl: Embedded Image

From the condition 2., we also see that for all x ∈ Ω, if xi ≤ δl: Embedded Image which implies: Embedded Image

Then, we obtain that for any admissible trajectory Embedded Image (i.e with a velocity in Ωv(ϕ(t)) at all time) and such that ϕ(0) = x and ϕ(T) = y, if yi < xi ≤ δl we have: Embedded Image where we denote {[tl,k, tr,k]}k=1, …,l the l intervals on which the velocity Embedded Image and ϕi(t) < δl on the interval [0, T]. As we now by assumption that ϕi(0) = xi ≤ δl and ϕi(T) = yi < ϕi(0), this set of intervals cannot be empty.

Moreover, for every k = 1, · · ·, l, we have by assumption: Embedded Image.

Then: Embedded Image

As by definition, for every Embedded Image on [tr,k, tl,k+1], we have ϕi(tl,k+1) ≥ ϕi(tr,k) (because Embedded Image on [tr,k, tl,k+1]). Finally, we obtain: Embedded Image which implies: Embedded Image

This last inequality combined with (45) allows to conclude: Embedded Image

Thus, if δl satisfies conditions 1. and 2., and fixing Embedded Image, for every x, y ∈ Ω such that yi < xi ≤ δl we have: Embedded Image

(ii) Denoting Embedded Image (which exists by assumption), we chose this time the real number δr in order to satisfy these these two conditions:

  1. Embedded Image,

  2. Embedded Image,

and we fix Embedded Image. The rest consists in applying exactly the same reasoning than for the proof of (i) in a neighborhood of 1 instead of 0.

We deduce immediately the following corollary:

Corollary 2.

∀x ∈ Ω, Embedded Image.

Let us denote V ∈ C1(Ω, ℝ) a solution of the equation (25), which satisfies the condition (C). From the proof of Theorem 4.(v), we know that for any attractor Embedded Image, there exists a ball Embedded Image, where Embedded Image is the basin of attraction of Embedded Image for the system defined by the drift Embedded Image. Moreover, as V decreases on trajectories solutions of this system, the set Embedded Image is necessarily stable: we have Embedded Image.

We deduce that: Embedded Image and in that case Embedded Image. If there existed Embedded Image, we would have, by continuity of V and from Corollary 2: Embedded Image

It would imply that Embedded Image, which is impossible when ∂Zi ≠ ∂Ω, which is necessarily the case when there is more than one attractor.

Thus, there exists at least one point xi on the boundary ∂Zi \ ∂Ω, such that for any neighborhood of Embedded Image, there exists a fluctuation trajectory starting inside and converging to xi.

We recall that we assume that Embedded Image. Then we have Embedded Image and there exists Zj such that: xi ∈ ∂Zi ∩ ∂Zj = ∂Zi \ Rij. We obtain, by continuity of V: Embedded Image

It remains to prove that under the assumption (A) of the theorem, Embedded Image. On one hand, from Theorem 4.(iii), V decreases on the relaxation trajectories. On the other hand, if every relaxation trajectories starting in ∂Zi ∩ ∂Zj stay inside, they necessarily converge from any point of ∂Zi ∩ ∂Zj to a saddle point (also in ∂Zi ∩ ∂Zj). Then, the minimum of V in ∂Zi ∩ ∂Zj is reached on the minimum of V on Embedded Image (the set of all the saddle points in ∂Zi ∩ ∂Zj), which is Embedded Image. Thus, if every relaxation trajectories starting in ∂Zi ∩ ∂Zj stay inside, then XZ ∈ ∂Zj implies Embedded Image. The theorem is proved.

I Algorithm to find the saddle points

We develop a simple algorithm using the Lagrangian associated to the fluctuation trajectories (28) to find the saddle points of the deterministic system (4). This Lagrangian is a nonnegative function which vanishes only at the equilibria of this system. Then, if there exists a saddle point connecting two attractors, this function will vanish at this point. Starting on a small neighborhood of the first attractor, we follow the direction of the second one until reaching a maximum on the line (see 15a). Then, we follow different lines, in the direction of each other attractor for which the Lagrangian function decreases (at least, the direction of the second attractor (see 15b)), until reaching a local minimum. We then apply a gradient descent to find a local minimum (see 15c). If this minimum is equal to 0, this is a saddle point, if not we repeat the algorithm from this local minimum until reaching a saddle point or an attractor. Repeating this operation for any ordered couple of attractors Embedded Image, we are likely to find most of the saddle points of the system. This method is described in pseudo-code in Algorithm 1.

Figure 15:
  • Download figure
  • Open in new tab
Figure 15:

Saddle-point algorithm between two attractors. The color map corresponds to the Lagrangian function associated to the fluctuation trajectories.

Algorithm 1

Find the list of saddle points: list-saddle-points

Figure
  • Download figure
  • Open in new tab

J Applicability of the method for non-symmetric networks and more than two genes

We present in Figures 16b and 17b an analogy of Figure 8, which was presented for the toggle-switch network, for two non-symmetric networks of respectively 3 and 4 genes. The networks are presented on the left-hand side of Figures 16b and 17b: the red arrows represent the inhibitions and the green arrows represent the activations between genes. A typical random trajectory for each network is presented in Figures 16a and 17a.

We recall that we build the LDP approximation (in red) by using the cost of the trajectories satisfying the system (28) between the attractors and the saddle points of the system (4). The cost of these trajectories is known to be optimal when there exists a solution V of the equation (25) which verifies the relations (26), which can generally happen only under symmetry conditions. This is not the case nor for the 3 genes network of Figure 16b when there is no symmetry between the interactions, neither for the 4 genes network of Figure 17b. Then, we could expect that these LDP approximations would be far from the Monte-Carlo and AMS computations, especially for the 4 genes network, since we have no symmetry between the interactions, not only in value but also in sign. However, we observe that the approximations given by our method seem to remain relatively accurate.

Figure 16:
  • Download figure
  • Open in new tab
Figure 16:

16a: A random trajectory associated to the non-symmetric toggle-switch network of 3 genes with ε = 1/8. The network is associated to 2 attractors only: Z−−− all the genes inactive, and Z+−− gene 1 active and gene 2 and gene 3 inactive due to the inhibitions. 16b: Analogy of Figure 8 between Z−−− and Z+−−. We see that the analytical approximations of the transition rates are very accurate although we had no theoretical evidence for the trajectories computed by the method presented in Section 4.3 to be optimal.

Figure 17:
  • Download figure
  • Open in new tab
Figure 17:

17a: A random trajectory associated to the non-symmetric 4 genes network with ε = 1/8. The network is associated to 3 attractors: Z−−−− all the genes inactive, Z+++− genes 1-2-3 active and gene 4 inactive, and Z++++ all the genes active. 17b: Analogy of Figure 8 between Z++++ and Z+++−. The analytical approximations seem to become accurate from ε ≃ 1/9.

Footnotes

  • New version of the manuscript

References

  1. [1].↵
    J. C. Mar. “The rise of the distributions: why non-normality is important for understanding the transcriptome and beyond”. In: Biophys Rev. (2019), pp. 89–94.
  2. [2].↵
    A. F. Coskun, U. Eser, and S. Islam. “Cellular identity at the single-cell level”. In: Mol Biosyst 12.10 (2016), pp. 2965–2979.
    OpenUrl
  3. [3].↵
    N. Moris and A. M. Arias. “The Hidden Memory of Differentiating Cells”. In: Cell Syst 5.3 (2017), pp. 163–164.
    OpenUrl
  4. [4].↵
    H. Mohammed et al. “Single-Cell Landscape of Transcriptional Heterogeneity and Cell Fate Decisions during Mouse Early Gastrulation”. In: Cell Rep 20.5 (2017), pp. 1215–1228.
    OpenUrlCrossRef
  5. [5].↵
    V. Antolovic et al. “Generation of Single-Cell Transcript Variability by Repression”. In: Curr Biol 27.12 (2017), 1811–1817 e3.
    OpenUrlCrossRef
  6. [6].↵
    S. Semrau et al. “Dynamics of lineage commitment revealed by single-cell transcriptomics of differentiating embryonic stem cells”. In: Nat Commun 8.1 (2017), pp. 1–16.
    OpenUrlCrossRefPubMed
  7. [7].↵
    M. Mojtahedi et al. “Cell Fate Decision as High-Dimensional Critical State Transition”. In: PLoS Biol 14.12 (2016), e2000640.
    OpenUrlCrossRefPubMed
  8. [8].↵
    A. Richard et al. “Single-Cell-Based Analysis Highlights a Surge in Cell-to-Cell Molecular Variability Preceding Irreversible Commitment in a Differentiation Process”. In: PLoS Biol 14.12 (2016), e1002585.
    OpenUrlCrossRef
  9. [9].↵
    A. Moussy et al. “Integrated time-lapse and single-cell transcription studies highlight the variable and dynamic nature of human hematopoietic cell fate commitment”. In: PLoS Biol 15.7 (2017), e2001867.
    OpenUrl
  10. [10].↵
    A. Guillemin et al. “Drugs modulating stochastic gene expression affect the erythroid differentiation process”. In: PLOS ONE 14.11 (2019), e0225166.
    OpenUrl
  11. [11].↵
    M.P. Stumpf et al. “Stem Cell Differentiation as a Non-Markov Stochastic Process”. In: Cell Systems 5 (2017), pp. 268–282.
    OpenUrl
  12. [12].↵
    C. H. Waddington. The strategy of the genes. Routledge, 1957.
  13. [13].↵
    S. Huang and D. E. Ingber. “A non-genetic basis for cancer progression and metastasis: self-organizing attractors in cell regulatory networks”. In: Breast disease 26.1 (2007), pp. 27–54.
    OpenUrl
  14. [14].↵
    N. Moris, C. Pina, and A. M. Arias. “Transition states and cell fate decisions in epigenetic landscapes”. In: Nat Rev Genet 17.11 (2016), pp. 693–703.
    OpenUrlCrossRefPubMed
  15. [15].↵
    E. Braun. “The unforeseen challenge: from genotype-to-phenotype in cell populations”. In: Rep Prog Phys 78.3 (2015), p. 036602.
    OpenUrl
  16. [16].↵
    S. A. Morris. “The evolving concept of cell identity in the single cell era”. In: Development 146.12 (2019), pp. 1–5.
    OpenUrlCrossRef
  17. [17].↵
    M. Bizzarri et al. “Gravity Constraints Drive Biological Systems Toward Specific Organization Patterns: Commitment of cell specification is constrained by physical cues”. In: Bioessays (2017).
  18. [18].↵
    K. R. Moon et al. “Manifold learning-based methods for analyzing single-cell RNA-sequencing data”. In: Current Opinion in Systems Biology 7 (2018), pp. 36–46.
  19. [19].↵
    S. Kauffman. “A proposal for using the ensemble approach to understand genetic regulatory networks”. In: Journal of theoretical biology 230.4 (2004), pp. 581–590.
    OpenUrlCrossRefPubMedWeb of Science
  20. [20].↵
    S. Huang et al. “Cell fates as high-dimensional attractor states of a complex gene regulatory network”. In: Physical review letters 94.12 (2005), p. 128701.
    OpenUrlCrossRefPubMed
  21. [21].↵
    H. Clevers et al. “What Is Your Conceptual Definition of “Cell Type” in the Context of a Mature Organism?” In: Cell Systems 4 (2017), pp. 255–259.
    OpenUrl
  22. [22].↵
    J. Wang et al. “The potential landscape of genetic circuits imposes the arrow of time in stem cell differentiation”. In: Biophysical journal 99.1 (2010), pp. 29–39.
    OpenUrlCrossRefPubMedWeb of Science
  23. [23].↵
    J. Wang et al. “Quantifying the Waddington landscape and biological paths for development and differentiation”. In: Proceedings of the National Academy of Sciences 108.20 (2011), pp. 8257–8262.
    OpenUrlAbstract/FREE Full Text
  24. [24].↵
    J. X. Zhou et al. “Quasi-potential landscape in complex multi-stable systems”. In: Journal of the Royal Society Interface 9.77 (2012), pp. 3539–3553.
    OpenUrl
  25. [25].↵
    Y. T. Lin and T. Galla. “Bursting noise in gene expression dynamics: linking microscopic and mesoscopic models”. In: Journal of The Royal Society Interface 13.114 (2016), p. 20150772.
    OpenUrl
  26. [26].↵
    C. Lv et al. “Constructing the energy landscape for genetic switching system driven by intrinsic noise”. In: PLoS one 9.2 (2014).
  27. [27].↵
    P. C. Bressloff. Stochastic processes in cell biology. Vol. 41. Springer, 2014.
  28. [28].↵
    U. Herbach et al. “Inferring gene regulatory networks from single-cell data: a mechanistic approach”. In: BMC Systems Biology 11.1 (2017), p. 105. issn: 1752-0509.
    OpenUrl
  29. [29].↵
    P. C. Bressloff and O. Faugeras. “On the Hamiltonian structure of large deviations in stochastic hybrid systems”. In: Journal of Statistical Mechanics: Theory and Experiment 2017.3 (2017), p. 033206.
    OpenUrl
  30. [30].↵
    M. SH. Ko. “A stochastic model for gene induction”. In: Journal of theoretical biology 153.2 (1991), pp. 181–194.
    OpenUrlCrossRefPubMedWeb of Science
  31. [31].↵
    J. Peccoud and B. Ycart. “Markovian modeling of gene-product synthesis”. In: Theoretical population biology 48.2 (1995), pp. 222–234.
    OpenUrlCrossRefWeb of Science
  32. [32].↵
    D. M Suter et al. “Mammalian genes are transcribed with widely different bursting kinetics”. In: science 332.6028 (2011), pp. 472–474.
    OpenUrlAbstract/FREE Full Text
  33. [33].↵
    G. C. Papanicolaou. “Asymptotic analysis of transport processes”. In: Bulletin of the American Mathematical Society 81.2 (1975), pp. 330–392.
    OpenUrl
  34. [34].↵
    J. M. Newby and J. P. Keener. “An asymptotic analysis of the spatially inhomogeneous velocity-jump process”. In: Multiscale Modeling & Simulation 9.2 (2011), pp. 735–765.
    OpenUrl
  35. [35].↵
    A. Faggionato, D. Gabrielli, and M. R. Crivellari. “Non-equilibrium thermodynamics of piecewise deterministic Markov processes”. In: Journal of Statistical Physics 137.2 (2009), p. 259.
    OpenUrl
  36. [36].↵
    K. Pakdaman, M. Thieullen, and G. Wainrib. “Asymptotic expansion and central limit theorem for multiscale piecewise-deterministic Markov processes”. In: Stochastic Processes and their Applications 122.6 (2012), pp. 2292–2318.
    OpenUrl
  37. [37].↵
    B. K. Chu et al. “Markov State Models of gene regulatory networks”. In: BMC systems biology 11.1 (2017), p. 14.
    OpenUrl
  38. [38].↵
    T. G Kurtz and J. Swanson. “Finite Markov chains coupled to general Markov processes and an application to metastability”. In: arXiv preprint arXiv:1906.03212 (2019).
  39. [39].↵
    M. I. Freidlin and A. D. Wentzell. Random perturbations of dynamical systems. Third. Vol. 260. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Springer, Heidelberg, 2012.
  40. [40].↵
    F. Cérou et al. “A multiple replica approach to simulate reactive trajectories”. In: The Journal of chemical physics 134.5 (2011), p. 054108.
    OpenUrlPubMed
  41. [41].↵
    C-E. Bréhier et al. “Unbiasedness of some generalized adaptive multilevel splitting algorithms”. In: The Annals of Applied Probability 26.6 (2016), pp. 3559–3601.
    OpenUrl
  42. [42].↵
    A. Dembo and O. Zeitouni. Large Deviations Techniques and Applications. 1998.
  43. [43].↵
    Y. Kifer. Large deviations and adiabatic transitions for dynamical systems and Markov processes in fully coupled averaging. American Mathematical Soc., 2009.
  44. [44].↵
    J. E. Cohen. “Convexity of the dominant eigenvalue of an essentially nonnegative matrix”. In: Proceedings of the American Mathematical Society 81.4 (1981), pp. 657–658.
    OpenUrl
  45. [45].↵
    F. Bouchet and J. Reygner. “Generalisation of the Eyring–Kramers transition rate formula to irreversible diffusion processes”. In: Annales Henri Poincaré. Vol. 17. 12. Springer. 2016, pp. 3499–3532.
    OpenUrl
  46. [46].↵
    N. Berglund. “Kramers’ law: Validity, derivations and generalisations”. In: arXiv preprint arXiv:1106.5799 (2011).
  47. [47].↵
    F. Bouchet et al. “Large deviations in fast-slow systems”. In: J. Stat. Phys. 162.4 (2016), pp. 793–812.
    OpenUrl
  48. [48].↵
    P. Zhou and T. Li. “Construction of the landscape for multi-stable systems: Potential landscape, quasi-potential, A-type integral and beyond”. In: The Journal of chemical physics 144.9 (2016), p. 094109.
    OpenUrlCrossRef
  49. [49].↵
    A. Fathi. “Weak KAM theorem in Lagrangian dynamics preliminary version number 10”. In: by CUP (2008).
  50. [50].↵
    C-E. Bréhier and T. Leliévre. “On a new class of score functions to estimate tail probabilities of some stochastic processes with adaptive multilevel splitting”. In: Chaos: An Interdisciplinary Journal of Nonlinear Science 29.3 (2019), p. 033126.
    OpenUrl
  51. [51].↵
    M. Heymann and E. Vanden-Eijnden. “The geometric minimum action method: A least action principle on the space of curves”. In: Communications on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences 61.8 (2008), pp. 1052–1117.
    OpenUrl
  52. [52].↵
    Y. Li, J. Duan, and X. Liu. “Machine learning framework for computing the most probable paths of stochastic dynamical systems”. In: Physical Review E 103.1 (2021), p. 012124.
    OpenUrl
  53. [53].↵
    R. D. Brackston, A. Wynn, and M. P. Stumpf. “Construction of quasipotentials for stochastic dynamical systems: An optimization approach”. In: Physical Review E 98.2 (2018), p. 022136.
    OpenUrl
  54. [54].↵
    N. P. Gao et al. “Universality of cell differentiation trajectories revealed by a reconstruction of transcriptional uncertainty landscapes from single-cell transcriptomic data”. In: bioRxiv (2020).
  55. [55].↵
    A. Richard et al. “Erythroid differentiation displays a peak of energy consumption concomitant with glycolytic metabolism rearrangements”. In: PLoS One 14.9 (2019), e0221472.
    OpenUrl
  56. [56].↵
    P. Pearce et al. “Learning dynamical information from static protein and sequencing data”. In: Nature communications 10.1 (2019), pp. 1–8.
    OpenUrlCrossRef
  57. [57].↵
    Z. Ma and A. Leijon. “Beta mixture models and the application to image classification”. In: 2009 16th IEEE International Conference on Image Processing (ICIP). IEEE. 2009, pp. 2045–2048.
  58. [58].↵
    A. Pratapa et al. “Benchmarking algorithms for gene regulatory network inference from single-cell transcriptomic data”. In: Nat Methods 17.2 (2020), pp. 147–154.
    OpenUrlCrossRef
  59. [59].↵
    J. C. Wheat et al. “Single-molecule imaging of transcription dynamics in somatic stem cells”. In: Nature (2020).
  60. [60].↵
    J. X. Zhou et al. “Nonequilibrium population dynamics of phenotype conversion of cancer cells”. In: PLoS One 9.12 (2014), e110714.
    OpenUrlCrossRef
  61. [61].↵
    P. B. Gupta et al. “Stochastic state transitions give rise to phenotypic equilibrium in populations of cancer cells”. In: Cell 146.4 (2011), pp. 633–44.
    OpenUrlCrossRefPubMedWeb of Science
  62. [62].↵
    M. Tong et al. “Transcriptomic but not genomic variability confers phenotype of breast cancer stem cells”. In: Cancer Commun (Lond) 38.1 (2018), p. 56.
    OpenUrl
  63. [63].↵
    V. Shahrezaei and P. S. Swain. “Analytical distributions for stochastic gene expression”. In: Proceedings of the National Academy of Sciences 105.45 (2008), pp. 17256–17261.
    OpenUrlAbstract/FREE Full Text
  64. [64].↵
    C. Albayrak et al. “Digital quantification of proteins and mRNA in single mammalian cells”. In: Molecular cell 61.6 (2016), pp. 914–924.
    OpenUrl
  65. [65].↵
    G-W. Li and X. S. Xie. “Central dogma at the single-molecule level in living cells”. In: Nature 475.7356 (2011), pp. 308–315.
    OpenUrlCrossRefPubMedWeb of Science
  66. [66].↵
    B. Schwanhäusser et al. “Global quantification of mammalian gene expression control”. In: Nature 473.7347 (2011), p. 337.
    OpenUrlCrossRefPubMedWeb of Science
  67. [67].↵
    T. Dean and P. Dupuis. “Splitting for rare event simulation: A large deviation approach to design and analysis”. In: Stochastic processes and their applications 119.2 (2009), pp. 562–587.
    OpenUrl
  68. [68].↵
    A. Budhiraja and P. Dupuis. “Multilevel Splitting”. In: Analysis and Approximation of Rare Events. Springer, 2019, pp. 439–469.
  69. [69].↵
    E. Gobet. “Weak approximation of killed diffusion using Euler schemes”. In: Stochastic Process. Appl. 87.2 (2000), pp. 167–197. issn: 0304-4149.
    OpenUrl
  70. [70].↵
    L. C. Evans. Partial differential equations. Providence, R.I.: American Mathematical Society, 2010. isbn: 97808218497430821849743.
Back to top
PreviousNext
Posted October 01, 2021.
Download PDF
Email

Thank you for your interest in spreading the word about bioRxiv.

NOTE: Your email address is requested solely to identify you as the sender of this article.

Enter multiple addresses on separate lines or separate them with commas.
Reduction of a Stochastic Model of Gene Expression: Lagrangian Dynamics Gives Access to Basins of Attraction as Cell Types and Metastabilty
(Your Name) has forwarded a page to you from bioRxiv
(Your Name) thought you would like to see this page from the bioRxiv website.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Share
Reduction of a Stochastic Model of Gene Expression: Lagrangian Dynamics Gives Access to Basins of Attraction as Cell Types and Metastabilty
Elias Ventre, Thibault Espinasse, Charles-Edouard Bréhier, Vincent Calvez, Thomas Lepoutre, Olivier Gandrillon
bioRxiv 2020.09.04.283176; doi: https://doi.org/10.1101/2020.09.04.283176
Digg logo Reddit logo Twitter logo Facebook logo Google logo LinkedIn logo Mendeley logo
Citation Tools
Reduction of a Stochastic Model of Gene Expression: Lagrangian Dynamics Gives Access to Basins of Attraction as Cell Types and Metastabilty
Elias Ventre, Thibault Espinasse, Charles-Edouard Bréhier, Vincent Calvez, Thomas Lepoutre, Olivier Gandrillon
bioRxiv 2020.09.04.283176; doi: https://doi.org/10.1101/2020.09.04.283176

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Subject Area

  • Systems Biology
Subject Areas
All Articles
  • Animal Behavior and Cognition (3513)
  • Biochemistry (7358)
  • Bioengineering (5334)
  • Bioinformatics (20290)
  • Biophysics (10032)
  • Cancer Biology (7753)
  • Cell Biology (11323)
  • Clinical Trials (138)
  • Developmental Biology (6442)
  • Ecology (9962)
  • Epidemiology (2065)
  • Evolutionary Biology (13340)
  • Genetics (9363)
  • Genomics (12594)
  • Immunology (7717)
  • Microbiology (19055)
  • Molecular Biology (7452)
  • Neuroscience (41084)
  • Paleontology (300)
  • Pathology (1232)
  • Pharmacology and Toxicology (2140)
  • Physiology (3169)
  • Plant Biology (6867)
  • Scientific Communication and Education (1275)
  • Synthetic Biology (1899)
  • Systems Biology (5320)
  • Zoology (1089)