1. Introduction
In this paper we revisit the notion of the so-called skeletal decomposition of superprocesses. It is well known that when the survival probability is not 0 or 1, then non-trivial infinite genealogical lines of descent, which we call prolific, can be identified on the event of survival. By now it is also well understood that the process itself can be decomposed along its prolific genealogies, where non-prolific mass is immigrated in a Poissonian way along the stochastically ‘thinner’ prolific skeleton. This fundamental phenomenon was first studied by Evans and O’Connell [Reference Evans and O’Connell15] for superprocesses with quadratic branching mechanism. They showed that the distribution of the superprocess at time $t\geq 0$ can be written as the sum of two independent processes. The first is a copy of the original process conditioned on extinction, while the second process is understood as the superposition of mass that has immigrated continuously along the trajectories of a dyadic branching particle diffusion, which is initiated from a Poisson number of particles. This distributional decomposition was later extended to the spatially dependent case by Engländer and Pinsky [Reference Engländer and Pinsky11].
A pathwise decomposition for superprocesses with general branching mechanism was provided by Berestycki, Kyprianou, and Murillo-Salas [Reference Berestycki, Kyprianou and Murillo-Salas2]. Here the role of the skeleton is played by a branching particle diffusion that has the same motion generator as the superprocess, and the immigration is governed by three independent Poisson point processes. The first one results in what we call continuous immigration along the skeleton, where the so-called excursion measure plays the central role, and it assigns zero initial mass to the immigration process. The second point process discontinuously grafts independent copies of the original process conditioned on extinction onto the path of the skeleton. Finally, additional copies of the original process conditioned on extinction are immigrated off the skeleton at its branch points, where the initial mass of the immigrant depends on the number of offspring at the branch point. The spatially dependent version of this decomposition was considered in [Reference Kyprianou, Pérez and Ren23] and [Reference Eckhoff, Kyprianou and Winkel10].
Other examples of skeletal decompositions for superprocesses include [Reference Salisbury and Verzani33], [Reference Etheridge and Williams13], [Reference Kyprianou and Ren22], [Reference Murillo-Salas and Pérez28], and [Reference Fekete, Palau, Pardo and Pérez17].
In a previous article [Reference Fekete, Fontbona and Kyprianou16] we developed a stochastic differential equation (SDE) approach to study the skeletal decomposition of continuous-state branching processes (CSBPs). These decompositions were by no means new; prolific genealogies for both supercritical and subcritical CSBPs had been described, albeit in the latter case we have to be careful what we mean by ‘prolific’. In particular, in [Reference Bertoin, Fontbona and Martínez3], [Reference Duquesne and Winkel5], and [Reference Kyprianou and Ren22] specifically CSBPs were considered, but since the total mass process of a superprocess with spatially independent branching mechanism is a CSBP, skeletal decompositions for CSBPs also appear as a special case of some of the previously mentioned results.
The results in [Reference Fekete, Fontbona and Kyprianou16] were motivated by the work of Duquesne and Winkel [Reference Duquesne and Winkel5] and Duquesne and Le Gall [Reference Duquesne and Le Gall4]. Duquesne and Winkel, in the context of Lévy trees, provided a parametric family of decompositions for finite-mean supercritical CSBPs that satisfy Grey’s condition. They showed that one can find a decomposition of the CSBP for a whole family of embedded skeletons, where the ‘thinnest’ one is the prolific skeleton with all the infinite genealogical lines of descent, while the other embedded skeletons contain not only the infinite genealogies but also some finite ones grafted onto the prolific tree. On the other hand, Duquesne and Le Gall studied subcritical CSBPs, and using the height process gave a description of the genealogies that survive until some fixed time $T>0$. It is well known that a subcritical CSBP goes extinct almost surely, thus prolific individuals, in the classical sense, do not exist in the population. But since it is possible that the process survives until some fixed time T, individuals who have at least one descendant at time T can be found with positive probability. We call these individuals T-prolific.
The SDE approach provides a common framework for the parametric family of decompositions of Duquesne and Winkel, as well as for the time-inhomogeneous decompositions we get when we decompose the process along its T-prolific genealogies. We note that these finite-horizon decompositions exist for both supercritical and subcritical process. In the subcritical case the SDE representation can be used to observe the behaviour of the system when we condition on survival up to time T, then take T to infinity. Conditioning a subcritical CSBP to survive eternally results in what is known as a spine decomposition, where independent copies of the original process are grafted onto one infinite line of descent, which we call the spine (for more details we refer the reader to [Reference Roelly-Coppoletta and Rouault32], [Reference Lambert24], [Reference Lambert25], [Reference Fittipaldi and Fontbona18], and [Reference Abraham and Delmas1]). And indeed, in [Reference Fekete, Fontbona and Kyprianou16] we see how the skeletal representation becomes, in the sense of weak convergence, a spinal decomposition when conditioning on survival, and in particular how the skeleton thins down to become the spine as $T\rightarrow\infty$.
In this paper our objective is to demonstrate the robustness of this aforementioned method by expanding the SDE approach to the spatial setting of superprocesses. We consider supercritical superprocesses with space-dependent branching mechanism, but in future work we hope to extend results to the time-inhomogeneous case of subcritical processes.
The rest of this paper is organised as follows. In the remainder of this section we introduce our model and fix some notation. Then in Section 2 we remind the reader of some key existing results relevant to the subsequent exposition; in particular, we recall the details of the skeletal decomposition of superprocesses with spatially dependent branching mechanism, as appeared in [Reference Kyprianou, Pérez and Ren23] and [Reference Eckhoff, Kyprianou and Winkel10]. The main result of the paper is stated in Section 3, where we reformulate the result of Section 2 by writing down a coupled SDE, whose second coordinate corresponds to the skeletal process, while the first coordinate describes the evolution of the total mass in system. In Sections 4, 5, and 6 we give the proof of our results.
Superprocess. Let E be a domain of $\mathbb{R}^d$ and let $\mathcal{M}( E)$ denote the space of finite Borel measures on E. Furthermore, let $\mathcal{M}( E)^\circ\,:\!=\mathcal{M}( E)\setminus\{ 0\}$, where 0 is the null measure. We are interested in a strong Markov process X on E taking values in $\mathcal{M}( E)$. The process is characterised by two quantities $\mathcal{P}$ and $\psi$. Here $(\mathcal{P}_t)_{t\geq 0}$ is the semigroup of an $\mathbb{R}^d$-valued diffusion killed on exiting E, and $\psi$ is the so-called branching mechanism. The latter takes the form
where $\alpha$ and $\beta\geq 0$ are bounded measurable mappings from E to $\mathbb{R}$ and $[0,\infty)$ respectively, and $(u\wedge u^2)m(x,{\textnormal d} u)$ is a bounded kernel from E to $(0,\infty)$.
For technical reasons we assume that $\mathcal{P}$ is a Feller semigroup whose generator takes the form
where $a\colon E\rightarrow \mathbb{R}^{d\times d}$ is the diffusion matrix that takes values in the set of symmetric, positive definite matrices, and $b\colon E\rightarrow \mathbb{R}^d$ is the drift term.
Then the one-dimensional distributions of X can be characterised as follows. For all $\mu\in \mathcal{M}( E)$ and $f\in B^+( E)$, where $B^+(E)$ denotes the non-negative measurable functions on E, we have
where $u_f(x,t)$ is the unique non-negative, locally bounded solution to the integral equation
Here we use the notation
For each $\mu\in \mathcal{M}( E)$ we let $\mathbb{P}_\mu$ denote the law of the process X issued from $X_0=\mu$. The process $(X,\mathbb{P}_\mu)$ is called a $(\mathcal{P},\psi)$-superprocess.
For more details of the above see Fitzsimmons [Reference Fitzsimmons19]; for a general overview of superprocesses we refer the reader to the books of Dynkin [Reference Dynkin7, Reference Dynkin8], Etheridge [Reference Etheridge12], Le Gall [Reference Le Gall26], and Li [Reference Li27]
Next we recall the SDE representation of $(X,\mathbb{P}_\mu)$ (for more details see Chapter 7 of [Reference Li27]). Recall that m was previously defined in (1.1). We assume that it satisfies the integrability condition
Let $C_0( E)^+$ denote the space of non-negative continuous functions on E vanishing at infinity. We assume that $\alpha$ and $\beta$ are continuous, furthermore $x\mapsto(u \wedge u^2)m(x,{\textnormal d} u)$ is continuous in the sense of weak convergence, and
maps $C_0( E)^+$ into itself.
Next define $\Delta X_s=X_s-X_{s-}$. As a random measure difference, if $s>0$ is such that $\Delta X_s\neq 0$, it can be shown that $\Delta X_s = u_s \delta_{x_s}$ for some $u_s\in(0,\infty)$ and $x_s\in E$. Suppose that for the countable set of times, say $(s_i, i\in\mathbb{N})$, that $\Delta X_{s_i}\neq 0$, $i\in\mathbb{N}$, we enumerate the pairs $((u_i,x_i), i\in\mathbb{N})$. We say that $N({\textnormal d} s, {\textnormal d} x, {\textnormal d} u)$, $s\geq 0$ is the optional random measure on $[0,\infty)\times E\times(0,\infty)$, which can otherwise be identified as $\sum_{i\in\mathbb{N}} \delta_{(s_i, x_i, u_i)} ({\textnormal d} s, {\textnormal d} x, {\textnormal d} u)$. Let $\hat{N}( {\textnormal d} s,{\textnormal d} x, {\textnormal d} u)$ denote the predictable compensator of $N({\textnormal d} s,{\textnormal d} x, {\textnormal d} u)$. It can be shown that $\hat{N}( {\textnormal d} s,{\textnormal d} x, {\textnormal d} u)={\textnormal d} s K(X_{s-}, {\textnormal d} x, {\textnormal d} u)$, where, given $\mu\in\mathcal{M}(E)$,
If we denote the compensated measure by $\tilde{N}$, then for any $f\in D_0(\mathcal{L})$ (the set of functions in $C_0(E)$ that are also in the domain of $\mathcal{L}$) we have
where $t\mapsto M_t^c({\kern1.3pt}f)$ is a continuous local martingale with quadratic variation $2\langle \beta f^2, X_{t-}\rangle \,{\textnormal d} t$ and
is a purely discontinuous local martingale. Here and throughout the paper, we prefer to write $\langle{\kern1.3pt} f, u\delta_x \rangle$ in place of uf(x) as a reminder that it is the increment of the process $\langle{\kern1.3pt} f, X_t\rangle$, $t\geq 0$.
The representation (1.4) is what we will use in Section 3 when developing the SDE approach to the skeletal decomposition of $(X,\mathbb{P}_\mu)$. However, before we proceed with this line of analysis, we first need to recall the details of this skeletal decomposition, as it not only motivates our results but also proves to be helpful in understanding the structure of our SDE.
2. Skeletal decomposition
Recall that the main idea behind the skeletal decomposition is that under certain conditions we can identify prolific genealogies in the population, and by immigrating non-prolific mass along the trajectories of these prolific genealogies, we can recover the law of the original superprocess. The infinite genealogies are described by a Markov branching process whose initial state is given by a Poisson random measure, while traditionally the immigrants are independent copies of the original process conditioned to become extinct.
In this section we first characterise the two components, then explain how to construct the skeletal decomposition from these building blocks. The results of this section are lifted from [Reference Kyprianou, Pérez and Ren23] and [Reference Eckhoff, Kyprianou and Winkel10].
As we have mentioned, the skeleton is often constructed using the event of extinction, that is, the event $\mathcal{E}_{\textrm{fin}}=\{\langle 1, X_t\rangle=0$ for some $t>0\}$. This guides the skeleton particles into regions where the survival probability is high. If we write $w(x)=-\log\mathbb{P}_{\delta_x}(\mathcal{E_{\textrm{fin}}})$, and assume that $\mu\in\mathcal{M}(E)$ is such that $\langle w,\mu\rangle<\infty$, then it is not hard to see that $\mathbb{P}_\mu(\mathcal{E}_{\textrm{fin}})=\exp \lbrace -\langle w, \mu\rangle \rbrace.$ Furthermore, by conditioning $\mathcal{E}_{\textrm{fin}}$ on $\mathcal{F}_t\,:\!=\sigma(X_s,s\leq t)$, we get that
Eckhoff, Kyprianou, and Winkel [Reference Eckhoff, Kyprianou and Winkel10] point out that in order to construct a skeletal decomposition along those genealogies that avoid the behaviour specified by w (in this case ‘extinction’), all we need is that the function w gives rise to a multiplicative martingale $(({\textrm{e}}^{-\langle w,X_t\rangle}, t\geq 0),\mathbb{P}_\mu)$. In particular, a skeletal decomposition is given for any choice of a martingale function w which satisfies the following conditions:
• for all $x\in E$ we have $w(x)>0$ and $\sup_{x\in E} w(x)<\infty $, and
-
• $\mathbb{E}_\mu({\textrm{e}}^{-\langle w,X_t\rangle})={\textrm{e}}^{-\langle w,\mu \rangle}$ for all $\mu\in\mathcal{M}_c(E)$, $t\geq 0$ (here $\mathcal{M}_c(E)$ denotes the set of finite, compactly supported measures on E).
The condition $w(x)>0$ implicitly hides the notion of supercriticality, as it ensures that survival happens with positive probability. Note, however, that ‘survival’ can be interpreted in many different ways. For example, the choice of $\mathcal{E}_{\textrm{fin}}$ results in skeleton particles that are simply part of some infinite genealogical line of descent, but we could also define surviving genealogies as those that visit a compact domain in E infinitely often.
Remark 1. Kyprianou, Pérez, and Ren [Reference Kyprianou, Pérez and Ren23] and Eckhoff et al. [Reference Eckhoff, Kyprianou and Winkel10] have shown the existence of the skeletal decomposition under a slightly more general set-up, where w is only locally bounded from above. Note, however, that their proof consists of first establishing dealing with the case when w is uniformly bounded, and then appealing to a localisation argument to relax this to the aforesaid local boundedness. Our SDE approach requires the case of uniform boundedness, but a localisation process can in principle be used to relax the assumption as in the aforementioned literature.
We will also make the additional assumption that w is in the domain of the generator $\mathcal{L}$. This is predominantly because of the use of partial differential equations in our analysis rather than integral equations.
2.1. Skeleton
First we identify the branching particle system that takes the role of the skeleton in the decomposition of the superprocess. In general, a Markov branching process $Z=(Z_t,t\geq 0)$ takes values in $\mathcal{M}_a(E)$ (the set of finite, atomic measures in E), and it can be characterised by the pair $(\mathcal{P},F)$, where $\mathcal{P}$ is the semigroup of a diffusion and F is the branching generator which takes the form
Here q is a bounded, measurable mapping from E to $[0,\infty)$, and $\{ p_n(x),n\geq 0\}$, $x\in E$ are measurable sequences of probability distributions. For $\nu\in\mathcal{M}_a(E)$ we denote the law of the process Z issued from $\nu$ by $\textbf{P}_\nu$. Then we can describe $(Z,\textbf{P}_\nu)$ as follows. We start with initial state $Z_0=\nu$. Particles move according to $\mathcal{P}$, and at a spatially dependent rate $q(x)\,{\textnormal d} t$ a particle is killed and is replaced by n offspring with probability $p_n(x)$. The offspring particles then behave independently and according to the same law as their parent.
In order to specify the parameters of Z we first need to introduce some notation. Let $\xi=(\xi_t,t\geq 0)$ be the diffusion process on $E\cup \{\dagger\}$ (the one-point compactification of E with a cemetery state) corresponding to $\mathcal{P}$, and let us denote its probabilities by $\{\Pi_x,x\in E\}$. (Note that the previously defined martingale function w can be extended to $E\cup \{\dagger\}$ by defining $w(\dagger)=0$). Then, for all $x\in E$,
is a positive local martingale, and hence a supermartingale. (To see why this is true we refer the reader to the discussion in Section 2.1.1 of [Reference Eckhoff, Kyprianou and Winkel10].) Now let $\tau_E=\inf\{ t>0\colon \xi_t\in\{\dagger\}\}$, and consider the following change of measure:
which uniquely determines a family of (sub)probability measures $\{\Pi_x^w,x\in E\}$ (see e.g. [Reference Ethier and Kurtz14]).
If we let $\mathcal{P}^w$ denote the semigroup of the $E\cup\{\dagger\}$-valued process whose probabilities are $\{\Pi^w_x,x\in E\}$, then it can be shown that the generator corresponding to $\mathcal{P}^w$ is given by
where $\mathcal{L}_0^w u=w^{-1}\mathcal{L}(wu)$ whenever u is in the domain of $\mathcal{L}$. Note that $\mathcal{L}^w$ is also called an h-transform of the generator $\mathcal{L}$ with $h=w$. The theory of h-transforms for measure-valued diffusions was developed in [Reference Engländer and Pinsky11].
Intuitively, if
defines a martingale function with the previously introduced conditions for some tail event $\mathcal{E}$, then the motion associated with $\mathcal{L}^w$ forces the particles to avoid the behaviour specified by $\mathcal{E}$. In particular, when $\mathcal{E}=\mathcal{E}_{\textrm{fin}}$ then $\mathcal{P}^w$ encourages $\xi$ to visit domains where the global survival rate is high.
Now we can characterise the skeleton process of $(X,\mathbb{P}_\mu)$ associated with w. In particular, $Z=(Z_t,t\geq 0)$ is a Markov branching process with diffusion semigroup $\mathcal{P}^w$ and branching generator
where
and $p_0(x)=p_1(x)=0$, and for $n\geq 2$
Here we used the notation
We refer to the process Z as the $(\mathcal{P}^w, F)$ skeleton.
2.2. Immigration
Next we characterise the process that we immigrate along the previously introduced branching particle system. To this end let us define the function
which can be written as
where
and
Note that under our assumptions $\psi^*$ is a branching mechanism of the form (1.1). We denote the probabilities of the $(\mathcal{P}, \psi^*)$-superprocess by $(\mathbb{P}^*_\mu)_{\mu\in\mathcal{M}(E)}$.
If $\mathcal{E}$ is the event associated with w (see (2.1)), and $\langle w,\mu\rangle<\infty$, then we have
In particular, when $\mathcal{E}=\mathcal{E}_{\textrm{fin}}$, then $\mathbb{P}^*_\mu$ is the law of the superprocess conditioned to become extinct.
2.3. Skeletal path decomposition
Here we give the precise construction of the skeletal decomposition that we introduced in a heuristic way at the beginning of this section. Let $\mathbb{D}([0,\infty)\times \mathcal{M}(E))$ denote the space of measure valued càdlàg function. Suppose that $\mu\in\mathcal{M}(E)$, and let Z be a $(\mathcal{P}^w,F)$-Markov branching process with initial configuration consisting of a Poisson random field of particles in E with intensity $w(x)\mu({\textnormal d} x)$. Next, dress the branches of the spatial tree that describes the trajectory of Z in such a way that a particle at the space-time position $(x,t)\in E\times[0,\infty)$ has a $\mathbb{D}([0,\infty)\times \mathcal{M}(E))$-valued trajectory grafted onto it, say $\omega = (\omega_t, t\geq0)$, with rate
Here $\mathbb{Q}^*_x$ is the excursion measure on the space $\mathbb{D}([0,\infty)\times \mathcal{M}(E))$ which satisfies
for $x\in E$, $t\geq 0$, and $f\in B^+_b(E)$ (the space of non-negative, bounded measurable functions on E), where $u^*_f(x,t)$ is the unique solution to (1.3) with the branching mechanism $\psi$ replaced by $\psi^*$. (For more details of excursion measures see [Reference Dynkin and Kuznetsov9].) Moreover, when a particle in Z dies and gives birth to $n\geq 2$ offspring at spatial position $x\in E$, with probability $\eta_n(x,{\textnormal d} y)\mathbb{P}^*_{y\delta_x}({\textnormal d} \omega)$ an additional $\mathbb{D}([0,\infty)\times \mathcal{M}(E))$-valued trajectory, $\omega$, is grafted onto the space-time branching point, where
Overall, we have three different types of immigration processes that contribute to the dressing of the skeleton. In particular, the first term of (2.5) is what we call ‘continuous immigration’ along the skeleton, while the second term is referred to as the ‘discontinuous immigration’, and finally (2.6) corresponds to the so-called ‘branch-point immigration’.
Now we define $\Lambda_t$ as the total mass from the dressing present at time t together with the mass present at time t of an independent copy of $(X,\mathbb{P}_\mu^*)$ issued at time 0. We denote the law of $(\Lambda, Z)$ by $\textbf{P}_\mu$. Then Kyprianou et al. [Reference Kyprianou, Pérez and Ren23] showed that $(\Lambda,\textbf{P}_\mu)$ is Markovian and has the same law to $(X,\mathbb{P}_\mu)$. Furthermore, under $\textbf{P}_\mu$, conditionally on $\Lambda_t$, the measure $Z_t$ is a Poisson random measure with intensity $w(x)\Lambda_t({\textnormal d} x)$.
3. SDE representation of the dressed tree
Recall that our main motivation is to reformulate the skeletal decomposition of superprocesses using the language of SDEs. Thus, in this section, after giving an SDE representation of the skeletal process, we derive the coupled SDE for the dressed skeleton, which simultaneously describes the evolution of the skeleton and the total mass in the system.
3.1. SDE of the skeleton
We use the arguments on page 3 of [Reference Xiong34] to derive the SDE for the branching particle diffusion, that will act as the skeleton. Let $(\xi_t,t\geq 0)$ be the diffusion process corresponding to the Feller semigroup $\mathcal{P}$. Since the generator of the motion is given by (1.2), the process $\xi$ satisfies
where $\sigma\colon \mathbb{R}^d\rightarrow\mathbb{R}^d$ is such that $\sigma(x)\sigma^{\texttt{T}}(x)=a(x)$ (where $\texttt{T}$ denotes matrix transpose), and $(B_t,t\geq 0)$ is a d-dimensional Brownian motion (see e.g. Chapter 1 of [Reference Pinsky30]).
It is easy to verify that if $(\tilde{\xi}_t,t\geq 0)$ is the diffusion process under $\mathcal{P}^w$, then it satisfies
where $\nabla w$ is the gradient of w. To simplify computations, define the function $\tilde{b}$ on E given by
For $h\in C^2_b(E)$ (the space of bounded, twice differentiable continuous functions on E), using Itô’s formula (see e.g. Section 8.3 of [Reference øksendal29]), we get
where $x^{\texttt{T}}$ denotes the transpose of x, Tr is the trace operator, and $H_h$ is the Hessian of h with respect to x, that is,
Next, summing over all the particles live at time t, the collection of which we denote by $\mathcal{I}_t$, gives
where for each $\alpha$, $B^\alpha$ is an independent copy of B, and $\xi^\alpha$ is the current position of individual $\alpha\in\mathcal{I}_t$.
If an individual branches at time t, then we have
Here $k_\alpha$ is the number of children of individual $\alpha$, which has distribution $\{p_k,k=0,1,\dots\}$.
Simple algebra shows that
thus by combining (3.1) and (3.2) we get
where $V_t^c$ is a continuous local martingale given by
and $N^\dagger_s$ is an optional random measure on $[0,\infty)\times E\times \mathbb{N}$ with predictable compensator of the form $\hat{N}^\dagger({\textnormal d} s, {\textnormal d} x, {\textnormal d} \{k\}) = {\textnormal d} s K^\dagger(Z_{s-}, {\textnormal d} x, {\textnormal d} \{k\})$ such that, for $\mu\in\mathcal{M}(E)$,
where q, $p_k(x)$ are given by (2.2), (2.3) and $\#$ is the counting measure. The reader will note that, for a (random) measure $M\in\mathcal{M}_a(E)$, we regularly interchange the notion of $\sum_{k\in \mathbb{N}} \cdot$ with $\int_{\mathbb{N}}\cdot M({\textnormal d}\{k\})$.
Note that from (3.4) it is easy to see that the quadratic variation of $V_t^c$ is
3.2. Thinning of the SDE
Now we will see how to modify the SDE given by (1.4) in order to separate out the different types of immigration processes. We use ideas developed in [Reference Fekete, Fontbona and Kyprianou16].
Recall that the SDE describing the superprocess $(X,\mathbb{P}_\mu)$ takes the following form:
Here $M_t^c({\kern1.3pt}f)$ is as in (1.4), and $N( {\textnormal d} s, {\textnormal d} x, {\textnormal d} u)$ is an optional random measure on $[0,\infty)\times E\times (0,\infty)$ such that, given $\mu\in\mathcal{M}(E)$, it has predictable compensator given by $\hat{N}( {\textnormal d} s, {\textnormal d} x, {\textnormal d} u)={\textnormal d} s K(X_{s-}, {\textnormal d} x, {\textnormal d} u)$, where
Moreover, $\tilde{N}( {\textnormal d} s,{\textnormal d} x, {\textnormal d} u)$ is the associated compensated version of $N( {\textnormal d} s, {\textnormal d} x, {\textnormal d} u)$. Let
denote some enumeration of the atoms of $N( {\textnormal d} s,{\textnormal d} x, {\textnormal d} u)$. Next we introduce independent marks to the atoms of N, that is, we define the random measure
whose predictable compensator ${\textnormal d} s\mathcal{K}(X_{s-},{\textnormal d} x, {\textnormal d} u,{\textnormal d}\{k\})$ has the property that, for $\mu\in\mathcal{M}(E)$,
Now we can define three random measures by
Using Proposition 10.47 of [Reference Jacod21], we see that $N^0$, $N^1$, and $N^2$ are also optional random measures and their compensators ${\textnormal d} s K^0(X_{s-},{\textnormal d} x, {\textnormal d} u)$, ${\textnormal d} s K^1(X_{s-},{\textnormal d} x, {\textnormal d} u)$, and ${\textnormal d} sK^2(X_{s-},{\textnormal d} x, {\textnormal d} u)$ satisfy
for $\mu\in\mathcal{M}(E)$. Using these processes we can rewrite (3.5), so we get
where we have used the fact that
Recalling (2.4), we see that the first line of the right-hand side of (3.6) corresponds to the dynamics of a $(\mathcal{P},\psi^*)$-superprocess. Our aim now is to link the remaining three terms to the three types of immigration along the skeleton, and write down a system of SDEs that describe the skeleton and total mass simultaneously. Heuristically speaking, this system of SDEs will consist of (3.3) and a second SDE which looks a little bit like (3.6) (note that the latter has no dependence on the process Z as things stand). To some extent, we can think of the SDE (3.6) as what one might see when ‘integrating out’ (3.3) from the aforesaid second SDE in the coupled system; indeed, this will be one of our main conclusions.
3.3. Coupled SDE
Following the ideas of the previous sections we introduce the following driving sources of randomness that we will use in the construction of our coupled SDE. Our coupled system will describe the evolution of the pair of random measures $(\Lambda, Z) = ((\Lambda_t, Z_t), t\geq 0)$ on $\mathcal{M}(E)\times\mathcal{M}_a(E)$.
• Let ${\texttt N}^0({\textnormal d} s,{\textnormal d} x, {\textnormal d} u)$ be an optional random measure on $[0,\infty)\times E\times (0,\infty)$, which depends on $\Lambda$ with predictable compensator $\hat{{\texttt N}}^0({\textnormal d} s,{\textnormal d} x, {\textnormal d} u) = {\textnormal d} s\, \texttt{K}^0(\Lambda_{s-}, {\textnormal d} x, {\textnormal d} u)$, where, for $\mu\in \mathcal{M}(E)$,
\begin{equation*}\texttt{K}^0(\mu, {\textnormal d} x, {\textnormal d} u)= \mu({\textnormal d} x) \,{\textrm{e}}^{-w(x)u} m(x,{\textnormal d} u),\end{equation*}and $\tilde{{\texttt N}}^{0}({\textnormal d} s, {\textnormal d} x, {\textnormal d} u)$ is its compensated version.-
• Let ${\texttt N}^1({\textnormal d} s,{\textnormal d} x, {\textnormal d} u)$ be an optional random measure on $[0,\infty)\times E\times (0,\infty)$, dependent on Z, with predictable compensator $\hat{{\texttt N}}^1({\textnormal d} s,{\textnormal d} x, {\textnormal d} u) = {\textnormal d} s\, \texttt{K}^1(Z_{s-}, {\textnormal d} x, {\textnormal d} u)$ so that, for $\mu\in\mathcal{M}_a(E)$,
\begin{equation*}\texttt{K}^1(\mu, {\textnormal d} x, {\textnormal d} u)= \mu({\textnormal d} x) \,{\textrm{e}}^{-w(x)u} m(x,{\textnormal d} u){.}\end{equation*} -
• Define ${\texttt N}^2({\textnormal d} s,{\textnormal d}\rho, {\textnormal d} x, {\textnormal d} u)$ as an optional random measure on $[0,\infty)\times \mathbb{N}\times E\times (0,\infty)$ also dependent on Z, with predictable compensator
\begin{equation*}\hat{{\texttt N}}^2({\textnormal d} s,{\textnormal d}\{k\},{\textnormal d} x, {\textnormal d} u)= {\textnormal d} s\, \texttt{K}^2(Z_{s-}, {\textnormal d}\{k\}, {\textnormal d} x, {\textnormal d} u)\end{equation*}so that, for $\mu\in\mathcal{M}_a(E)$,\begin{equation*}\texttt{K}^2(\mu, {\textnormal d}\{k\}, {\textnormal d} x, {\textnormal d} u)= \mu({\textnormal d} x) q(x) p_k(x)\eta_k(x,{\textnormal d} u) \#({\textnormal d}\{k\}),\end{equation*}where q, $p_k({\textnormal d} x)$, and $\eta_k(x,{\textnormal d} u)$ are given by (2.2), (2.3), and (2.6).
Now we can state our main result.
Theorem 1. Consider the following system of SDEs for $f,h\in D_0(\mathcal{L})$:
inducing probabilities $\textbf{P}_{(\mu,\nu)}$, $\mu\in\mathcal{M}(E)$, $\nu\in\mathcal{M}_a(E)$, where $(U_t^c({\kern1.3pt}f),t\geq 0)$ is a continuous local martingale with quadratic variation $2 \langle\beta f^2,\Lambda_{t-}\rangle\,{\textnormal d} t$, and $(V_t^c(h),t\geq 0)$ is a continuous local martingale with quadratic variation $\langle (\nabla h)^{\texttt{T}} a\nabla h,Z_{t-}\rangle\,{\textnormal d} t$. (Note that $\partial_z\psi^*(x,0) = \psi'(x,w(x))$ is another way of identifying the drift term in the first integral above.) With an immediate abuse of notation, write $\textbf{P}_\mu = \textbf{P}_{(\mu, \textrm{Po}(w\mu))}$, where $\textrm{Po}(w\mu)$ is an independent Poisson random measure on E with intensity $w\mu$. Then we have the following.
(i) There exists a unique weak solution to the SDE (3.7) under each $\textbf{P}_{(\mu,\nu)}$.
-
(ii) Under each $\textbf{P}_\mu$, for $t\geq 0$, conditional on $\mathcal{F}_t^\Lambda=\sigma(\Lambda_s,s\leq t)$, $Z_t$ is a Poisson random measure with intensity $w(x)\Lambda_t({\textnormal d} x)$.
-
(iii) The process $(\Lambda_t,t\geq 0)$, with probabilities $\textbf{P}_{\mu}$, $\mu\in\mathcal{M}(E)$, is Markovian and a weak solution to (3.5).
The rest of the paper is dedicated to the proof of this theorem, which we split over several subsections.
4. Proof of Theorem 1 (i): existence
Consider the pair $(\Lambda, Z)$, where Z is a $(\mathcal{P}^w,F)$ branching Markov process with $Z_0=\nu$ for some $\nu\in\mathcal{M}_a(E)$, and whose jumps are coded by the coordinates of the random measure $\mathtt{N}^2$. Furthermore, we define $\Lambda_t=X^*_t+D_t$, where $X^*$ is an independent copy of the $(\mathcal{P},\psi^*)$-superprocess with initial value $X^*_0=\mu$, $\mu\in\mathcal{M}(E)$, and the process $(D_t,t\geq 0)$ is described by
where $f\in D_0(\mathcal{L})$ and, with a slight abuse of the notation that was introduced preceding Theorem 1,
• $\mathtt{N}^1$ is an optional random measure on $[0,\infty)\times E \times(0,\infty)\times\mathbb{D}([0,\infty)\times \mathcal{M}(E))$ with predictable compensator
\begin{equation*}\hat{\mathtt{N}}^1({\textnormal d} s,{\textnormal d} x, {\textnormal d} u,{\textnormal d}\omega)={\textnormal d} s Z_{s-}({\textnormal d} x)u\,{\textrm{e}}^{-w(x)u}m(x,{\textnormal d} u)\mathbb{P}_{u\delta_x}^*({\textnormal d}\omega),\end{equation*}-
• $\mathtt{N}^2$ is an optional random measure on $[0,\infty)\times \mathbb{N}\times E \times(0,\infty)\times \mathbb{D}([0,\infty)\times \mathcal{M}(E))$ with predictable compensator
\begin{equation*}\hat{\mathtt{N}}^2({\textnormal d} s,{\textnormal d}\{k\},{\textnormal d} x,{\textnormal d} u, {\textnormal d}\omega)={\textnormal d} s Z_{s-}({\textnormal d} x)q(x)p_k(x)\eta_k(x,{\textnormal d} u)\mathbb{P}_{u\delta_x}^*({\textnormal d}\omega)\#({\textnormal d} \{k\}),\end{equation*} -
• $\mathtt{N}^*$ is an optional random measure on $[0,\infty)\times \mathbb{D}([0,\infty)\times \mathcal{M}(E))$ with predictable compensator
(4.2) \begin{equation}\hat{\mathtt{N}}^*({\textnormal d} s,{\textnormal d}\omega)={\textnormal d} s\int_E 2\beta(x) Z_{s-}({\textnormal d} x)\mathbb{Q}^*_{x}({\textnormal d}\omega),\end{equation}where\begin{equation*}\mathbb{Q}_x^*(1-{\textrm{e}}^{-\langle{\kern1.3pt} f,\omega_t\rangle})=-\log\mathbb{E}_{\delta_x}^*({\textrm{e}}^{-\langle{\kern1.3pt} f, X_t\rangle})=u_f^*(x,t).\end{equation*}
Note that we have used $\cdot$ to denote marginalisation so, for example,
and, to be consistent with previous notation, we also have, for example,
We claim that the pair $(\Lambda, Z)$ as defined above solves the coupled system of SDEs (3.7). To see why, start by noting that Z solves the second coordinate component of (3.7) by definition of it being a spatial branching process; see (3.3). In dealing with the term ${\textnormal d} D_t$, we first note that the random measures $ \mathtt{N}^1$ and $ \mathtt{N}^2$ have finite activity through time, whereas $ \mathtt{N}^*$ has infinite activity. Suppose we write $\mathtt{I}^{(i)}_t$, $i =1,2,3$, for the three integrals on the right-hand side of (4.1), respectively. Taking the case of $\mathtt{I}^{(1)}_t$, if t is a jump time of $ \mathtt{N}^1({\textnormal d} t, {\textnormal d} x, {\textnormal d} u, {\textnormal d} \omega)$, then $\Delta \mathtt{I}^{(1)}_t = \langle{\kern1.3pt} f, \omega_0 \rangle,$ noting in particular that $ \omega_0 = u\delta_x$. A similar argument produces $\Delta \mathtt{I}^{(2)}_t = \langle{\kern1.3pt} f, \omega_0 \rangle = \langle{\kern1.3pt} f, u\delta_x\rangle $, when t is a jump time of $ \mathtt{N}^2({\textnormal d} t ,{\textnormal d} \{k\}, {\textnormal d} x, {\textnormal d} u, {\textnormal d}\omega)$. In contrast, on account of the excursion measures $(\mathbb{Q}^*_{x}, x\in E)$ having the property that $\mathbb{Q}^*_{x}(\omega_0 >0) =0$, we have $\Delta \mathtt{I}^{(3)}_t =0$. Nonetheless, the structure of the compensator (4.2) implies that there is a rate of arrival (of these zero contributions) given by
where we have used the fact that $\mathbb{E}^*_{\delta_x}[\langle{\kern1.3pt} f, X_t\rangle] = \mathbb{Q}^*_{x}(\langle{\kern1.3pt} f, \omega_t \rangle)$, $t\geq 0$; see e.g. [Reference Dynkin and Kuznetsov9].
Now suppose that t is not a jump time of $\mathtt{I}^{(i)}_t$, $i =1,2,3$. In that case, we note that $\langle{\kern1.3pt} f, \Lambda_{t}\rangle = \langle{\kern1.3pt} f, X^*_{t} \rangle + \langle{\kern1.3pt} f, D_{t} \rangle $ is simply the aggregation of mass that has historically immigrated and evolved under $\mathbb{P}^*$. As such (comparing with (1.4), for example),
where
and $U^c({\kern1.3pt}f)$ was defined immediately above Theorem 1. As such, we see from (4.3) that the pair $(\Lambda,Z)$ defined in this section provides a solution to (3.7).
5. Some integral and differential equations
The key part of our reasoning in proving parts (ii) and (iii) of Theorem 1 will be to show that
where X satisfies (3.5). Moreover, the key idea behind the proof of (5.1) is to fix $T>0$ and $f,h\in D_0(\mathcal{L})$, and choose time-dependent test functions $f^T$ and $h^T$ in such a way that the processes
and
have constant expectations on [0, T]. The test functions are defined as solutions to some partial differential equations with final value conditions $f^T(x,T)=f(x)$ and $h^T(x,T)=h(x)$. This, together with the fact that $\Lambda_0=X_0=\mu$, and that $Z_0$ is a Poisson random measure with intensity $w(x)\Lambda_0({\textnormal d} x)$, will then give us (5.1).
Thus, to prove (5.1) and hence parts (ii) and (iii) of Theorem 1, we need the existence of solutions of two differential equations. Recall from Section 2 that in the skeletal decomposition of superprocesses the total mass present at time t has two main components. The first one corresponds to an initial burst of subcritical mass, which is an independent copy of $(X,\mathbb{P}^*_\mu)$, and the second one is the accumulated mass from the dressing of the skeleton. As we will see in the next two results below, one can associate the first differential equation, i.e. the equation defining $f^T$, to $(X,\mathbb{P}^*_\mu)$, while the equation defining $h^T$ has an intimate relation to the dressed tree defined in the previous section.
Lemma 1. Fix $T>0$, and let $f\in D_0(\mathcal{L})$. Then the following differential equation has a unique non-negative solution:
where $\psi^*$ is given by (2.4).
Proof. Recall that $(X,\mathbb{P}^*_\mu)$ is a $(\mathcal{P}, \psi^*)$-superprocess, and as such its law can be characterised through an integral equation. More precisely, for all $\mu\in \mathcal{M}( E)$ and $f\in B^+( E)$, we have
where $u^*_f(x,t)$ is the unique non-negative solution to the integral equation
Li [Reference Li27, Theorem 7.11] showed that this integral equation is equivalent to the following differential equation:
Thus (5.6) also has a unique non-negative solution. If for each fixed $T>0$ we define $f^T(x,t)=u_f^*(x,T-t)$, then it is not hard to see that the lemma holds.
Theorem 2. Fix $T>0$, and take $f,h\in D_0(\mathcal{L})\cap B^+_b(E)$. If $f^T$ is the unique solution to (5.4), then the following differential equation has a unique non-negative solution:
where $\psi^*$ is given by (2.4), and w is a martingale function that satisfies the conditions in Section 2.
Proof. Recall the process (D, Z) constructed in Section 4. For every $\mu\in\mathcal{M}(E)$, $\nu\in\mathcal{M}_a(E)$, and $f,h\in B^+_b(E)$, we have
where $\textrm{exp}\{-v_{f,h}(x,t)\}$ is the unique [0, 1]-valued solution to the integral equation
and $u^*_f$ is the unique non-negative solution to (5.5). Indeed, this claim is a straightforward adaptation of the proof of Theorem 2 in [Reference Kyprianou, Pérez and Ren23], the details of which we leave to the reader. Note also that a similar statement has appeared in [Reference Berestycki, Kyprianou and Murillo-Salas2] in the non-spatial setting.
Next suppose that $f,h\in D_0(\mathcal{L})\cap B^+_b(E)$. We want to show that solutions to the integral equation (5.8) are equivalent to solutions of the following differential equation:
The reader will note that the statement and proof of this claim are classical. However, we include them here for the sake of completeness. One may find similar computations in the Appendix of [Reference Dynkin6], for example.
We first prove the claim that solutions to the integral equation (5.8) are solutions to the differential equation (5.9). To this end consider (5.8). Note that since $\mathcal{P}$ is a Feller semigroup the right-hand side is differentiable in t, and thus $v_{f,h}(x,t)$ is also differentiable in t. To find the differential version of the equation, we can use the standard technique of propagating the derivative at zero using the semigroup property of $v_{f,h}$ and $u^*_f$. Indeed, on one hand the semigroup property can easily be verified using
that is, we have $v_{u^*_f(\cdot,t),v_{f,h}(\cdot,t)}(\cdot,s)=v_{f,h}(\cdot,t+s)$, and $u^*_{u_f^*(\cdot,t)}(\cdot,s)=u_f^*(\cdot,t+s)$. This implies
and
providing that the two derivatives at zero exist from the right. One may similarly use the semigroup property, splitting at time s and $t-s$ to give the left derivatives at time $t>0$. On the other hand, differentiating (5.8) in t and taking $t\downarrow 0$ gives
which, recalling $v_{f,h}(x,0)=h(x)$ and $u_f^*(x,0)=f(x)$, can be rewritten as
Hence, combining the previous observations in (5.10) and (5.11), we get
To see why the differential equation (5.9) implies the integral equation (5.8), define
Then differentiating with respect to the time parameter gives
which we can then integrate over [0, t] to get (5.8).
To complete the proof we fix $T>0$ and define $h^T(x,t)\,:\!= v_{f,h}(x,T-t)$, and the result follows.
6. Proof of Theorem 1 (ii) and (iii)
The techniques we use here are similar in spirit to those in the proof of Theorem 2.1 in [Reference Fekete, Fontbona and Kyprianou16], in the sense that we use stochastic calculus to show the equality (5.1). However, what is new in the current setting is the use of the processes (5.2) and (5.3).
Fix $T>0$, and let $f^T$ be the unique non-negative solution to (5.4), and let $h^T$ be the unique non-negative solution to (5.7). Define $F_t^T\,:\!= {\textrm{e}} ^{-\langle{\kern1.3pt} f^T(\cdot,t),\Lambda_t\rangle-\langle h^T(\cdot,t),Z_t\rangle}$, $t\leq T$. Using stochastic calculus, we first verify that our choice of $f^T$ and $h^T$ results in the process $F_t^T$, $t\leq T$, having constant expectation on [0, T]. In the definition of $F^T$ both $\langle{\kern1.3pt} f^T(\cdot,t),\Lambda_t\rangle$ and $\langle h^T(\cdot,t),Z_t\rangle$ are semi-martingales, thus we can use Itô’s formula (see e.g. Theorem 32 in [Reference Protter31]) to get
where $\Delta \Lambda_t^{f^T}=\langle{\kern1.3pt} f^T(\cdot,t),\Lambda_t-\Lambda_{t-}\rangle$, and to avoid heavy notation we have written $\Lambda_t^{f^T}$ instead of $\langle{\kern1.3pt} f^T(\cdot,t),\Lambda_t\rangle$ and $Z_t^{h^T}$ instead of $\langle h^T(\cdot,t),Z_t\rangle$. Note that without the movement Z is a pure jump process, and since the interaction between $\Lambda$ and Z is limited to the time of the immigration events, we have $[\Lambda^{f^T},Z^{h^T}]_t^c=0$. Taking advantage of
we may thus write in integral form
To simplify the notation we used the fact that both $f^T(x,t)$ and $h^T(x,t)$ are continuous in t, thus $f^T(x,t)=f^T(x,t-)$ and $h^T(x,t)=h^T(x,t-)$.
We can split up the last term, that is, the sum of discontinuities according to the optional random measure in (3.7) responsible for this discontinuity. Thus, writing $\Delta^{(i)}$, $i=0,1,2$, to mean an increment coming from each of the three random measures,
Next, plugging in ${\textnormal d}\Lambda_s^{f^T}$ and ${\textnormal d} Z_s^{h^T}$ gives
where $M_t^{\textrm{loc}}$ is a local martingale corresponding to the terms $U_t^c({\kern1.3pt}f^T)$, $V_t^c(h^T)$ and the integral with respect to the random measure $\tilde{{\texttt N}}^{0}$ in (3.7). Note that the two terms with the time-derivative are due to the extra time dependence of the test functions in the integrals $\langle{\kern1.3pt} f^T(\cdot,s),\Lambda_s\rangle$ and $\langle h^T(\cdot,s),Z_s\rangle$. In particular, a change in $\langle{\kern1.3pt} f^T(s,\cdot),\Lambda_s\rangle$ corresponds to either a change in $\Lambda_s$ or a change in $f^T(\cdot,s)$.
Next we show that the local martingale term is in fact a real martingale, which will then disappear when we take expectations. First note that due to the boundedness of the drift and diffusion coefficients of the branching mechanism, and the conditions we had on its Lévy measure, the branching of the superprocess can be stochastically dominated by a finite-mean CSBP. This means that the CSBP associated with the Esscher-transformed branching mechanism $\psi^*$ is almost surely finite on any finite time interval [0, T], and thus the function $f^T$ is bounded on [0, T]. Using the boundedness of $f^T$ and the drift coefficient $\beta$, the quadratic variation of the integral
can be bounded from above as follows:
where C, $\widehat{C}$, and $\widetilde{C}$ are finite constants. Since the function $x\mapsto {\textrm{e}}^{-\widetilde{C}x}x$ is bounded on $[0,\infty)$, the previous quadratic variation is finite, and so the process (6.2) is a martingale on [0, T].
To show the martingale nature of the stochastic integral
we note that due to construction, $h^T\in \mathcal{D}_0(\mathcal{L})$, and is bounded on [0, T]. Thus, $V_t^c(h^T)$ is in fact a martingale on [0, T], and since $F_{s-}\leq 1$, $s\in[0, T]$, the quadratic variation of (6.3) is also finite, which gives the martingale nature of (6.3) on [0, T].
Finally, we consider the integral
Note that for compactly supported $\mu\in\mathcal{M}(E)^\circ$,
where C, $\widetilde{C}$, $\widehat{C}$, and C ′ are finite constants. Thus $Q_t<\infty$ on [0, T], and we can refer to page 63 of [Reference Ikeda and Watanabe20] to conclude that the process (6.4) is indeed a martingale on [0, T].
Thus, after taking expectations and gathering terms in (6.1), we get
where
and
We can see immediately that $A(x,f^T(x,t))$ is exactly what we have on the right-hand side of (5.4). Furthermore, using the fact that
we can also verify that
that is, $B(x,h^T(x,t),f^T(x,t))$ is equal to the right-hand side of (5.7). Hence, recalling the defining equations of $f^T$ (5.4) and $h^T$ (5.7), we get that the last four terms of (6.5) cancel, and thus $\mathbb{E}_\mu[F_t^T]=\mathbb{E}_\mu[F_0^T]$ for $t\in[0, T]$, as required. In particular, using the boundary conditions for $f^T$ and $h^T$, we get
Note that by construction we can relate the right-hand side of this previous expression to the superprocess. In particular, using the Poissonian nature of $Z_0$, and the fact that $X_0=\Lambda_0=\mu$ is deterministic, we have
where $X_t$ is a solution to (3.6). Thus, by choosing the right test functions, we could equate the value of $F_t^T$ at T to its initial value, which in turn gave a connection with the superprocess. The next step is to show that the process
has constant expectation on [0, T], which would then allow us to deduce
To simplify the notation let $\kappa^T(x,t)\,:\!=f^T(x,t)+w(x)( 1-{\textrm{e}}^{-h^T(x,t)})$, and define $G_t^T\,:\!={\textrm{e}}^{-\langle \kappa^T(\cdot,t),X_t\rangle}$. As the argument here is the exact copy of the previous analysis, we only give the main steps of the calculus, and leave it to the reader to fill in the gaps.
Since $\langle \kappa^T(\cdot,t),X_t\rangle$, $t\leq T$, is a semi-martingale, we can use Itô’s formula to get
where $M_t^{\textrm{loc}}$ is a local martingale corresponding to the term $M_t^c({\kern1.3pt}f)$, and the integral with respect to the random measure $\tilde{N}^{0}$ in (3.6). Note that the reasoning that led to the martingale nature of the local martingale term of (6.1) can also be applied here, which gives that $M_t^{\textrm{loc}}$ in (6.11) is in fact a true martingale on [0, T], which we denote by $M_t$.
Next we plug in $\kappa^T$, and after some laborious amount of algebra get
Once again using the identity (6.8), and taking expectations, gives
where A and B are given by (6.6) and (6.7). Finally, noting
gives
which results in the cancellation of the last three terms in (6.12), and hence verifies the constant expectation of $G_t^T$ on [0, T]. In particular, we have proved that
In conclusion, combining the previous observations (6.9) and (6.10) with (6.13) gives
Since $T>0$ was arbitrary, the above equality holds for any time $T>0$.
Then we have the following implications. First, by setting $h=0$ we find that
which not only shows that under our conditions $(\Lambda_t,t\geq 0)$ is Markovian, but also that its semigroup is equal to the semigroup of $(X,\mathbb{P}_\mu)$, and hence proves that $(\Lambda_t,t\geq 0)$ is indeed a weak solution to (3.5).
Furthermore, choosing h and f not identical to zero, we get that the pair $(\Lambda_t,Z_t)$ under $\textbf{P}_{\mu}$ has the same law as $(X_t,\textrm{Po}(w(x)X_t({\textnormal d} x)))$ under $\mathbb{P}_\mu$, where $\textrm{Po}(w(x)X_t({\textnormal d} x))$ is an autonomously independent Poisson random measure with intensity $w(x)X_t({\textnormal d} x)$. Thus $Z_t$ given $\Lambda_t$ is indeed a Poisson random measure with intensity $w(x)\Lambda_t({\textnormal d} x)$.
7. Proof of Theorem 1 (i): uniqueness
If we review the calculations that led to (6.9), we observe that any solution $(\Lambda, Z)$ to the coupled SDE (3.7) has the property that, for $\mu\in\mathcal{M}(E)$ and $\nu\in\mathcal{M}_a(E)$,
Hence, since any two solutions to (3.7) are Markovian, the second equality above identifies their transitions as equal thanks to the uniqueness of the PDEs in Lemma 1 and Theorem 2. In other words, there is a unique weak solution to (3.7).
Acknowledgements
AEK acknowledges support from EPSRC grant EP/L002442/1. DF is supported by a scholarship from the EPSRC Centre for Doctoral Training, SAMBa. JF acknowledges support from Basal-Conicyt Centre for Mathematical Modelling AFB 170001 and Millennium Nucleus SMCDS. Part of this work was carried out whilst AEK was visiting the Centre for Mathematical Modelling, Universidad de Chile, and JF was visiting the Department of Mathematical Sciences at the University of Bath; each is grateful to the host institution of the other for their support. The authors are additionally grateful to two anonymous referees and an AE for their careful reading and assessment of an earlier version of this paper, which led to significant improvements.