IISA-2023
6/3/23
The raw EEG can indicate general state, e.g., awake or asleep, but it can’t tell us much about specific neural processes that underlie perception, cognition or emotion.
Solve the issues by
\begin{align*} y_i &= f(x_i) + \epsilon _i, ~~ \epsilon _i \sim N(0, \sigma^2), ~~ i = 1, \dots, n, \\ f'(t_m) &= 0, ~~m = 1, \dots, M, \end{align*}
[k_{01}(\bm{x}, \bm{t})]_{ij} := \frac{\partial }{\partial t_j}k(x_i, t_j), ~~[k_{11}(\bm{t}, \bm{t})]_{ij} := \frac{\partial ^2 }{\partial t_i \partial t_j}k(t_i, t_j)
k_{10}(\bm{x}, \bm{t}) = k_{01}^T(\bm{x}, \bm{t})
t_1, \dots, t_M are parameters to be estimated.
Challenge: Unknown number of stationary points M.
Use an univariate prior t \sim \pi(t) that corresponds to assuming that f has at least one stationary point, and utilize the posterior of t to infer all stationary points.
An univariate prior of t leads to
Efficient computation:
Avoids possible misspecification of M that distorts the fitted result.
\begin{align*} y_i &= f(x_i) + \epsilon_i, ~~ \epsilon_i \sim N(0, \sigma^2), ~~i = 1, \dots, n, \\ f &\sim GP(0, k(\cdot, \cdot; \boldsymbol \theta)) , ~~ f'(t) = 0,\\ t &\sim \pi(t), ~~ \sigma^2 \sim \pi(\sigma^2), ~~ \boldsymbol \theta\sim \pi(\boldsymbol \theta) \end{align*}
For iteration i = 1, 2, \dots until convergence:
Monte Carlo E-step: Metropolis-Hastings on t and Gibbs sampling on \sigma^2.
M-step: Given samples of t and \sigma^2, update \boldsymbol \theta to \hat{\boldsymbol \theta}^{(i+1)} by setting \begin{align*} \hat{\boldsymbol \theta}^{(i+1)} &= \arg \max_{\boldsymbol \theta} \frac{1}{J} \sum_{j=1}^J \log p(\bm{y} \mid t_j^{(i)}, (\sigma_j^2)^{(i)}, \boldsymbol \theta^{(i)}) \end{align*}
Estimates the amplitude and latency as well as the ERP waveform in a unified framework.
Supports a multi-factor ANOVA setting that is capable of examining how multiple factors or covariates affect ERP components’ latency and amplitude.
Bayesian approach having probabilistic statements about the amplitude and latency of ERP components.
The hierarchical nature of the model provides estimates at both the grand group and individual subject levels.
\begin{equation*} y_{igls} = f_{gls}(x_i) + \epsilon_{igls}, ~~ \epsilon_{igls} \stackrel{\rm iid}{\sim}N(0, \sigma^2). \end{equation*}
Time point i = 1, \dots, n
Factor A level g = 1, \dots, G
Factor B level l = 1, \dots, L
Subject s = 1, \dots, S_{gl}
\begin{equation*} f_{gls} \mid \bm{t}_{gls} \sim \text{DGP}(0, k(\cdot, \cdot; \boldsymbol \theta), \bm{t}_{gls}), \end{equation*} where \bm{t}_{gls} = \{t_{gls}^m\}_{m = 1}^M.
\begin{align*} t_{gls}^m \mid r_{gl}^m, \eta_{gl}^m &\stackrel{\rm iid}{\sim}\text{gbeta}\left(r_{gl}^m\eta_{gl}^m, \, (1 - r_{gl}^m)\eta_{gl}^m, \, a^{m}, \, b^{m}\right), \, s = 1, \dots, S_{gl}. \end{align*}
Subject-level t_{gls}^m \in [a^{m}, b^{m}]
Group-level r_{gl}^m \in (0, 1): location of prior mean in [a^{m}, b^{m}]
{\mathbb E}\left(t_{gls}^m\right) = (1 - r_{gl}^m)a^{m} + r_{gl}^mb^{m}
Linear model and link function \phi_{gl}^m \in (-\infty, \infty)
\begin{equation*} \phi_{gl}^m\left(r_{gl}^m\right) = \beta_0^m + \beta_{1, 1}^mz_{1, 1} + \cdots + \beta_{1, G-1}^mz_{1, G-1}+ \beta_{2, 1}^mz_{2, 1} + \cdots + \beta_{2, L-1}^mz_{2, L-1}. \end{equation*}
Accommodate any linear model with categorical and numerical variables
Being able to describe interactions
Any link function can be used, logit, probit, complementary log-log, etc.
\begin{align*} \beta_0^m &\sim N(\mu_0^m, (\sigma_0^m)^2),\\ \beta_{1, a}^m &\sim N(\mu_1^m, (\sigma_1^m)^2), \\ \beta_{2, b}^m &\sim N(\mu_2^m, (\sigma_2^m)^2), \\ \eta_{gl}^m &\sim Ga(\alpha_{\eta}^m, \beta_{\eta}^m),\\ \sigma^2 &\sim IG(\alpha_{\sigma}, \beta_{\sigma}),\\ \boldsymbol \theta&\sim \pi(\boldsymbol \theta) \end{align*}
Data are generated with \sigma = 0.25
Experiment on speech recognition
The N100 (60 ~ 140 msec) component captures phonological (syllable) representation.
P200 (140 ~ 220 msec) component represents higher-order perceptual processing.
Novel Bayesian model SLAM estimates the amplitude and latency of ERP components with uncertainty quantification.
Solves the search window specification problem by generating posterior distribution of latency.
Incorporates the ANOVA and possibly a latent generalized linear model structure in a single unified framework.
Examines both the subjects’ individual differences and group-level differences that facilitate comparing different characteristics or factors, such as age/gender.
Extensions could include treatments of other noise distributions, for example, the autoregressive correlation or trial variability.
Spatial modeling and mapping on electrodes would be another extension.