SlideShare ist ein Scribd-Unternehmen logo
1 von 50
Downloaden Sie, um offline zu lesen
Spike sorting: What is it? Why do we 
need it? Where does it come from? How 
is it done? How to interpret it? 
III. Improving sorting quality through 
stochastic modeling of spike trains. 
Christophe Pouzat 
Mathématiques Appliquées à Paris 5 (MAP5) 
Université Paris-Descartes and CNRS UMR 8145 
christophe.pouzat@parisdescartes.fr 
Thursday November 27, 2014
Where are we ? 
What makes data "difficult" 
An interlude to keep you motivated 
A more realistic model and its problems 
A analogy with statistical physics 
Back to our simulated data 
Application to real data
Yesterday’s summary
Spike amplitude dynamics (1) 
We did not deal with that 
Fig. 2 of Calvin (1972) Electroenceph clin Neurophysiol 34:94.
Spike amplitude dynamics (2) 
Fig. 3 of Pouzat (2005) Technique(s) for Spike Sorting. In 
Methods and Models in Neurophysics. Les Houches 2003. 
Elsevier.
Spike amplitude dynamics (3) 
We try a naive "exponential relaxation model" for these data: 
a (isi ) = p  (1    exp (  isi )) :
The exponential relaxation in only an 
approximation! 
Fig. 3A of Delescluse and Pouzat (2006) J Neurosci Methods 
150:16.
Neuronal discharges are not Poisson 
Fig. 7 of Kuffler, Fitzhugh and Barlow (1957) J Gen Physiol 
40:683. Evidence for serial correlation of the intervals was 
found (p 687). Hagiwara (1950, 1954) made similar reports.
Log-normal density 
Empirical ISI densities are better described by a log-normal 
density than by a Poisson density: 
ISI (ISI = isi ) = 
1 
isi f 
p 
2 
exp 
 
 
1 
2 
 
ln isi  ln s 
f 
2 
# 
where, s is a scale parameter (measured in sec) and f is a 
dimensionless shape parameter.
Log-normal density: exemples
What do we want? 
I Recognizing that the inter spike interval (ISI) distribution, 
and more generally the neuron’s stochastic intensity, 
provides potentially useful information for spike sorting, 
we would like a data generation model including this 
information. 
I We would also like a data generation model giving room 
for a description of the spikes amplitude dynamics during 
bursts.
Where are we ? 
What makes data difficult 
An interlude to keep you motivated 
A more realistic model and its problems 
A analogy with statistical physics 
Back to our simulated data 
Application to real data
Some simulated data 
Simulated data with 3 neurones on 2 recording sites.
The results we will get
Where are we ? 
What makes data difficult 
An interlude to keep you motivated 
A more realistic model and its problems 
A analogy with statistical physics 
Back to our simulated data 
Application to real data
First improved model 
We will assume that the neuron discharge independently and 
that: 
I Individual discharges are well approximated by a renewal 
process with a log-normal density. 
I The spike amplitude depends on the elapsed time since 
the last spike (of the same neuron); this dependence is 
moreover well approximated by an exponential relaxation. 
I The recording noise is white and Gaussian and is 
independent of the spikes.
Likelihood for single neuron data (1) 
We have: 
 (D j p; ; ; s; f ) = 
YN 
j=1 
isi (ij j s; f )  amp (aj j ij ; p; ; ) ; 
where 
amp (aj j ij ; p; ; ) = 
1 
(2) 
ns 
2 
 e12 
kajp(1exp(ij))k2 
:
Likelihood for single neuron data (2) 
The log-likelihood can be written as the sum of two terms: 
L(D j p; ; ; s; f ) = Lisi (D j s; f ) + Lamp (D j p; ;  ) 
where: 
Lisi (D j s; f ) = N  ln f  
XN 
j=1 
8 
: 
ln ij + 
1 
2 
2 
4 
ln 
 
ij 
s 
 
f 
29= 
3 
5 
; 
+Cst 
and: 
Lamp (D j p; ;  ) =  
1 
2 
XN 
j=1 
kaj  p  (1    exp (  ij ))k2+Cst :
Configuration 
We will use  for our model parameters vector, that is, for a 
model with K neurons: 
 = (P1;1; 1; S1; F1; : : : ;PK;K; K; SK; FK) : 
We will formalize our a priori ignorance of the origin of each 
spike by associating to each spike, j , a label Lj 2 f1; : : : ;Kg. 
lj = 3 means that event j was generated by neuron 3. 
We will use C for the configuration, that is, the random 
variable: 
C = (L1; : : : ; LN)T : 
With this formalism, spike sorting amounts to estimating the 
configuration.
Likelihood for data from several neurons 
If the configuration realization is known, the likelihood 
computation is easily done: 
 (D j c; ) = 
YK 
q=1 
YNq 
j=1 
isi (iq;j j sq; fq)amp 
 
aq;j j iq;j ; pq; q; q 
 
:
Bayesian formalism (1) 
I Being somewhat opportunist, we are going to formalize 
our ignorance on the configuration c and the model 
parameters  by viewing them as realizations of the 
random variables C and . 
I We then have a triplet of random variables D (whose 
realization is D), C and . 
I We have, by definition of the conditional probability: 
(D; c; ) = (c;  j D) (D) = (D j c; ) (c; ) : 
I Rearranging the last two members we get (Bayes rule or 
Bayes theorem): 
(c;  j D) = 
(D j c; ) (c; ) 
(D) 
:
Bayesian formalism (2) 
I In the literature, our (c; ) is called the a priori density 
and is written prior (c; ); that’s what we know about the 
parameters and configuration before observing the data. 
I Our (c;  j D) is called the a posteriori density and is 
written posterior (c;  j D); that’s what we know about the 
parameters and configuration after observing the data. 
I Bayes rule tells us how to update our knowledge once the 
data have been observed. 
I We know how to compute the likelihood (D j c; ).
Bayesian formalism (3) 
I We will write prior (c; ) = prior (c)prior () and take 
prior (c) uniform on C = f1; : : : ;KgN (notice that I’m 
cheating, I can know N only after seeing the data) and 
take 
prior () = 
YK 
q=1 
prior (sq)prior (fq)prior (pq)prior (q)prior (q) ; 
and each of the terms is going to be the density of a 
uniform distribution over a large domain. 
I The real problem is the denominator: 
(D) = 
X 
c2C 
Z 
 
(D; c; )d = 
X 
c2C 
Z 
 
(D j c; )prior (c; )d ; 
since it involves a summation over KN configurations and 
K is of the order of 10 and N of the order of 1000!
Bayesian formalism (4) 
I Assuming we find a way to circumvent the combinatorial 
explosion (KN) problem and to get an estimator, 
^(c;  j D), of: 
posterior (c;  j D) = 
(D j c; ) (c; ) 
(D) 
: 
I The estimated posterior configuration probability would 
be: 
^posterior (c j D) = 
Z 
 
^(c;  j D)d : 
I The spike sorting results could then be summarized by the 
Maximum A Posteriori (MAP) estimator: 
^c = arg min 
c2C 
^posterior (c j D) :
Bayesian formalism (5) 
I Clearly, a better use of the results would be for any 
statistics T(C) (think of the cross-correlogram between 
two spike trains) to use: 
[T(C) = 
X 
c2C 
T(c)^posterior (c j D) ; 
instead of: T(^c). 
I But to do all that we have to deal with the combinatorial 
explosion! 
I In such situations, a good strategy is too look at other 
domains to see if other people met a similar problem and 
found a solution. . .
Where are we ? 
What makes data difficult 
An interlude to keep you motivated 
A more realistic model and its problems 
A analogy with statistical physics 
Back to our simulated data 
Application to real data
A analogy with statistical physics (1) 
I Luckily for us, Physicists got a problem similar to ours at 
the turn of the 1950s and found a solution. 
I Let us write: 
E (c;  j D) = log [ (D j c; )  prior (c; )] ; 
notice that we know how to compute E (c;  j D). 
I Then: 
posterior (c;  j D) = 
exp [
E (c;  j D)] 
(D) 
; 
where
= 1,
is the inverse temperature for Physicists 
who write Z(
) = (D) and call it the partition function. 
I Written in this way, posterior (c;  j D), is nothing more 
than a Gibbs distribution.
A analogy with statistical physics (2) 
I Physicists don’t know how to compute the partition 
function Z(
) in many cases of interest. 
I But they need to compute quantities like: 
X 
c2C 
Z 
 
T(c; )posterior (c;  j D)d : 
I In 1953, Metropolis, Rosenbluth, Rosenbluth, Teller and 
Teller published Equation of State Calculations by Fast 
Computing Machines in Journal of Chemical Physics, 
proposing an algorithm for estimating this kind of integral 
without knowing the partition function.
A analogy with statistical physics (3) 
I In 1971, Hastings published Monte Carlo Sampling 
Methods Using Markov Chains and Their Applications in 
Biometrika proving that the previous algorithm was 
indeed doing what it was suppose to do. 
I This algorithm is know called the Metropolis-Hastings 
Algorithm.
Metropolis-Hastings in a simple setting (1) 
I We will generate a collection of states 
fx(1); x(2); : : : ; x(R)g according to a target distribution 
(x) (the x stand here for our previous c; ). 
I To that end we use a Markov chain which asymptotically 
reaches the stationary distribution (x). 
I A Markov chain is uniquely defined by its transition 
matrix P(X(k+1) = x0 j X(k) = x) and its initial state 
distribution.
Metropolis-Hastings in a simple setting (2) 
I A Markov chain has a unique stationary distribution when 
the following two conditions are met: 
1. A stationary distribution exists: a sufficient condition for 
that is the detailed balance that implies 
(x)P(x0 j x) = (x0)P(x j x0) : 
2. The stationary distribution is unique if the chain is 
aperiodic (the system does not return to the same state 
at fixed intervals) and positive recurrent (the expected 
number of steps for returning to the same state is finite). 
I Checking the detailed balance requires a knowledge of 
(x) up to a normalizing constant. 
I Constructing a transition matrix satisfying the detailed 
balance for a given target distribution also requires the 
knowledge of that distribution up to a normalizing 
constant.
Metropolis-Hastings in a simple setting (3) 
I The trick is to construct the transition matrix elements 
in two sub-steps: a proposal, g(x0 j x), and an 
acceptance-rejection, A(x0 j x), to get 
P(x0 j x) = g(x0 j x) A(x0 j x). 
I The detailed balance is satisfied if: 
(x)g(x0 j x)A(x0 j x) = (x0)g(x j x0)A(x j x0) 
that is if 
A(x0 j x) 
A(x j x0) 
= 
(x0)g(x j x0) 
(x)g(x0 j x) 
: 
I The Metropolis choice is: 
A(x0 j x) = min 
 
1; 
(x0)g(x j x0) 
(x)g(x0 j x) 
 
:
Metropolis-Hastings in a simple setting (4) 
I The Metropolis choice for A(x0 j x) requires only the 
knowledge of the target distribution up to a normalizing 
constant. 
I The proposal matrix g(x0 j x) is chosen to ensure positive 
recurrence and aperiodicity. 
I If g is positive recurrent and aperiodic and if 
g(x0 j x) = 0 implies g(x j x0) = 0 then we can show that 
P(x0 j x) is positive recurrent and aperiodic. 
I We then have a recipe to modify a positive recurrent 
and aperiodic transition matrix in order to have the 
stationary distribution of our choice while just knowing it 
up to a normalizing constant.
A first demo 
Show the demo on a toy example.
Where are we ? 
What makes data difficult 
An interlude to keep you motivated 
A more realistic model and its problems 
A analogy with statistical physics 
Back to our simulated data 
Application to real data
Back to our simulated data 
Simulated data with 3 neurones on 2 recording sites.
Energy evolution 
Energy evolution during 3  105 MC steps.
Posterior densities
Replica Exchange Method / Parallel Tempering 
REM / Tempering demo. . .
Energy evolution with REM 
Energy evolution during 32  104 MC steps, REM turned on 
after 104 steps.
Closer look at REM
Posterior densities
REM dynamics
Most likely configuration
Where are we ? 
What makes data difficult 
An interlude to keep you motivated 
A more realistic model and its problems 
A analogy with statistical physics 
Back to our simulated data 
Application to real data

Weitere ähnliche Inhalte

Was ist angesagt?

Physics Research Summer2009
Physics Research Summer2009Physics Research Summer2009
Physics Research Summer2009Ryan Melvin
 
Sienna 4 divideandconquer
Sienna 4 divideandconquerSienna 4 divideandconquer
Sienna 4 divideandconquerchidabdu
 
Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘 Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘 Jungkyu Lee
 
2012 mdsp pr13 support vector machine
2012 mdsp pr13 support vector machine2012 mdsp pr13 support vector machine
2012 mdsp pr13 support vector machinenozomuhamada
 
Cs229 notes10
Cs229 notes10Cs229 notes10
Cs229 notes10VuTran231
 
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...Matt Moores
 
Skiena algorithm 2007 lecture15 backtracing
Skiena algorithm 2007 lecture15 backtracingSkiena algorithm 2007 lecture15 backtracing
Skiena algorithm 2007 lecture15 backtracingzukun
 
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...Matt Moores
 
PR12-225 Discovering Physical Concepts With Neural Networks
PR12-225 Discovering Physical Concepts With Neural NetworksPR12-225 Discovering Physical Concepts With Neural Networks
PR12-225 Discovering Physical Concepts With Neural NetworksKyunghoon Jung
 
Using Sparse Matrix for the Contact Calculation_ZhanWang
Using Sparse Matrix for the Contact Calculation_ZhanWangUsing Sparse Matrix for the Contact Calculation_ZhanWang
Using Sparse Matrix for the Contact Calculation_ZhanWangZhan Wang
 
Bayesian Inference and Uncertainty Quantification for Inverse Problems
Bayesian Inference and Uncertainty Quantification for Inverse ProblemsBayesian Inference and Uncertainty Quantification for Inverse Problems
Bayesian Inference and Uncertainty Quantification for Inverse ProblemsMatt Moores
 
machinelearning project
machinelearning projectmachinelearning project
machinelearning projectLianli Liu
 
SPSF04 - Euler and Runge-Kutta Methods
SPSF04 - Euler and Runge-Kutta MethodsSPSF04 - Euler and Runge-Kutta Methods
SPSF04 - Euler and Runge-Kutta MethodsSyeilendra Pramuditya
 

Was ist angesagt? (20)

Physics Research Summer2009
Physics Research Summer2009Physics Research Summer2009
Physics Research Summer2009
 
Rand final
Rand finalRand final
Rand final
 
Sienna 4 divideandconquer
Sienna 4 divideandconquerSienna 4 divideandconquer
Sienna 4 divideandconquer
 
Ma2520962099
Ma2520962099Ma2520962099
Ma2520962099
 
Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘 Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘
 
Signal Processing Assignment Help
Signal Processing Assignment HelpSignal Processing Assignment Help
Signal Processing Assignment Help
 
Diffusion Homework Help
Diffusion Homework HelpDiffusion Homework Help
Diffusion Homework Help
 
2012 mdsp pr13 support vector machine
2012 mdsp pr13 support vector machine2012 mdsp pr13 support vector machine
2012 mdsp pr13 support vector machine
 
Cs229 notes10
Cs229 notes10Cs229 notes10
Cs229 notes10
 
Data Analysis Homework Help
Data Analysis Homework HelpData Analysis Homework Help
Data Analysis Homework Help
 
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
 
Skiena algorithm 2007 lecture15 backtracing
Skiena algorithm 2007 lecture15 backtracingSkiena algorithm 2007 lecture15 backtracing
Skiena algorithm 2007 lecture15 backtracing
 
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...
 
PR12-225 Discovering Physical Concepts With Neural Networks
PR12-225 Discovering Physical Concepts With Neural NetworksPR12-225 Discovering Physical Concepts With Neural Networks
PR12-225 Discovering Physical Concepts With Neural Networks
 
Using Sparse Matrix for the Contact Calculation_ZhanWang
Using Sparse Matrix for the Contact Calculation_ZhanWangUsing Sparse Matrix for the Contact Calculation_ZhanWang
Using Sparse Matrix for the Contact Calculation_ZhanWang
 
Fdtd ppt for mine
Fdtd ppt   for mineFdtd ppt   for mine
Fdtd ppt for mine
 
Fdtd
FdtdFdtd
Fdtd
 
Bayesian Inference and Uncertainty Quantification for Inverse Problems
Bayesian Inference and Uncertainty Quantification for Inverse ProblemsBayesian Inference and Uncertainty Quantification for Inverse Problems
Bayesian Inference and Uncertainty Quantification for Inverse Problems
 
machinelearning project
machinelearning projectmachinelearning project
machinelearning project
 
SPSF04 - Euler and Runge-Kutta Methods
SPSF04 - Euler and Runge-Kutta MethodsSPSF04 - Euler and Runge-Kutta Methods
SPSF04 - Euler and Runge-Kutta Methods
 

Ähnlich wie Spike sorting: What is it? Why do we need it? Where does it come from? How is it done? How to interpret it?

Metodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang LandauMetodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang Landauangely alcendra
 
Machine learning (11)
Machine learning (11)Machine learning (11)
Machine learning (11)NYversity
 
InternshipReport
InternshipReportInternshipReport
InternshipReportHamza Ameur
 
Linear regression [Theory and Application (In physics point of view) using py...
Linear regression [Theory and Application (In physics point of view) using py...Linear regression [Theory and Application (In physics point of view) using py...
Linear regression [Theory and Application (In physics point of view) using py...ANIRBANMAJUMDAR18
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Alexander Litvinenko
 
An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...Alexander Decker
 
Modelos de Redes Neuronales
Modelos de Redes NeuronalesModelos de Redes Neuronales
Modelos de Redes NeuronalesCarlosNigerDiaz
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component AnalysisSumit Singh
 
SPSF02 - Graphical Data Representation
SPSF02 - Graphical Data RepresentationSPSF02 - Graphical Data Representation
SPSF02 - Graphical Data RepresentationSyeilendra Pramuditya
 
Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Fabian Pedregosa
 
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language ModelsAnomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language ModelsCynthia Freeman
 

Ähnlich wie Spike sorting: What is it? Why do we need it? Where does it come from? How is it done? How to interpret it? (20)

Linear algebra havard university
Linear algebra havard universityLinear algebra havard university
Linear algebra havard university
 
Metodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang LandauMetodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang Landau
 
Machine learning (11)
Machine learning (11)Machine learning (11)
Machine learning (11)
 
Technical
TechnicalTechnical
Technical
 
InternshipReport
InternshipReportInternshipReport
InternshipReport
 
ICLR 2018
ICLR 2018ICLR 2018
ICLR 2018
 
Linear regression [Theory and Application (In physics point of view) using py...
Linear regression [Theory and Application (In physics point of view) using py...Linear regression [Theory and Application (In physics point of view) using py...
Linear regression [Theory and Application (In physics point of view) using py...
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
 
An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...An investigation of inference of the generalized extreme value distribution b...
An investigation of inference of the generalized extreme value distribution b...
 
AI Lesson 29
AI Lesson 29AI Lesson 29
AI Lesson 29
 
Lesson 29
Lesson 29Lesson 29
Lesson 29
 
kactl.pdf
kactl.pdfkactl.pdf
kactl.pdf
 
Modelos de Redes Neuronales
Modelos de Redes NeuronalesModelos de Redes Neuronales
Modelos de Redes Neuronales
 
By BIRASA FABRICE
By BIRASA FABRICEBy BIRASA FABRICE
By BIRASA FABRICE
 
A bit about мcmc
A bit about мcmcA bit about мcmc
A bit about мcmc
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component Analysis
 
SPSF02 - Graphical Data Representation
SPSF02 - Graphical Data RepresentationSPSF02 - Graphical Data Representation
SPSF02 - Graphical Data Representation
 
Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3
 
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language ModelsAnomaly Detection in Sequences of Short Text Using Iterative Language Models
Anomaly Detection in Sequences of Short Text Using Iterative Language Models
 
Unit 3
Unit 3Unit 3
Unit 3
 

Mehr von NeuroMat

Gromov and the ”ergo-brain”
Gromov and the ”ergo-brain”Gromov and the ”ergo-brain”
Gromov and the ”ergo-brain”NeuroMat
 
Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...
Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...
Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...NeuroMat
 
Hidden context tree modeling of EEG data
Hidden context tree modeling of EEG dataHidden context tree modeling of EEG data
Hidden context tree modeling of EEG dataNeuroMat
 
Functional Regression Analysis
Functional Regression AnalysisFunctional Regression Analysis
Functional Regression AnalysisNeuroMat
 
Goodness–of–fit tests for regression models: the functional data case
Goodness–of–fit tests for regression models: the functional data caseGoodness–of–fit tests for regression models: the functional data case
Goodness–of–fit tests for regression models: the functional data caseNeuroMat
 
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03NeuroMat
 
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02NeuroMat
 
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01NeuroMat
 
MEPHunter: Making things easier for signal visualization and analysis
MEPHunter: Making things easier for signal visualization and analysisMEPHunter: Making things easier for signal visualization and analysis
MEPHunter: Making things easier for signal visualization and analysisNeuroMat
 
Neuroscience Experiments System - NES - Versão 0.1
Neuroscience Experiments System - NES - Versão 0.1Neuroscience Experiments System - NES - Versão 0.1
Neuroscience Experiments System - NES - Versão 0.1NeuroMat
 
Spike sorting: What is it? Why do we need it? Where does it come from? How is...
Spike sorting: What is it? Why do we need it? Where does it come from? How is...Spike sorting: What is it? Why do we need it? Where does it come from? How is...
Spike sorting: What is it? Why do we need it? Where does it come from? How is...NeuroMat
 
Desafios matemáticos e computacionais da neurociência
Desafios matemáticos e computacionais da neurociênciaDesafios matemáticos e computacionais da neurociência
Desafios matemáticos e computacionais da neurociênciaNeuroMat
 
Introdução elementar à modelagem estocástica de cadeias simbólicas
Introdução elementar à modelagem estocástica de cadeias simbólicasIntrodução elementar à modelagem estocástica de cadeias simbólicas
Introdução elementar à modelagem estocástica de cadeias simbólicasNeuroMat
 
Data Provenance and Scientific Workflow Management
Data Provenance and Scientific Workflow ManagementData Provenance and Scientific Workflow Management
Data Provenance and Scientific Workflow ManagementNeuroMat
 

Mehr von NeuroMat (14)

Gromov and the ”ergo-brain”
Gromov and the ”ergo-brain”Gromov and the ”ergo-brain”
Gromov and the ”ergo-brain”
 
Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...
Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...
Perspectives on Applications of a Stochastic Spiking Neuron Model to Neural N...
 
Hidden context tree modeling of EEG data
Hidden context tree modeling of EEG dataHidden context tree modeling of EEG data
Hidden context tree modeling of EEG data
 
Functional Regression Analysis
Functional Regression AnalysisFunctional Regression Analysis
Functional Regression Analysis
 
Goodness–of–fit tests for regression models: the functional data case
Goodness–of–fit tests for regression models: the functional data caseGoodness–of–fit tests for regression models: the functional data case
Goodness–of–fit tests for regression models: the functional data case
 
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 03
 
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 02
 
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01
Introdução ao Armazenamento de Dados de Experimentos em Neurociência - Parte 01
 
MEPHunter: Making things easier for signal visualization and analysis
MEPHunter: Making things easier for signal visualization and analysisMEPHunter: Making things easier for signal visualization and analysis
MEPHunter: Making things easier for signal visualization and analysis
 
Neuroscience Experiments System - NES - Versão 0.1
Neuroscience Experiments System - NES - Versão 0.1Neuroscience Experiments System - NES - Versão 0.1
Neuroscience Experiments System - NES - Versão 0.1
 
Spike sorting: What is it? Why do we need it? Where does it come from? How is...
Spike sorting: What is it? Why do we need it? Where does it come from? How is...Spike sorting: What is it? Why do we need it? Where does it come from? How is...
Spike sorting: What is it? Why do we need it? Where does it come from? How is...
 
Desafios matemáticos e computacionais da neurociência
Desafios matemáticos e computacionais da neurociênciaDesafios matemáticos e computacionais da neurociência
Desafios matemáticos e computacionais da neurociência
 
Introdução elementar à modelagem estocástica de cadeias simbólicas
Introdução elementar à modelagem estocástica de cadeias simbólicasIntrodução elementar à modelagem estocástica de cadeias simbólicas
Introdução elementar à modelagem estocástica de cadeias simbólicas
 
Data Provenance and Scientific Workflow Management
Data Provenance and Scientific Workflow ManagementData Provenance and Scientific Workflow Management
Data Provenance and Scientific Workflow Management
 

Kürzlich hochgeladen

TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...Nguyen Thanh Tu Collection
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentationcamerronhm
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseAnaAcapella
 
Understanding Accommodations and Modifications
Understanding  Accommodations and ModificationsUnderstanding  Accommodations and Modifications
Understanding Accommodations and ModificationsMJDuyan
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptxMaritesTamaniVerdade
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...christianmathematics
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...Poonam Aher Patil
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxVishalSingh1417
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin ClassesCeline George
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Jisc
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsMebane Rash
 
Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfSherif Taha
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.christianmathematics
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17Celine George
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Association for Project Management
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxAmanpreet Kaur
 

Kürzlich hochgeladen (20)

TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentation
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please Practise
 
Understanding Accommodations and Modifications
Understanding  Accommodations and ModificationsUnderstanding  Accommodations and Modifications
Understanding Accommodations and Modifications
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdf
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 

Spike sorting: What is it? Why do we need it? Where does it come from? How is it done? How to interpret it?

  • 1. Spike sorting: What is it? Why do we need it? Where does it come from? How is it done? How to interpret it? III. Improving sorting quality through stochastic modeling of spike trains. Christophe Pouzat Mathématiques Appliquées à Paris 5 (MAP5) Université Paris-Descartes and CNRS UMR 8145 christophe.pouzat@parisdescartes.fr Thursday November 27, 2014
  • 2. Where are we ? What makes data "difficult" An interlude to keep you motivated A more realistic model and its problems A analogy with statistical physics Back to our simulated data Application to real data
  • 4. Spike amplitude dynamics (1) We did not deal with that Fig. 2 of Calvin (1972) Electroenceph clin Neurophysiol 34:94.
  • 5. Spike amplitude dynamics (2) Fig. 3 of Pouzat (2005) Technique(s) for Spike Sorting. In Methods and Models in Neurophysics. Les Houches 2003. Elsevier.
  • 6. Spike amplitude dynamics (3) We try a naive "exponential relaxation model" for these data: a (isi ) = p (1 exp ( isi )) :
  • 7. The exponential relaxation in only an approximation! Fig. 3A of Delescluse and Pouzat (2006) J Neurosci Methods 150:16.
  • 8. Neuronal discharges are not Poisson Fig. 7 of Kuffler, Fitzhugh and Barlow (1957) J Gen Physiol 40:683. Evidence for serial correlation of the intervals was found (p 687). Hagiwara (1950, 1954) made similar reports.
  • 9. Log-normal density Empirical ISI densities are better described by a log-normal density than by a Poisson density: ISI (ISI = isi ) = 1 isi f p 2 exp 1 2 ln isi ln s f 2 # where, s is a scale parameter (measured in sec) and f is a dimensionless shape parameter.
  • 11. What do we want? I Recognizing that the inter spike interval (ISI) distribution, and more generally the neuron’s stochastic intensity, provides potentially useful information for spike sorting, we would like a data generation model including this information. I We would also like a data generation model giving room for a description of the spikes amplitude dynamics during bursts.
  • 12. Where are we ? What makes data difficult An interlude to keep you motivated A more realistic model and its problems A analogy with statistical physics Back to our simulated data Application to real data
  • 13. Some simulated data Simulated data with 3 neurones on 2 recording sites.
  • 14. The results we will get
  • 15. Where are we ? What makes data difficult An interlude to keep you motivated A more realistic model and its problems A analogy with statistical physics Back to our simulated data Application to real data
  • 16. First improved model We will assume that the neuron discharge independently and that: I Individual discharges are well approximated by a renewal process with a log-normal density. I The spike amplitude depends on the elapsed time since the last spike (of the same neuron); this dependence is moreover well approximated by an exponential relaxation. I The recording noise is white and Gaussian and is independent of the spikes.
  • 17. Likelihood for single neuron data (1) We have: (D j p; ; ; s; f ) = YN j=1 isi (ij j s; f ) amp (aj j ij ; p; ; ) ; where amp (aj j ij ; p; ; ) = 1 (2) ns 2 e12 kajp(1exp(ij))k2 :
  • 18. Likelihood for single neuron data (2) The log-likelihood can be written as the sum of two terms: L(D j p; ; ; s; f ) = Lisi (D j s; f ) + Lamp (D j p; ; ) where: Lisi (D j s; f ) = N ln f XN j=1 8 : ln ij + 1 2 2 4 ln ij s f 29= 3 5 ; +Cst and: Lamp (D j p; ; ) = 1 2 XN j=1 kaj p (1 exp ( ij ))k2+Cst :
  • 19. Configuration We will use for our model parameters vector, that is, for a model with K neurons: = (P1;1; 1; S1; F1; : : : ;PK;K; K; SK; FK) : We will formalize our a priori ignorance of the origin of each spike by associating to each spike, j , a label Lj 2 f1; : : : ;Kg. lj = 3 means that event j was generated by neuron 3. We will use C for the configuration, that is, the random variable: C = (L1; : : : ; LN)T : With this formalism, spike sorting amounts to estimating the configuration.
  • 20. Likelihood for data from several neurons If the configuration realization is known, the likelihood computation is easily done: (D j c; ) = YK q=1 YNq j=1 isi (iq;j j sq; fq)amp aq;j j iq;j ; pq; q; q :
  • 21. Bayesian formalism (1) I Being somewhat opportunist, we are going to formalize our ignorance on the configuration c and the model parameters by viewing them as realizations of the random variables C and . I We then have a triplet of random variables D (whose realization is D), C and . I We have, by definition of the conditional probability: (D; c; ) = (c; j D) (D) = (D j c; ) (c; ) : I Rearranging the last two members we get (Bayes rule or Bayes theorem): (c; j D) = (D j c; ) (c; ) (D) :
  • 22. Bayesian formalism (2) I In the literature, our (c; ) is called the a priori density and is written prior (c; ); that’s what we know about the parameters and configuration before observing the data. I Our (c; j D) is called the a posteriori density and is written posterior (c; j D); that’s what we know about the parameters and configuration after observing the data. I Bayes rule tells us how to update our knowledge once the data have been observed. I We know how to compute the likelihood (D j c; ).
  • 23. Bayesian formalism (3) I We will write prior (c; ) = prior (c)prior () and take prior (c) uniform on C = f1; : : : ;KgN (notice that I’m cheating, I can know N only after seeing the data) and take prior () = YK q=1 prior (sq)prior (fq)prior (pq)prior (q)prior (q) ; and each of the terms is going to be the density of a uniform distribution over a large domain. I The real problem is the denominator: (D) = X c2C Z (D; c; )d = X c2C Z (D j c; )prior (c; )d ; since it involves a summation over KN configurations and K is of the order of 10 and N of the order of 1000!
  • 24. Bayesian formalism (4) I Assuming we find a way to circumvent the combinatorial explosion (KN) problem and to get an estimator, ^(c; j D), of: posterior (c; j D) = (D j c; ) (c; ) (D) : I The estimated posterior configuration probability would be: ^posterior (c j D) = Z ^(c; j D)d : I The spike sorting results could then be summarized by the Maximum A Posteriori (MAP) estimator: ^c = arg min c2C ^posterior (c j D) :
  • 25. Bayesian formalism (5) I Clearly, a better use of the results would be for any statistics T(C) (think of the cross-correlogram between two spike trains) to use: [T(C) = X c2C T(c)^posterior (c j D) ; instead of: T(^c). I But to do all that we have to deal with the combinatorial explosion! I In such situations, a good strategy is too look at other domains to see if other people met a similar problem and found a solution. . .
  • 26. Where are we ? What makes data difficult An interlude to keep you motivated A more realistic model and its problems A analogy with statistical physics Back to our simulated data Application to real data
  • 27. A analogy with statistical physics (1) I Luckily for us, Physicists got a problem similar to ours at the turn of the 1950s and found a solution. I Let us write: E (c; j D) = log [ (D j c; ) prior (c; )] ; notice that we know how to compute E (c; j D). I Then: posterior (c; j D) = exp [
  • 28. E (c; j D)] (D) ; where
  • 29. = 1,
  • 30. is the inverse temperature for Physicists who write Z(
  • 31. ) = (D) and call it the partition function. I Written in this way, posterior (c; j D), is nothing more than a Gibbs distribution.
  • 32. A analogy with statistical physics (2) I Physicists don’t know how to compute the partition function Z(
  • 33. ) in many cases of interest. I But they need to compute quantities like: X c2C Z T(c; )posterior (c; j D)d : I In 1953, Metropolis, Rosenbluth, Rosenbluth, Teller and Teller published Equation of State Calculations by Fast Computing Machines in Journal of Chemical Physics, proposing an algorithm for estimating this kind of integral without knowing the partition function.
  • 34. A analogy with statistical physics (3) I In 1971, Hastings published Monte Carlo Sampling Methods Using Markov Chains and Their Applications in Biometrika proving that the previous algorithm was indeed doing what it was suppose to do. I This algorithm is know called the Metropolis-Hastings Algorithm.
  • 35. Metropolis-Hastings in a simple setting (1) I We will generate a collection of states fx(1); x(2); : : : ; x(R)g according to a target distribution (x) (the x stand here for our previous c; ). I To that end we use a Markov chain which asymptotically reaches the stationary distribution (x). I A Markov chain is uniquely defined by its transition matrix P(X(k+1) = x0 j X(k) = x) and its initial state distribution.
  • 36. Metropolis-Hastings in a simple setting (2) I A Markov chain has a unique stationary distribution when the following two conditions are met: 1. A stationary distribution exists: a sufficient condition for that is the detailed balance that implies (x)P(x0 j x) = (x0)P(x j x0) : 2. The stationary distribution is unique if the chain is aperiodic (the system does not return to the same state at fixed intervals) and positive recurrent (the expected number of steps for returning to the same state is finite). I Checking the detailed balance requires a knowledge of (x) up to a normalizing constant. I Constructing a transition matrix satisfying the detailed balance for a given target distribution also requires the knowledge of that distribution up to a normalizing constant.
  • 37. Metropolis-Hastings in a simple setting (3) I The trick is to construct the transition matrix elements in two sub-steps: a proposal, g(x0 j x), and an acceptance-rejection, A(x0 j x), to get P(x0 j x) = g(x0 j x) A(x0 j x). I The detailed balance is satisfied if: (x)g(x0 j x)A(x0 j x) = (x0)g(x j x0)A(x j x0) that is if A(x0 j x) A(x j x0) = (x0)g(x j x0) (x)g(x0 j x) : I The Metropolis choice is: A(x0 j x) = min 1; (x0)g(x j x0) (x)g(x0 j x) :
  • 38. Metropolis-Hastings in a simple setting (4) I The Metropolis choice for A(x0 j x) requires only the knowledge of the target distribution up to a normalizing constant. I The proposal matrix g(x0 j x) is chosen to ensure positive recurrence and aperiodicity. I If g is positive recurrent and aperiodic and if g(x0 j x) = 0 implies g(x j x0) = 0 then we can show that P(x0 j x) is positive recurrent and aperiodic. I We then have a recipe to modify a positive recurrent and aperiodic transition matrix in order to have the stationary distribution of our choice while just knowing it up to a normalizing constant.
  • 39. A first demo Show the demo on a toy example.
  • 40. Where are we ? What makes data difficult An interlude to keep you motivated A more realistic model and its problems A analogy with statistical physics Back to our simulated data Application to real data
  • 41. Back to our simulated data Simulated data with 3 neurones on 2 recording sites.
  • 42. Energy evolution Energy evolution during 3 105 MC steps.
  • 44. Replica Exchange Method / Parallel Tempering REM / Tempering demo. . .
  • 45. Energy evolution with REM Energy evolution during 32 104 MC steps, REM turned on after 104 steps.
  • 50. Where are we ? What makes data difficult An interlude to keep you motivated A more realistic model and its problems A analogy with statistical physics Back to our simulated data Application to real data
  • 51. A Hidden-Markov model for actual discharges Fig. 1 of Delescluse and Pouzat (2006) J Neurosci Methods 150:16.
  • 52. MCMC algorithm applied to a single neuron discharge Fig. 2 of Delescluse and Pouzat (2006).
  • 53. Sorting with a reference Fig. 4 of Delescluse and Pouzat (2006).
  • 54. Bibliography I Hammersley, J.M. and Handscomb, D.C. (1964) Monte Carlo Methods. Methuen. Must read! Buy it second hand or find it on the web. I Liu, Jun S. (2001) Monte Carlo Strategies in Scientific Computing. Springer. Modern equivalent of the previous one, lots of applications in Biology. I MacKay, David JC (2003) Information theory, inference, and learning algorithms. CUP. Clear and fun! Available online: http://www.inference.phy.cam.ac.uk/mackay/itila/. I Neal, Radford M (1993) Probabilistic Inference Using Markov Chain Monte Carlo Methods. Technical Report CRG-TR-93-1, Dept. of Computer Science, University of Toronto. I learned MCMC with that. Available online: http: //www.cs.toronto.edu/~radford/papers-online.html.
  • 55. That’s all for today! I want to thank: I Antonio Galvès for inviting me to give these lectures. I João Alexandre Peschanski and Simone Harnik for taking care of the transmission of these lectures. I You guys for listening!