# -*- coding: utf-8 -*-
"""
CuBIC is a statistical method for the detection of higher order of
correlations in parallel spike trains based on the analysis of the
cumulants of the population count.
.. autosummary::
:toctree: _toctree/cubic
cubic
Examples
--------
Homogeneous Poisson random spike trains population count histogram third
cumulant is explained by the first correlation order (xi=1).
Given a list of spike trains, the analysis comprises the following steps:
1) compute the population histogram (PSTH) with the desired bin size
>>> import numpy as np
>>> import quantities as pq
>>> from elephant import statistics
>>> from elephant.cubic import cubic
>>> from elephant.spike_train_generation import StationaryPoissonProcess
>>> np.random.seed(10)
>>> spiketrains = [StationaryPoissonProcess(rate=10*pq.Hz,
... t_stop=10 * pq.s).generate_spiketrain() for _ in range(20)]
>>> pop_count = statistics.time_histogram(spiketrains, bin_size=0.1 * pq.s)
2) apply CuBIC to the population count
>>> xi, p_val, kappa, test_aborted = cubic(pop_count, alpha=0.05)
>>> xi
1
>>> p_val # doctest: +SKIP
[0.43014065113883904]
>>> kappa # doctest: +SKIP
[20.1, 22.656565656565657, 27.674706246134818]
:copyright: Copyright 2014-2024 by the Elephant team, see `doc/authors.rst`.
:license: BSD, see LICENSE.txt for details.
"""
from __future__ import division, print_function, unicode_literals
import math
import numpy as np
import warnings
import scipy.special
import scipy.stats
__all__ = [
"cubic"
]
# Based on matlab code by Benjamin Staude
# Adaptation to python by Pietro Quaglio and Emiliano Torre
[docs]
def cubic(histogram, max_iterations=100, alpha=0.05):
r"""
Performs the CuBIC analysis :cite:`cubic-Staude2010_327` on a population
histogram, calculated from a population of spiking neurons.
The null hypothesis :math:`H_0: k_3(data)<=k^*_{3,\xi}` is iteratively
tested with increasing correlation order :math:`\xi` until it is possible
to accept, with a significance level `alpha`, that :math:`\hat{\xi}` is
the minimum order of correlation necessary to explain the third cumulant
:math:`k_3(data)`.
:math:`k^*_{3,\xi}` is the maximized third cumulant, supposing a Compound
Poisson Process (CPP) model for correlated spike trains (see the paper)
with maximum order of correlation equal to :math:`\xi`.
Parameters
----------
histogram : neo.AnalogSignal
The population histogram (count of spikes per time bin) of the entire
population of neurons.
max_iterations : int, optional
The maximum number of iterations of the hypothesis test. Corresponds
to the :math:`\hat{\xi_{\text{max}}}` in :cite:`cubic-Staude2010_327`.
If it is not possible to compute the :math:`\hat{\xi}` before
`max_iterations` iteration, the CuBIC procedure is aborted.
Default: 100
alpha : float, optional
The significance level of the hypothesis tests performed.
Default: 0.05
Returns
-------
xi_hat : int
The minimum correlation order estimated by CuBIC, necessary to
explain the value of the third cumulant calculated from the population.
p : list
The ordered list of all the p-values of the hypothesis tests that have
been performed. If the maximum number of iteration `max_iterations` is
reached, the last p-value is set to -4.
kappa : list
The list of the first three cumulants of the data.
test_aborted : bool
Whether the test was aborted because reached the maximum number of
iteration, `max_iterations`.
"""
if alpha < 0 or alpha > 1:
raise ValueError(f'the significance level alpha ({alpha}) has to be '
f'in [0, 1] range')
if not isinstance(max_iterations, int) or max_iterations < 0:
raise ValueError(f"'max_iterations' ({max_iterations}) has to be a "
"positive integer")
# dict of all possible rate functions
try:
histogram = histogram.magnitude
except AttributeError:
pass
L = len(histogram)
# compute first three cumulants
kappa = _kstat(histogram)
xi_hat = 1
xi = 1
pval = 0.
p = []
test_aborted = False
# compute xi_hat iteratively
while pval < alpha:
xi_hat = xi
if xi > max_iterations:
warnings.warn(f'Test aborted after ximax={max_iterations} '
f'iterations with p-value={pval}')
test_aborted = True
break
# compute p-value
pval = _H03xi(kappa, xi, L)
p.append(pval)
xi = xi + 1
return xi_hat, p, kappa, test_aborted
def _H03xi(kappa, xi, L):
"""
Computes the p_value for testing the :math:`H_0: k_3(data)<=k^*_{3,\\xi}`
hypothesis of CuBIC in the stationary rate version
Parameters
-----
kappa : list
The first three cumulants of the populaton of spike trains
xi : int
The the maximum order of correlation :math:`\\xi` supposed in the
hypothesis for which is computed the p value of :math:`H_0`
L : float
The length of the orginal population histogram on which is performed
the CuBIC analysis
Returns
-----
p : float
The p-value of the hypothesis tests
"""
# Check the order condition of the cumulants necessary to perform CuBIC
if kappa[1] < kappa[0]:
raise ValueError(f"The null hypothesis H_0 cannot be tested: the "
f"population count histogram variance ({kappa[1]}) "
f"is less than the mean ({kappa[0]}). This can "
f"happen when the spike train population is not "
f"large enough or the bin size is small.")
else:
# computation of the maximized cumulants
kstar = [_kappamstar(kappa[:2], i, xi) for i in range(2, 7)]
k3star = kstar[1]
# variance of third cumulant (from Stuart & Ord)
sigmak3star = math.sqrt(
kstar[4] / L + 9 * (kstar[2] * kstar[0] + kstar[1] ** 2) /
(L - 1) + 6 * L * kstar[0] ** 3 / ((L - 1) * (L - 2)))
# computation of the p-value (the third cumulant is supposed to
# be gaussian distributed)
p = 1 - scipy.stats.norm(k3star, sigmak3star).cdf(kappa[2])
return p
def _kappamstar(kappa, m, xi):
"""
Computes maximized cumulant of order m
Parameters
-----
kappa : list
The first two cumulants of the data
xi : int
The :math:`\\xi` for which is computed the p value of :math:`H_0`
m : float
The order of the cumulant
Returns
-----
k_out : list
The maximized cumulant of order m
"""
if xi == 1:
kappa_out = kappa[1]
else:
kappa_out = \
(kappa[1] * (xi ** (m - 1) - 1) -
kappa[0] * (xi ** (m - 1) - xi)) / (xi - 1)
return kappa_out
def _kstat(data):
"""
Compute first three cumulants of a population count of a population of
spiking
See http://mathworld.wolfram.com/k-Statistic.html
Parameters
-----
data : numpy.ndarray
The population histogram of the population on which are computed
the cumulants
Returns
-----
moments : list
The first three unbiased cumulants of the population count
"""
if len(data) == 0:
raise ValueError('The input data must be a non-empty array')
moments = [scipy.stats.kstat(data, n=n) for n in [1, 2, 3]]
return moments