Author: | Thomas V. Wiecki, Imri Sofer, Mads L. Pedersen, Michael J. Frank |
---|---|
Contact: | thomas.wiecki@gmail.com, imri_sofer@brown.edu, madslupe@gmail.com, michael_frank@brown.edu |
Web site: | http://ski.clps.brown.edu/hddm_docs |
Github: | http://github.com/hddm-devs/hddm |
Mailing list: | https://groups.google.com/group/hddm-users/ |
Copyright: | This document has been placed in the public domain. |
License: | HDDM is released under the BSD 2 license. |
Version: | 0.7.8 |
HDDM is a python toolbox for hierarchical Bayesian parameter estimation of the Drift Diffusion Model (via PyMC). Drift Diffusion Models are used widely in psychology and cognitive neuroscience to study decision making.
Check out the tutorial on how to get started. Further information can be found below as well as in the howto section and the documentation.
- Uses hierarchical Bayesian estimation (via PyMC) of DDM parameters to allow simultaneous estimation of subject and group parameters, where individual subjects are assumed to be drawn from a group distribution. HDDM should thus produce better estimates when less RT values are measured compared to other methods using maximum likelihood for individual subjects (i.e. DMAT or fast-dm).
- Heavily optimized likelihood functions for speed (Navarro & Fuss, 2009).
- Flexible creation of complex models tailored to specific hypotheses (e.g. estimation of separate drift-rates for different task conditions; or predicted changes in model parameters as a function of other indicators like brain activity).
- Estimate trial-by-trial correlations between a brain measure (e.g. fMRI BOLD) and a diffusion model parameter using the HDDMRegression model.
- Built-in Bayesian hypothesis testing and several convergence and goodness-of-fit diagnostics.
- As of version 0.7.1 HDDM includes modules for analyzing reinforcement learning data with the reinforcement learning drift diffusion model (RLDDM), including a module for estimating the impact of continuous regressors onto RLDDM parameters, and a reinforcement learning (RL) model. See tutorial for the RLDDM and RL modules here: https://nbviewer.jupyter.org/github/hddm-devs/hddm/blob/master/hddm/examples/RLHDDMtutorial.ipynb
A recent paper by Roger Ratcliff quantitatively compared DMAT, fast-dm, and EZ, and concluded: "We found that the hierarchical diffusion method [as implemented by HDDM] performed very well, and is the method of choice when the number of observations is small."
Find the paper here: http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4517692/
The following is a minimal python script to load data, run a model and examine its parameters and fit.
import hddm # Load data from csv file into a NumPy structured array data = hddm.load_csv('simple_difficulty.csv') # Create a HDDM model multi object model = hddm.HDDM(data, depends_on={'v':'difficulty'}) # Create model and start MCMC sampling model.sample(2000, burn=20) # Print fitted parameters and other model statistics model.print_stats() # Plot posterior distributions and theoretical RT distributions model.plot_posteriors() model.plot_posterior_predictive()
For more information about the software and theories behind it, please see the main publication.
As of release 0.6.0, HDDM is compatible with Python 3 which we encourage.
The easiest way to install HDDM is through Anaconda (available for Windows, Linux and OSX):
- Download and install Anaconda.
- In a shell (Windows: Go to Start->Programs->Anaconda->Anaconda command prompt) type:
conda install -c pymc hddm
If you want to use pip instead of conda, type:
pip install pandas pip install pymc pip install kabuki pip install hddm
This might require super-user rights via sudo. Note that this installation method is discouraged as it leads to all kinds of problems on various platforms.
If you are having installation problems please contact the mailing list.
And if you're a mac user, check out this thread for advice on installation.
If HDDM was used in your research, please cite the publication:
Wiecki TV, Sofer I and Frank MJ (2013). HDDM: Hierarchical Bayesian estimation of the Drift-Diffusion Model in Python. Front. Neuroinform. 7:14. doi: 10.3389/fninf.2013.00014
HDDM has been used in over 400 published papers.
James Rowe (Cambridge University): "The HDDM modelling gave insights into the effects of disease that were simply not visible from a traditional analysis of RT/Accuracy. It provides a clue as to why many disorders including PD and PSP can give the paradoxical combination of akinesia and impulsivity. Perhaps of broader interest, the hierarchical drift diffusion model turned out to be very robust. In separate work, we have found that the HDDM gave accurate estimates of decision parameters with many fewer than 100 trials, in contrast to the hundreds or even thousands one might use for ‘traditional’ DDMs. This meant it was realistic to study patients who do not tolerate long testing sessions."
Check out the tutorial on how to get started. Further information can be found in howto and the documentation.
Join our low-traffic mailing list.