Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Bayesian Modeling and Probabilistic Programming in Python

License

NotificationsYou must be signed in to change notification settings

pymc-devs/pymc

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

PyMC logo

Build StatusCoverageNumFOCUS_badgeBinderDockerhubDOIzenodoConda Downloads

PyMC (formerly PyMC3) is a Python package for Bayesian statistical modelingfocusing on advanced Markov chain Monte Carlo (MCMC) and variational inference (VI)algorithms. Its flexibility and extensibility make it applicable to alarge suite of problems.

Check out thePyMC overview, orone ofthe many examples!For questions on PyMC, head on over to ourPyMC Discourse forum.

Features

  • Intuitive model specification syntax, for example,x ~ N(0,1)translates tox = Normal('x',0,1)
  • Powerful sampling algorithms, such as theNo U-TurnSampler, allow complex modelswith thousands of parameters with little specialized knowledge offitting algorithms.
  • Variational inference:ADVIfor fast approximate posterior estimation as well as mini-batch ADVIfor large data sets.
  • Relies onPyTensor which provides:
    • Computation optimization and dynamic C or JAX compilation
    • NumPy broadcasting and advanced indexing
    • Linear algebra operators
    • Simple extensibility
  • Transparent support for missing value imputation

Linear Regression Example

Plant growth can be influenced by multiple factors, and understanding these relationships is crucial for optimizing agricultural practices.

Imagine we conduct an experiment to predict the growth of a plant based on different environmental variables.

importpymcaspm# Taking draws from a normal distributionseed=42x_dist=pm.Normal.dist(shape=(100,3))x_data=pm.draw(x_dist,random_seed=seed)# Independent Variables:# Sunlight Hours: Number of hours the plant is exposed to sunlight daily.# Water Amount: Daily water amount given to the plant (in milliliters).# Soil Nitrogen Content: Percentage of nitrogen content in the soil.# Dependent Variable:# Plant Growth (y): Measured as the increase in plant height (in centimeters) over a certain period.# Define coordinate values for all dimensions of the datacoords={"trial":range(100),"features": ["sunlight hours","water amount","soil nitrogen"],}# Define generative modelwithpm.Model(coords=coords)asgenerative_model:x=pm.Data("x",x_data,dims=["trial","features"])# Model parametersbetas=pm.Normal("betas",dims="features")sigma=pm.HalfNormal("sigma")# Linear modelmu=x @betas# Likelihood# Assuming we measure deviation of each plant from baselineplant_growth=pm.Normal("plant growth",mu,sigma,dims="trial")# Generating data from model by fixing parametersfixed_parameters= {"betas": [5,20,2],"sigma":0.5,}withpm.do(generative_model,fixed_parameters)assynthetic_model:idata=pm.sample_prior_predictive(random_seed=seed)# Sample from prior predictive distribution.synthetic_y=idata.prior["plant growth"].sel(draw=0,chain=0)# Infer parameters conditioned on observed datawithpm.observe(generative_model, {"plant growth":synthetic_y})asinference_model:idata=pm.sample(random_seed=seed)summary=pm.stats.summary(idata,var_names=["betas","sigma"])print(summary)

From the summary, we can see that the mean of the inferred parameters are very close to the fixed parameters

Paramsmeansdhdi_3%hdi_97%mcse_meanmcse_sdess_bulkess_tailr_hat
betas[sunlight hours]4.9720.0544.8665.0660.0010.001300312571
betas[water amount]19.9630.05119.87220.0620.0010.001311216581
betas[soil nitrogen]1.9940.0551.8992.1070.0010.001322115591
sigma0.5110.0370.4380.5750.0010294515221
# Simulate new data conditioned on inferred parametersnew_x_data=pm.draw(pm.Normal.dist(shape=(3,3)),random_seed=seed,)new_coords=coords| {"trial": [0,1,2]}withinference_model:pm.set_data({"x":new_x_data},coords=new_coords)pm.sample_posterior_predictive(idata,predictions=True,extend_inferencedata=True,random_seed=seed,   )pm.stats.summary(idata.predictions,kind="stats")

The new data conditioned on inferred parameters would look like:

Outputmeansdhdi_3%hdi_97%
plant growth[0]14.2290.51513.32515.272
plant growth[1]24.4180.51123.42825.326
plant growth[2]-6.7470.511-7.740-5.797
# Simulate new data, under a scenario where the first beta is zerowithpm.do(inference_model, {inference_model["betas"]:inference_model["betas"]* [0,1,1]},)asplant_growth_model:new_predictions=pm.sample_posterior_predictive(idata,predictions=True,random_seed=seed,   )pm.stats.summary(new_predictions,kind="stats")

The new data, under the above scenario would look like:

Outputmeansdhdi_3%hdi_97%
plant growth[0]12.1490.51511.19313.135
plant growth[1]29.8090.50828.83230.717
plant growth[2]-0.1310.507-1.1210.791

Getting started

If you already know about Bayesian statistics:

Learn Bayesian statistics with a book together with PyMC

Audio & Video

  • Here is aYouTube playlist gathering several talks on PyMC.
  • You can also find all the talks given atPyMCon 2020here.
  • The"Learning Bayesian Statistics" podcast helps you discover and stay up-to-date with the vast Bayesian community. Bonus: it's hosted by Alex Andorra, one of the PyMC core devs!

Installation

To install PyMC on your system, follow the instructions on theinstallation guide.

Citing PyMC

Please choose from the following:

  • DOIpaperPyMC: A Modern and Comprehensive Probabilistic Programming Framework in Python, Abril-Pla O, Andreani V, Carroll C, Dong L, Fonnesbeck CJ, Kochurov M, Kumar R, Lao J, Luhmann CC, Martin OA, Osthege M, Vieira R, Wiecki T, Zinkov R. (2023)
  • DOIzenodo A DOI for all versions.
  • DOIs for specific versions are shown on Zenodo and underReleases

Contact

We are usingdiscourse.pymc.io as our main communication channel.

To ask a question regarding modeling or usage of PyMC we encourage posting to our Discourse forum under the“Questions” Category. You can also suggest feature in the“Development” Category.

You can also follow us on these social media platforms for updates and other announcements:

To report an issue with PyMC please use theissue tracker.

Finally, if you need to get in touch for non-technical information about the project,send us an e-mail.

License

Apache License, Version2.0

Software using PyMC

General purpose

  • Bambi: BAyesian Model-Building Interface (BAMBI) in Python.
  • calibr8: A toolbox for constructing detailed observation models to be used as likelihoods in PyMC.
  • gumbi: A high-level interface for building GP models.
  • SunODE: Fast ODE solver, much faster than the one that comes with PyMC.
  • pymc-learn: Custom PyMC models built on top of pymc3_models/scikit-learn API

Domain specific

  • Exoplanet: a toolkit for modeling of transit and/or radial velocity observations of exoplanets and other astronomical time series.
  • beat: Bayesian Earthquake Analysis Tool.
  • CausalPy: A package focussing on causal inference in quasi-experimental settings.
  • PyMC-Marketing: Bayesian marketing toolbox for marketing mix modeling, customer lifetime value, and more.

Please contact us if your software is not listed here.

Papers citing PyMC

See Google Scholarhere andhere for a continuously updated list.

Contributors

See theGitHub contributorpage. Also read ourCode of Conduct guidelines for a better contributing experience.

Support

PyMC is a non-profit project under NumFOCUS umbrella. If you want to support PyMC financially, you can donatehere.

Professional Consulting Support

You can get professional consulting support fromPyMC Labs.

Sponsors

NumFOCUS

PyMCLabs

Mistplay

ODSC

Thanks to our contributors

contributors


[8]ページ先頭

©2009-2025 Movatter.jp