<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>http://mw.hh.se/wg211/index.php?action=history&amp;feed=atom&amp;title=WG211%2FM21Taha</id>
	<title>WG211/M21Taha - Revision history</title>
	<link rel="self" type="application/atom+xml" href="http://mw.hh.se/wg211/index.php?action=history&amp;feed=atom&amp;title=WG211%2FM21Taha"/>
	<link rel="alternate" type="text/html" href="http://mw.hh.se/wg211/index.php?title=WG211/M21Taha&amp;action=history"/>
	<updated>2026-04-05T21:08:50Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.43.5</generator>
	<entry>
		<id>http://mw.hh.se/wg211/index.php?title=WG211/M21Taha&amp;diff=2340&amp;oldid=prev</id>
		<title>Walid: Bean Machine</title>
		<link rel="alternate" type="text/html" href="http://mw.hh.se/wg211/index.php?title=WG211/M21Taha&amp;diff=2340&amp;oldid=prev"/>
		<updated>2022-08-16T04:26:10Z</updated>

		<summary type="html">&lt;p&gt;Bean Machine&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 06:26, 16 August 2022&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l2&quot;&gt;Line 2:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 2:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Presented by: Walid Taha, Meta&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Presented by: Walid Taha, Meta&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Abstract and slides by: Rodrigo de Salvo Braz, Meta&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Abstract and slides by: Rodrigo de Salvo Braz, Meta&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Bean Machine (https://beanmachine.org/) (BM) is a new probabilistic programming system (https://research.facebook.com/blog/2021/12/introducing-bean-machine-a-probabilistic-programming-platform-built-on-pytorch/) recently beta-released as open-source by Meta (formerly Facebook). Probabilistic programming languages (PPLs) provide a way for experts to specify a Bayesian probabilistic model using a programming language, but without having to concern themselves with, or even know, how the probabilistic inference will be performed. While probabilistic programming is a growing field, Bean Machine provides a novel combination of features: experts can declaratively specify their models in Python (including the use of PyTorch tensors), use a variety of inference methods (including Hamiltonian Monte Carlo, No U-Turn Sample (NUTS) and Newtonian Monte Carlo), use composable inference to apply different methods to different parts of a model, work with hybrid models (containing both discrete and continuous variables) and, for a certain subset of models, use a very fast C++ inference layer. Bean Machine is already being used in production for Meta internal applications, but there are many exciting future plans for making it even better. In this talk I will give examples, describe applications, and discuss some of these future directions.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Bean Machine (https://beanmachine.org/) (BM) is a new probabilistic programming system (https://research.facebook.com/blog/2021/12/introducing-bean-machine-a-probabilistic-programming-platform-built-on-pytorch/) recently beta-released as open-source by Meta (formerly Facebook). Probabilistic programming languages (PPLs) provide a way for experts to specify a Bayesian probabilistic model using a programming language, but without having to concern themselves with, or even know, how the probabilistic inference will be performed. While probabilistic programming is a growing field, Bean Machine provides a novel combination of features: experts can declaratively specify their models in Python (including the use of PyTorch tensors), use a variety of inference methods (including Hamiltonian Monte Carlo, No U-Turn Sample (NUTS) and Newtonian Monte Carlo), use composable inference to apply different methods to different parts of a model, work with hybrid models (containing both discrete and continuous variables) and, for a certain subset of models, use a very fast C++ inference layer. Bean Machine is already being used in production for Meta internal applications, but there are many exciting future plans for making it even better. In this talk I will give examples, describe applications, and discuss some of these future directions.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Walid</name></author>
	</entry>
	<entry>
		<id>http://mw.hh.se/wg211/index.php?title=WG211/M21Taha&amp;diff=2339&amp;oldid=prev</id>
		<title>Walid: Bean Machine</title>
		<link rel="alternate" type="text/html" href="http://mw.hh.se/wg211/index.php?title=WG211/M21Taha&amp;diff=2339&amp;oldid=prev"/>
		<updated>2022-08-16T04:25:17Z</updated>

		<summary type="html">&lt;p&gt;Bean Machine&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;Bean Machine&lt;br /&gt;
&lt;br /&gt;
Presented by: Walid Taha, Meta&lt;br /&gt;
Abstract and slides by: Rodrigo de Salvo Braz, Meta&lt;br /&gt;
&lt;br /&gt;
Bean Machine (https://beanmachine.org/) (BM) is a new probabilistic programming system (https://research.facebook.com/blog/2021/12/introducing-bean-machine-a-probabilistic-programming-platform-built-on-pytorch/) recently beta-released as open-source by Meta (formerly Facebook). Probabilistic programming languages (PPLs) provide a way for experts to specify a Bayesian probabilistic model using a programming language, but without having to concern themselves with, or even know, how the probabilistic inference will be performed. While probabilistic programming is a growing field, Bean Machine provides a novel combination of features: experts can declaratively specify their models in Python (including the use of PyTorch tensors), use a variety of inference methods (including Hamiltonian Monte Carlo, No U-Turn Sample (NUTS) and Newtonian Monte Carlo), use composable inference to apply different methods to different parts of a model, work with hybrid models (containing both discrete and continuous variables) and, for a certain subset of models, use a very fast C++ inference layer. Bean Machine is already being used in production for Meta internal applications, but there are many exciting future plans for making it even better. In this talk I will give examples, describe applications, and discuss some of these future directions.&lt;/div&gt;</summary>
		<author><name>Walid</name></author>
	</entry>
</feed>