Skip to content

Commit de8cb58

Browse files
committed
updating
1 parent 14a3224 commit de8cb58

File tree

8 files changed

+142
-608
lines changed

8 files changed

+142
-608
lines changed

doc/pub/week13/html/week13-bs.html

Lines changed: 0 additions & 79 deletions
Original file line numberDiff line numberDiff line change
@@ -132,14 +132,6 @@
132132
('Diffusion models', 2, None, 'diffusion-models'),
133133
('Original idea', 2, None, 'original-idea'),
134134
('Diffusion learning', 2, None, 'diffusion-learning'),
135-
('Diffusion models, basics', 2, None, 'diffusion-models-basics'),
136-
('Problems with probabilistic models',
137-
2,
138-
None,
139-
'problems-with-probabilistic-models'),
140-
('Diffusion models', 2, None, 'diffusion-models'),
141-
('Original idea', 2, None, 'original-idea'),
142-
('Diffusion learning', 2, None, 'diffusion-learning'),
143135
('Mathematics of diffusion models',
144136
2,
145137
None,
@@ -273,11 +265,6 @@
273265
<!-- navigation toc: --> <li><a href="#diffusion-models" style="font-size: 80%;">Diffusion models</a></li>
274266
<!-- navigation toc: --> <li><a href="#original-idea" style="font-size: 80%;">Original idea</a></li>
275267
<!-- navigation toc: --> <li><a href="#diffusion-learning" style="font-size: 80%;">Diffusion learning</a></li>
276-
<!-- navigation toc: --> <li><a href="#diffusion-models-basics" style="font-size: 80%;">Diffusion models, basics</a></li>
277-
<!-- navigation toc: --> <li><a href="#problems-with-probabilistic-models" style="font-size: 80%;">Problems with probabilistic models</a></li>
278-
<!-- navigation toc: --> <li><a href="#diffusion-models" style="font-size: 80%;">Diffusion models</a></li>
279-
<!-- navigation toc: --> <li><a href="#original-idea" style="font-size: 80%;">Original idea</a></li>
280-
<!-- navigation toc: --> <li><a href="#diffusion-learning" style="font-size: 80%;">Diffusion learning</a></li>
281268
<!-- navigation toc: --> <li><a href="#mathematics-of-diffusion-models" style="font-size: 80%;">Mathematics of diffusion models</a></li>
282269
<!-- navigation toc: --> <li><a href="#chains-of-vaes" style="font-size: 80%;">Chains of VAEs</a></li>
283270
<!-- navigation toc: --> <li><a href="#mathematical-representation" style="font-size: 80%;">Mathematical representation</a></li>
@@ -1316,72 +1303,6 @@ <h2 id="code-in-pytorch-for-vaes" class="anchor">Code in PyTorch for VAEs </h2>
13161303
</div>
13171304

13181305

1319-
<!-- !split -->
1320-
<h2 id="diffusion-models-basics" class="anchor">Diffusion models, basics </h2>
1321-
1322-
<p>Diffusion models are inspired by non-equilibrium thermodynamics. They
1323-
define a Markov chain of diffusion steps to slowly add random noise to
1324-
data and then learn to reverse the diffusion process to construct
1325-
desired data samples from the noise. Unlike VAE or flow models,
1326-
diffusion models are learned with a fixed procedure and the latent
1327-
variable has high dimensionality (same as the original data).
1328-
</p>
1329-
1330-
<!-- !split -->
1331-
<h2 id="problems-with-probabilistic-models" class="anchor">Problems with probabilistic models </h2>
1332-
1333-
<p>Historically, probabilistic models suffer from a tradeoff between two
1334-
conflicting objectives: \textit{tractability} and
1335-
\textit{flexibility}. Models that are \textit{tractable} can be
1336-
analytically evaluated and easily fit to data (e.g. a Gaussian or
1337-
Laplace). However, these models are unable to aptly describe structure
1338-
in rich datasets. On the other hand, models that are \textit{flexible}
1339-
can be molded to fit structure in arbitrary data. For example, we can
1340-
define models in terms of any (non-negative) function \( \phi(\boldsymbol{x}) \)
1341-
yielding the flexible distribution \( p\left(\boldsymbol{x}\right) =
1342-
\frac{\phi\left(\boldsymbol{x} \right)}{Z} \), where \( Z \) is a normalization
1343-
constant. However, computing this normalization constant is generally
1344-
intractable. Evaluating, training, or drawing samples from such
1345-
flexible models typically requires a very expensive Monte Carlo
1346-
process.
1347-
</p>
1348-
1349-
<!-- !split -->
1350-
<h2 id="diffusion-models" class="anchor">Diffusion models </h2>
1351-
<p>Diffusion models have several interesting features</p>
1352-
<ul>
1353-
<li> extreme flexibility in model structure,</li>
1354-
<li> exact sampling,</li>
1355-
<li> easy multiplication with other distributions, e.g. in order to compute a posterior, and</li>
1356-
<li> the model log likelihood, and the probability of individual states, to be cheaply evaluated.</li>
1357-
</ul>
1358-
<!-- !split -->
1359-
<h2 id="original-idea" class="anchor">Original idea </h2>
1360-
1361-
<p>In the original formulation, one uses a Markov chain to gradually
1362-
convert one distribution into another, an idea used in non-equilibrium
1363-
statistical physics and sequential Monte Carlo. Diffusion models build
1364-
a generative Markov chain which converts a simple known distribution
1365-
(e.g. a Gaussian) into a target (data) distribution using a diffusion
1366-
process. Rather than use this Markov chain to approximately evaluate a
1367-
model which has been otherwise defined, one can explicitly define the
1368-
probabilistic model as the endpoint of the Markov chain. Since each
1369-
step in the diffusion chain has an analytically evaluable probability,
1370-
the full chain can also be analytically evaluated.
1371-
</p>
1372-
1373-
<!-- !split -->
1374-
<h2 id="diffusion-learning" class="anchor">Diffusion learning </h2>
1375-
1376-
<p>Learning in this framework involves estimating small perturbations to
1377-
a diffusion process. Estimating small, analytically tractable,
1378-
perturbations is more tractable than explicitly describing the full
1379-
distribution with a single, non-analytically-normalizable, potential
1380-
function. Furthermore, since a diffusion process exists for any
1381-
smooth target distribution, this method can capture data distributions
1382-
of arbitrary form.
1383-
</p>
1384-
13851306
<!-- !split -->
13861307
<h2 id="diffusion-models-basics" class="anchor">Diffusion models, basics </h2>
13871308

doc/pub/week13/html/week13-reveal.html

Lines changed: 0 additions & 72 deletions
Original file line numberDiff line numberDiff line change
@@ -1348,78 +1348,6 @@ <h2 id="diffusion-learning">Diffusion learning </h2>
13481348
</p>
13491349
</section>
13501350

1351-
<section>
1352-
<h2 id="diffusion-models-basics">Diffusion models, basics </h2>
1353-
1354-
<p>Diffusion models are inspired by non-equilibrium thermodynamics. They
1355-
define a Markov chain of diffusion steps to slowly add random noise to
1356-
data and then learn to reverse the diffusion process to construct
1357-
desired data samples from the noise. Unlike VAE or flow models,
1358-
diffusion models are learned with a fixed procedure and the latent
1359-
variable has high dimensionality (same as the original data).
1360-
</p>
1361-
</section>
1362-
1363-
<section>
1364-
<h2 id="problems-with-probabilistic-models">Problems with probabilistic models </h2>
1365-
1366-
<p>Historically, probabilistic models suffer from a tradeoff between two
1367-
conflicting objectives: \textit{tractability} and
1368-
\textit{flexibility}. Models that are \textit{tractable} can be
1369-
analytically evaluated and easily fit to data (e.g. a Gaussian or
1370-
Laplace). However, these models are unable to aptly describe structure
1371-
in rich datasets. On the other hand, models that are \textit{flexible}
1372-
can be molded to fit structure in arbitrary data. For example, we can
1373-
define models in terms of any (non-negative) function \( \phi(\boldsymbol{x}) \)
1374-
yielding the flexible distribution \( p\left(\boldsymbol{x}\right) =
1375-
\frac{\phi\left(\boldsymbol{x} \right)}{Z} \), where \( Z \) is a normalization
1376-
constant. However, computing this normalization constant is generally
1377-
intractable. Evaluating, training, or drawing samples from such
1378-
flexible models typically requires a very expensive Monte Carlo
1379-
process.
1380-
</p>
1381-
</section>
1382-
1383-
<section>
1384-
<h2 id="diffusion-models">Diffusion models </h2>
1385-
<p>Diffusion models have several interesting features</p>
1386-
<ul>
1387-
<p><li> extreme flexibility in model structure,</li>
1388-
<p><li> exact sampling,</li>
1389-
<p><li> easy multiplication with other distributions, e.g. in order to compute a posterior, and</li>
1390-
<p><li> the model log likelihood, and the probability of individual states, to be cheaply evaluated.</li>
1391-
</ul>
1392-
</section>
1393-
1394-
<section>
1395-
<h2 id="original-idea">Original idea </h2>
1396-
1397-
<p>In the original formulation, one uses a Markov chain to gradually
1398-
convert one distribution into another, an idea used in non-equilibrium
1399-
statistical physics and sequential Monte Carlo. Diffusion models build
1400-
a generative Markov chain which converts a simple known distribution
1401-
(e.g. a Gaussian) into a target (data) distribution using a diffusion
1402-
process. Rather than use this Markov chain to approximately evaluate a
1403-
model which has been otherwise defined, one can explicitly define the
1404-
probabilistic model as the endpoint of the Markov chain. Since each
1405-
step in the diffusion chain has an analytically evaluable probability,
1406-
the full chain can also be analytically evaluated.
1407-
</p>
1408-
</section>
1409-
1410-
<section>
1411-
<h2 id="diffusion-learning">Diffusion learning </h2>
1412-
1413-
<p>Learning in this framework involves estimating small perturbations to
1414-
a diffusion process. Estimating small, analytically tractable,
1415-
perturbations is more tractable than explicitly describing the full
1416-
distribution with a single, non-analytically-normalizable, potential
1417-
function. Furthermore, since a diffusion process exists for any
1418-
smooth target distribution, this method can capture data distributions
1419-
of arbitrary form.
1420-
</p>
1421-
</section>
1422-
14231351
<section>
14241352
<h2 id="mathematics-of-diffusion-models">Mathematics of diffusion models </h2>
14251353

doc/pub/week13/html/week13-solarized.html

Lines changed: 0 additions & 74 deletions
Original file line numberDiff line numberDiff line change
@@ -159,14 +159,6 @@
159159
('Diffusion models', 2, None, 'diffusion-models'),
160160
('Original idea', 2, None, 'original-idea'),
161161
('Diffusion learning', 2, None, 'diffusion-learning'),
162-
('Diffusion models, basics', 2, None, 'diffusion-models-basics'),
163-
('Problems with probabilistic models',
164-
2,
165-
None,
166-
'problems-with-probabilistic-models'),
167-
('Diffusion models', 2, None, 'diffusion-models'),
168-
('Original idea', 2, None, 'original-idea'),
169-
('Diffusion learning', 2, None, 'diffusion-learning'),
170162
('Mathematics of diffusion models',
171163
2,
172164
None,
@@ -1242,72 +1234,6 @@ <h2 id="code-in-pytorch-for-vaes">Code in PyTorch for VAEs </h2>
12421234
</div>
12431235

12441236

1245-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1246-
<h2 id="diffusion-models-basics">Diffusion models, basics </h2>
1247-
1248-
<p>Diffusion models are inspired by non-equilibrium thermodynamics. They
1249-
define a Markov chain of diffusion steps to slowly add random noise to
1250-
data and then learn to reverse the diffusion process to construct
1251-
desired data samples from the noise. Unlike VAE or flow models,
1252-
diffusion models are learned with a fixed procedure and the latent
1253-
variable has high dimensionality (same as the original data).
1254-
</p>
1255-
1256-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1257-
<h2 id="problems-with-probabilistic-models">Problems with probabilistic models </h2>
1258-
1259-
<p>Historically, probabilistic models suffer from a tradeoff between two
1260-
conflicting objectives: \textit{tractability} and
1261-
\textit{flexibility}. Models that are \textit{tractable} can be
1262-
analytically evaluated and easily fit to data (e.g. a Gaussian or
1263-
Laplace). However, these models are unable to aptly describe structure
1264-
in rich datasets. On the other hand, models that are \textit{flexible}
1265-
can be molded to fit structure in arbitrary data. For example, we can
1266-
define models in terms of any (non-negative) function \( \phi(\boldsymbol{x}) \)
1267-
yielding the flexible distribution \( p\left(\boldsymbol{x}\right) =
1268-
\frac{\phi\left(\boldsymbol{x} \right)}{Z} \), where \( Z \) is a normalization
1269-
constant. However, computing this normalization constant is generally
1270-
intractable. Evaluating, training, or drawing samples from such
1271-
flexible models typically requires a very expensive Monte Carlo
1272-
process.
1273-
</p>
1274-
1275-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1276-
<h2 id="diffusion-models">Diffusion models </h2>
1277-
<p>Diffusion models have several interesting features</p>
1278-
<ul>
1279-
<li> extreme flexibility in model structure,</li>
1280-
<li> exact sampling,</li>
1281-
<li> easy multiplication with other distributions, e.g. in order to compute a posterior, and</li>
1282-
<li> the model log likelihood, and the probability of individual states, to be cheaply evaluated.</li>
1283-
</ul>
1284-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1285-
<h2 id="original-idea">Original idea </h2>
1286-
1287-
<p>In the original formulation, one uses a Markov chain to gradually
1288-
convert one distribution into another, an idea used in non-equilibrium
1289-
statistical physics and sequential Monte Carlo. Diffusion models build
1290-
a generative Markov chain which converts a simple known distribution
1291-
(e.g. a Gaussian) into a target (data) distribution using a diffusion
1292-
process. Rather than use this Markov chain to approximately evaluate a
1293-
model which has been otherwise defined, one can explicitly define the
1294-
probabilistic model as the endpoint of the Markov chain. Since each
1295-
step in the diffusion chain has an analytically evaluable probability,
1296-
the full chain can also be analytically evaluated.
1297-
</p>
1298-
1299-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1300-
<h2 id="diffusion-learning">Diffusion learning </h2>
1301-
1302-
<p>Learning in this framework involves estimating small perturbations to
1303-
a diffusion process. Estimating small, analytically tractable,
1304-
perturbations is more tractable than explicitly describing the full
1305-
distribution with a single, non-analytically-normalizable, potential
1306-
function. Furthermore, since a diffusion process exists for any
1307-
smooth target distribution, this method can capture data distributions
1308-
of arbitrary form.
1309-
</p>
1310-
13111237
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
13121238
<h2 id="diffusion-models-basics">Diffusion models, basics </h2>
13131239

doc/pub/week13/html/week13.html

Lines changed: 0 additions & 74 deletions
Original file line numberDiff line numberDiff line change
@@ -236,14 +236,6 @@
236236
('Diffusion models', 2, None, 'diffusion-models'),
237237
('Original idea', 2, None, 'original-idea'),
238238
('Diffusion learning', 2, None, 'diffusion-learning'),
239-
('Diffusion models, basics', 2, None, 'diffusion-models-basics'),
240-
('Problems with probabilistic models',
241-
2,
242-
None,
243-
'problems-with-probabilistic-models'),
244-
('Diffusion models', 2, None, 'diffusion-models'),
245-
('Original idea', 2, None, 'original-idea'),
246-
('Diffusion learning', 2, None, 'diffusion-learning'),
247239
('Mathematics of diffusion models',
248240
2,
249241
None,
@@ -1319,72 +1311,6 @@ <h2 id="code-in-pytorch-for-vaes">Code in PyTorch for VAEs </h2>
13191311
</div>
13201312

13211313

1322-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1323-
<h2 id="diffusion-models-basics">Diffusion models, basics </h2>
1324-
1325-
<p>Diffusion models are inspired by non-equilibrium thermodynamics. They
1326-
define a Markov chain of diffusion steps to slowly add random noise to
1327-
data and then learn to reverse the diffusion process to construct
1328-
desired data samples from the noise. Unlike VAE or flow models,
1329-
diffusion models are learned with a fixed procedure and the latent
1330-
variable has high dimensionality (same as the original data).
1331-
</p>
1332-
1333-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1334-
<h2 id="problems-with-probabilistic-models">Problems with probabilistic models </h2>
1335-
1336-
<p>Historically, probabilistic models suffer from a tradeoff between two
1337-
conflicting objectives: \textit{tractability} and
1338-
\textit{flexibility}. Models that are \textit{tractable} can be
1339-
analytically evaluated and easily fit to data (e.g. a Gaussian or
1340-
Laplace). However, these models are unable to aptly describe structure
1341-
in rich datasets. On the other hand, models that are \textit{flexible}
1342-
can be molded to fit structure in arbitrary data. For example, we can
1343-
define models in terms of any (non-negative) function \( \phi(\boldsymbol{x}) \)
1344-
yielding the flexible distribution \( p\left(\boldsymbol{x}\right) =
1345-
\frac{\phi\left(\boldsymbol{x} \right)}{Z} \), where \( Z \) is a normalization
1346-
constant. However, computing this normalization constant is generally
1347-
intractable. Evaluating, training, or drawing samples from such
1348-
flexible models typically requires a very expensive Monte Carlo
1349-
process.
1350-
</p>
1351-
1352-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1353-
<h2 id="diffusion-models">Diffusion models </h2>
1354-
<p>Diffusion models have several interesting features</p>
1355-
<ul>
1356-
<li> extreme flexibility in model structure,</li>
1357-
<li> exact sampling,</li>
1358-
<li> easy multiplication with other distributions, e.g. in order to compute a posterior, and</li>
1359-
<li> the model log likelihood, and the probability of individual states, to be cheaply evaluated.</li>
1360-
</ul>
1361-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1362-
<h2 id="original-idea">Original idea </h2>
1363-
1364-
<p>In the original formulation, one uses a Markov chain to gradually
1365-
convert one distribution into another, an idea used in non-equilibrium
1366-
statistical physics and sequential Monte Carlo. Diffusion models build
1367-
a generative Markov chain which converts a simple known distribution
1368-
(e.g. a Gaussian) into a target (data) distribution using a diffusion
1369-
process. Rather than use this Markov chain to approximately evaluate a
1370-
model which has been otherwise defined, one can explicitly define the
1371-
probabilistic model as the endpoint of the Markov chain. Since each
1372-
step in the diffusion chain has an analytically evaluable probability,
1373-
the full chain can also be analytically evaluated.
1374-
</p>
1375-
1376-
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1377-
<h2 id="diffusion-learning">Diffusion learning </h2>
1378-
1379-
<p>Learning in this framework involves estimating small perturbations to
1380-
a diffusion process. Estimating small, analytically tractable,
1381-
perturbations is more tractable than explicitly describing the full
1382-
distribution with a single, non-analytically-normalizable, potential
1383-
function. Furthermore, since a diffusion process exists for any
1384-
smooth target distribution, this method can capture data distributions
1385-
of arbitrary form.
1386-
</p>
1387-
13881314
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
13891315
<h2 id="diffusion-models-basics">Diffusion models, basics </h2>
13901316

0 Bytes
Binary file not shown.

0 commit comments

Comments
 (0)