Skip to content

Commit

Permalink
update
Browse files Browse the repository at this point in the history
  • Loading branch information
MJ10 committed Dec 11, 2024
1 parent b1f3518 commit c05256a
Showing 1 changed file with 5 additions and 5 deletions.
10 changes: 5 additions & 5 deletions ft.html
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@
MathJax.Hub.Config({tex2jax: {inlineMath: [['$','$'], ['\\(','\\)']]}});
</script>
<script type="text/javascript"
src="./static/js/index.js">
src="./static/js/mathjax.js">
</script>
</head>
<body>
Expand Down Expand Up @@ -226,7 +226,7 @@ <h3 class="title is-4" id="setup">Setup</h3>


<div class="content has-text-justified">
<p> Given a diffusion model prior $p(\mathbf{x})$ and a black-box likelihood function $r(\mathbf{x})$, our goal is to sample from the posterior $p^{\text{post}}(\mathbf{x}) \propto p(\mathbf{x}) r(\mathbf{x})$. Conventional approaches often rely on heuristic guidance, leading to bias or restricted applicability. In contrast, we derive a principled, unbiased objective for posterior sampling, rooted in the Generative Flow Network (GFlowNet) perspective, which ensures improved mode coverage and asymptotic correctness without requiring data or approximations.
<p> Given a diffusion model prior \( p(\mathbf{x}) \) and a black-box likelihood function \( r(\mathbf{x}) \), our goal is to sample from the posterior \( p^{\text{post}}(\mathbf{x}) \propto p(\mathbf{x}) r(\mathbf{x}) \). Conventional approaches often rely on heuristic guidance, leading to bias or restricted applicability. In contrast, we derive a principled, unbiased objective for posterior sampling, rooted in the Generative Flow Network (GFlowNet) perspective, which ensures improved mode coverage and asymptotic correctness without requiring data or approximations.

</p>
</div>
Expand All @@ -248,7 +248,7 @@ <h2 class="title is-3" id="method">Relative trajectory balance</h2>
<img src="./static/images/example.png" alt="GMM Example" style="width: 100%;"/>
</center>
<p>
The <b>Relative Trajectory Balance (RTB)</b> objective ensures that the ratio of the forward trajectory probabilities under the posterior model $p_\phi^{\text{post}}$ to the prior model $p_\theta$ is proportional to the constraint function $r(\mathbf{x})$. This is achieved by minimizing the loss:
The <b>Relative Trajectory Balance (RTB)</b> objective ensures that the ratio of the forward trajectory probabilities under the posterior model \( p_\phi^{\text{post}} \) to the prior model \( p_\theta \) is proportional to the constraint function \( r(\mathbf{x}) \). This is achieved by minimizing the loss:
</p>
<div class="box" style="background-color:aliceblue">
<div class="content has-text-justified">
Expand All @@ -258,7 +258,7 @@ <h2 class="title is-3" id="method">Relative trajectory balance</h2>
</div>
<div class="content has-text-justified">
<p>
Here, $Z_{\phi}$ is a learnable normalization constant. Satisfying the RTB constraint (minimizing loss to 0) for all diffusion trajectories facilitates unbiased sampling from the desired posterior distribution $p^{\text{post}}(\mathbf{x}) \propto p_\theta(\mathbf{x}) r(\mathbf{x})$.
Here, \( Z_{\phi} \) is a learnable normalization constant. Satisfying the RTB constraint (minimizing loss to 0) for all diffusion trajectories facilitates unbiased sampling from the desired posterior distribution \( p^{\text{post}}(\mathbf{x}) \propto p_\theta(\mathbf{x}) r(\mathbf{x}) \).
</p>
</div>

Expand All @@ -275,7 +275,7 @@ <h2 class="title is-3" id="results">Empirical Results</h2>

<h3 class="title is-4" id="results">Unconditional Image</h3>
<p>
We fine-tune unconditional diffusion priors for class-conditional generation on MNIST and CIFAR-10 datasets. Starting with pretrained unconditional models $p_\theta(x)$, we apply the RTB objective to adapt the priors to sample from posteriors conditioned on class labels $c$. This is achieved by incorporating class-specific constraints $r(x) = p(c | x)$ during fine-tuning. In the figure, we observe some of our results. RTB effectively balances reward maximization and sample diversity, finetuning both for single class conditions, or multimodal distributions (e.g. even numbers).
We fine-tune unconditional diffusion priors for class-conditional generation on MNIST and CIFAR-10 datasets. Starting with pretrained unconditional models \( p_\theta(x) \), we apply the RTB objective to adapt the priors to sample from posteriors conditioned on class labels \( c \). This is achieved by incorporating class-specific constraints \( r(x) = p(c | x) \) during fine-tuning. In the figure, we observe some of our results. RTB effectively balances reward maximization and sample diversity, finetuning both for single class conditions, or multimodal distributions (e.g. even numbers).
</p>
<div class="content has-text-justified">
<center>
Expand Down

0 comments on commit c05256a

Please sign in to comment.