MarkTechPost@AI 05月21日 15:15
Sampling Without Data is Now Scalable: Meta AI Releases Adjoint Sampling for Reward-Driven Generative Modeling
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Meta AI推出了Adjoint Sampling,一种新型学习算法,旨在仅使用标量奖励信号训练生成模型。它基于随机最优控制的理论框架,将训练过程重构为受控扩散过程的优化任务。与标准生成模型不同,它不需要显式数据,而是通过使用奖励函数迭代地改进样本来学习生成高质量的样本,该奖励函数通常来自物理或化学能量模型。Adjoint Sampling擅长于只能访问未归一化能量函数的场景,它可以生成与此能量定义的目标分布对齐的样本,从而绕过了诸如重要性采样或MCMC之类的校正方法,这些方法在计算上非常密集。

🚀Adjoint Sampling算法基于随机微分方程(SDE),模拟样本轨迹的演变过程。通过学习控制漂移u(x,t),使这些轨迹的最终状态逼近期望分布,例如玻尔兹曼分布。

🔄Reciprocal Adjoint Matching (RAM)是其关键创新,作为一种损失函数,仅使用样本轨迹的初始和最终状态即可实现基于梯度的更新,避免了通过整个扩散路径的反向传播,从而大大提高了计算效率。

🧪Adjoint Sampling在合成和真实世界的任务中均取得了最先进的结果。在Double-Well (DW-4), Lennard-Jones (LJ-13 and LJ-55)等合成基准测试中,它显著优于DDS和PIS等基线,尤其是在能源效率方面。

🧬该算法在SPICE-MACE-OFF数据集上使用eSEN能量模型进行了大规模分子构象生成评估。Adjoint Sampling(尤其是带有预训练的笛卡尔变体)在所有指标上均超过了广泛使用的基于化学的基线RDKit ETKDG,实现了高达96.4%的召回率和0.60 Å的平均RMSD。

Data Scarcity in Generative Modeling

Generative models traditionally rely on large, high-quality datasets to produce samples that replicate the underlying data distribution. However, in fields like molecular modeling or physics-based inference, acquiring such data can be computationally infeasible or even impossible. Instead of labeled data, only a scalar reward—typically derived from a complex energy function—is available to judge the quality of generated samples. This presents a significant challenge: how can one train generative models effectively without direct supervision from data?

Meta AI Introduces Adjoint Sampling, a New Learning Algorithm Based on Scalar Rewards

Meta AI tackles this challenge with Adjoint Sampling, a novel learning algorithm designed for training generative models using only scalar reward signals. Built on the theoretical framework of stochastic optimal control (SOC), Adjoint Sampling reframes the training process as an optimization task over a controlled diffusion process. Unlike standard generative models, it does not require explicit data. Instead, it learns to generate high-quality samples by iteratively refining them using a reward function—often derived from physical or chemical energy models.

Adjoint Sampling excels in scenarios where only an unnormalized energy function is accessible. It produces samples that align with the target distribution defined by this energy, bypassing the need for corrective methods like importance sampling or MCMC, which are computationally intensive.

Source: https://arxiv.org/abs/2504.11713

Technical Details

The foundation of Adjoint Sampling is a stochastic differential equation (SDE) that models how sample trajectories evolve. The algorithm learns a control drift u(x,t)u(x, t)u(x,t) such that the final state of these trajectories approximates a desired distribution (e.g., Boltzmann). A key innovation is its use of Reciprocal Adjoint Matching (RAM)—a loss function that enables gradient-based updates using only the initial and final states of sample trajectories. This sidesteps the need to backpropagate through the entire diffusion path, greatly improving computational efficiency.

By sampling from a known base process and conditioning on terminal states, Adjoint Sampling constructs a replay buffer of samples and gradients, allowing multiple optimization steps per sample. This on-policy training method provides scalability unmatched by previous approaches, making it suitable for high-dimensional problems like molecular conformer generation.

Moreover, Adjoint Sampling supports geometric symmetries and periodic boundary conditions, enabling models to respect molecular invariances like rotation, translation, and torsion. These features are crucial for physically meaningful generative tasks in chemistry and physics.

Performance Insights and Benchmark Results

Adjoint Sampling achieves state-of-the-art results in both synthetic and real-world tasks. On synthetic benchmarks such as the Double-Well (DW-4), Lennard-Jones (LJ-13 and LJ-55) potentials, it significantly outperforms baselines like DDS and PIS, especially in energy efficiency. For example, where DDS and PIS require 1000 evaluations per gradient update, Adjoint Sampling only uses three, with similar or better performance in Wasserstein distance and effective sample size (ESS).

In a practical setting, the algorithm was evaluated on large-scale molecular conformer generation using the eSEN energy model trained on the SPICE-MACE-OFF dataset. Adjoint Sampling, especially its Cartesian variant with pretraining, achieved up to 96.4% recall and 0.60 Å mean RMSD, surpassing RDKit ETKDG—a widely used chemistry-based baseline—across all metrics. The method generalizes well to the GEOM-DRUGS dataset, showing substantial improvements in recall while maintaining competitive precision.

The algorithm’s ability to explore the configuration space broadly, aided by its stochastic initialization and reward-based learning, results in greater conformer diversity—critical for drug discovery and molecular design.

Conclusion: A Scalable Path Forward for Reward-Driven Generative Models

Adjoint Sampling represents a major step forward in generative modeling without data. By leveraging scalar reward signals and an efficient on-policy training method grounded in stochastic control, it enables scalable training of diffusion-based samplers with minimal energy evaluations. Its integration of geometric symmetries and its ability to generalize across diverse molecular structures position it as a foundational tool in computational chemistry and beyond.


Check out the Paper, Model on Hugging Face and GitHub Page. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 95k+ ML SubReddit and Subscribe to our Newsletter.

The post Sampling Without Data is Now Scalable: Meta AI Releases Adjoint Sampling for Reward-Driven Generative Modeling appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Adjoint Sampling 生成模型 Meta AI 奖励驱动学习
相关文章