Add model card for DiffGap
#1
by nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,9 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
# DiffGap
|
| 2 |
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
-
|
| 6 |
-
- ckpts: baseline checkpoints and ours
|
| 7 |
-
- metrics: the sampling and evaluation results
|
| 8 |
|
| 9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
pipeline_tag: other
|
| 3 |
+
---
|
| 4 |
+
|
| 5 |
# DiffGap
|
| 6 |
|
| 7 |
+
This repository contains the official checkpoints and metrics for **DiffGap**, as introduced in the paper [Bridging the Gap between Learning and Inference for Diffusion-Based Molecule Generation](https://huggingface.co/papers/2411.05472).
|
| 8 |
+
|
| 9 |
+
DiffGap is a diffusion-based framework that integrates adaptive sampling and pseudo-molecule estimation to bridge the gap between training objectives and inference dynamics in 3D molecule generation. By dynamically aligning intermediate denoising steps with realistic generation trajectories, DiffGap enables the diffusion model to adapt to input biases in advance during the training phase.
|
| 10 |
+
|
| 11 |
+
- **Repository:** [https://github.com/neusymlab/DiffGap](https://github.com/neusymlab/DiffGap)
|
| 12 |
+
- **Paper:** [https://huggingface.co/papers/2411.05472](https://huggingface.co/papers/2411.05472)
|
| 13 |
+
|
| 14 |
+
## Repository Structure
|
| 15 |
+
|
| 16 |
+
The repository includes:
|
| 17 |
+
- **ckpts**: Baseline checkpoints and DiffGap model weights.
|
| 18 |
+
- **metrics**: Sampling and evaluation results.
|
| 19 |
+
|
| 20 |
+
## Usage
|
| 21 |
+
|
| 22 |
+
The official implementation uses a `pipeline.py` script to wrap the processes of training, sampling, and evaluation.
|
| 23 |
+
|
| 24 |
+
```bash
|
| 25 |
+
# Basic command structure
|
| 26 |
+
python -m pipeline <configs> <sampling_results> [train|sample|eval] [-c resume_from_checkpoint_for_training]
|
| 27 |
+
|
| 28 |
+
# Example: Run the whole pipeline
|
| 29 |
+
# python -m pipeline configs/training.yml sampling_results/reproduce
|
| 30 |
+
|
| 31 |
+
# Example: Run sampling from a specific config
|
| 32 |
+
# python -m pipeline configs/sampling.yml sampling_results/reproduce sample
|
| 33 |
+
|
| 34 |
+
# Example: Run evaluation
|
| 35 |
+
# python -m pipeline "no matter" sampling_results/reproduce eval
|
| 36 |
+
```
|
| 37 |
+
|
| 38 |
+
For sampling on the PDBbind dataset with the DiffGap configuration:
|
| 39 |
+
```bash
|
| 40 |
+
python pipeline.py configs/gbd_pdbbind.yaml sampling_results/binddm_pdbbind sample -p PDBbind_refined_2020_test
|
| 41 |
+
```
|
| 42 |
|
| 43 |
+
## Citation
|
|
|
|
|
|
|
| 44 |
|
| 45 |
+
```bibtex
|
| 46 |
+
@misc{liu2024gapdiff,
|
| 47 |
+
title={Bridging the Gap between Learning and Inference for Diffusion-Based Molecule Generation},
|
| 48 |
+
author={Peidong Liu and Wenbo Zhang and Xue Zhe and Jiancheng Lv and Xianggen Liu},
|
| 49 |
+
year={2024},
|
| 50 |
+
eprint={2411.05472},
|
| 51 |
+
archivePrefix={arXiv},
|
| 52 |
+
primaryClass={cs.LG},
|
| 53 |
+
url={https://arxiv.org/abs/2411.05472},
|
| 54 |
+
}
|
| 55 |
+
```
|