Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DPO #31

Merged
merged 11 commits into from
Jun 5, 2024
Merged

DPO #31

Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
33 changes: 33 additions & 0 deletions scripts/dpo/config_denovo.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,33 @@
# Logging configuration
experiment_name: acegen
agent_name: dpo
log_dir: results # Directory to save the results
logger_backend: null # wandb, tensorboard, or null
seed: 101 # multiple seeds can be provided as a list to multiple experiments sequentially e.g. [101, 102, 103]

# Environment configuration
num_envs: 128 # Number of smiles to generate in parallel
total_smiles: 20_000 # Total number of smiles to generate

# Scoring function
molscore: MolOpt
molscore_include: ["Albuterol_similarity"]
custom_task: null # Requires molscore to be set to null

# Promptsmiles configuration
prompt: null # e.g. c1ccccc # Fix the beginning of the generated molecules

# Model architecture
model: gru # gru, lstm, or gpt2
# The default prior varies for each model. Refer to the README file in the root directory for more information.
# The default vocabulary varies for each prior. Refer to the README file in the root directory for more information.
custom_model_factory: null # Path to a custom model factory (e.g. my_module.create_model)

# Optimizer configuration
lr: 0.0001
eps: 1.0e-08
weight_decay: 0.0

# DPO configuration
beta: 0.1
num_updates: 3
36 changes: 36 additions & 0 deletions scripts/dpo/config_fragment.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,36 @@
# Logging configuration
experiment_name: acegen
agent_name: dpo
log_dir: results # Directory to save the results
logger_backend: null # wandb, tensorboard, or null
seed: 101 # multiple seeds can be provided as a list to multiple experiments sequentially e.g. [101, 102, 103]

# Environment configuration
num_envs: 128 # Number of smiles to generate in parallel
total_smiles: 20_000 # Total number of smiles to generate

# Scoring function
molscore: MolOpt
molscore_include: ["Celecoxxib_rediscovery"]
custom_task: null # Requires molscore to be set to null

# Promptsmiles configuration
promptsmiles: c1(C)ccc(*)cc1.NS(=O)(=O)(*)
promptsmiles_optimize: True
promptsmiles_shuffle: True
promptsmiles_multi: False

# Model architecture
model: gru # gru, lstm, or gpt2
# The default prior varies for each model. Refer to the README file in the root directory for more information.
# The default vocabulary varies for each prior. Refer to the README file in the root directory for more information.
custom_model_factory: null # Path to a custom model factory (e.g. my_module.create_model)

# Optimizer configuration
lr: 0.0001
eps: 1.0e-08
weight_decay: 0.0

# DPO configuration
beta: 0.1
num_updates: 1
36 changes: 36 additions & 0 deletions scripts/dpo/config_scaffold.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,36 @@
# Logging configuration
experiment_name: acegen
agent_name: dpo
log_dir: results # Directory to save the results
logger_backend: null # wandb, tensorboard, or null
seed: 101 # multiple seeds can be provided as a list to multiple experiments sequentially e.g. [101, 102, 103]

# Environment configuration
num_envs: 128 # Number of smiles to generate in parallel
total_smiles: 20_000 # Total number of smiles to generate

# Scoring function
molscore: LibINVENT_Exp1
molscore_include: ["DRD2_SelRF_SubFilt_DF"]
custom_task: null # Requires molscore to be set to null

# Promptsmiles configuration
promptsmiles: N1(*)CCN(CC1)CCCCN(*)
promptsmiles_optimize: True
promptsmiles_shuffle: True
promptsmiles_multi: False

# Model architecture
model: gru # gru, lstm, or gpt2
# The default prior varies for each model. Refer to the README file in the root directory for more information.
# The default vocabulary varies for each prior. Refer to the README file in the root directory for more information.
custom_model_factory: null # Path to a custom model factory (e.g. my_module.create_model)

# Optimizer configuration
lr: 0.0001
eps: 1.0e-08
weight_decay: 0.0

# DPO configuration
beta: 0.1
num_updates: 1
Loading