| ### model | |
| model_name_or_path: Qwen/Qwen2-7B-Instruct | |
| new_special_tokens: <design_start>,<design_end>,<design_body>,<molecule>,<retro_start>,<retro_end>,<retro_body>,<rollback_start>,<rollback_end> | |
| graph_decoder_path: saves/graph_decoder | |
| graph_encoder_path: saves/graph_encoder | |
| graph_predictor_path: saves/graph_predictor | |
| adapter_name_or_path: saves/Qwen2-7B-Instruct-Adapter | |
| graph_lm_connector_path: saves/Qwen2-7B-Instruct-Adapter/connector | |
| ### generation | |
| max_length: 512 | |
| max_new_tokens: 128 | |
| temperature: 0.6 | |
| top_p: 0.9 | |
| ### method | |
| finetuning_type: lora | |
| do_train: false | |
| flash_attn: disabled | |
| learned_query_size: 8 | |
| ### dataset | |
| dataset: molqa_drug_examples | |
| template: qwen | |
| cutoff_len: 128 | |
| overwrite_cache: true | |
| preprocessing_num_workers: 16 | |
| output_dir: null | |
| bf16: true | |
| pure_bf16: true | |
| ### eval | |
| per_device_eval_batch_size: 6 | |
| report_to: 'none' |