FastESMFold / config.json
lhallee's picture
Upload config
3bcee3d verified
{
"architectures": [
"FastEsmForProteinFolding"
],
"attention_probs_dropout_prob": 0.0,
"attn_backend": "sdpa",
"auto_map": {
"AutoConfig": "modeling_fast_esmfold.FastEsmFoldConfig",
"AutoModel": "modeling_fast_esmfold.FastEsmForProteinFolding"
},
"classifier_dropout": null,
"dtype": "float32",
"emb_layer_norm_before": false,
"esmfold_config": {
"bypass_lm": false,
"embed_aa": true,
"esm_ablate_pairwise": false,
"esm_ablate_sequence": false,
"esm_input_dropout": 0,
"esm_type": "esm2_3B",
"fp16_esm": false,
"lddt_head_hid_dim": 128,
"trunk": {
"chunk_size": null,
"cpu_grad_checkpoint": false,
"dropout": 0,
"layer_drop": 0,
"max_recycles": 4,
"num_blocks": 48,
"pairwise_head_width": 32,
"pairwise_state_dim": 128,
"position_bins": 32,
"sequence_head_width": 32,
"sequence_state_dim": 1024,
"structure_module": {
"dropout_rate": 0.1,
"epsilon": 1e-08,
"inf": 100000.0,
"ipa_dim": 16,
"num_angles": 7,
"num_blocks": 8,
"num_heads_ipa": 12,
"num_qk_points": 4,
"num_resnet_blocks": 2,
"num_transition_layers": 1,
"num_v_points": 8,
"pairwise_dim": 128,
"resnet_dim": 128,
"sequence_dim": 384,
"trans_scale_factor": 10
}
},
"use_esm_attn_map": false
},
"hidden_act": "gelu",
"hidden_dropout_prob": 0.0,
"hidden_size": 2560,
"initializer_range": 0.02,
"intermediate_size": 10240,
"is_folding_model": true,
"layer_norm_eps": 1e-05,
"mask_token_id": 32,
"max_position_embeddings": 1026,
"model_type": "fast_esmfold",
"num_attention_heads": 40,
"num_hidden_layers": 36,
"pad_token_id": 1,
"position_embedding_type": "rotary",
"token_dropout": true,
"transformers_version": "4.57.6",
"ttt_config": {
"ags": 4,
"batch_size": 4,
"lora_alpha": 32.0,
"lora_rank": 8,
"lr": 0.0004,
"mask_ratio": 0.15,
"steps": 30
},
"use_cache": true,
"vocab_list": [
"<cls>",
"<pad>",
"<eos>",
"<unk>",
"L",
"A",
"G",
"V",
"S",
"E",
"R",
"T",
"I",
"D",
"P",
"K",
"Q",
"N",
"F",
"Y",
"M",
"H",
"W",
"C",
"X",
"B",
"U",
"Z",
"O",
".",
"-",
"<null_1>",
"<mask>"
],
"vocab_size": 33
}