Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
61 changes: 61 additions & 0 deletions src/inference_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -669,6 +669,67 @@ def inference(in_pdb_1, in_pdb_2):

return {"energy": min_energy.item()}

def inference_multiple_poses(in_pdb_1, in_pdb_2, num_samples=40, output_dir="output_poses"):
# Ensure output directory exists
os.makedirs(output_dir, exist_ok=True)

# Set device
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")

# Load ESM model
esm_model, alphabet = esm.pretrained.esm2_t33_650M_UR50D()
batch_converter = alphabet.get_batch_converter()
esm_model = esm_model.to(device).eval()

# Load score model
model = Score_Model.load_from_checkpoint(
str(Path("./checkpoints/dips/model_0.ckpt")),
map_location=device,
)
model.to(device).eval()

# Load PDBs
receptor = get_info_from_pdb(in_pdb_1)
ligand = get_info_from_pdb(in_pdb_2)

# Prepare inputs
inputs = {"receptor": receptor, "ligand": ligand}
batch = get_batch_from_inputs(inputs, batch_converter, esm_model, device)
batch = {k: v.to(device) if isinstance(v, torch.Tensor) else v for k, v in batch.items()}

# Define parameters
num_steps = 40
use_clash_force = False

all_outputs = []

# Run sampling
for i in range(num_samples):
rec_pos, lig_pos, rot_update, tr_update, outputs = Euler_Maruyama_sampler(
model=model,
batch=batch.copy(),
num_steps=num_steps,
device=device,
use_clash_force=use_clash_force,
)

lig_aa_coords = modify_aa_coords(ligand["aa_coords"], inputs["ligand"]["bb_coords"], rot_update, tr_update)
rec_structure = receptor["structure"]
lig_structure = ligand["structure"]
lig_structure.coord = lig_aa_coords

complex_structure = combine_atom_arrays(rec_structure, lig_structure)

file = PDBFile()
file.set_structure(complex_structure)
output_pdb_path = os.path.join(output_dir, f"output_pose_{i + 1}.pdb")
file.write(output_pdb_path)

all_outputs.append({"pose_id": i + 1, "energy": outputs["energy"].item(), "pdb_path": output_pdb_path})

return all_outputs


if __name__ == "__main__":
# Initialize the parser
parser = argparse.ArgumentParser(description="A description of what your program does")
Expand Down
13 changes: 13 additions & 0 deletions src/inference_multiple_poses.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
import argparse
from inference_base import inference_multiple_poses

def parse_args():
parser = argparse.ArgumentParser(description="Process two required PDB files.")
parser.add_argument("pdb_1", type=str, help="Path to the first PDB file")
parser.add_argument("pdb_2", type=str, help="Path to the second PDB file")
parser.add_argument("--num_samples", type=int, default=40, help="Number of output poses/samples, default=40")
return parser.parse_args()

if __name__ == "__main__":
args = parse_args()
inference_multiple_poses(args.pdb_1, args.pdb_2, args.num_samples)