Practical Guide To Large Scale Docking
Practical Guide To Large Scale Docking
https://doi.org/10.1038/s41596-021-00597-z
used. Docking software described in the outlined protocol (DOCK3.7) is made freely available for academic research to
explore new hits for a range of targets.
Introduction
Screening chemical libraries using biophysical assays has long been the dominant approach to dis-
cover new chemotypes for chemical biology and drug discovery. High-throughput screening (HTS) of
libraries of 500,000 to 3 million molecules has been used since the 1990s1, and multiple drugs have
had their origins in this technique2. While the libraries physically screened in HTS were an enormous
expansion on those used by classical, pre-molecular pharmacology3, they nevertheless represent only
a tiny fraction of possible ‘drug-like’ molecules4. DNA-encoded libraries5, where molecules are
synthesized on DNA that encodes their chemistry, begin to address this problem by offering
investigators libraries of 108 molecules, sometimes more, in a single, highly compact format; and
multiple such libraries can be used in a single campaign. However, as DNA-encoded libraries are
restricted to reactions on DNA, reaction chemistries are limited to aqueous solutions, thereby limiting
the type of chemical reactions and subsequent chemical libraries available with this technology6.
Computational approaches using virtual libraries are an attractive way to explore a much larger
chemical space7. Large numbers of molecules—certainly into the tens of billions, and likely many
more—may be enumerated in a virtual library. Naturally, very few of these compounds can ever be
actually synthesized because of time, cost and storage limitations, but one can imagine a computa-
tional method to prioritize those that should be pursued. In practice, this idea has had two limitations
that have prevented wide-scale adoption: the virtual libraries have rarely been carefully curated for
true synthetic accessibility8, and there were well-founded concerns that computational methods, such
as molecular docking, were not accurate enough to prioritize true hits within this large space9.
In the last several years, however, two advances have at least partly addressed these problems:
First, several vendors and academic laboratories have introduced ‘make-on-demand’ libraries
based on relatively simple two- or three-component reactions where the final product is readily
purified in high yields10. At Enamine, a pioneer in this area, >140 reactions may be used to synthesize
products from among >120,000 distinct and often highly stereogenic building blocks, leading to a
1
Department of Pharmaceutical Chemistry, University of California–San Francisco, San Francisco, CA, USA. 2Science for Life Laboratory, Department of
Cell and Molecular Biology, Uppsala University, Uppsala, Sweden. 3NCI RAS Initiative, Cancer Research Technology Program, Frederick National
Laboratory for Cancer Research, Leidos Biomedical Research, Inc, Frederick, MD, USA. 4These authors contributed equally: Brian J. Bender,
Stefan Gahbauer. ✉e-mail: [email protected]
remarkably diverse and, critically, pragmatically accessible library of currently over 29 billion
molecules11.
Second, structure-based molecular docking, for all of its problems, has proven able to prioritize
among these ultra-large libraries, if not at the tens of billion molecule level, then at the 0.1–1.4 billion
molecule level12–14, finding unusually potent and selective molecules against several unrelated targets
(Table 1). Indeed, simulations and proof-of-concept experiments suggest that, at least for now, as the
libraries get bigger, docking results and experimental molecular efficacies improve12.
If docking ultra-large libraries brings new opportunities, it also brings new challenges. Docking
tests the fit of each library molecule in a protein binding site in a process that often involves sampling
hundreds-of-thousands to millions of possible configurations. Each molecule is scored for fit using
one of several different scoring functions15–18. To be feasible for a billion-molecule library on
moderately sized computer clusters (e.g., 500–1,000 cores), this calculation must consume not much
more than 1 s/molecule/core (1 ms/configuration).
This need for speed means that the calculation cannot afford the level of detail and number of
interaction terms that would be necessary to achieve chemical accuracy. For instance, docking
typically undersamples conformational states, ignores important terms (e.g., ligand strain) and
approximates terms that it does include (e.g., fixed potentials)19,20.
Owing to these approximations and neglected terms, docking energies have known errors, and the
method cannot even reliably rank order molecules from a large library screen21,22. What it can hope
to do, however, is separate a tiny fraction of plausible ligands from the much larger number of library
molecules that are unlikely to bind a target. This level of prioritization is enhanced with the careful
implementation of best-practice guidelines and controls. It is the goal of this essay to provide
investigators with such best practices and control calculations for ultra-large library docking, though
they equally apply to modest-sized library docking (Table 1). These will not ensure the success of a
prospective docking campaign—the only true test of the method—but they may eliminate some of the
more common reasons for failure.
We begin by describing protocols and controls that can be used across docking programs and that
are general to the field (Fig. 1). There are by now multiple widely used and effective docking
programs23–31, employing different strategies for sampling ligand orientations and ligand con-
formations in the protein site, for handling protein flexibility, and for scoring ligand fit once they have
been docked. Notwithstanding these differences, there are strategies and controls that may be used
across docking programs, including how to prepare protein sites for docking calculations, bench-
marking controls to investigate whether one can identify known ligands from among a large library,
and controls to investigate whether one’s calculations contain biases towards particular types of
interactions. Since the true success of a virtual screen is the experimental confirmation of docking
hits, we also propose a set of control assays to validate initial in vitro results. We then turn to
protocols that are specific to the docking program we use in our own laboratory, DOCK3.7—these
necessarily get into fine details, and will be of most interest to investigators wanting to use this
particular program.
Screen target Library size/type Best hit Docking rank Docking program Computing time Hit rate (hits/
molecules
tested)
D4 dopamine 138 million ZINC621433144 Top 0.07% DOCK3.7 4.2 × 104 core h 24% (58/238)
receptors12,37 (make-on-demand) EC50 = 180 pM (1 s/molecule)
N
HN
N
O
AmpC 99 million ZINC339204163 Top 0.00001% DOCK3.779 4.1 × 104 core h 11% (5/44)
β-lactamase12,119 (make-on-demand) Ki = 1.3 µM
O H
N OH
S
O O
OH
Cl
69,000 Compound 3 Not reported DOCK3.5120 Not reported 12.5% (2/16)
(HTS library) Ki = 37 µM
O O
OH
HO
OH
O O
MT1 Melatonin 150 million ZINC442850041 Top DOCK3.779 45,020 core h 39% (15/38)
receptor13 (make-on-demand) EC50 = 470 pM 0.005%
N
O
O N
14
KEAP1 1 billion iKeap1 Top 0.0001% QuickVina 2121 30 s/molecule 11.7% (69/
(make-on-demand) Kd = 114 nM (initial screen) 590)
Smina Vinardo122
O and AutoDock
N N O
S Vina23
O
(secondary
N N O screen)
O S O
Cl Table continued
Table 1 (continued)
Screen target Library size/type Best hit Docking rank Docking program Computing time Hit rate (hits/
molecules
tested)
HisG124 500,000 Compound 4 Within GOLD125 (initial ~60 h on 350 14% (7/50)
(commercially Ki = 2 µM top 0.2% screen) FLEXX24 networked PCs
available) (secondary
screen of
top 1%)
N
N
H
N S
N S
O N
Dipeptidyl 190,000 Compound 1 Within GLIDE25 (initial Not reported 15% (15/99)
peptidase IV (commercially IC50 = 5.77 µM top 0.05% screen)
(DPP4)126 available) AutoDock4 26
H (secondary
N screen of top 2%)
HO
N
N
N
O H
Docking rank indicates how highly ranked the best hit was among all virtually screened compounds. Hit rate denotes the fraction of docking hits confirmed in physical experiments.
2 723 K 8.8 M 32 M 48 M
cLogP
25
2.5 277 K 4.3 M 19 M 32 M
3 151 K 3.3 M 17 M 12 M
0
10–1 100 101 102 3.5 63 K 1.4 M 16 M 24 M
%decoys found
4 Virtual compound
Optimization and evaluation of
4.5 library
Concentration–response
virtual screening
curves
100 Compound 1 Docked compounds 109
Compound 2
80 Compound 3
Top 0.1% 105 Docking
60
Signal
program
Filter1
40
20 Filter2
Crystal structure
Cryo-EM structure 0 Hits 50–100
Homology model –9 –8 –7 –6 –5 –4 Large-scale
log(compound)
Hit-picking docking screen
Experimental tests
Fig. 1 | Large library docking workflow. The two required inputs for such a screen are the target structure and a screening database. Prior to using the
database, the target structure must be converted into a representation used by the docking software and the pocket should be optimized with control
calculations using retrospective analysis on known actives. After the prospective library has been docked, top-ranked hits can be filtered and selected
for experiment. Multiple assays and controls are typically necessary to confirm activity.
Modifying the high-resolution protein structure. It is not always a good idea to use the structure exactly
as it was found in the database.
● Dealing with mutations. For stability, crystallization and other biochemical reasons, high-resolution
protein structures are sometimes determined in a mutant form; such mutations should be reverted to
the wild type especially if they are within the ligand site to be targeted. Missing side chains and loops in
the experimental structures should be added as well if they are close to the binding site, while those that
are weakly defined by the experimental observables (e.g., low occupancy, high displacement parameters
(B values), poor electron density) should be examined critically
● What about water molecules? When the resolution permits, water molecules can also be included in
the target preparation, often treating them as nondisplaceable parts of the protein structure. Typically,
water molecules enclosed in the targeted binding pocket or involved in interactions between the
● Hydrogen atoms. Due to the resolution limits of most experimentally determined structures, hydrogen atoms
are often unresolved. The position of many hydrogen atoms can be readily modeled according to holonomic
constraints (e.g., backbone amide hydrogen atoms). For those that are not, such as hydroxyl protons on serine
and tyrosine residues, imidazole hydrogens on histidine, and the adjustment of frequently erroneous terminal
amide groups for glutamine and asparagine residues46, programs like Reduce47 (default in DOCK3.7), Maestro
(Schrödinger)48, PropKa49 or Chimera50 can be used to protonate the target of interest. While these automated
protocols usually produce reasonable protonation states, special care should be taken for the residues within
the ligand binding pocket or residues that form a catalytic/enzymatic site. Lastly, although less frequently
encountered, glutamic and aspartic acids can adopt protonated forms under specific circumstances51
● In summary, the protonation of the protein structure is critical for the success of docking to more accurately
depict the Van der Waals (VDW) surface and dipole moments of the binding pocket
Homology modeling
When no experimental structure has been determined for the target protein, structural models can be
generated if a template structure with high sequence identity is known. Common programs used for
homology modeling include Modeller52, Rosetta53, ICM27 and I-Tasser54. These two principles can
improve the chances of success:
● Typically, the higher the sequence identity between the target and the template, the better the accuracy
of the model55. Particular focus should be given to identity within the target binding pocket; if there is
a choice, choose the template that has the highest identity in the binding pocket
● Incorporation of a ligand during the modeling process or ligand-steered homology modeling
approaches56,57 will help prevent the pocket from collapsing inward, and will better orient the side
chains of binding residues32,58,59
When it is unknown how a ligand binds within the pocket, orthogonal experimental data can
guide the modeling such as iterating between docking and modeling60,61. In the case of MRGPRX262
and GPR6863, for instance, the authors predicted multiple binding poses of a known active ligand and
used mutagenesis and binding assays to test these predictions. A binding mode and receptor structure
were identified that was consistent with the mutagenesis data and used for subsequent preparation in
the prospective screen. Despite many difficulties, homology models have been successful in identi-
fying novel ligands from prospective docking campaigns41,62–76; though it is also true that, given the
choice, most investigators will prefer to use a well-determined experimental structure.
Control calculations
Docking undersamples ligand–protein configurations and conformations, and its scoring of these
configurations for fit remains highly approximate. Unlike methods like quantum mechanics, or
certain lattice calculations in statistical mechanics, docking has surrendered ‘ground truth’ to be able
to pragmatically search among a large and growing chemical space. Accordingly, control calculations
are critical to the success of a docking campaign. As with experimental controls, they do not ensure
prospective success, but they do guard against obvious sources of failure, and can help one under-
stand where things have gone wrong if they do. Through a key control, we assess whether the
prepared binding pocket and docking parameters can prioritize known ligands over presumed
inactive molecules. In an optimized binding pocket, these known actives should rank higher against a
background of decoy molecules in a retrospective screen, and reasonable poses should be predicted.
As it is more likely to know true actives than true inactives, it is common practice to use property-
matched decoys77, which are compounds that have similar physical properties as the actives, but
unrelated topologies, and so are presumed inactive. The DUDE-Z server (tldr.docking.org) was built
specifically to generate decoys for a given list of ligands by matching the following physical properties:
molecular weight, hydrophobicity (LogP), charge, the number of rotatable bonds, and the number of
hydrogen bond donors and acceptors78.
The performance of the parameterized binding pocket to enrich known ligands over decoys can be
evaluated by receiver-operator characteristic (ROC) curves, quantifying the true positive rate as a
function of false positive rate (Fig. 2)17,77,79–81. The area under the curve (AUC) is a well-regarded
75 75
%ligands found
%ligands found
50 50
Early enrichment
25 25
0 0
0 25 50 75 100 10–1 100 101 102
%decoys found
%decoys found (log)
Fig. 2 | Enrichment of actives against decoys. a, ROC curves for two models used for retrospective docking screens plotting the rate of true positives
found against decoys found. The AUC can be used to describe the ability of the models to identify true positive, known ligands against a background of
decoys. In this format, the two models have similar AUCs, suggesting similar performance. b, Semilogarithmic ROC curves focus on the early
enrichment, i.e., determine if true positives are identified within the e.g. top 10% (gray area) of docked decoys. The LogAUC is calculated as the
difference between the semilogarithmic AUC of the model and the random semilogarithmic AUC (dashed line). In this format, it is clear that model 2
outperforms model 1 in early enrichment with a LogAUC value more than double of model 1.
Fig. 3 | Control sets for retrospective docking calculations. For DUDE-Z decoys, properties of the decoys are either forced to match (green) or be
different (red). Properties that are neither selected for or against are highlighted in yellow. In the Extrema set, the charge state is explicitly sampled.
metric to monitor the performance of a virtual screen by a single number17 (Fig. 2a). The log
transformation of the false positive rate enhances the effect of early enrichment for true positives17
(Fig. 2b). This is important because, in a docking campaign with hundreds of millions of molecules,
only compounds ranked within the top 0.1% are often closely evaluated (see ranks in Table 1).
For example, if a retrospective docking challenge shows that known actives are only identified starting
around the tenth percentile, novel actives may be missed in a prospective screen. In this setting,
higher LogAUC values correspond to better discrimination between actives and inactives and provide
a sanity check on the ability of the docking parameters to identify actives.
A second criterion is the pose fidelity of the docked ligands to their experimental structures.
The validity of predicted binding poses can be assessed qualitatively by visual inspection of
reported key interactions between protein and ligand or, in the best case, quantitatively by calculating
root mean square deviations between predicted and experimentally determined poses82. During pocket
modeling and docking parameter optimization, one will often insist that the retrospective controls lead
to both high LogAUC values and good pose fidelity; often there will be some trade-off between the two.
While calibrating the scoring functions, it is also important to monitor the contributions of each energy
term to the total score and ensure they match the properties of the binding pocket. If one term
dominates, the scoring may have been overoptimized to that term, while other protein–ligand inter-
actions are underweighted, leading to dominance by a certain type of molecule. For instance, if the
docking score of a polar solvent-exposed pocket is inappropriately dominated by VDW energies, large
molecules may score high due to nonspecific surface contacts with the target protein.
In addition to property-matched decoys, other chemical matter can be used to evaluate different
aspects of the docking model (Fig. 3). A test set including molecules with extreme physical properties
Prospective screen
Once the docking model is calibrated, large libraries of molecules can be virtually screened against the
target protein. For this virtual screen, it makes sense to focus on compounds that are readily available for
testing. The ZINC20 database (http://zinc20.docking.org/) enumerates over 14 billion commercially
available chemical products, of which ~700 million are available with calculated 3D conformer libraries
ready for docking. Most of the enumerated compounds belong to the make-on-demand libraries of
Enamine and WuXi. Further, ZINC20 allows one to preselect subsets of molecules for docking, reducing
computation time. For instance, ZINC20 allows users to download ready-to-dock subsets of molecules
within user-defined ranges of molecular weight (MWT), LogP and net charge, as well as predefined
sets such as fragments (MWT ≤ 250 amu) or lead-like molecules (250 ≤ MWT < 350 amu; LogP ≤ 3.5).
The result of a prospective screen is a list of molecules rank-ordered by docking score.
Hit-picking
A well-controlled docking calculation can concentrate likely ligands among the top-ranked molecules.
But even if it was able to do so among the top 0.1% of the ranked library, in a screen of 1 billion
molecules this would still leave 1 million molecules to consider, and with the errors inherent in
docking, many of these will be false positives. Accordingly, we rarely pick the top N-ranked com-
pounds by docking to test experimentally, but rather will use additional filters to identify promising
hits within the top scoring 300,000–1,000,000 molecules. These filters can catch problematic features
missed by the primary docking function, ensure dissimilarity to known ligands and promote diversity
among the prioritized compounds.
Compounds may be filtered for both positive and negative features84 (Table 2). For instance, one
may insist that a docked orientation has favored interactions with key residues. Conversely, molecules
with strained conformations should be discarded. Molecules with unsatisfied, buried hydrogen-bond
donors and acceptors may also be deprioritized. Compounds with metabolic liabilities85,86 or that
closely resemble colloidal aggregators87 can also be filtered out, despite otherwise favorable scores.
Further, as closely related compounds will likely dock in similar poses with similar score, we typically
cluster compounds by 2D structure similarity after all other filters have been used and only select the
best scoring cluster representative for testing. In such a large dataset, clustering can be computa-
tionally expensive, so reduced scaffold clustering such as Bemis–Murcko88 analysis is useful to
efficiently parse the compounds. In principle, many of these filters could be included directly in the
scoring functions, but balancing them against other scoring terms can demand extensive optimization
and will likely increase the compute time, which will become a hindrance for ultra-large library
docking. Lastly, visual examination of the docked poses has been useful for selection of compounds to
purchase. Following the criteria in Table 2, we typically inspect up to 5,000 compounds visually after
applying automated filtering and clustering steps.
Docking score Docking scores are approximations of interaction energies and therefore cannot be used for absolute
ranking of hits. Its main purpose is to identify compounds that are likely to bind from the large
background of compounds that are unlikely to bind
Broken molecules Errors in 3D building can result in incorrectly built compounds. Common errors include improper
tautomerization or protonation states owing to the difficulty in predicting pKa. Such ‘broken’ molecules
should be deprioritized during visual inspection
Internal strain Often internal strain is not included in energy functions but is an important concern since a ligand may
score highly if it adopts a strained conformation that increases its contacts with the receptor (e.g.,
nonplanar amide). Visual examination can catch these strained compounds, but various tools have been
developed to systematically analyze hundreds or thousands of compounds making them useful in this
setting127–129. We recently proposed a computational tool that estimates torsion strain energies based on
experimentally determined torsional populations in the Cambridge Structural Dataset129,130
Interaction patterns Specific interactions to key residues are main features of promising candidates. Hydrogen bonds between
ligands and side chains can be computationally identified by simple distance cut-offs (<3.5 Å). For
example, hydrogen bonding to kinase hinge residues131, or salt bridges to the conserved aspartic acid
D3.32 in aminergic GPCRs132, offer reasonable anchor points for interaction filters. In a similar fashion, a
pharmacophore filter can be used to search for any atoms in an unexplored subpocket within the larger
binding pocket. Molecules that score well but form only one key interaction may be deprioritized.
‘Floating’ molecules that may bury an energetically frustrated water should also be eliminated
Unsatisfied hydrogen bond donors Molecules with unsatisfied hydrogen bond donors or acceptors, especially in hydrophobic pockets of the
and acceptors site, often pay a high desolvation cost; not all scoring functions are designed to filter these out explicitly.
Therefore, we suggest keeping the number of unsatisfied acceptors below 3 and the number of
unsatisfied donors to 1 or below, as burying a donor can incur a greater penalty than burying
an acceptor133
Novelty filter One of the advantages of large-scale docking with make-on-demand chemical libraries is the ability to
screen for novel chemistry. Therefore, if the goal of the project is to identify novel scaffolds for hits, it is
useful to filter out docking hits that resemble known actives. This can be accomplished using any number
of computational methods that compare the 2D or 3D topologies of hit compounds. In our hands, we
typically remove hits that have an ECFP4 Tanimoto cutoff of 0.35 or higher to any known active12,13
Scaffold clustering From the initial docking screen, a set of diverse scaffolds should be selected for experimental testing to
cover a larger subset of chemical space. Clustering metrics to identify diverse hits may include ECFP4134
fingerprints, Bemis-Murcko88 scaffolds, bcl::Cluster similarity135 or Atom Counts136
100
80
Signal
60
Primary screen
40
20
0
Compounds
Potential hits
Concentration-response
Off-target inhibition
curves
100 No detergent 100 Compound 1 22.0
With detergent Compound 2 20.0 NMR
NMR
80 80 Compound 3 18.0
Activity (%)
16.0
Hit validation
Signal
14.0
60 60 12.0
10.0
40 40 8.0
6.0
20 20 4.0
2.0
0.0
0 0
10.0 8.0 6.0 4.0 2.0 0.0
–9 –8 –7 –6 –5 –4 –9 –8 –7 –6 –5 –4 δ/ppm
log(compound) log(compound)
Confirmed hits
Analog-by-catalog
Affinity of purified stereoisomers Experimental high-resolution structure
focused SAR libraries
binding pose confirmation
Hit optimization and
Docking
Crystal structure
Fig. 4 | Suggested experimental validation of docking hits. In general, a primary screen will use a limited number of compound concentrations to test
for activity at a target. Compounds that pass a set threshold of activity in the primary screen will be moved forward to secondary confirmation of
activity that is not attributed to colloidal aggregation. Identity of the compound should be confirmed if it passes these stages and before proceeding to
optimization by stereoisomer purification, selection of analogs and/or experimental structure determination.
synthesis and downstream purification can result in mismatches between the in silico predicted
compound and the in vitro tested compound. For example, in the screen against the severe acute
respiratory syndrome coronavirus 2 (SARS-CoV-2) macrodomain protein, a number of purine
derivatives with an N9 substitution were requested but N3-substituted compounds were synthesized
instead83. Accordingly, it is worth fully characterizing promising compounds before costly lead
optimization is undertaken; we ourselves have made the mistake of designing analogs based on an in
silico starting scaffold that was different than the true hit delivered from the chemists. At a minimum,
the identity and purity of the compounds should be determined. High-resolution mass spectrometry
or quantitative 1D proton nuclear magnetic resonance spectroscopy can be used to detect gross
variation between the tested sample and the expected molecule; these data can be obtained from most
vendors, or the experiments could be performed independently.
The most direct experiment to test a docking pose prediction is the determination of an experi-
mental protein structure in complex with the docking hit (Fig. 4). Such structures illuminate crucial
Covalent Presence of an electrophile Incubation of ligand with the target typically increases activity. For soluble
Useful if sought, but if unintended (e.g., Michael acceptor, α- proteins, pre-incubate ligand and target at a concentration higher than the
are an artifact halo-ketone, activated nitrile) EC50, and then dilute to below the EC50. If irreversible, activity will reflect
the pre-incubation concentration. If reversible, ligand activity will diminish
over time
Redox cycling Almost always Several chemotypes (e.g., Test in the presence and absence of a reductant or radical trap, like
an artifact rhodanines) β-mercaptoenthanol or ascorbic acid
Spectroscopic interference Assay Multiple chemotypes, often Spectroscopic interference will change linearly with compound
disruption via light absorption or highly conjugated concentration, following Beer’s law, rather than sigmoidally as in a binding
fluorescence137–140, or inhibition isotherm. Molecules that might be inhibiting a reporter enzyme used in the
of a reporter enzyme, like assay (e.g., luciferase) may be tested for inhibition against it
luciferase141
Colloidal aggregation Likely the Multiple chemotypes. It is Run the assay in the presence and absence of a non-ionic detergent like
most common single mechanism difficult to predict which Triton X-100 (for soluble proteins) or Tween-80 (for cellular assays). Right
of artifact in early discovery compounds will aggregate. shifts of two- to threefold or greater in EC50 on detergent addition suggest
LogP > 3.5 and similarity to compound aggregation. Detergent solutions should be made fresh daily.
known aggregators can Control for the effect of the detergent on a well-behaved, nonaggregating
capture some of these, ligand, which should be unaffected93,142,143
but many will be missed Test for the formation of particles >30 nm in radii by dynamic light
(http://advisor.bkslab.org/) scattering92,144
For assays that will not tolerate detergent, an inert protein can be added to
prophylactically coat the colloids, protecting the assayed target from
adsorption. BSA is commonly used, though most inert soluble proteins will
work. The inert protein should be present in mg/mL concentrations92,95
Colloids are liquid droplets of pure compound, and can typically be pelleted
in a microfuge after a 20 min spin at low temperature. Draw off the
supernatant and test for activity—if it is much reduced, the compound is
likely an aggregator32,96
As with covalent inhibitors, pre-incubation typically increases the activity of
a colloidal aggregator. Unlike covalent inhibitors, incubation at higher
enzyme concentrations will typically slow the onset of aggregate-based
inhibition, eventually eliminating it almost entirely92
Counter-screen against unrelated off-target model enzymes, with and
without detergent. Common enzymes include malate dehydrogenase
(commercially available) and AmpC β-lactamase (freely available from
the authors)145
details of ligand recognition, including adjustments of binding site residues or the docking hit in the
binding pocket12,100,101. The key question to answer with high-resolution structures is whether
docking worked for the right reasons, i.e., does the predicted binding mode agree with an experi-
mental structure? Previous studies suggest that hits from virtual screening generally compare fairly
well with experimental high-resolution structures; i.e., key anchor points are predicted cor-
rectly12,83,102,103. However, as docking is typically performed against rigid protein structures, con-
formational changes, especially in flexible loops, will complicate pose prediction104. Water-mediated
protein–ligand interactions are generally difficult to explore de novo45,105, though experimentally
determined and structurally conserved water molecules can be included in virtual screens100. Lastly,
there are few cases where docking hits showing on-target activity revealed binding positions at
unexpected and nontargeted subsites. For example, the β2 adrenergic receptor allosteric modulator
AS408 was predicted to bind to an extracellular subpocket but crystallographically determined to bind
to the membrane-facing surface of the receptor106. In contrast, the pose of an inverse agonist also
identified at the β2 adrenergic receptor from in silico docking39 was confirmed by X-ray crystal-
lography107, demonstrating the variability in outcomes even at the same target.
If the protein target can be readily expressed and purified in milligram quantities, crystallography
and cryo electron microscopy can guide hit discovery early in the campaign. In campaigns against
targets where high-resolution structures are more challenging to obtain (e.g., for transmembrane
receptors), an experimental protein–ligand complex might only be achievable for the most potent hit
compound. Nonetheless, it is usually worth the effort to confirm the predicted binding site and ligand
pose, or to identify unexpected interactions between the discovered hit and the target106.
precalculated 3D conformer libraries (flexibases, db2 files)9,99. During conformer library building,
each molecule is divided into its rigid fragments, and different conformations of rigid fragment
substituents are generated.
Ligand poses are evaluated using a physics-based scoring function (Escore) combining VDW,
electrostatic (ES) and ligand desolvation (lig_desol) energy terms (Equation 1):
In order to allow rapid scoring of new poses (typically 1,000 poses per molecule per second),
contributions of the target protein pocket are mapped onto pregenerated scoring grids. The VDW
surface of the binding pocket is converted into a scoring grid by ChemGrid15. Electrostatic potentials
within the binding pocket are estimated by numerically solving the Poisson–Boltzmann equation
using QNIFFT16,110. Context-dependent desolvation energy scoring grids, both polar and apolar, are
generated by the Solvmap program17. Thereby, ligand desolvation energies are computed as the sum
of the atomic desolvation of each ligand atom scaled by the extent to which it is buried within the
binding site. Atomic desolvation energies for ligand atoms are calculated during ligand building using
AMSOL111 and are included in the ligand conformer library. Both, the electrostatic and ligand
desolvation scoring grids depend on the dielectric boundary between the low dielectric protein
(εr = 2) and high dielectric solvent (εr = 80) environments. Consequently, these scoring grids can be
fine-tuned by modulating the protein–solvent dielectric interface (see below Steps 41–44).
The following protocol (Fig. 5) prepares a starting structure (Steps 1–4), generates the scoring
grids and matching spheres (Steps 5–10), and collects a set of control ligands for model evaluation
(Steps 11–33). Control ligands are docked retrospectively (Steps 34–58) to determine the model’s
ability to identify actives from a pool of inactives. Optimization of sampling (Steps 59–64) and
scoring (Steps 65–77) is evaluated with the same controls. Model biases are evaluated using an
Extrema control set (Steps 78–88) to examine for charge preferences in the scoring grids and with a
small-scale in-stock screen (Steps 89–97) to ensure a computationally expensive large-scale pro-
spective screen (Steps 98–103) is likely to yield interesting chemical matter. Hit picking (Steps
104–108), though critical to success of these prospective experiments, is beyond the scope of the
protocol as it is highly user- and target-specific, and we suggest referring to Table 2 for insight.
Additionally, we caution against regarding hits from a screen as true positives until common artifacts
are ruled out (Box 1). All controls in silico and in vitro are applicable to any docking program or
computer-aided drug discovery campaign. The details of grid preparation and modification are
DOCK3.7-specific, but the principles are transferrable to other software.
Materials
Software
● DOCK3.7.5: apply for a license from http://dock.docking.org/Online_Licensing/dock_license_applica
tion.html. Licenses are free for nonprofit academic research. Once your application is approved, you
will be directed to a download for the source code. The code should run without issue on most Linux
environments, but can be optimized by recompiling with gfortran if needed. Questions related to
installation can be addressed to [email protected]
● Python: the current code uses python2.7, which will also need to be installed. Additional python
dependencies that are required for running scripts can be found in the file $DOCKBASE/install/
environ/python/requirements.txt
● AMSOL: free academic licenses can be obtained from https://comp.chem.umn.edu/amsol/
● (Optional) 3D ligand building software: if interested in 3D ligand building in-house (not necessary for
this protocol), licenses will also need to be obtained for ChemAxon (https://chemaxon.com/), OpenEye
Omega (https://www.eyesopen.com/omega) and Corina (https://www.mn-am.com/products/corina).
We note that, for many campaigns, 3D molecular structures with all necessary physical properties may
be downloaded directly from ZINC20. This tutorial makes use of a webserver for 3D ligand building
that is suitable for small control sets. Please apply for an account at https://tldr.docking.org/, which is
free of charge
● Chimera: this application is recommended for grid visualization with the VolumeViewer feature (see
Troubleshooting). The ViewDock feature, also within Chimera, is useful to examine the docking results
Extrema set
In-stock set
Test post-docking filters
Known inactives
Zinc database
Steps 104–108
Table 2 Hit picking
Post-docking filters
Visual hit picking
Experimental testing
Fig. 5 | Outline of the procedure for DOCK3.7 virtual ligand discovery campaigns. Collecting and preparing materials (blue) requires obtaining a
structure or model and ligand control sets and setting them up for retrospective control calculations (yellow). In each control calculation, modifications
may demand returning to a previous step and reoptimizing. In the absence of known actives for robust retrospective analysis, one may jump to testing
the prospective performance with a small library. With a final setup, large-scale prospective screening (orange) can proceed, followed by in vitro
testing of docking hits (green). The numbers refer to steps described in the Procedure.
Equipment
● Hardware: initial tests of this tutorial (grid preparation and small retrospective screens) can be
performed on a single workstation, but more intensive docking will need access to a high-performance
computing cluster (we regularly use 1,000 cores)
● Queuing system: DOCK3.7 comes with submission scripts for SGE, Slurm and PBS job schedulers.
If using a different scheduler, scripts may need to be adapted for your specific computing cluster
● Protein structure: structures can be downloaded from www.rcsb.org or generated by the user in the
docking grids and optimized docking grids can be downloaded from http://files.docking.org/dock/
mt1_protocol.tar.gz. The example dataset uses the MT1 structure (PDB: 6ME3) co-crystallized with
2-phenylmelatonin
Equipment setup
Define the pathway to critical software
The environment will need to be defined in order to run the following commands correctly. The two
most important paths that need to be defined are to DOCKBASE and python. We provide an envir-
onment script in the example dataset that will need to be modified for a user’s settings. Always run the
following command with updated paths specific to your file directory before working with DOCK3.7:
source env.csh
Procedure
Section 1: set up binding pocket for docking calculations ● Timing 30 min to 1 h
1 Download the structure from the PDB or use in-house structures. Preference should be given to
structures of high resolution and with a ligand bound in the site that will be docked into. Apo
structures are useable, but tend to yield poorer results due to unconstrained binding site geometries32.
2 Visualize the binding pocket in PyMOL or Chimera to decide on the relevance of protein cofactors.
● Delete components from the structure that do not contribute to ligand binding such as lipids,
removed, and the resulting chain break should be either capped or the loop remodeled
● Protein cofactors such as heme or ions should also be kept or discarded depending on their
decisions. For the example MT1, we deleted the fusion protein, capped non-native termini and
reverted two mutations in the binding pocket (6me3_cleaned.pdb)
3 Save the rec.pdb file. This file will contain any component that will remain static during the
docking such as structural waters and cofactors.
4 Save the xtal-lig.pdb file. This file will contain the atoms of the bound ligand, which will be
used to generate the matching spheres that guide ligand sampling in the pocket.
5 Generate the scoring grids and matching spheres with blastermaster from these two inputs
(rec.pdb and xtal-lig.pdb)
$DOCKBASE/proteins/blastermaster/blastermaster.py --addhOptions="
-HIS -FLIPs " –v
? TROUBLESHOOTING
The outputs are new rec.pdb and xtal-lig.pdb files, as well as a rec.crg.pdb file inside
a working directory. HID, HIE and HIP naming is generated automatically from the protonation
state of the HIS residues and again should be checked for accuracy.
Note: this script requires a path to the Chimera binary file and may need to be modified by the user.
9 Generate scoring grids and matching spheres from the new files (rec.pdb, xtal-lig.pdb, and
working/rec.crg.pdb), and examine the outputs as in Step 6:
$DOCKBASE/proteins/blastermaster/blastermaster.py --addNOhydrogensflag –v
Section 2: collect and build the control ligand set ● Timing Minutes to hours for ligand
building, depending on the number of molecules
11 Collect the known actives (positive controls) for retrospective analysis. For a given target, these can
be found in the scientific literature, patent literature or public databases such as IUPHAR/BPS114,
ChEMBL115 or ZINC9,99,116, or available in-house.
same chemical series. For a rigorous control analysis, redundant (i.e., highly similar) compounds should
be clustered and the most potent compound selected.
12 Sort all knowns by potency.
13 Cluster these based on 2D similarity using any preferred method. We suggest calculating clusters
based on ECFP4 Tanimoto similarities with a cutoff of 0.35.
14 Use the most potent compound from each cluster as the representative of that scaffold. It is best to
refine these actives to a set that is representative of the chemical space intended for prospective
screening (i.e., with limits on molecular weight) so that the retrospective analysis will match the
prospective aims.
15 Find out whether the known ligands of the target are neutral or charged as one criteria of the
docking parameter calibration is the ability to score ligands with corresponding charges well.
16 Save a final list of actives’ SMILES as ligands.smi. Typically, 10–30 diverse actives represent a
good control set. For targets with less than this value, the controls are still useful but may not be as
informative. For the example of MT1 we extracted a set of 28 agonist and antagonists from the
IUPHAR/BPS database13.
23 The results are in a build3d_results/ directory with two subdirectories failed/ and
finished/ indicating the build status of the compounds. In example data provided, all ligands
were built to completion.
24 To build a split database index file, or path file to the ligands, use the following command:
33 Collect output files. At this point, four files should be generated: actives.sdi, actives.id,
decoys.sdi and decoys.id.
this number of heavy atoms it will not be docked. For retrospective calculations, we want all
ligands to be docked and scored regardless of size, so we use a large value here
● Set the mol2_maximum_cutoff to 100. This value is a cutoff for saving 3D coordinates of
docked poses. If a ligand scores worse (i.e., more positive) than this cutoff, the pose will not be
saved. Setting a low value is useful in large-scale prospective screens to save on disk space and
computation time, but for retrospective screens we want all information saved for analysis
37 Set up the docking directory
$DOCKBASE/docking/setup/setup_db2_zinc15_file_number.py ./ controls
controls.sdi 1 count
This script will separate the controls.sdi file into one directory called controls0000. For
this first calculation, the size of the sdi file is manageable on a single core, though splitting it will be
faster (i.e., 10). For control calculations with 10,000–100,000 ligands and for large-scale prospective
screens, the sdi file should be split into multiple directories and the jobs distributed over multiple
cores. A dirlist file is generated that lists all of the split directories prepared
for docking.
38 Dock the control set. At this point, it is possible to submit the data to a computer cluster or do the
calculation on a single computer.
(A) Docking without submitting to a cluster
(i) Move into the controls0000 directory. You will find INDOCK (pointing to the
dockfiles/ directory) and a split_database_index file. These are all the inputs
that the DOCK program needs to run the calculation.
(ii) Run the docking calculation:
$DOCKBASE/docking/DOCK/bin/dock.csh
? TROUBLESHOOTING
(iii) The output files from the docking program are OUTDOCK, listing docking scores of
successfully docked molecules, computational performance or potential error messages,
and a zipped mol2 file (test.mol2.gz) containing the 3D poses of docked molecules.
(iv) Move back into the directory that contains dirlist when done.
(B) Docking with submitting to a cluster
(i) In the directory that contains the dirlist (created in the previous step), run one of the
following commands depending on cluster architecture:
● SGE: $DOCKBASE/docking/submit/submit.csh
● Slurm: $DOCKBASE/docking/submit/submit_slurm.csh
● PBS: $DOCKBASE/docking/submit/submit_pbs.csh
CRITICAL STEP If you use another cluster scheduler, adopt these scripts to your needs.
c
39 When the job has completed, the last line of the OUTDOCK file should contain an elapsed time. If
this is not the last line of the OUTDOCK file, an error has occurred (see Troubleshooting) and the
following script will not run.
? TROUBLESHOOTING
40 Extract the scores of all docked compounds
The arguments are the dirlist, the name of the file to be written (extract_all.txt) and
the max energy to be kept (this value should match the mol2_maximum_cutoff value in the
INDOCK file). The important output file for future analysis is the extract_all.sort.uniq.
txt file, containing the rank-ordered list of docked compounds with only the highest score for
each compound.
41 Get the poses of the docked compounds
The arguments are the path where the docking is located (./), the name of the extract_all.
sort.uniq.txt file, the number of poses to get (10000, i.e., set to larger than the number of
compounds docked for retrospective calculations since we want to get all), the name of the output
mol2 file (poses.mol2), and the name of the input mol2 files (test.mol2.gz), containing
3D coordinates of predicted poses from the docking calculation (located in the
controls0000 directory).
42 Get the poses of just the actives
The arguments are the file containing active IDs, the pose file containing all of the compounds both
actives and decoys, and the name of the output file to be written.
43 Calculate the LogAUC early enrichment
Inputs are the working directory (‘.’ if in the directory where the extract_all file is located) and
the two files containing the IDs of actives and decoys. The output is a roc.txt and roc_own.
txt file. Both files report an AUC and LogAUC and contain the grid points for plotting receiver
operator curve (ROC) plots. The roc.txt file is calculated over all inputs in the ID files, and the
roc_own.txt file is calculated only over the compounds that successfully docked.
44 To generate a plot (roc_own.png) of the roc_own.txt file:
Inputs are the poses files and the name of the output file to be written. The output charges file is
a list of DOCK scores and the charge of the ligand with that score.
46 To generate a plot of total DOCK score by ligand charge (charge_distributions_vs_
energy.png) run:
47 Plot the contribution of each score term in the energy function outlined in Equation 1
(energy_distributions.png):
48 Collate the output files. The important outputs of these steps are: actives.mol2, roc_own.
png, energy_distributions.png and charge_distributions_vs_energy.png.
%ligands found
Gln181
50
25
0
10–1 100 101 102
%decoys found (log)
–20
–30
Score
Score
–20 –40
–50
Default extrema
Optimized extrema
–40
Fig. 6 | Controls for docking optimization. a, The receptor (blue) is shown with the crystallized ligand (orange).
Docked control actives are shown in green and yield similar poses and interactions as the crystal ligand. The two
residues, Asn162 and Gln181, that have their dipoles artificially increased (‘polarized’) to enhance the weight of polar
interactions are shown hydrogen bonding to the crystal ligand. b, A log-transformed ROC plot is shown comparing
the rate of identifying ligands versus decoys. A random selection would follow the dashed black line. The area under
this dashed line is subtracted from the values reported for LogAUC such that a curve above the line would have a
positive LogAUC, a curve below the line would have a negative LogAUC, and a curve following the dashed line would
yield a LogAUC value of zero. Shown are the curves for the default settings and optimized settings for either the
DUDE-Z control set and the Extrema control set. In both cases, the overall LogAUC value increases and the early
enrichment improves. c, The energy distribution breakdown shows the individual score terms for each scored
molecule in the docked setup. Based on this breakdown, it is clear that VDW interactions primarily drive ligand
recognition. However, in the optimized setup in which electrostatic spheres with a radius of 1.9 that extend the
dielectric boundary are used, the electrostatic score term shifts to more negative values. The desolvation spheres at
the dielectric boundary in the optimized setup, with a radius of 0.1, have only minor effects on the ligand desolvation
score term. d, In the Extrema challenge, the top-ranking ligands are plotted by their charge and DOCK score. In the
Default settings, there is a preference for neutral ligands followed closely by monocations. The Optimized settings
enhance the preference for neutrals.
53 Under the ‘Column’ tab, several details about the docked ligands can be listed, e.g., Total Energy,
Electrostatic or Van der Waals terms.
54 In the Tools tab, under Structure/Binding Analysis, ‘Find HBonds’ can be used to identify hydrogen
bonds between the protein and ligand molecules.
55 Evaluate the docked poses, asking the following questions:
● Do the ligands occupy the part of the pocket as expected?
● Do they make the types of interactions anticipated from what is known about the ligands and
the pocket?
● Are there aberrant or unsatisfied interactions? For example, in MT , it is known from the crystal
1
structure that ligands can form hydrogen bonds with Asn162 and Gln181. Issues with binding
poses are usually a sampling problem, but scoring terms such as electrostatics can influence
specific interactions
● If many actives do not dock, is there a reasonable explanation (e.g., selected controls are larger than
the pocket volume as has been observed when attempting to dock antagonists into agonist pockets)?
Optimization of sampling is performed with a matching sphere scan as detailed in Steps 59–64
python $DOCKBASE/proteins/optimization/scramble-matching-spheres.py
-i $PWD/dockfiles -o $PWD -s 0.5 -n 50
The arguments are: -i, the path to the dockfiles/ directory; -o, the path to where new docking
directories should be created; -s, the maximum distance to move a matching sphere
(0.5 recommended); -n, the number of perturbed sphere combinations to create. For a single
receptor target, we recommend 50–100 sets of sphere sets.
61 Each new directory should contain an INDOCK file and dockfiles/ directory. Check that the
matching_spheres.sph file in the dockfiles/ directory is unique to each new directory created.
Fig. 7 | Matching and dielectric boundary spheres drive changes to sampling and scoring in DOCK3.7. a, the crystal
ligand is shown as orange sticks in the receptor pocket (gray). Matching spheres derived from the coordinates of the
crystal ligand are shown in yellow and remain fixed during sphere perturbation. Random spheres (blue) are
calculated with the program SphGen, and a set of spheres are selected that are near the crystal ligand. In a matching
sphere scan, only the random spheres are perturbed and a new set is obtained (green). b, the crystal ligand (orange)
is again shown in the context of the receptor binding pocket (gray). Dielectric boundary spheres (cyan) cover the
binding surface around the crystal ligand to alter the electrostatic or desolvation potentials at the boundary between
solvent and protein.
$DOCKBASE/proteins/optimization/boundary-sphere-scan.sh -b $PWD -v
This version of the script will run on an SGE cluster. Modify the submission script and scheduler as
needed for different cluster architecture. To run on a local machine, use the script:
$DOCKBASE/proteins/optimization/boundary-sphere-scan_no_cluster.sh
-b $PWD -v
This script generates a set of combinations of boundary modifying spheres across both ligand
desolvation and low dielectric spheres. These directories have everything needed to run individual
CB CB
CG CG HD21
HD21
0D1
ND2 ND2
0D1
HD22
HD22
Fig. 8 | Polarizing effects specific atoms’ electrostatic potential. In contrast to global modifications to the electrostatic potential with the
incorporation of thin spheres, polarizing allows for very specific modifications to a residue’s charge status. A canonical asparagine (ASN) from the
prot.table.ambcrg.ambH file is shown with its polarized version ASM in which the carbonyl becomes more electronegative while the amide
hydrogens become more electropositive to maintain the overall charge. The electrostatic potential corresponding to each atom is shown as spheres,
with red corresponding to negative charge and blue corresponding to positive charge.
retrospective calculations except for the controls.sdi file that will need to be copied into each
one. Repeat the retrospective calculations from Steps 37–48.
68 Following retrospective calculations across the combinations of dielectric boundary sphere radii,
select a combination that ideally improves LogAUC (Fig. 6b) and helps to balance the score
distributions in the energy_distribution.png plot (Fig. 6c). If there are a number of
equally performing combinations, subsequent control calculations can help identify the best set for
your system (see Steps 78–97). In the case of MT1, we chose electrostatic spheres of radius 1.9 and
desolvation spheres of radius 0.1 as they increased LogAUC and enhanced the charge distribution
of compounds to favor neutrals to match known MT1 ligands (Fig. 6d).
necessary to modify the partial charges of a particular residue. For example, in the case of MT1, the
docking scores are largely dominated by VDW interactions (Fig. 6c). However, two residues Gln181 and
Asn162 form hydrogen bonds with the known ligands. For these specific interactions, a global
modification to the dielectric boundary (Steps 65–68) may not be sufficient and, instead, a local
modification to partial charges as outlined in Steps 69–77 may enhance favorable scores for these
interactions. For targets where this is not necessary, proceed to Step 78.
69 Make a new directory for polarizing residues. Copy over rec.pdb, xtal-lig.pdb, working/
rec.crg.pdb, working/amb.crg.oxt and working/prot.table.ambcrg.ambH
from the previous docking setup.
70 In the rec.crg.pdb file, rename the residues to be polarized with a unique three-letter amino
acid code. For example, in the MT1 receptor, we chose to polarize GLN181 and ASN162 to
enhance the polar interactions between actives and these hydrogen bonding residues. Accordingly,
we renamed GLN181 as GLD and ASN162 as ASM (see Fig. 8).
c
ambcrg.ambH files located withing the working/ directory. These files contain default partial
charges for all amino acids and some special residues such as ions. Make the following changes in both
amb.crg.oxt and prot.table.ambcrg.ambH files.
71 Create new sections for the polarized residues in the two files by copying the standard residue
partial charges.
72 Rename the polarized residues to match the three-letter codes used in the rec.crg.pdb file. The
names are all capitalized in the prot.table.ambcrg.ambH file as shown in Fig. 8 and all
lowercase in the amb.crg.oxt file.
73 Redistribute the charge around the atom of interest, making sure that the net charge remains the
same. We suggest testing modifications of 0.2 or 0.4 charge units for a given atom. For example, for
an ASN-to-ASM change, where we want to enhance the electronegativity of the sidechain carbonyl
(OD1), increase the charge by 0.4 units, resulting in a change from −0.470 to −0.870. As −0.4
charge was added to the residue, +0.4 must be distributed to other atoms in the residue to maintain
the net charge. One option is to add +0.2 to each of the sidechain amide hydrogens (HD21 and
HD22) as in Fig. 8, though the charge could have been distributed to the backbone amide or
another atom not involved in binding.
$DOCKBASE/proteins/blastermaster/blastermaster.py --addNOhydrogensflag
--chargeFile=amb.crg.oxt --vdwprottable=prot.table.ambcrg.ambH –v
? TROUBLESHOOTING
76 Rerun control calculations as in Step 37–48. When examining the new poses, consider the following questions:
● Does the polarized residue promote or discourage the desired interaction?
77 Rerun the dielectric boundary modifying sphere scan in Steps 65–68. We recommend doing this
because the new polarized residue(s) will have altered the electrostatics grid. It may be necessary to
test different combinations and strengths (0.2, 0.4) of the polarized residues to identify the best-
performing set of parameters that enhances correct interactions, improves poses and ideally
improves LogAUC. However, due to the modification to the electrostatic potential introduced by
these changes, it is important to ensure these improvements do not come at the cost of biasing the
screen towards overly charged molecules. This bias can be checked in the next section Steps 78–88.
Section 7: Extrema charge control calculations ● Timing 2–3 h for compound building
Generate the Extrema set
CRITICAL Compounds in the extrema set come from the area of chemical space occupied by the
c
known actives (molecular weight and LogP) but distinct charges (−2, −1, 0, +1, +2). These decoys will
be used to measure biases introduced into the scoring function owing to modifications to the
electrostatic potential.
78 On tldr.docking.org, navigate to the ‘extrema’ application.
79 Upload the SMILES of your active ligands used in section II to the webpage.
80 Select 500 or 1,000 molecules per charge, and click the ‘Go’ button.
81 After a few minutes, a results file will be available to download and extract. Within the file should be
five SMILES files grouped by charge. Extract the SMILES file from the download:
unzip extrema_results.zip
84 When all sets are built and downloaded, combine the path to all compounds in a split database
index file along with the known actives.
$DOCKBASE/docking/setup/setup_zinc15_file_number.py ./ extrema
extrema.sdi 100 count
$DOCKBASE/docking/setup/setup_zinc15_file_number.py ./ instock
instock.sdi 1000 count
If there are any issues, determine if it is a sampling or scoring problem and alter the previous steps
(matching spheres or dielectric boundary modifying spheres, respectively) to optimize. If the top
poses capture expected interactions, then it is worth moving into a large-scale prospective screen.
97 Choose settings for the large-scale docking experiment. If multiple docking parameters were tested
in this section (i.e., section 8), one method for selecting the best settings for large-scale docking is to
choose the parameters that yield the greatest number of viable compounds after all of the post-
docking filters have been applied. This will ensure a large number of hits in the large-scale
campaign. Alternatively, one can purchase hits from the different screens and see which setup leads
to more true actives in an experimental setting. Even if just a few low-potency hits are obtained at
this point, it may suggest which interactions are critical for activity in the pocket.
Toggle buttons to
select/deselect
columns/rows
Total count of
selected compounds
Fig. 9 | Navigating the ZINC20 Tranche Viewer. Several options are available at http://zinc20.docking.org/tranches/home/ for selecting different
subsets of ligands for virtual screening. Important criteria such as selecting between 2D/3D, purchasability, charge, molecular weight and logP are
highlighted. To download compounds, different methods such as downloading as an index file or directly downloading with cURL and
WGET are shown.
binding pocket117. To obtain a split database index (sdi) file for a predownloaded ZINC20 database,
use the ‘DOCK Database Index’ download format in the tranche viewer (Fig. 9) and provide a path
to the downloaded database in the ‘ZINC DB Root’ field.
102 Split the docking campaign over 1,000 or more jobs. Typically, ~100 million compounds can be
screened per day on an academic 1,000 core cluster of recent vintage (as of 2020).
$DOCKBASE/docking/setup/setup_zinc15_file_number.py ./ largescale
docking zinc20library.sdi 1000 count
105 Collect between 300,000 and 1,000,000 poses from the best-scoring compounds.
106 Use any number of post-docking filters including but not limited to those described in Table 2.
107 From the compounds that remain after filtering, visually examine up to 5,000 molecules depending
on how viable the docked compounds look. We recommend that visual examination is done by
more than one person, because of the different knowledge base that each individual brings to the
selection process (i.e., a medicinal chemist will prioritize different features over a molecular
biologist and vice versa)19. This step should be done before prioritizing compounds for purchase, as
we have found human-picked compounds have better efficacy than compounds obtained from fully
automated hit-picking12. By using post-docking filters as in Step 106 prior to visual examination,
one can search more deeply through the list of top ranked molecules to identify molecules
for testing.
108 Choose 50–200 compounds for wet-lab testing depending on price, capability of testing and
confidence of success as informed from retrospective calculations.
Troubleshooting
Structure preparation
Alternative conformations or missing side chains should be completely modeled before starting the
grid generation (blastermaster.py). Otherwise, any superfluous or missing atoms will result in
erroneous VDW surface or partial charge calculations (see below).
Blastermaster
Section 1, Step 5
All input and output files as well as options and flags of the blastermaster program are listed in the
Blastermaster Guide (see Supplementary Information). Of important note is that blastermaster
requires protein and ligand structures to be provided as PDB files called rec.pdb (and working/
rec.crg.pdb in case a protonated structure is used) and xtal-lig.pdb. Alternative file names
will not be recognized in the default settings.
Section 1, Step 10
If blastermaster does not successfully finish the grid generation, log files in the working directory
will yield the required information to backtrace the error (see Blastermaster Guide in Supplementary
Information). We suggest running blastermaster in the verbose mode (-v) as it allows to easily detect
at which step the program failed.
In our experience, most protein input structures can be converted into complete docking grids with
the default parameters provided in DOCKBASE. For some modifications such as capped termini or
structural waters, parameters are included in the provided files (e.g., amb.crg.oxt) but may use
slightly different atom names compared with default names from modeling programs such as Chi-
mera, PyMol or Maestro. In these cases, we suggest adapting the corresponding atom names in the
input protein structure files (rec.pdb, rec.crg.pdb) rather than adding more atom types to
the default parameter files. Common naming errors occur for disulfide bonded cysteines (CYX in
prot.table.ambcrg.ambH) and water atoms (HOH, TIP, WAT and SPC in prot.table.
ambcrg.ambH).
The qnifft program may crash if too many atoms (>50,000) are present in rec.crg.pdb. While
we do rarely encounter this problem, particularly large proteins or other molecular complexes may
have to be reduced in atom number (e.g., removing fusion proteins or distal monomers). Typically,
protein segments >20 Å away from the binding pocket are unlikely to influence the resulting
docking grids.
In certain cases, blastermaster may be able to calculate all grids, even though some parameters
might be missing. For example, if nonstandard amino acids are used, the protonation state as well as
partial charges or VDW surfaces may not be computed correctly, but will not cause blastermaster
to crash.
To check if all force field parameters were assigned to the protein structure correctly, we suggest
visual inspection of docking grids (Chimera). The scoring grids (vdw.vdw, trim.electro-
statics.phi, as well as ligand.desolv.heavy and ligand.desolv.hydrogen) are
found in the dockfiles directory. The following scripts can be run in the dockfiles directory to
convert the grids into dx files for visualization.
Convert VDW grid:
python $DOCKBASE/proteins/grid_visualization/create_VDW_DX.py
Input: trim.electrostatics.phi
Output: trim.electrostatics.dx
Convert desolvation grid:
python $DOCKBASE/proteins/grid_visualization/create_LigDeSolv_DX.py
To visualize the grids, first open the rec.crg.pdb file in Chimera. The dx files can be opened with
the Volume Viewer application. The vdw.dx file should resemble the surface of the receptor. The
ligdesolv.dx file demonstrates various solvation levels of the pocket (the grid itself is a
continuum). By changing the level of the representation, the volume should fill the pocket illustrating
the ligand moving from a less solvated state to the most solvated state. For the trim.electro-
statics.dx file, it is best to set the minimum and maximum levels to −100 and 100, respectively.
Then, set the colors to red for −100 and blue for 100. This now shows regions of positive partial
charges (blue) and negative partial charges (red). For all grids, ensure that the representations match
what is expected for the binding pocket and that there are no regions of missing grid points. If there
are obvious holes in any of the grids, in close proximity to the ligand, certain residue types were not
identified correctly and corresponding parameters could not be assigned.
A list of all partial charges assigned to rec.crg.pdb is given in working/qnifft.atm
(column 11). If the total charge of the system does not sum up to an integer value, certain partial
charges may not have been found in $DOCKBASE/proteins/defaults/amb.crg.oxt.
Furthermore, if many partial charges for a given residue (or other structural components such as
metals or water molecules) are 0, proper parameters were likely not assigned. Specific errors in partial
charge assignment can be found in working/OUTPARM.
Docking
Section 3, Step 38A(ii)
The DOCK executable ($DOCKBASE/docking/DOCK/bin/dock64) will only read a file called
INDOCK.
Section 4, Step 64
If ligands dock in poses far from the binding pocket, visualize the matching spheres to ensure they
occupy the same area as the xtal-lig. The matching_spheres.sph file in the dockfiles
directory contains the set of matching spheres. To view this file in a protein visualization application,
they first must be converted to a pdb file.
Input: matching_spheres.sph
Output: matching_spheres.pdb
The output file, matching_spheres.pdb, can be visualized with the receptor and ligand in either
Chimera or PyMOL. The matching spheres should align to the ligand’s heavy atoms, and additional
random spheres should be scattered around the ligand as in Fig. 7a.
If many ligands fail to dock (>50%), a number of issues may be at fault. It is common for 10–20%
of ligands to fail to dock in retrospective calculations because of mismatches between the pocket and
ligand properties. In the OUTDOCK file, two common statements may indicate these issues: ‘Grids
too small’ and ‘Skip size’. The ‘Grids too small’ statement indicates that the ligands do
not fit in the binding pocket. The ‘Skip size’ statement is used when the number of atoms in the
ligand exceeds the atom_maximum value in the INDOCK file. Adjusting this number to a higher
value will ensure that larger ligands will be docked and scored. If the OUTDOCK file shows many
compounds getting scored but the poses are not being saved, the mol2_maximum_cutoff value
may be too stringent. Relaxing this term will save more poses, but be careful to not go too high as the
size of the output files will correspondingly increase and may be an issue for disk storage capacity.
Additional INDOCK parameters can be adjusted according to the INDOCK Guide available in the
Supplementary Information.
Anticipated results
At the end of this protocol, a receptor will have been converted into a dock-readable binding pocket,
the system optimized on retrospective control calculations, and the system screened prospectively
against a large chemical library. While this protocol has been specific to a single target that yielded
successful results13, we have equally applied this protocol to a number of targets with general success
demonstrating its broad applicability12,83,118. We have attempted to curate all of the relevant steps,
though expert intuition brought by each user for their target will result in a different selection of final
hits for purchase. For these intuition-based steps, we recommend multiple people review the data
before making the final decision regarding which compounds to buy.
● The compound library did not contain enough examples of the class of compound most likely to dock
● Retrospective optimization was not possible due to lack of known chemical matter
● The experimental structure contained ambiguities inside chain or loop placement due to poor electron
For example, targets that bind dianions or dications will suffer from much smaller library sizes,
while targets that bind large protein ligands may be difficult to modulate with small molecules. As in
experimental biology, no set of retrospective controls will ever guarantee prospective success, and the
many approximations in docking ensure that we still measure success by hit rates.
Nevertheless, these controls and optimization steps can reduce obvious sources of failure, and
allow one to better design a subsequent campaign. The experience of the field by now suggests that
the approach is likely enough to succeed to be worth the investment, while the novel ligands
that result can bring genuinely new biological insight to a field, a longstanding goal of the structure-
based enterprise.
Reporting Summary
Further information on research design is available in the Nature Research Reporting Summary
linked to this article.
Data availability
An example set of files used in this protocol, including ligand and decoy sets, default docking grids and
optimized docking grids, can be downloaded from http://files.docking.org/dock/mt1_protocol.tar.gz. The
example dataset uses the MT1 structure (PDB: 6ME3) co-crystallized with 2-phenylmelatonin.
Software availability
DOCK3.7 can be downloaded after applying for a license from http://dock.docking.org/Online_
Licensing/index.htm. Licenses are free for nonprofit research.
References
1. Mayr, L. M. & Bojanic, D. Novel trends in high-throughput screening. Curr. Opin. Pharmacol. 9,
580–588 (2009).
2. Keserü, G. M. & Makara, G. M. The influence of lead discovery strategies on the properties of drug
candidates. Nat. Rev. Drug Discov. 8, 203–212 (2009).
3. Keiser, M. J., Irwin, J. J. & Shoichet, B. K. The chemical basis of pharmacology. Biochemistry 49,
10267–10276 (2010).
4. Bohacek, R. S., McMartin, C. & Guida, W. C. The art and practice of structure-based drug design: a
molecular modeling perspective. Med. Res. Rev. 16, 3–50 (1996).
5. Brenner, S. & Lerner, R. A. Encoded combinatorial chemistry. Proc. Natl Acad. Sci. USA 89, 5381–5383 (1992).
6. Fitzgerald, P. R. & Paegel, B. M. DNA-encoded chemistry: drug discovery from a few good reactions. Chem.
Rev. https://doi.org/10.1021/acs.chemrev.0c00789 (2020).
7. Grebner, C. et al. Virtual screening in the Cloud: how big is big enough? J. Chem. Inf. Model 60, 24 (2020).
8. Davies, E. K., Glick, M., Harrison, K. N. & Richards, W. G. Pattern recognition and massively distributed
computing. J. Comput. Chem. 23, 1544–1550 (2002).
9. Sterling, T. & Irwin, J. J. ZINC 15—ligand discovery for everyone. J. Chem. Inf. Model. 55,
2324–2337 (2015).
10. Patel, H. et al. SAVI, in silico generation of billions of easily synthesizable compounds through expert-
system type rules. Sci. Data 7, 384 (2020).
11. Grygorenko, O. O. et al. Generating multibillion chemical space of readily accessible screening compounds.
iScience 23, 101681 (2020).
12. Lyu, J. et al. Ultra-large library docking for discovering new chemotypes. Nature 566, 224–229 (2019).
13. Stein, R. M. et al. Virtual discovery of melatonin receptor ligands to modulate circadian rhythms. Nature
579, 609–614 (2020).
14. Gorgulla, C. et al. An open-source drug discovery platform enables ultra-large virtual screens. Nature 580,
663 (2020).
15. Meng, E. C., Shoichet, B. K. & Kuntz, I. D. Automated docking with grid‐based energy evaluation.
J. Comput. Chem. 13, 505–524 (1992).
16. Sharp, K. A., Friedman, R. A., Misra, V., Hecht, J. & Honig, B. Salt effects on polyelectrolyte-ligand binding:
comparison of Poisson–Boltzmann, and limiting law/counterion binding models. Biopolymers 36,
245–262 (1995).
17. Mysinger, M. M. & Shoichet, B. K. Rapid context-dependent ligand desolvation in molecular docking.
J. Chem. Inf. Model. 50, 1561–1573 (2010).
18. Adeshina, Y. O., Deeds, E. J. & Karanicolas, J. Machine learning classification can reduce false positives in
structure-based virtual screening. Proc. Natl Acad. Sci. USA 117, 18477–18488 (2020).
19. Irwin, J. J. & Shoichet, B. K. Docking screens for novel ligands conferring new biology. J. Med. Chem. 59,
4103–4120 (2016).
20. Mobley, D. L. & Dill, K. A. Binding of small-molecule ligands to proteins: “what you see” is not always
“what you get. Structure 17, 489–498 (2009).
21. Bissantz, C., Folkers, G. & Rognan, D. Protein-based virtual screening of chemical databases. 1. Evaluation
of different docking/scoring combinations. J. Med. Chem. 43, 4759–4767 (2000).
22. Tirado-Rives, J. & Jorgensen, W. L. Contribution of conformer focusing to the uncertainty in predicting free
energies for protein-ligand binding. J. Med. Chem. 49, 5880–5884 (2006).
23. Trott, O. & Olson, A. J. AutoDock Vina: improving the speed and accuracy of docking with a new scoring
function, efficient optimization, and multithreading. J. Comput. Chem. 31, 455–461 (2009).
24. Kramer, B., Rarey, M. & Lengauer, T. Evaluation of the FLEXX incremental construction algorithm for
protein-ligand docking. Proteins 37, 228–241 (1999).
25. Halgren, T. A. et al. Glide: a new approach for rapid, accurate docking and scoring. 2. Enrichment factors in
database screening. J. Med. Chem. 47, 1750–1759 (2004).
26. Morris, G. M. et al. Automated docking using a Lamarckian genetic algorithm and an empirical binding free
energy function. J. Comput. Chem. 19, 1639–1662 (1998).
27. Abagyan, R., Totrov, M. & Kuznetsov, D. ICM—a new method for protein modeling and design: appli-
cations to docking and structure prediction from the distorted native conformation. J. Comput. Chem. 15,
488–506 (1994).
28. Goodsell, D. S. & Olson, A. J. Automated docking of substrates to proteins by simulated annealing. Proteins
8, 195–202 (1990).
29. Mcgann, M. FRED pose prediction and virtual screening accuracy. J. Chem. Inf. Model 51, 578–596 (2011).
30. Jones, G., Willett, P., Glen, R. C., Leach, A. R. & Taylor, R. Development and validation of a genetic
algorithm for flexible docking. J. Mol. Biol. 267, 727–748 (1997).
31. Corbeil, C. R., Williams, C. I. & Labute, P. Variability in docking success rates due to dataset preparation.
J. Comput. Aided Mol. Des. 26, 775–786 (2012).
32. McGovern, S. L. & Shoichet, B. K. Information decay in molecular docking screens against Holo, Apo, and
modeled conformations of enzymes. J. Med. Chem. 46, 2895–2907 (2003).
33. Rueda, M., Bottegoni, G. & Abagyan, R. Recipes for the selection of experimental protein conformations for
virtual screening. J. Chem. Inf. Model. 50, 186–193 (2010).
34. Kuntz, I. D., Blaney, J. M., Oatley, S. J., Langridge, R. & Ferrin, T. E. A geometric approach to
macromolecule-ligand interactions. J. Mol. Biol. 161, 269–288 (1982).
35. Halgren, T. A. Identifying and characterizing binding sites and assessing druggability. J. Chem. Inf. Model.
49, 377–389 (2009).
36. Ngan, C. H. et al. FTMAP: extended protein mapping with user-selected probe molecules.
Nucleic Acids Res. 40, W271–W275 (2012).
37. Wang, S. et al. D4 dopamine receptor high-resolution structures enable the discovery of selective agonists.
Science 358, 381–386 (2017).
38. Katritch, V. et al. Structure-based discovery of novel chemotypes for adenosine A(2A) receptor antagonists.
J. Med. Chem. 53, 1799 (2010).
39. Kolb, P. et al. Structure-based discovery of beta2-adrenergic receptor ligands. Proc. Natl Acad. Sci. USA 106,
6843–6848 (2009).
40. De Graaf, C. et al. Crystal structure-based virtual screening for fragment-like ligands of the human hista-
mine H 1 receptor. J. Med. Chem. 54, 8195–8206 (2011).
41. Mysinger, M. M. et al. Structure-based ligand discovery for the protein–protein interface of chemokine
receptor CXCR4. Proc. Natl Acad. Sci. USA 109, 5517–5522 (2012).
42. Powers, R. A., Morandi, F. & Shoichet, B. K. Structure-based discovery of a novel, noncovalent inhibitor of
AmpC β-lactamase. Structure 10, 1013–1023 (2002).
43. Zarzycka, B. et al. Discovery of small molecule CD40–TRAF6 inhibitors. J. Chem. Inf. Model. 55,
294–307 (2015).
44. Huang, N. & Shoichet, B. K. Exploiting ordered waters in molecular docking. J. Med. Chem. 51,
4862–4865 (2008).
45. Balius, T. E. et al. Testing inhomogeneous solvation theory in structure-based ligand discovery. Proc. Natl
Acad. Sci. USA 114, E6839–E6846 (2017).
46. Weichenberger, C. X. & Sippl, M. J. NQ-Flipper: recognition and correction of erroneous asparagine and
glutamine side-chain rotamers in protein structures. Nucleic Acids Res. 35, W403–W406 (2007).
47. Word, J. M., Lovell, S. C., Richardson, J. S. & Richardson, D. C. Asparagine and glutamine: using hydrogen
atom contacts in the choice of side-chain amide orientation. J. Mol. Biol. 285, 1735–1747 (1999).
78. Stein, R. M. et al. Property-unmatched decoys in docking benchmarks. J. Chem. Inf. Model. 61,
699–714 (2020).
79. Coleman, R. G., Carchia, M., Sterling, T., Irwin, J. J. & Shoichet, B. K. Ligand pose and orientational
sampling in molecular docking. PLoS One 8, e75992 (2013).
80. Huang, N., Shoichet, B. K. & Irwin, J. J. Benchmarking sets for molecular docking. J. Med. Chem. 49,
6789–6801 (2006).
81. Jain, A. N. & Nicholls, A. Recommendations for evaluation of computational methods. J. Comput. Aided
Mol. Des. 22, 133–139 (2008).
82. Allen, W. J. & Rizzo, R. C. Implementation of the Hungarian algorithm to account for ligand symmetry and
similarity in structure-based design. J. Chem. Inf. Model. 54, 518–529 (2014).
83. Schuller, M. et al. Fragment binding to the Nsp3 macrodomain of SARS-CoV-2 identified through crys-
tallographic screening and computational docking. Sci. Adv. 7, eabf8711 (2021).
84. Fischer, A., Smieško, M., Sellner, M. & Lill, M. A. Decision making in structure-based drug discovery: visual
inspection of docking results. J. Med. Chem. 64, 2489–2500 (2021).
85. Kirchmair, J. et al. Predicting drug metabolism: experiment and/or computation? Nat. Rev. Drug Discov. 14,
387–404 (2015).
86. Kirchmair, J. et al. Computational prediction of metabolism: sites, products, SAR, P450 enzyme dynamics,
and mechanisms. J. Chem. Inf. Model. 52, 617–648 (2012).
87. Irwin, J. J. et al. An aggregation advisor for ligand discovery. J. Med. Chem. 58, 7076–7087 (2015).
88. Bemis, G. W. & Murcko, M. A. The properties of known drugs. 1. Molecular frameworks. J. Med. Chem. 39,
2887–2893 (1996).
89. Jadhav, A. et al. Quantitative analyses of aggregation, autofluorescence, and reactivity artifacts in a screen
for inhibitors of a thiol protease. J. Med. Chem. 53, 37–51 (2010).
90. Capuzzi, S. J., Muratov, E. N. & Tropsha, A. Phantom PAINS: problems with the utility of alerts for Pan-
Assay Interference Compound S. J. Chem. Inf. Model. 57, 417–427 (2017).
91. Baell, J. B. & Holloway, G. A. New substructure filters for removal of pan assay interference
compounds (PAINS) from screening libraries and for their exclusion in bioassays. J. Med. Chem. 53,
2719–2740 (2010).
92. McGovern, S. L., Caselli, E., Grigorieff, N. & Shoichet, B. K. A common mechanism underlying promiscuous
inhibitors from virtual and high-throughput screening. J. Med. Chem. 45, 1712–1722 (2002).
93. Feng, B. Y. et al. A high-throughput screen for aggregation-based inhibition in a large compound library.
J. Med. Chem. 50, 2385–2390 (2007).
94. Ganesh, A. N. et al. Colloidal drug aggregate stability in high serum conditions and pharmacokinetic
consequence. ACS Chem. Biol. 14, 751–757 (2019).
95. Coan, K. E. D. & Shoichet, B. K. Stoichiometry and physical chemistry of promiscuous aggregate-based
inhibitors. J. Am. Chem. Soc. 130, 9606–9612 (2008).
96. Coan, K. E. D., Maltby, D. A., Burlingame, A. L. & Shoichet, B. K. Promiscuous aggregate-based inhibitors
promote enzyme unfolding. J. Med. Chem. 52, 2067–2075 (2009).
97. Wolan, D. W., Zorn, J. A., Gray, D. C. & Wells, J. A. Small-molecule activators of a proenzyme. Science 326,
853–858 (2009).
98. Zorn, J. A., Wolan, D. W., Agard, N. J. & Wells, J. A. Fibrils colocalize caspase-3 with procaspase-3 to foster
maturation. J. Biol. Chem. 287, 33781–33795 (2012).
99. Irwin, J. J. et al. ZINC20—a free ultralarge-scale chemical database for ligand discovery. J. Chem. Inf. Model.
60, 6065–6073 (2020).
100. Teotico, D. G. et al. Docking for fragment inhibitors of AmpC -lactamase. Proc. Natl Acad. Sci. USA 106,
7455–7460 (2009).
101. Chen, Y. & Shoichet, B. K. Molecular docking and ligand specificity in fragment-based inhibitor discovery.
Nat. Chem. Biol. 5, 358–364 (2009).
102. Kolb, P. & Irwin, J. J. Docking screens: right for the right reasons? Curr. Top. Med. Chem. 9, 755–770 (2009).
103. Wu, Y., Lou, L. & Xie, Z.-R. A pilot study of all-computational drug design protocol–from structure
prediction to interaction analysis. Front. Chem. 8, 81 (2020).
104. Greenidge, P. A., Kramer, C., Mozziconacci, J. C. & Sherman, W. Improving docking results via reranking of
ensembles of ligand poses in multiple X-ray protein conformations with MM-GBSA. J. Chem. Inf. Model.
54, 2697–2717 (2014).
105. Mahmoud, A. H., Masters, M. R., Yang, Y. & Lill, M. A. Elucidating the multiple roles of hydration for
accurate protein-ligand binding prediction via deep learning. Commun. Chem. 3, 19 (2020).
106. Liu, X. et al. An allosteric modulator binds to a conformational hub in the β2 adrenergic receptor.
Nat. Chem. Biol. 16, 749–755 (2020).
107. Wacker, D. et al. Conserved binding mode of human β 2 adrenergic receptor inverse agonists and antagonist
revealed by X-ray crystallography. J. Am. Chem. Soc. 132, 11443–11445 (2010).
108. Manglik, A. et al. Structure-based discovery of opioid analgesics with reduced side effects. Nature 537,
185–190 (2016).
109. Ewing, T. J. A. & Kuntz, I. D. Critical evaluation of search algorithms for automated molecular docking and
database screening. J. Comput. Chem. 18, 1175–1189 (1997).
110. Gallagher, K. & Sharp, K. Electrostatic contributions to heat capacity changes of DNA-ligand binding.
Biophys. J. 75, 769–776 (1998).
142. Sassano, M. F., Doak, A. K., Roth, B. L. & Shoichet, B. K. Colloidal aggregation causes inhibition
of G protein-coupled receptors. J. Med. Chem. 56, 2406–2414 (2013).
143. Owen, S. C., Doak, A. K., Wassam, P., Shoichet, M. S. & Shoichet, B. K. Colloidal aggregation affects the
efficacy of anticancer drugs in cell culture. ACS Chem. Biol. 7, 1429–1435 (2012).
144. McLaughlin, C. K. et al. Stable colloidal drug aggregates catch and release active enzymes. ACS Chem. Biol.
11, 992–1000 (2016).
145. McGovern, S. L. & Shoichet, B. K. Kinase inhibitors: not just for kinases anymore. J. Med. Chem. 46,
1478–1483 (2003).
Acknowledgements
This work was supported by NIH grants R35GM122481 (to B.K.S.) and GM133836 (to J.J.I.). J.C. was supported by grants from the
Swedish Research Council (2017-04676) and the European Research Council (ERC) under the European Union’s Horizon 2020 research
and innovation programme (grant agreement: 715052). B.J.B. was partly supported by an NIH NRSA fellowship F32GM136062. C.M.W.
was partly supported by the National Institutes of Health Training Grant T32 GM007175, NSF GRFP and UCSF Discovery Fellowship.
We thank members of the Shoichet lab for feedback on the procedures described.
Author contributions
B.J.B. and S.G. wrote the manuscript with additional input from all authors. B.J.B., S.G., A.L., J.L., C.W., R.M.S., E.F., T.E.B., J.C., J.J.I. and
B.K.S. developed the protocol. B.J.B., S.G., A.L., J.L., C.W., R.M.S. and T.E.B. contributed scripts. E.F. tested the protocol. Research was
supervised by J.C., J.J.I. and B.K.S.
Competing interests
B.K.S. and J.J.I. are founders of Blue Dolphin Lead Discovery LLC, which undertakes fee-for-service ligand discovery.
Additional information
Supplementary information The online version contains supplementary material available at https://doi.org/10.1038/s41596-021-00597-z.
Correspondence and requests for materials should be addressed to Brian K. Shoichet.
Peer review information Nature Protocols thanks Claudio Cavasotto, Vincent Zoete and the other, anonymous reviewer(s) for their
contribution to the peer review of this work.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Related links
Key references using this protocol
Stein, R. M. et al. Nature 579, 609–614 (2020): https://doi.org/10.1038/s41586-020-2027-0
Lyu, J. et al. Nature 566, 224–229 (2019): https://www.nature.com/articles/s41586-019-0917-9
Schuller, M. et al. Sci. Adv. 7, eabf8711 (2021): https://advances.sciencemag.org/content/7/16/eabf8711
Reporting Summary
Nature Research wishes to improve the reproducibility of the work that we publish. This form provides structure for consistency and transparency
in reporting. For further information on Nature Research policies, see our Editorial Policies and the Editorial Policy Checklist.
Statistics
For all statistical analyses, confirm that the following items are present in the figure legend, table legend, main text, or Methods section.
n/a Confirmed
The exact sample size (n) for each experimental group/condition, given as a discrete number and unit of measurement
A statement on whether measurements were taken from distinct samples or whether the same sample was measured repeatedly
The statistical test(s) used AND whether they are one- or two-sided
Only common tests should be described solely by name; describe more complex techniques in the Methods section.
For null hypothesis testing, the test statistic (e.g. F, t, r) with confidence intervals, effect sizes, degrees of freedom and P value noted
Give P values as exact values whenever suitable.
For Bayesian analysis, information on the choice of priors and Markov chain Monte Carlo settings
For hierarchical and complex designs, identification of the appropriate level for tests and full reporting of outcomes
Estimates of effect sizes (e.g. Cohen's d, Pearson's r), indicating how they were calculated
Our web collection on statistics for biologists contains articles on many of the points above.
Data analysis DOCK3.7.5 can be downloaded after applying for a license from http://dock.docking.org/Online_Licensing/index.htm. Licenses are free for
non-profit research.
For manuscripts utilizing custom algorithms or software that are central to the research but not yet described in published literature, software must be made available to editors and
reviewers. We strongly encourage code deposition in a community repository (e.g. GitHub). See the Nature Research guidelines for submitting code & software for further information.
Data
Policy information about availability of data
All manuscripts must include a data availability statement. This statement should provide the following information, where applicable:
- Accession codes, unique identifiers, or web links for publicly available datasets
April 2020
An example set of files used in this protocol including ligand and decoy sets, default docking grids, and optimized docking grids can be downloaded from http://
files.docking.org/dock/mt1_protocol.tar.gz. The example dataset uses the MT1 structure (PDB: 6ME3) co-crystallized with 2-phenylmelatonin.
1
nature research | reporting summary
Field-specific reporting
Please select the one below that is the best fit for your research. If you are not sure, read the appropriate sections before making your selection.
Life sciences Behavioural & social sciences Ecological, evolutionary & environmental sciences
For a reference copy of the document with all sections, see nature.com/documents/nr-reporting-summary-flat.pdf
Replication N/A
Randomization N/A
Blinding N/A
April 2020