FTMap: fast and free* druggable hotspot prediction

*free to academics

FTMap is a useful and fast online tool that attempts to mimic experimental fragment-screening methodologies (SAR-by-NMR and X-ray crystallography) using in silico methods.   The algorithm is based on the premise that ligand binding sites in proteins often show “hotspots” that contribute most of the free energy to binding.

Often, fragment screening will identify these hotspots when clusters of different types of fragments all bind to the same subsite of a larger binding site.   In fact, x-ray crystallography studies of protein structures solved in a variety of organic solvents demonstrate that small organic fragments often form clusters in active sites.

In the FTMap approach, small organic probes are used for an initial rigid-body docking against the entire protein surface.  The “FT” of FTMap stands for the use of fast Fourier transform (FFT) methods to quickly sample billions of probe positions while calculating accurate energies based on a robust energy expression.

Following docking of each probe, thousands of poses are energy-minimized and clustered based on proximity.  The clusters are then ranked for lowest energy.   Consensus sites (“hot spots”) are determined by looking for overlapping clusters of different types of probes within several angstroms of each other.   If several consensus sites appear near each other on the protein surface, that is a strong indication of a potentially druggable binding site.

What is tidy data?

tidy data
A warehouse of tidy data (in paper form).

What is “tidy” data?

What is meant by the term “tidy” data, as opposed to “messy” data?  In my last post I listed five of the most common problems encountered with messy datasets.  Logically, “tidy” data must not have any of these problems.  So just what does tidy data look like?

Let’s take a look at an example of tidy data.  Below are the first 20 lines from R’s built-in “airquality” dataset:

Fig 1.  Air quality dataset is messy data.
Figure 1. The “airquality” dataset.

According to R programmer and professor of statistics Hadley Wickham, tidy data can be defined as the following:

1)  Each variable forms a column

2) Each observation forms a row

3) Each type of observational unit forms a table

That’s it.  “Airquality” is tidy because each row corresponds to one month/day combination and the four measured weather variables (ozone, solar, wind, and temp) on that day.

What about messy data?

Let’s see an example of a messy weather dataset for a counterexample (data examples are from this paper by H. Wickham):

Figure 2.  A messy weather dataset.  Not all columns are shown for the sake of clarity.
Figure 2. A messy weather station dataset.  Not all columns are shown for the sake of clarity.

There are multiple “messy” data problems with this table.  First, identifying variables like day of the month are stored in column headers (“d1”, “d2”, etc…), not in rows.  Second, there are a lot of missing values, complicating analysis and making it harder to read the table.  Third, the column “element” consists of variable names (“tmin” and “tmax”) violating rule 1 of tidy data.

How to use R tools to transform this table into tidy form is beyond the scope of this post, so I will just show the tidy version of this dataset in Figure 3.

Screen shot 2014-08-01 at 1.55.23 PM
Figure 3. The weather station data in tidy form.

Each column now forms a unique variable.  The date information has been condensed into a more compact form and each row contains the measurements for only one day.  The two variables in the “element” column are now forming their own columns, “tmax” and “tmin.”  With the data in this form it is far easier to prepare plots, aggregate the data, and perform statistical analysis.

 

 

 

 

 

Five Common Problems with Messy Data

6957593947_75f7aaecd0_zReal world datasets are often quite messy and not well-organized for available data analysis tools.  The data scientist’s job often begins with whipping these messy datasets into shape for analysis.

Listed below are five of the most common problems with messy datasets, according to an excellent paper on “tidy data” by Hadley Wickham:

1) Column headers are variables, not variable names

Tabular data falls into this type, where columns are variables themselves.  For example,  a table with median income by percentile in columns and US states in rows. 

2) Multiple variables are stored in one column

An example here would be storing data in columns that combine two variables, like gender and age range.  Better to make two separate columns for gender and age range.

3) Variables are stored in both rows and columns

The most complex form of messy data.   For example, a dataset in which measurements from a weather station are stored according to date and time, with the various measurment types (temp, pressure, etc…) in a column called “measurements”.  

4) Multiple types of observational units are stored in the same table

A dataset that combines multiple unrelated observations or facts into one table.   For example, a clinical trial dataset that includes both treatment outcomes and diet choices into one large table by patient and date. 

5) A single observational unit stored in multiple tables

Measurements recorded in different tables split up by person, location, or time.  For example, a separate table of an individual’s medical history for each year of their life. 

Isotopic labeling of proteins in non-bacterial expression systems

As therapeutic proteins gain importance alongside traditional small molecule drugs, there is increasing interest in using NMR methods to examine their structure, dynamics, and stability/aggregation in solution.

Modern heteronuclear NMR of proteins relies on isotopically-labeled samples containing NMR active nuclei in the peptide backbone, sidechains, or both.

Although isotopic-labeling of recombinant protein is typically carried out in E. Coli expression systems, many biotherapeutic proteins must be expressed in eukaryotic systems to insure proper folding and/or post-translational modifications.   In practice, this means overexpression in either yeast, insect or mammalian cells.

Increased interest in attaining labeled protein samples for analysis by NMR is leading to better commercial availability of isotopically-labeled expression media and improved vectors for overexpression in non-bacterial systems.

Comprehensive reviews of state-of-the-art protocols and procedures for expression of isotopically-labeled proteins in non-standard systems are available here: yeast, insect cells, and mammalian cells.

 

 

Virtual screening capability for under $5K?

Many early stage companies may be missing out on the value that docking can provide at the validated hit and hit-to-lead stages of development, where structure/activity relationships (SAR) can help guide chemistry development of lead compounds.

While docking large HTS libraries with millions of compounds may require specialized CPU clusters, docking of small libraries (i.e., thousands of compounds) and SAR compounds from experimental assays is readily achievable in short time frames with a relatively inexpensive Intel Xeon workstation.

Following an initial investment in the workstation and software, follow-on costs are minimal (e.g., electricity, IT support and data backup). Turnaround times may be faster than with CRO services.  Also, sensitive IP data is also protected by being retained onsite and not transmitted over the internet.

Equipment / cost breakdown:

Software:

AutoDock Vina (non-restrictive commercial license)       cost: free

Accurate (benchmarked against 6 other commercial docking programs)

Compatible with AutoDock tools

Optimized for speed (orders of magnitude faster than previous generation)

Parallelized code for multi-core systems

AutoDock Tools (non-restrictive commercial license)      cost: free

PyMol Incentive (commercial license)                                        cost ~$90 / mo

Visualize docking results, free plugin can allow Vina to be run within PyMol GUI

Fedora Linux                                                                                              cost: free

Hardware:

HP Z620 Workstation (stock configuration)                          cost: $2999

2 GHz (6 Core) Intel Xeon E5-2620 2GHz

USB keyboard and mouse                                                                  cost: $50

Dell Ultrasharp 27” LED monitor                                                  cost: $649

1TB USB HD for data backup                                                          cost: $150

IT support for initial setup ~ 4 hours                                           cost: $400

Total initial capital expenditure:                                                  ~$4350