top of page

General Workflow:

First is data Input. Users can upload a diverse array of data generated by various mass spectrometry (MS) techniques. This includes datasets such as deuterium uptake curves derived from hydrogen-deuterium exchange mass spectrometry (HDX-MS), comprehensive lists of cross-linked peptides obtained through cross-linking mass spectrometry (XL-MS), arrival time distributions measured in ion mobility mass spectrometry (IM-MS), and extensive native mass spectra that provide critical insights into the structural characteristics of proteins.

Then, through Data Integration and Alignment, ConformAIt integrates all incoming datasets and aligns them to a standardized reference protein sequence or structural framework. This alignment may utilize advanced models derived from AlphaFold or precise structures elucidated through X-ray crystallography. The system diligently identifies and highlights discrepancies or inconsistencies between the experimental data and the static models, ensuring that all variations are thoroughly examined and accurately addressed.

At last, we use super computer to do the AI-Powered Modelling. Leveraging advanced neural networks rigorously trained on a comprehensive range of known protein structures and their dynamic behaviours, ConformAIt excels at generating an ensemble of conformations that satisfy all specified experimental constraints. The platform is proficient in predicting complex phenomena, including:
- Misfolding intermediates: For instance, it effectively models the "extended monomeric" species related to α1-antitrypsin deficiency, capturing the nuances of protein misfolding.
- Allosteric changes: The platform can detect conformational shifts induced by ligand interactions, as evidenced by insights derived from HDX-MS studies of molecular chaperones such as GroEL.
- Oligomerization pathways: By integrating constraints from XL-MS with data on oligomeric states gathered through IM-MS, the platform reconstructs the intricate pathways leading to protein oligomerization. (2, 5-7)

 

Theory at the back: AI Model Model Used to do the prediction

Regarding AI model design, interpretability, validation, uncertainty quantification, and computational requirements, the ConformAIt platform is fundamentally an integrative, multi-modal AI system designed to synthesize heterogeneous data from various structural mass spectrometry (MS) techniques. While the documentation does not explicitly detail the specific underlying neural network architectures (e.g., geometric deep learning, diffusion models), its design philosophy is clearly articulated through its workflow and capabilities.
The core design principle is data-driven, constraint-based ensemble modeling. The AI models are trained on a corpus of experimental MS data—including Collision-Induced Unfolding (CIU) fingerprints (arrival time vs. collision voltage landscapes), Collision Cross Section (CCS) distributions from Ion Mobility-MS (IM-MS), deuterium uptake rates from Hydrogen-Deuterium Exchange MS (HDX-MS), and spatial distance restraints from Cross-Linking MS (XL-MS). (1)

The model's primary task is to generate an ensemble of protein conformations that collectively satisfy all input experimental constraints simultaneously. This approach moves beyond predicting a single static structure (like AlphaFold) to modeling the dynamic energy landscape and populations of states (native, intermediates, misfolded).

The philosophy is to use experimental data as the "ground truth" to guide, validate, and refine predictions, making dynamic modeling more experimentally anchored and less computationally prohibitive than brute-force molecular dynamics simulations.

 

Model Interpretability and Data Point Attribution
A key strength of ConformAIt is its focus on interpretability and mechanistic insight, which is crucial for scientific discovery. The platform is designed not as a black box but as a tool for hypothesis generation and testing. The AI system can indeed highlight which specific experimental data points exert a decisive influence on the final predicted conformational ensemble. For instance:
-For HDX-MS data: The model can identify and report which specific peptide regions or residues show significant protection or deprotection changes between conditions (e.g., wild-type vs. mutant, apo vs. ligand-bound). It can attribute the stabilization or destabilization of a particular predicted conformational state directly to the altered deuterium uptake kinetics in those regions1.
-For XL-MS data: The platform can pinpoint which specific cross-linked residue pairs (e.g., Lys-Lys pairs within a defined distance) provided the critical spatial restraints that guided the docking of subunits or the folding of a domain. It can report the satisfaction or violation scores for each cross-link constraint in the final model, indicating which data points were most influential1.
-For CIU-IM-MS data: The software determines unfolding threshold voltages and monitors population shifts. It can correlate features in the CIU fingerprint (e.g., the emergence of a new arrival time peak at a specific collision voltage) with the predicted formation of a specific unfolding intermediate in the structural ensemble1.

This capability transforms the platform from a mere predictor into an analytical partner, allowing researchers to understand why the model predicts a certain conformational change, based directly on their experimental observations.

How to Convert Your Mass Spectrometry Data into Structural Mechanisms Using ConformAIt?

Step 1. Data Upload: Supports raw or processed mass spectrometry data (HDX-MS, XL-MS, Native MS, IM-MS, etc.)

Step 2. AI integrated modeling: automatically aligning multiple technical data to generate conformational sets that meet experimental constraints.

 

Step 3. Visualization and Verification: Output dynamic trajectories, free energy landscape maps, key flexible regions, and drug binding sites

Note: Technical compatibility- Supports mainstream mass spectrometer data formats such as Thermo, Agilent, Waters, Sciex, etc.

Quantification of Prediction Uncertainty and Confidence
ConformAIt addresses the critical issue of model ambiguity and uncertainty. The platform generates quantitative metrics and confidence scores for its predictions.
-Ensemble Diversity and Population Weights: The output is not a single structure but an ensemble. The model reports the relative population or probability of each major conformational state (e.g., 60% Native, 30% Intermediate I, 10% Intermediate II) under the given experimental conditions. The spread of the ensemble itself represents uncertainty.
-Constraint Satisfaction Scores: For each experimental constraint (e.g., an XL-MS distance, an HDX protection factor), the model calculates a satisfaction score or a residual. The overall goodness-of-fit metric across all integrated data points indicates the confidence in the ensemble model. Poor fit in certain regions highlights areas of high uncertainty or potential conflict in the data.
-Confidence Scores for States: The platform can assign confidence scores for predicted conformational states. This likely derives from the consistency of that state across multiple computational sampling trajectories and its ability to simultaneously satisfy a high proportion of the diverse experimental inputs.
 

Computational Resource Estimates for a Typical Task
The following estimates can be inferred as a cloud-based SaaS platform with high-throughput compatibility.
E.g. For a typical analysis task involving a 50 kDa protein:
1. Data Input: Integrating HDX-MS data (time-course for ~50 peptides), XL-MS data (~20-50 cross-links), and Native IM-MS/CIU data (CCS & unfolding fingerprint).
2. Compute Time: Generating a dynamic conformational ensemble from this multi-technique data is computationally intensive. A reasonable estimate for the AI-driven modeling and analysis pipeline on a cloud instance would range from ~4 to 12 GPU hours. This includes time for data alignment, constraint processing, ensemble sampling, refinement, and validation scoring. Simpler analyses (e.g., comparing two conditions) would be faster, while de novo modeling of large complexes would take longer.
3. Cloud Resource Consumption: The platform likely utilizes high-memory VM instances with a dedicated GPU (e.g., equivalent to an NVIDIA A10G or V100). The total cost for a single project would depend on the cloud provider but could be estimated in the range of ~$20 - $80 per run, assuming on-demand pricing.
4. Dataset Scale Upper Limit: The platform is designed for high-throughput compatibility1. It can handle datasets from "small peptide constructs to large complexes," such as the 11-subunit TRAP complex mentioned1. Practical upper limits would be governed by cloud instance memory and are likely in the range of multi-megadalton complexes (e.g., ribosomal subunits, viral capsids). The limiting factor for users would typically be the generation of the experimental MS data itself, not the platform's processing capacity. (2, 4, 5)

In summary, ConformAIt's AI philosophy centers on experimental-data-constrained ensemble modeling to reveal protein dynamics. Its strength lies in interpretable, attribution-aware predictions that can be validated by orthogonal biophysics, with quantified uncertainties, delivered via a scalable cloud platform designed for practical research throughput.

 

Note: The core description of ConformAIt's workflow, data integration, comparison to other methods, and stated capabilities (like confidence scores, high-throughput handling) are all derived only for BIOC0015 Coursework.​ The estimation time is based on an average number of a typical AI platform working data.

This is a fake website, it is only used for UCL BIOC0015 coursework.

Contact Us  for more information

Privacy

Refund

Order

  • Facebook
  • Instagram
  • X
  • TikTok

500 Terry Francine Street, 6th Floor, San Francisco, CA 94158

123-456-7890

ConformAlt

© 2035 by ConformAlt. Powered and secured by Wix

bottom of page