Explore Alternative Metrics for Comprehensive Evaluation of Quantitative Performance when using Mixed-Species Datasets #13
hollenstein
announced in
Hackathon proposals
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Title
Alternative Metrics for Quantitative Proteomics Evaluation
Abstract
Proteomics experiments typically aim to not only identify but also quantify protein content across samples. Assessing quantitative performance is crucial for evaluating instruments, acquisition methods, and data processing algorithms. Benchmarking is typically done using mixed-species proteome samples with varying amounts of the individual species, to create a quantification ground truth. Evaluation of the benchmark results, however, often focuses solely on few global accuracy and precision metrics.
In this project, I propose to explore alternative and more detailed metrics for evaluating the quantitative performance of mixed-species datasets. To investigate the characteristics and usefulness of these metrics, we will implement and compare different algorithms for summarizing ions to proteins. We will use a collection of ground truth datasets with different characteristics: measured on various MS instruments, acquired in both DIA and DDA mode, and analyzed with multiple programs. To easily apply the protein summarization algorithms across all datasets and automatically calculate various performance metrics, we will create a flexible data processing pipeline in Python.
This project aims to provide more detailed performance assessments in quantitative proteomics that will facilitate method development and the evaluation of strengths and weaknesses of data acquisition and processing pipelines.
Project Plan
During the hackathon we will focus on the four major tasks outlined below. Before we start, we will decide on common interfaces for the modules in the Python pipeline. This will allow working on tasks 1-3 in parallel. At the end, in task 4, we will integrate all components and use the results to discuss the usefulness of the different metrics, and if time permits, plan how to improve existing metrics and think of potential additional metrics.
Note: An alternative approach for evaluating the performance metrics is to consider only one protein summarization algorithm, but add varying amounts of noise and biases to the reported ion intensities. One could then investigate how well the effects of the manipulation are reflected in the calculated metrics. If time permits, we might implement additional modules for the processing pipeline that introduce such errors to the ion intensities, and include them in step 4 when running the pipeline and for characterizing the metrics.
After the Hackathon
Technical Details
Contact information
David M. Hollenstein
University of Vienna, Austria
Mass Spectrometry Facility of the Max Perutz Labs (Part of Vienna BioCenter)
[email protected]
Beta Was this translation helpful? Give feedback.
All reactions