Hello,

Setting: We have a rather large dataset consisting of somewhat 500 samples containing metabolomics data. The data was obtained by analysing dry capillary blood using HPLC methods, yielding relative spectral intensity of approximately 650 different metabolites.

Since the relevant databases rarely contain dry blood-derived metabolomics datasets we are currently not entirely sure on how to normalise our dataset. This is as the common suspects, such as normalisation by weight, median, mean either seem very counterintuitive or hardly change the data at all.

Fortunately, we are in possession of various laboratory blood-derived values (Creatinine etc.), sort of as a ground truth value for most of our samples, which were obtained at the same timepoints to the dry blood. We were wondering if those would be useful in finding an adequate parameter to normalise our data with. For instance, Creatinine could be correlated to the amount of "water" and thus dilution of blood samples. Since we measure dry blood, the water would evaporate and therefore, using Creatinine to normalise our data would help us dampen this effect.

It would be much appreciated if anyone could shed some light on their usual practice when it comes to normalising metabolomics data for which there is no significant standard procedure?

More Philipp Brunnbauer's questions See All
Similar questions and discussions