In the attached study which seeks to identify immunological markers associated with Chronic Fatigue Syndrome, the authors use sera from 21 cases and 21 controls, and screen it with a 125,000 random 12-mer array. They identify a 25 peptide signature that is specific (93%) and sensitive (98%) for CFS. These peptides relate to self-antigens and endogenous retroviruses, and a small but eclectic mix of bacterial and viral proteins.
Could a statistician explain how you can demonstrate that (in this and any other comics studies looking for biomarkers or signatures associated with a specific condition or outcome), of all the trillions of possible signatures, this observation of a 25 peptide signature is not just a random observation. If you anonymised cases and controls and randomly assigned them into one of two groups, and maybe bootstrapped this 100 times, I think that there are so many possible signatures (combinations of 1,2,3,4,5...25...n peptides, among a total population of 125,000), that with such a small sample of just 42 participants, you will likely always identify some kind of differential signature between two random groups.
In this kind of analysis how is significance testing done, and how does it account for the very high total number of possible observations that would be deemed significant?