Anyone doing a study should do a preliminary literature review beforehand to see what's been done already, so such collation should already exist in research teams.
There are lots of review articles that are intended to collate and review evidence on a particular theme.
That's true, but maybe because the number of different things tested is so large, (for example, 445 chemicals in just the NIH CSF metabolomics study), and there are so many studies, it's hard to get a bird's eye view of the entire ME/CFS field. So while a researcher studying metabolomics might have a pretty good idea of the data in that field, it might be hard to stay current on everything else, to make connections to other fields.
Even in a given field, I think the advantage would be that this is exhaustive, including every single test, while a review would necessarily do a bit of summarizing, I think.
There are thousands of medical and biomedical papers published every year, so the task is either massive or has to be restricted to specific aspects and confined only to ME/CFS, such as CPET.
The idea is just for ME/CFS, though if people wanted to put in the work, I suppose it could be more general. When I search "chronic fatigue syndrome" in Pubmed, I get 8,425 total results. That's high, but I think not impossible. Of course a good bit higher if including long COVID probably.
Is there already such a collation done for ME/CFS?
If there is, I assume it'd be in static form and can't continue to be updated, at least with no delay. Though maybe something similar does exist.
How far back in history would the search go?
All the way back to the earliest ME/CFS (or the alternate names more frequently used then) studies, ideally.
Does it require expert knowledge to be able to use the resulting compilation in any meaningful way?
The core concept's motivation, I was just thinking of this as more of a fun tool for me to be able to follow the most promising areas. If some random chemical, like X-29542 is repeatedly low in the body in every test, I can set a Pubmed alert for that chemical to see what happens next.
But I think there is potential for it to be actually useful as a quick overview of everything for a researcher, to see at a glance every paper in every field which has had a test be replicated multiple times. If anything catches their eye, they might read the paper and gain some insights.
Or if they are reading a paper in a different, unfamiliar field and want to quickly check a few of the reported tests that might be related to their field, without spending time reading through multiple studies and reviews, they could just quickly get the info on the website.
Also, as new papers come out, they can be immediately added, which might help since reviews take some time.
There are thousands of things that can be measured in humans - from metabolomics, proteomics, genomics ..... would such a website resource just end up with hundreds of short unsorted lists?
The main goal would be to concentrate the fraction of tests that have been looked at multiple times to see those results. The majority of tests would be mostly useless for seeing where the science is if they've only been tested once so far, but they'd be there, ready to be added to when new studies are done.
The main concept would be one master list that includes tests from all fields. If, for example, serotonin has been high in 10 studies, and low in 1, it might be near the top.
Though I do imagine a tagging system as well. Maybe each test could be tagged with the field or organ, for example "exercise" or "brain", and you could filter by those if desired.
The vast majority of studies only report summary data (mean, sd, p-value) or in some cases just that the measurement wasn't significant. So the overview of data that is publicly available likely be a small subset of what has been published and tested.
It's true, it could only use what is publicly available, but it might still have value from what is published.
I was thinking at minimum this would store any test that at least says significance of a test or has a p-value, and use <0.05 if they didn't specify a cutoff themselves. And if significant, has whether the ME/CFS group's result is higher or lower than the other group.
In addition there is often a problem with the quality of data. If one does not take this into account, the overview of data may be misleading.
Quality is often so poor that data is worthless, how would such a compilation deal with that?
I don't think it could deal with curation on this scale. But I think it might have value just through scope, of which bad studies might only spoil a small portion of the data. And if interested in a specific test, it'd be easy to follow through to all studies to see them.