Meta-analyses and systematic reviews seem the shortcut to academic success as they usually have a better chance of getting published in accredited journals, be read more, and bring home a lot of citations. Interestingly enough, apart from being time-consuming, they are very easy; they are actually nothing but carefully followed protocols of online data collection and statistical analysis, if any.

The point is that most of this can be easily done (at least in theory) by a simple computer algorithm. A combination of if/thenstatements would simply allow the software to decide on the statistical parameters to be used, not to mention more advanced approaches that can be available to expert systems.

The only part needing a much more advanced algorithm like a very good artificial intelligence is the part that is supposed to search the articles, read them, accurately understand them, include/exclude them accordingly, and extract data from them. It seems that today’s level of AI is becoming more and more sufficient for this purpose. AI can now easily read papers and understand them quite accurately. So AI programs that can either do the whole meta-analysis themselves, or do the heavy lifting and let the human check and polish/correct the final results are on the rise. All needed would be the topic of the meta-analysis. The rest is done automatically or semi-automatically.

We can even have search engines that actively monitor academic literature, and simply generate the end results (i.e., forest plots, effect sizes, risk of bias assessments, result interpretations, etc.), as if it is some very easily done “search result”. Humans then can get back to doing more difficult research instead of putting time on searching and doing statistical analyses and writing the final meta-analysis paper. At least, such search engines can give a pretty good initial draft for humans to check and polish them.

When we ask a medical question from a search engine, it will not only give us a summary of relevant results (the way the currently available LLM chatbots do) but also will it calculate and produce an initial meta-analysis for us based on the available scientific literature. It will also warn the reader that the results are generated by AI and should not be deeply trusted, but can be used as a rough guess. This is of course needed until the accuracy of generative AI surpasses that of humans.

It just needs some enthusiasts with enough free time and resources on their hands to train some available open-source, open-parameter LLMs to do this specific task. Maybe even big players are currently working on this concept behind the scene to optimize their propriety LLMs for meta-analysis generation.

Any thoughts would be most welcome.

Vahid Rakhshan

More Vahid Rakhshan's questions See All
Similar questions and discussions