As with all of the other steps in the review process, best practice for data extraction is to have two independent reviewers extract data from each article.
Beyond that, here are some tips for a smooth data extraction process.
Data extraction is the process of taking the most important characteristics of each study in the review and placing them into a chart, so that the details can be easily compared across studies. Typically, at least two people independently extract the data and then compare the results afterward, resolving any conflicts in the data that may have arisen in the process. Their data is then combined to create one final table.
While there are no prescribed set of study characteristics to include in a review, the most frequently extracted data items are things like:
There are many different ways to extract data. You can use a form that you build in Microsoft Forms, Google Forms, RedCap or in a tool like Covidence. (See the Review Tools and Applications tab.) You can enter the data directly into a spreadsheet program like Excel and then blend the charts later by copying and pasting. No matter how you extract the data, the important thing is to describe how you did it in the methods section and to reduce bias in the results by having the extractors work independently.
As mentioned earlier, you should also pilot your data extraction tool. To do this, have your extractors work on 2-3 articles to start. Compare answers and ensure that the results are similar. If your tool is lacking interrater reliability, it may be a sign that the extractors aren't understanding the data items the same way, and that definitions may need to be clarified.
More Advice
Examples of Evidence Tables
Data Extraction Templates
Yes! There are several tools that can help with data extraction. However, as with all AI products, please always verify the results-it is not a substitute for human reviewers.
Many of the tools recommended on the following tabs have been reviewed for their usefulness (Blaizot, et al, 2022; Khalil, 2022).
Covidence
Covidence helps automate several stages of a systematic and scoping review. Covidence uses both "study data from external repositories" and a "large language model extraction from full text articles" to collect data.
Learn More About Covidence
Systematic Review Data Repository (SRDR+)
The Systematic Review Data Repository (SRDR+) is a collaborative, Web-based repository of systematic review data. This resource serves as both an archive and data extraction tool and is shared among organizations and individuals producing systematic reviews worldwide, enabling the creation of a central database of systematic review data which may be critiqued, updated, and augmented on an ongoing basis. This database is freely accessible to facilitate evidence reviews and thus improve and speed up policy-making with regards to healthcare.
Learn More About How to Use SRDR+ for Data Extraction
DistillerSR
DistillerSR is similar to Covidence in that it is a review management tool. DistillerSR "automates the management of literature collection, screening, and assessment using AI and intelligent workflows."
Regarding the data-extraction component, DistillerSR says, "Simplify data extraction through templates and configurable forms. Extract data easily with in-form validations and calculations, and easily capture repeating, complex data sets."
Please note that DistillerSR is a paid subscription.
Below is a video walkthrough from Evidence Partners on how to export your references and data from DistillerSR
MGH RedCap
Our final tool for data extraction that we wanted to introduce is MGH RedCap. While this is not an AI tool, it is a software application that can be used for electronic collection and management of research and clinical study data. More often used for surveys, but some of our faculty and students have used it to build data extraction forms.