Case Study

Using DataSeer’s NLP to pre-fill the MDAR checklist lowered friction for authors while improving the quality of reporting. Instead of treating compliance as a manual, after-the-fact task, it was embedded into the workflow. Authors shifted from generating responses to validating them. To move to high throughput, Science needs to further automate the process with DataSeer. The goal is that well-designed automation will both enhance author experience and raise reporting standards, rather than forcing a trade-off between the two. This is especially important for levelling the playing field for authors with differing resources. 

The Open Science Metrics (OSI) analysis moved beyond policy enforcement to real evaluation. By quantifying data, code, and protocol sharing across 2,680 articles, the pilot revealed both strengths (e.g., universal data availability statements, strong overall data sharing rates) and specific gaps (e.g., limited sharing of plotting data and executable code). This kind of portfolio-level visibility enables publishers to identify where policies are working and where they need to evolve. 

Crucially, the pilot didn’t end with insights. It informed concrete policy changes. Science is strengthening requirements around sharing underlying data for figures and aims to expand automated MDAR across journals. This illustrates a broader takeaway: pilots should be designed not just to test tools, but to generate evidence that drives editorial policy. Iterative, data-informed policy development helps publishers stay credible with researchers while meaningfully advancing reproducibility.