Reproducible research in Price Statistics: Call for action!

Presented at the 2026 Ottawa Group (Warsaw, Poland)

Serge Goussev
Steve Martin
Claude Lamboray
Christophe Bontemps
Tanya Flower
Ben Hillman
Caroline White

2026-05-14

Why should you care?

Right now, it is a pain to replicate most research in price statistics

  • Say you found a cool methodological paper with an empirical section. Let’s try to
    • understand it / learn from it by trying it;
    • peer review it;
    • evaluate it on your own data and maybe use it in production;
    • extend it, there is a cool change that may make it better;
  • Can we do any of the above right now in price statistics without a large amount of effort?

Replication crisis is not new, it’s actually well acknowledged in science

  • Ioannidis (2005), Baker (2016), and many others put the replication crisis on the map.

  • There have been a number of notable movements that have sought to help solve this problem. For example:

    • FAIR—Findable, Accessible, Interoperable, Reusable—a set of principles for open science (Wilkinson et al. 2016).

    • Turing way—a set of guides and processes to follow that (Community 2025).

  • A lot of progress has been made! For example a recent survey of reproducibility in Economics (at least for journals with strict data sharing mandates and policies) reproducibility is as high as 85% (Brodeur et al. 2026).

(Check out the references for more info)

Baker, Monya. 2016. 1,500 Scientists Lift the Lid on Reproducibility. Nature Publishing Group UK London.
Brodeur, Abel, Derek Mikola, Nikolai Cook, et al. 2026. “Reproducibility and Robustness of Economics and Political Science Research.” Nature 652 (8108): 151–56.
Community, The Turing Way. 2025. The Turing Way: A Handbook for Reproducible, Ethical and Collaborative Research. V. 1.2.3. Zenodo, released April. https://doi.org/10.5281/zenodo.15213042.
Ioannidis, John P. A. 2005. “Why Most Published Research Findings Are False.” PLOS Medicine 2 (8): null. https://doi.org/10.1371/journal.pmed.0020124.
Wilkinson, Mark D, Michel Dumontier, IJsbrand Jan Aalbersberg, et al. 2016. “The FAIR Guiding Principles for Scientific Data Management and Stewardship.” Scientific Data 3 (1): 1–9.

Progress in price statistics is lower, but it doesn’t have to be this way!

  • In Price Statistics, we can adopt common practices to improve replicability (and ideally aim for reproducibility).

We can move from a “trust me, it works on my data” toward “here is exactly how I did what I did, and here is everything you need to check my work.”

  • The movement is actually already underway! For example increasing adoption of Reproducible Analytical Pipelines (RAP) is a step in the right direction!

  • What is left is to confirm discipline specific research practices.

Who are we?

A bit about us

  • As a workstream in the UN Task Team, we are a group of price statisticians from NSOs and research institutes around the world.

  • In 2024, we formed the FAIR/Reproducibility workstream to instil reproducibility within the discipline.

  • We introduced the topic to the community a year ago (at 2025 CPI Expert Group).

How do we help?

  • We see two main blockers for reproducibility in price statistics:

    • Open datasets are few and far between, or are undocumented! To get to reproducibility, we all need access to high value open datasets. Without that, it is natural to default to internal data at our NSOs that is already processed, classified, and ready to go.

    • Unclear how to be reproducible! There is a lot of guidance out there, but what works for our discipline? The learning curve is non-trivial—so why bother if the “how to do this” path isn’t clear, and the mission isn’t obvious?

  • We are working to solve both by:

    • Developing and supporting a curated catalogue of good research datasets.

    • Providing guidance on practices that apply to our discipline.

How can you share your research findings in a reproducible way?

Prioritize reproducibility or replicability

Reproducibility quadrant
  • We should aim for reproducibility for methodological studies—i.e. if the analysis (i.e. the code) is public and the data is open.

  • Replicability can be the backup—if the same data can’t be used or to try other data with the same code.

  • Without either, generalizability is the default. Generalizability however means that research is:

    • Slower—it takes much more time and resources to reach consensus

    • Sensitive to local patterns—an observed trend may be sensitive to local conditions (data, geography, etc).

    • More inclined to novelty over robustness or production benefits—the breadth of options to consider is large but work is left on NSOs to operationalize.

Keep an eye on the how the research compendium helps structure the objects of research

Research objects
  • The input data is used to answer research questions. Output data is produced by your research project (such as final indices).

  • The broader (structured) digital object that tracks all the objects of research (code, doc, output, etc) is the research compendium:

    • The active version can be on GitHub.

    • Final version is archived.

  • Track software used in the computational environment file (no more “it works on my device”).

  • Store the source for your paper in your research compendium.

Adopt the pace that works for you

  • Apply the Reproducible Analytical Pipelines (RAP) framework to research.

    • It’s structured, robust, and is increasingly adopted!
  • Get started by:

    • Using open tools—i.e., R and python
    • Use git to version the project / research compendium.
    • Explain to users how to reproduce the project.
    • Make the compendium available (i.e. on GitHub).
    • If you are trying a methodology, try it on open data.
  • As you start—there is a 3 step maturity model that can be followed. Step 1 is very approachable.

Structure your research compendium

  • Add coherent structure to the compendium!

    • (input) data folders (even though you should ignore the data itself) make it clear how it runs end-to-end;

    • code is grouped;

    • document the process—from the project design to project documentation (such as the paper);

    • (output) data for data the code produces;

    • choose a licence to tell others how to use your materials;

    • include a file that lists your computational environment;

    • carefully track what packages are used;

Choose the most applicable input data for the project

  • There are 3 types of (input) data:

    • Open data—common types include Dominick’s or Turvey.

    • Proprietary data—commonly used ones include Nielsen and IRI.

    • Sensitive data—the internal data holdings in NSOs.

  • Open data is the default, but other data could be okay (next slide).

  • Irrespective of the data type, structure the research compendium in a standardized way and ignore the data files themselves (i.e. add *.csv to your .gitignore file).

Use open datasets if you can

  • We document discipline specific open datasets in the catalogue.

The catalogue documents each dataset in a standardized way

The datasets themselves still belong to the owners.

Ideally, each dataset is in a data repository (e.g., Zenodo)

You can still be reproducible if you work with sensitive data

  • If you still prefer to use proprietary or internal data, you can modify lightly to make your project reproducible.

    • Publish dummy files—to make sure the code runs end-to-end.
    • Create synthetic data—to ensure make a rich dataset that can then be used for research. Publish it separately from your compendium so that others can also use it.
    • Modify the original data—to help publicly release a copy of the data.
  • As with the previous guide, make sure to .gitignore the data itself!

When you are finalizing your paper, lean on metadata to cite things properly

  • Archive your research compendium (immutable copy tied to your submission).

  • Cite the objects of research using its findability metadata:

    • Find each object with its (permanent) unique digital identifier (typically a DOI).

    • Cite the unique identifier assigned to author (sign up to ORCID if you haven’t yet).

  • Metadata standards (i.e. SDMX/DDI/etc) typically applies to publishing new input data.

Are you convinced?

Takeaways—what you can do if you are convinced!

Great!

Goussev, S., Martin, S., Lamboray, C., Bontemps, C., Flower, T., Hillman, B., White, C., & Mehroff, J. (2026). Price Statistics Reproducibility Project (v0.1). Zenodo. https://doi.org/10.5281/zenodo.19779579

Not convinced?

Tell us what is in your way to research in a reproducible way?

https://forms.gle/DZRzqNbPHCfgdReLA

TO DO - figure out QR code!

What’s happening next?

What is on the docket for our workstream

  • Add more open data to the catalogue—ideally with your help! Add new datasets that can be used.
  • Investigate adding proprietary datasets to the catalogue—for those cases where open data is not enough.
  • Flesh out further guidance on topics that are most valuable.
  • Work with CPI EG and OG to instill reproducibitliy within the conference process.
  • Perhaps host a training session if it’s of value.

Questions?

More than anything—are you convinced?