Z-Curve.2.0

Update 11/22
A year went by, and we are still trying to publish Z-Curve.2.0. Psychological Methods is not interested in a method that can detect publication bias and shows how low power in psychological science is. Also, we learned that statistics journals couldn’t care less about psychology and their attempts to improve research practices. Our latest attempt will be a submission to Research Synthesis Methods, which just published a review on methods for the detection of publication bias (link). Wish us luck. However, even if we are lucky, z-curve.2.0 will have to play catch up with p-curve, which has become a popular tool to report confidence intervals of mean power based on the p-curve.4.06 app. Nobody seems to care that the performance of this app has never been evaluated. We thought it would be useful to do so and found that the 95% confidence interval does not even reach 50% coverage across a wide range of conditions. However, this is probably a feature rather than a bug for meta-analysists who want to claim that effects survive a bias-correction test, Power-Posing works!!!

For anybody interested in actual bias correction, here is the latest version of our z-curve.2.0 manuscript (pdf).

If you want to use z-curve, we recommend installing the r-package zcurve from a Cran server and run zcurve in R (https://cran.r-project.org/web/packages/zcurve/index.html).

Posted 01/20

The new year and decade just started and I am excited to announce that publication of a pre-print that introduces Z-Curve.2.0 (Bartoš & Schimmack, 2020; Preprint). The ms. “Z-Curve.2.0: Estimating Replication Rates and Discovery Rates” is the product of a nearly year-long collaboration with František Bartoš.

Last year, František emailed me to introduce a new way of estimating z-curves finite mixture models with an EM-algorithm. We started working together on evaluating the density approach from Brunner & Schimmack (2019) and the EM algorithm. In the end, the EM algorithm performs a bit better although the density approach leaves some wiggle room to improve coverage for confidence intervals. Both methods produce useful estimates and confidence intervals with good coverage.

The collaboration with František was amazing and provides another example of the power of social media. Not only does it allow fast exchange of ideas, it also makes it possible to collaborate with people you might never meet otherwise. Just like I met Rickard Carlsson in person only several years after we became Facebook friends and started a journal together, I still have to meet František in person (hopefully this year).

František also created an R-package for z-curve ( Z-Curve Package ). We are pleased to make this package publicly available. Please try it out and give feedback so that we can improve it before František submits it to the R-team as an official package that can be downloaded.

Here is the abstract of the ms. and a figure that was created with the zcurve package.

Abstract

This article introduces z-curve.2.0 as a method that estimates the expected replication rate and the expected discovery rate based on the test-statistics of studies selected for significance. Z-curve.2.0 extends the work by Brunner and Schimmack (2019) in several ways. First, we show that a new estimation method using expectation-maximization outperforms the kernel-density approach of z-curve.1.0. Second, we examine the coverage of bootstrapped confidence intervals to provide information about the uncertainty in z-curve estimates. Third, we extended z-curve to estimate the number of all studies that were conducted, including studies with non-significant results that may not have been reported, solely on the basis of significant results. This allows us to estimate the expected discovery rate (EDR); that is, the percentage of significant results that were obtained in all studies. EDR can be used to assess the size of the file-drawer, estimate the maximum number of false positive results, and may provide a better estimate of the success rate in actual replication studies than the expected replication rate because exact replications are impossible.


Keywords: Power estimation, Post-hoc power analysis, Publication bias, Expectancy Maximization, Z-curve, Replicability, Meta-Science, Expected Replication Rate, Expected Discovery Rate, False Discovery Rate, File-Drawer

Example -Figure created with graph.zcurve package.

Data are original test-statistics of 90 studies with good replication studies from the Open Science Collaboration(OSC) rep. project (OSC, 2015). Publication bias is indicated by the observed discovery rate (85/90 = 94% significant results) when the z-curve estimate of the expected discovery rate is only 39%. The expected replication rate of 62% successful replications is based on the assumption that studies can be replicated exactly. However, with contextual sensitivity, the expected discovery rate is a better estimate of the success rate in replication studies and it is more in line with the actual success rate (well, failure rate, really) in the OSC project.

12 thoughts on “Z-Curve.2.0

  1. Dear Uli

    As someone completely new to R, I was wondering wether there was some sort of guide on how to use Zcurve?

    Kind Regards,
    Dio

    1. After you install R or probably R-studio, you would really need to run only a few lines of syntax.

      install.packages(c(“zcurve”,”readxl”))
      #just select a close cran mirror from the pulldown menu
      name.of.excel.file = “
      name.of.directory = “

      setwd(name.of.directory)
      data = read.excel(name.of.excel.file)
      #assuming column name for the z-scores is z”
      zcurve(data$z)

Leave a Reply