To the Editor P values are often misused and misinterpreted in the medical literature. Common mistakes include assuming P values measure the probability of a hypothesis being true (ie, indicate uncertainty) and quantify the strength of an observed effect (ie, conflating small P values with true effects). 1 As a solution, Tignanelli et al 2 strongly recommend the use of the Fragility Index (FI) and Fragility Quotient (FQ) in the reporting and appraisal of surgical randomized clinical trials (RCTs).While we agree that it is problematic to use a threshold P value of .05 in evaluating results of RCTs, the FI/FQ is not the solution for several reasons. First, as previously demonstrated, the FI almost perfectly correlates, albeit inversely, with the P value. 3 As a result, comparisons with larger P values (closer to .05) will be categorized as fragile while comparisons with smaller P values (eg, P < .001) will be deemed robust. This language is misleading, as it is widely accepted that P values should not be interpreted as a measure of the strength of an effect. 1 Any effect, no matter how small, can produce a small P value if the sample size is large enough, and large effects may produce unimpressive P values if the sample size is small. Similarly, the FI is a function of the sample size. 3 Randomized clinical trials with larger sample sizes are less likely to have a small FI. To address this issue, Tignanelli et al 2 present the FQ, which is obtained by dividing the FI by the sample size. 4 However, the authors provide no guidance for the interpretation of FQ values, which are less intuitive than the FI. Moreover, the FI is also correlated with the event rate of the outcome of interest. By virtue of this, results for outcomes with low event rates (eg, mortality) will be fragile, even when effects might be clinically meaningful. Lastly, the FI is at odds with the principles of RCT design. To conserve resources and minimize potential harm to patients, superiority RCTs are carefully designed to recruit the fewest patients necessary to detect a minimal clinically important difference. Considering these worthy objectives, it is unsurprising that many RCTs demonstrate fragility. Alternative solutions to improve the appraisal of surgical RCTs include the use of confidence intervals, as these may more directly indicate the size of an effect and its associated uncertainty, 1 and the interpretation of the observed differences in the context of absolute estimates and measures of clinical significance, such as minimal clinically important differences.