Artificial intelligent assistant

Showing that an increase in uncertainty is significant I have a linear model $y = ax+b$ and I estimate the coefficients $a$ and $b$ in the ordinary way. I have found out that all of my values of $x$ were systematically overestimated, and also that they have known uncertainties that were heretofore not propagated. Using total least squares (Deming regression), I find new estimates $\tilde a$ and $\tilde b$. These estimates have a much larger uncertainty ($15\%$ greater), but overlap with the previous estimates of $a$ and $b$. Since the distributions of the estimates overlap, a referee has said that the change is not significant, and has asked me if I can show that the increase in uncertainty is significant. How can I determine this? Also, due to regression dilution, the estimated central value of $\tilde a$ is larger than $a$. Is there a one-sided test I can use to show that the means are different, even if their distributions overlap? Any advice appreciated. Thanks!

I think you ca use the $F$ test for equivalence of variance. I.e., $H_0: \sigma^2_{a^1} /\sigma^2_{a^0} \le 1$, thus checking the estimated ratio where $$ \hat{\sigma}^2_{a^0} = \frac{MSE}{\sum(x_i - \bar x)^2} $$ against $1-\alpha$ quantaile of $F_{(n_1 -2, n_0 - 2)}$ will give you the answer.

xcX3v84RxoQ-4GxG32940ukFUIEgYdPy 3c0ac8a46acfeaba4f7632c78c4e1dc3