Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] What is the Powered effect in the results? #74

Open
LibiSC opened this issue Sep 24, 2022 · 7 comments
Open

[Question] What is the Powered effect in the results? #74

LibiSC opened this issue Sep 24, 2022 · 7 comments

Comments

@LibiSC
Copy link

LibiSC commented Sep 24, 2022

If you could tell me. I can't fully understand it from the code

@iampelle
Copy link
Contributor

If you have set the power to, say 80%, then if the true difference between control and treatment was equal to the powered effect, then you would detect that difference (I.e. get a statistically significant result) 80% of the time.

This has a couple of implications:

  • If the power is set to 50% and the measured effect equals the powered effect, then the lower confidence bound will be exactly zero, i.e. the powered effect is the smallest difference that will be statistically significant (but only when the power is set to 50%).
  • The higher you set the power, the larger the powered effect will be since higher power basically means that you want to be more certain that the null is rejected when the true difference equals a specific alternative hypothesis.

Hope this help!

@LibiSC
Copy link
Author

LibiSC commented Sep 25, 2022

Thanks for the explanation. To be sure I understand the practical use, so if abs(difference)>=powered_effect then I can say my test has achieved or surpassed the power set as objective right?

@jpzhangvincent
Copy link

Just curious, is there a documentation website for this package?

@iampelle
Copy link
Contributor

Thanks for the explanation. To be sure I understand the practical use, so if abs(difference)>=powered_effect then I can say my test has achieved or surpassed the power set as objective right?

Well, yes, but the usual way to use power is to use it when you, before launching the experiment, calculate the sample size. The power is one of the parameters that determine the required sample size, along with the significance level alpha, the minimum detectable effect size (MDE) and the variance of the metric in question.

The powered effect in the dataframe you get from the difference methods is the currently powered effect, based on the current sample size, which might be larger or smaller then the sample size required to power the MDE you set before launching the experiment.

Hope that clarifies it further!

@iampelle
Copy link
Contributor

Just curious, is there a documentation website for this package?

Not at the moment no. The only documentation we currently have is the doc strings and the example notebooks in the repo.

@jpzhangvincent
Copy link

jpzhangvincent commented Jan 10, 2023

It would be good to have a more comprehensive documentation website.

Another question, is the "non_inferiority_margin" equivalent to "minimum detectable effect size (MDE)"? Is it in absolute or relative sense? I didn't see clear documentation on that.

@iampelle
Copy link
Contributor

It would be good to have a more comprehensive documentation website.

Yes, totally agree! I will see if we can find time to improve the documentation.

Another question, is the "non_inferiority_margin" equivalent to "minimum detectable effect size (MDE)"? Is it in absolute or relative sense? I didn't see clear documentation on that.

MDE are used to say what size improvement you want to be able to detect. NIMs are used to say what the largest deterioration of a metric you can accept, e.g. "I want to be sure (e.g. with 99% confidence) that latency doesn't increase more than 5%". Bot are given in relative terms, relative to the average. So, for example, if you have average latency of 200ms and then providing a NIM of 0.05 means that you want to make sure it doesn't go above 210ms.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants