-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Power Analysis Simulation comparison between optimal designs with zero priors and true priors #26
Comments
I haven't looked too closely at everything above, but if you're using priors in the creation of the design with |
That's precisely the issue I addressed in the update :) |
And you're still getting that the DB design has worse performance? That's certainly not what would be expected. |
I believe so! If you compare the plots in the update vs the original post, it seems like the SE has increased for any given N, for most attributes. The only difference was fixing my oversight between the two runs, so the updated plots are now with priors included in Edit: And that's just the comparison between runs. Within runs, it also seems like including priors in most cases increases SE vs N compared to zero priors. |
My only other thought as to why this may be occuring is because you've used restricted profiles. Does this still occur without restrictions? I'd like to see a simple example (maybe the apple one I use in the documentation) where a DB and DZ design are compared where the only differences are the priors used to simulate the choices. Btw, I believe you should be using the same priors to simulate choices for both designs. |
I don't know much about the structure of the apples dataset but I can definitely look into that! Currently, I'm comparing the power analysis simulation for: Are you suggesting that I should compare the following instead? |
Yes. If you optimize a DZ design, then it will perform best when the true parameters are zero. So your simulation where you use 0 for the priors of the simulated choices is as optimistic as it could get. That design will perform worse if the true parameters are not zero (which they most likely aren't). Instead, I tend to use a fixed set of parameters when simulation choices because those are the true parameters that I expect I will encounter when respondents complete my survey. I definitely don't expect they will have all 0s. So I would use the same priors when simulating choices regardless of the design. |
It seems fair to use zero priors but at least provide best-guesses for priors in choice simulation. When priors are provided, |
You know what, you're totally right! Your comment made me realize that the code I used before to make these figures was left over from a very long time ago where I was only checking to make sure that the standard errors fell with increasing N (just to visually see if things were working as expected). I never updated the code to plot the p values. Looking at standard errors is not really that informative - we should be looking at p-value vs. N to assess sample size requirements. This has been this way for so long now I'm embarrassed I missed it! Alright, I'm going to fix this. I think |
Frankly I'm glad I wasn't embarrassing myself by raising an irrelevant concern! Thank you for addressing this! Just a couple of thoughts:
|
By summary table I meant return all of the information in the summary tables. Right now I am thinking of re-structuring what is returned from Unfortunately classes are now starting for me, and I have a very heavy teaching load this fall. So this is probably all going to have to wait for the time being. I also use this package in the class I teach, so I'm hesitant to make considerable modifications so as to avoid introducing bugs. For sure all of this would be a much needed improvement to the package. In doing some preliminary changes to look at p-values vs. N I'm getting strange results where the p-values are all very high, even at high sample sizes. Good chance I just haven't got the code right yet though. |
EDIT: I made a mistake by not including the true priors in
cbc_choices
. I will not touch this post because I think it is still interesting to see howcbc_choices
behaves. I added the priors and ran the power analysis simulation again. The new results as well as the section of the code that was updated are in the second comment. In summary, it seems like "fixing" the code by including true priors, actually made the simulation results even more conservative.Background: I designed a DZ-Efficient DCE (Bayesian D-Efficient with zero priors) using {cbcTools}. Using
cbc_choices
andcbc_power
, I performed a power analysis simulation to determine an optimal sample size. Based on the output, it seemed like with about 180 participants I should be able to achieve standard errors of less than 0.05 for the model coefficients. Based on the recommended best practices in the literature, I conducted a pilot study to estimate the true priors, with 1/10th of the sample size (N=18). After analyzing the results, I found that almost all estimated coefficients had achieved very high statistical significance. For some, the p-value was <0.001, and the rest were less than 0.05, except for one factor that did not achieve significance (however, based on other indicators in the study, this factor truly did not impact the decisions of my respondents).The discrepancy between the simulation results and true results is understandable, since zero priors were identified for the optimization. With zero priors, cbc_choices has no way of knowing how the choices are likely to be made. As a result, the estimated impact of the sample size on the standard errors for coefficients is on the "very conservative" side. To explore this discrepancy further, I compared the results of power analysis simulation, with zero priors, and with the true priors from the pilot study.
I will be renaming some attributes and levels for simplification and readability, otherwise, everything else will remain true to the actual procedures and findings.
Study Parameters:
Attributes and Levels:
Cost ($20, $25, $30)
,Type (A, B)
,UserRatings (3.2 Stars, 4.0 Stars, 4.8 Stars)
,Information (Unavailable, Low, Medium, High)
Note:
Type
andInformation
are categorical attributes,Cost
andUserRatings
are numerical attributesChoice Matrix Shape: 16 choice tasks, each with 4 profiles (16x4)
No Choice Option: False
Pilot Results:
These coefficient estimates will be used as true priors.
Power Analysis Simulation Comparison (Results):
This is very interesting! My estimation was that including true priors will improve things drastically. Evidently, it only made it worse! Also noteworthy, the DB-Error also went from 0.34 (zero priors) to 0.54 (true priors)!
Power Analysis Simulation Comparison (Code):
The text was updated successfully, but these errors were encountered: