You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
codon_probs_of_parent_scaled_rates_and_sub_probs can be lifted from the neutral_codon notebook. We can keep this in this file for now, but eventually it may want to land in epam.molevol.
Objects
CodonProbBurrito
__init__ will just hand arguments off to Burrito superclass, but we'll want to do all of the trimming-to-codon-boundary stuff before doing so.
we will need branch length optimization functions.
find_optimal_branch_lengths prepares the data etc for branch length optimization
_find_optimal_branch_length does the per-branch work
note that the key for this function is to be able to define log_pcp_probability and then just hand it off to the generic optimize_branch_length routine. Just to get things going, let's have our log probability just be the probability of getting a codon mutation vs no.
Once we have a multihit model we will want to use the probability of getting mutations of various numbers of hits.
Eventually we'll want to have a multihit model in there, and fit it, but for now let's just do branch length optimization. As part of that next step, we'll also need to process the data into hit classes to be able to calculate loss.
CodonProbDataset
I think we'll need to have one of these, which is going to have some similarities to DNSMDataset in that it's going to have per-site SHM rates which we'll use to calculate things for our codon models. These will input into the init.
The text was updated successfully, but these errors were encountered:
Free functions
codon_probs_of_parent_scaled_rates_and_sub_probs
can be lifted from theneutral_codon
notebook. We can keep this in this file for now, but eventually it may want to land inepam.molevol
.Objects
CodonProbBurrito
__init__
will just hand arguments off toBurrito
superclass, but we'll want to do all of the trimming-to-codon-boundary stuff before doing so.find_optimal_branch_lengths
prepares the data etc for branch length optimization_find_optimal_branch_length
does the per-branch worklog_pcp_probability
and then just hand it off to the genericoptimize_branch_length
routine. Just to get things going, let's have our log probability just be the probability of getting a codon mutation vs no.Eventually we'll want to have a multihit model in there, and fit it, but for now let's just do branch length optimization. As part of that next step, we'll also need to process the data into hit classes to be able to calculate loss.
CodonProbDataset
I think we'll need to have one of these, which is going to have some similarities to
DNSMDataset
in that it's going to have per-site SHM rates which we'll use to calculate things for our codon models. These will input into the init.The text was updated successfully, but these errors were encountered: