Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Formalise AD integration status, rewrite AD page correspondingly #595

Open
wants to merge 6 commits into
base: main
Choose a base branch
from

Conversation

penelopeysm
Copy link
Member

@penelopeysm penelopeysm commented Mar 28, 2025

This is an initial attempt to put down in words what @willtebbutt and I think our approach to integrating AD backends should be going forward.

@penelopeysm penelopeysm changed the title AD formalism Formalise AD integration status, rewrite AD page correspondingly Mar 28, 2025
@penelopeysm penelopeysm force-pushed the py/ad branch 2 times, most recently from 23469c6 to e594515 Compare March 29, 2025 17:02
Copy link
Contributor

Preview the changes: https://turinglang.org/docs/pr-previews/595
Please avoid using the search feature and navigation bar in PR previews!

Copy link
Member

@willtebbutt willtebbutt left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Some thoughts.

Comment on lines +76 to +78
**Tier 3** is the same as Tier 2, but in addition to that, we formally also take responsibility for ensuring that the backend works with Turing models.
If you submit an issue about using Turing with a Tier 3 library, we will actively try to make it work.
Realistically, this is only possible for AD backends that are actively maintained by somebody on the Turing team, such as Mooncake.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What are the limits to how far we're willing to take this? Per our discussion yesterday, if someone does something really non-differentiable (e.g. a custom ccall), we're not going to try and add support for their proposal.

Maybe "we will actively try to make it work" could be extended to say how we'll try to make it work? e.g. if someone encounters a bug, we'll fix it, but if they're doing something unusual we might suggest a more standard way to go about it that avoids the problem they're seeing entirely.

Comment on lines +192 to +214
Firstly, you could broaden the type of the container:

```{julia}
@model function forwarddiff_working1()
x = Real[0.0, 1.0]
a ~ Normal()
x[1] = a
b ~ MvNormal(x, I)
end
sample(forwarddiff_working1(), NUTS(; adtype=AutoForwardDiff()), 10)
```

Or, you can pass a type as a parameter to the model:

```{julia}
@model function forwarddiff_working2(::Type{T}=Float64) where T
x = T[0.0, 1.0]
a ~ Normal()
x[1] = a
b ~ MvNormal(x, I)
end
sample(forwarddiff_working2(), NUTS(; adtype=AutoForwardDiff()), 10)
```
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Would it be helpful for users to make it clear that the second option here is highly preferable to the first in general, and that the first should only be used if the second doesn't work for some reason?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, definitely, and a link to https://discourse.julialang.org/t/vector-real-vector-float64-methoderror/25926/5 might also be helpful (it helped me back in the day)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants