We need the notion of convergent functions -- which may expose
convergent behavior to callers -- and convergent calls, which are calls
where we would like to preserve convergent behavior in the callee, if
possible.
In CUDA device mode, all calls and functions are convergent. The
optimizer can then strip this away under some circumstances.