@@ -92,7 +92,7 @@ If set to `true`, record per-method-instance timings within type inference in th
92
92
__set_measure_typeinf (onoff:: Bool ) = __measure_typeinf__[] = onoff
93
93
const __measure_typeinf__ = RefValue {Bool} (false )
94
94
95
- function finish! (interp:: AbstractInterpreter , caller:: InferenceState )
95
+ function finish! (interp:: AbstractInterpreter , caller:: InferenceState , validation_world :: UInt )
96
96
result = caller. result
97
97
opt = result. src
98
98
if opt isa OptimizationState
@@ -108,12 +108,7 @@ function finish!(interp::AbstractInterpreter, caller::InferenceState)
108
108
ci = result. ci
109
109
# if we aren't cached, we don't need this edge
110
110
# but our caller might, so let's just make it anyways
111
- if last (result. valid_worlds) >= get_world_counter ()
112
- # TODO : this should probably come after all store_backedges (after optimizations) for the entire graph in finish_cycle
113
- # since we should be requiring that all edges first get their backedges set, as a batch
114
- result. valid_worlds = WorldRange (first (result. valid_worlds), typemax (UInt))
115
- end
116
- if last (result. valid_worlds) == typemax (UInt)
111
+ if last (result. valid_worlds) >= validation_world
117
112
# if we can record all of the backedges in the global reverse-cache,
118
113
# we can now widen our applicability in the global cache too
119
114
store_backedges (ci, edges)
@@ -202,7 +197,14 @@ function finish_nocycle(::AbstractInterpreter, frame::InferenceState)
202
197
if opt isa OptimizationState # implies `may_optimize(caller.interp) === true`
203
198
optimize (frame. interp, opt, frame. result)
204
199
end
205
- finish! (frame. interp, frame)
200
+ validation_world = get_world_counter ()
201
+ finish! (frame. interp, frame, validation_world)
202
+ if isdefined (frame. result, :ci )
203
+ # After validation, under the world_counter_lock, set max_world to typemax(UInt) for all dependencies
204
+ # (recursively). From that point onward the ordinary backedge mechanism is responsible for maintaining
205
+ # validity.
206
+ ccall (:jl_promote_ci_to_current , Cvoid, (Any, UInt), frame. result. ci, validation_world)
207
+ end
206
208
if frame. cycleid != 0
207
209
frames = frame. callstack:: Vector{AbsIntState}
208
210
@assert frames[end ] === frame
@@ -236,10 +238,19 @@ function finish_cycle(::AbstractInterpreter, frames::Vector{AbsIntState}, cyclei
236
238
optimize (caller. interp, opt, caller. result)
237
239
end
238
240
end
241
+ validation_world = get_world_counter ()
242
+ cis = CodeInstance[]
239
243
for frameid = cycleid: length (frames)
240
244
caller = frames[frameid]:: InferenceState
241
- finish! (caller. interp, caller)
245
+ finish! (caller. interp, caller, validation_world)
246
+ if isdefined (caller. result, :ci )
247
+ push! (cis, caller. result. ci)
248
+ end
242
249
end
250
+ # After validation, under the world_counter_lock, set max_world to typemax(UInt) for all dependencies
251
+ # (recursively). From that point onward the ordinary backedge mechanism is responsible for maintaining
252
+ # validity.
253
+ ccall (:jl_promote_cis_to_current , Cvoid, (Ptr{CodeInstance}, Csize_t, UInt), cis, length (cis), validation_world)
243
254
resize! (frames, cycleid - 1 )
244
255
return nothing
245
256
end
0 commit comments