6 inline static code_block_visitor<code_workhorse> make_code_visitor(factor_vm *parent)
8 return code_block_visitor<code_workhorse>(parent,code_workhorse(parent));
11 full_collector::full_collector(factor_vm *parent_) :
12 collector<tenured_space,full_policy>(
14 parent_->data->tenured,
15 full_policy(parent_)),
16 code_visitor(make_code_visitor(parent_)) {}
18 void full_collector::trace_code_block(code_block *compiled)
20 data_visitor.visit_code_block_objects(compiled);
21 data_visitor.visit_embedded_literals(compiled);
22 code_visitor.visit_embedded_code_pointers(compiled);
25 void full_collector::trace_context_code_blocks()
27 code_visitor.visit_context_code_blocks();
30 void full_collector::trace_object_code_block(object *obj)
32 code_visitor.visit_object_code_block(obj);
35 /* After a sweep, invalidate any code heap roots which are not marked,
36 so that if a block makes a tail call to a generic word, and the PIC
37 compiler triggers a GC, and the caller block gets gets GCd as a result,
38 the PIC code won't try to overwrite the call site */
39 void factor_vm::update_code_roots_for_sweep()
41 std::vector<code_root *>::const_iterator iter = code_roots.begin();
42 std::vector<code_root *>::const_iterator end = code_roots.end();
44 mark_bits<code_block> *state = &code->allocator->state;
46 for(; iter < end; iter++)
48 code_root *root = *iter;
49 code_block *block = (code_block *)(root->value & -data_alignment);
50 if(root->valid && !state->marked_p(block))
55 /* After a compaction, invalidate any code heap roots which are not
56 marked as above, and also slide the valid roots up so that call sites
57 can be updated correctly. */
58 void factor_vm::update_code_roots_for_compaction()
60 std::vector<code_root *>::const_iterator iter = code_roots.begin();
61 std::vector<code_root *>::const_iterator end = code_roots.end();
63 mark_bits<code_block> *state = &code->allocator->state;
65 for(; iter < end; iter++)
67 code_root *root = *iter;
68 code_block *block = (code_block *)(root->value & -data_alignment);
70 /* Offset of return address within 16-byte allocation line */
71 cell offset = root->value - (cell)block;
73 if(root->valid && state->marked_p((code_block *)root->value))
75 block = state->forward_block(block);
76 root->value = (cell)block + offset;
83 void factor_vm::collect_mark_impl(bool trace_contexts_p)
85 full_collector collector(this);
89 code->clear_mark_bits();
90 data->tenured->clear_mark_bits();
92 collector.trace_roots();
95 collector.trace_contexts();
96 collector.trace_context_code_blocks();
99 while(!mark_stack.empty())
101 cell ptr = mark_stack.back();
102 mark_stack.pop_back();
106 code_block *compiled = (code_block *)(ptr - 1);
107 collector.trace_code_block(compiled);
111 object *obj = (object *)ptr;
112 collector.trace_object(obj);
113 collector.trace_object_code_block(obj);
117 data->reset_generation(data->tenured);
118 data->reset_generation(data->aging);
119 data->reset_generation(&nursery);
120 code->clear_remembered_set();
123 void factor_vm::collect_sweep_impl()
125 current_gc->event->started_data_sweep();
126 data->tenured->sweep();
127 current_gc->event->ended_data_sweep();
129 update_code_roots_for_sweep();
131 current_gc->event->started_code_sweep();
132 code->allocator->sweep();
133 current_gc->event->ended_code_sweep();
136 void factor_vm::collect_full(bool trace_contexts_p)
138 collect_mark_impl(trace_contexts_p);
139 collect_sweep_impl();
140 if(data->low_memory_p())
142 current_gc->op = collect_compact_op;
143 current_gc->event->op = collect_compact_op;
144 collect_compact_impl(trace_contexts_p);
146 code->flush_icache();