5 profiling_sample_count profiling_sample_count::record_counts() volatile {
7 profiling_sample_count returned(sample_count, gc_sample_count,
8 jit_sample_count, foreign_sample_count,
9 foreign_thread_sample_count);
10 atomic::fetch_subtract(&sample_count, returned.sample_count);
11 atomic::fetch_subtract(&gc_sample_count, returned.gc_sample_count);
12 atomic::fetch_subtract(&jit_sample_count, returned.jit_sample_count);
13 atomic::fetch_subtract(&foreign_sample_count, returned.foreign_sample_count);
14 atomic::fetch_subtract(&foreign_thread_sample_count,
15 returned.foreign_thread_sample_count);
19 void profiling_sample_count::clear() volatile {
23 foreign_sample_count = 0;
24 foreign_thread_sample_count = 0;
28 profiling_sample::profiling_sample(profiling_sample_count const& counts, cell thread,
29 cell callstack_begin, cell callstack_end)
30 : counts(counts), thread(thread),
31 callstack_begin(callstack_begin),
32 callstack_end(callstack_end) { }
34 void factor_vm::record_sample(bool prolog_p) {
35 profiling_sample_count counts = sample_counts.record_counts();
39 // Appends the callstack, which is just a sequence of quotation or
40 // word references, to sample_callstacks.
41 cell begin = sample_callstacks.size();
43 bool skip_p = prolog_p;
44 auto recorder = [&](cell frame_top, cell size, code_block* owner, cell addr) {
48 sample_callstacks.push_back(owner->owner);
50 iterate_callstack(ctx, recorder);
51 cell end = sample_callstacks.size();
54 cell thread = special_objects[OBJ_CURRENT_THREAD];
55 samples.push_back(profiling_sample(counts, thread, begin, end));
58 void factor_vm::set_sampling_profiler(fixnum rate) {
59 bool running_p = (atomic::load(&sampling_profiler_p) != 0);
60 if (rate > 0 && !running_p)
61 start_sampling_profiler(rate);
62 else if (rate == 0 && running_p)
63 end_sampling_profiler();
66 void factor_vm::start_sampling_profiler(fixnum rate) {
67 samples_per_second = rate;
68 sample_counts.clear();
69 // Release the memory consumed by collecting samples.
71 samples.shrink_to_fit();
72 sample_callstacks.clear();
73 sample_callstacks.shrink_to_fit();
75 samples.reserve(10 * rate);
76 sample_callstacks.reserve(100 * rate);
77 atomic::store(&sampling_profiler_p, true);
78 start_sampling_profiler_timer();
81 void factor_vm::end_sampling_profiler() {
82 atomic::store(&sampling_profiler_p, false);
83 end_sampling_profiler_timer();
87 void factor_vm::primitive_sampling_profiler() {
88 set_sampling_profiler(to_fixnum(ctx->pop()));
92 void factor_vm::primitive_get_samples() {
93 if (atomic::load(&sampling_profiler_p) || samples.empty()) {
94 ctx->push(false_object);
96 data_root<array> samples_array(allot_array(samples.size(), false_object),
98 std::vector<profiling_sample>::const_iterator from_iter = samples.begin();
101 for (; from_iter != samples.end(); ++from_iter, ++to_i) {
102 data_root<array> sample(allot_array(7, false_object), this);
104 set_array_nth(sample.untagged(), 0,
105 tag_fixnum(from_iter->counts.sample_count));
106 set_array_nth(sample.untagged(), 1,
107 tag_fixnum(from_iter->counts.gc_sample_count));
108 set_array_nth(sample.untagged(), 2,
109 tag_fixnum(from_iter->counts.jit_sample_count));
110 set_array_nth(sample.untagged(), 3,
111 tag_fixnum(from_iter->counts.foreign_sample_count));
112 set_array_nth(sample.untagged(), 4,
113 tag_fixnum(from_iter->counts.foreign_thread_sample_count));
115 set_array_nth(sample.untagged(), 5, from_iter->thread);
117 cell callstack_size =
118 from_iter->callstack_end - from_iter->callstack_begin;
119 data_root<array> callstack(allot_array(callstack_size, false_object),
122 std::vector<cell>::const_iterator callstacks_begin =
123 sample_callstacks.begin(),
126 from_iter->callstack_begin,
129 from_iter->callstack_end;
132 for (; c_from_iter != c_from_iter_end; ++c_from_iter, ++c_to_i)
133 set_array_nth(callstack.untagged(), c_to_i, *c_from_iter);
135 set_array_nth(sample.untagged(), 6, callstack.value());
137 set_array_nth(samples_array.untagged(), to_i, sample.value());
139 ctx->push(samples_array.value());