]> gitweb.factorcode.org Git - factor.git/commitdiff
VM: Refactor gc* to Factor style
authorErik Charlebois <erikcharlebois@gmail.com>
Sun, 12 May 2013 02:04:08 +0000 (22:04 -0400)
committerErik Charlebois <erikcharlebois@gmail.com>
Sun, 12 May 2013 17:24:44 +0000 (13:24 -0400)
vm/gc.cpp
vm/gc.hpp
vm/gc_info.cpp
vm/gc_info.hpp

index 0e3d87836d1cffc8ce6c78b95a7fa1d8d458519a..8303741c88b11396ad79a41846cf35bb1e9c377d 100644 (file)
--- a/vm/gc.cpp
+++ b/vm/gc.cpp
 #include "master.hpp"
 
-namespace factor
-{
-
-gc_event::gc_event(gc_op op_, factor_vm *parent) :
-       op(op_),
-       cards_scanned(0),
-       decks_scanned(0),
-       code_blocks_scanned(0),
-       start_time(nano_count()),
-       card_scan_time(0),
-       code_scan_time(0),
-       data_sweep_time(0),
-       code_sweep_time(0),
-       compaction_time(0)
-{
-       data_heap_before = parent->data_room();
-       code_heap_before = parent->code_room();
-       start_time = nano_count();
+namespace factor {
+
+gc_event::gc_event(gc_op op_, factor_vm* parent)
+    : op(op_),
+      cards_scanned(0),
+      decks_scanned(0),
+      code_blocks_scanned(0),
+      start_time(nano_count()),
+      card_scan_time(0),
+      code_scan_time(0),
+      data_sweep_time(0),
+      code_sweep_time(0),
+      compaction_time(0) {
+  data_heap_before = parent->data_room();
+  code_heap_before = parent->code_room();
+  start_time = nano_count();
 }
 
-void gc_event::started_card_scan()
-{
-       temp_time = nano_count();
-}
+void gc_event::started_card_scan() { temp_time = nano_count(); }
 
-void gc_event::ended_card_scan(cell cards_scanned_, cell decks_scanned_)
-{
-       cards_scanned += cards_scanned_;
-       decks_scanned += decks_scanned_;
-       card_scan_time = (cell)(nano_count() - temp_time);
+void gc_event::ended_card_scan(cell cards_scanned_, cell decks_scanned_) {
+  cards_scanned += cards_scanned_;
+  decks_scanned += decks_scanned_;
+  card_scan_time = (cell)(nano_count() - temp_time);
 }
 
-void gc_event::started_code_scan()
-{
-       temp_time = nano_count();
-}
+void gc_event::started_code_scan() { temp_time = nano_count(); }
 
-void gc_event::ended_code_scan(cell code_blocks_scanned_)
-{
-       code_blocks_scanned += code_blocks_scanned_;
-       code_scan_time = (cell)(nano_count() - temp_time);
+void gc_event::ended_code_scan(cell code_blocks_scanned_) {
+  code_blocks_scanned += code_blocks_scanned_;
+  code_scan_time = (cell)(nano_count() - temp_time);
 }
 
-void gc_event::started_data_sweep()
-{
-       temp_time = nano_count();
-}
+void gc_event::started_data_sweep() { temp_time = nano_count(); }
 
-void gc_event::ended_data_sweep()
-{
-       data_sweep_time = (cell)(nano_count() - temp_time);
+void gc_event::ended_data_sweep() {
+  data_sweep_time = (cell)(nano_count() - temp_time);
 }
 
-void gc_event::started_code_sweep()
-{
-       temp_time = nano_count();
-}
+void gc_event::started_code_sweep() { temp_time = nano_count(); }
 
-void gc_event::ended_code_sweep()
-{
-       code_sweep_time = (cell)(nano_count() - temp_time);
+void gc_event::ended_code_sweep() {
+  code_sweep_time = (cell)(nano_count() - temp_time);
 }
 
-void gc_event::started_compaction()
-{
-       temp_time = nano_count();
-}
+void gc_event::started_compaction() { temp_time = nano_count(); }
 
-void gc_event::ended_compaction()
-{
-       compaction_time = (cell)(nano_count() - temp_time);
+void gc_event::ended_compaction() {
+  compaction_time = (cell)(nano_count() - temp_time);
 }
 
-void gc_event::ended_gc(factor_vm *parent)
-{
-       data_heap_after = parent->data_room();
-       code_heap_after = parent->code_room();
-       total_time = (cell)(nano_count() - start_time);
+void gc_event::ended_gc(factor_vm* parent) {
+  data_heap_after = parent->data_room();
+  code_heap_after = parent->code_room();
+  total_time = (cell)(nano_count() - start_time);
 }
 
-gc_state::gc_state(gc_op op_, factor_vm *parent) : op(op_)
-{
-       if(parent->gc_events)
-       {
-               event = new gc_event(op,parent);
-               start_time = nano_count();
-       }
-       else
-               event = NULL;
+gc_state::gc_state(gc_op op_, factor_vm* parent) : op(op_) {
+  if (parent->gc_events) {
+    event = new gc_event(op, parent);
+    start_time = nano_count();
+  } else
+    event = NULL;
 }
 
-gc_state::~gc_state()
-{
-       if(event)
-       {
-               delete event;
-               event = NULL;
-       }
+gc_state::~gc_state() {
+  if (event) {
+    delete event;
+    event = NULL;
+  }
 }
 
-void factor_vm::end_gc()
-{
-       if(gc_events)
-       {
-               current_gc->event->ended_gc(this);
-               gc_events->push_back(*current_gc->event);
-       }
+void factor_vm::end_gc() {
+  if (gc_events) {
+    current_gc->event->ended_gc(this);
+    gc_events->push_back(*current_gc->event);
+  }
 }
 
-void factor_vm::start_gc_again()
-{
-       end_gc();
-
-       switch(current_gc->op)
-       {
-       case collect_nursery_op:
-               /* Nursery collection can fail if aging does not have enough
-               free space to fit all live objects from nursery. */
-               current_gc->op = collect_aging_op;
-               break;
-       case collect_aging_op:
-               /* Aging collection can fail if the aging semispace cannot fit
-               all the live objects from the other aging semispace and the
-               nursery. */
-               current_gc->op = collect_to_tenured_op;
-               break;
-       default:
-               /* Nothing else should fail mid-collection due to insufficient
-               space in the target generation. */
-               critical_error("Bad GC op",current_gc->op);
-               break;
-       }
-
-       if(gc_events)
-               current_gc->event = new gc_event(current_gc->op,this);
+void factor_vm::start_gc_again() {
+  end_gc();
+
+  switch (current_gc->op) {
+    case collect_nursery_op:
+      /* Nursery collection can fail if aging does not have enough
+         free space to fit all live objects from nursery. */
+      current_gc->op = collect_aging_op;
+      break;
+    case collect_aging_op:
+      /* Aging collection can fail if the aging semispace cannot fit
+         all the live objects from the other aging semispace and the
+         nursery. */
+      current_gc->op = collect_to_tenured_op;
+      break;
+    default:
+      /* Nothing else should fail mid-collection due to insufficient
+         space in the target generation. */
+      critical_error("Bad GC op", current_gc->op);
+      break;
+  }
+
+  if (gc_events)
+    current_gc->event = new gc_event(current_gc->op, this);
 }
 
-void factor_vm::set_current_gc_op(gc_op op)
-{
-       current_gc->op = op;
-       if(gc_events) current_gc->event->op = op;
+void factor_vm::set_current_gc_op(gc_op op) {
+  current_gc->op = op;
+  if (gc_events)
+    current_gc->event->op = op;
 }
 
-void factor_vm::gc(gc_op op, cell requested_size, bool trace_contexts_p)
-{
-       FACTOR_ASSERT(!gc_off);
-       FACTOR_ASSERT(!current_gc);
-
-       /* Important invariant: tenured space must have enough contiguous free
-       space to fit the entire contents of the aging space and nursery. This is
-       because when doing a full collection, objects from younger generations
-       are promoted before any unreachable tenured objects are freed. */
-       FACTOR_ASSERT(!data->high_fragmentation_p());
-
-       current_gc = new gc_state(op,this);
-       atomic::store(&current_gc_p, true);
-
-       /* Keep trying to GC higher and higher generations until we don't run
-       out of space in the target generation. */
-       for(;;)
-       {
-               try
-               {
-                       if(gc_events) current_gc->event->op = current_gc->op;
-
-                       switch(current_gc->op)
-                       {
-                       case collect_nursery_op:
-                               collect_nursery();
-                               break;
-                       case collect_aging_op:
-                               /* We end up here if the above fails. */
-                               collect_aging();
-                               if(data->high_fragmentation_p())
-                               {
-                                       /* Change GC op so that if we fail again,
-                                       we crash. */
-                                       set_current_gc_op(collect_full_op);
-                                       collect_full(trace_contexts_p);
-                               }
-                               break;
-                       case collect_to_tenured_op:
-                               /* We end up here if the above fails. */
-                               collect_to_tenured();
-                               if(data->high_fragmentation_p())
-                               {
-                                       /* Change GC op so that if we fail again,
-                                       we crash. */
-                                       set_current_gc_op(collect_full_op);
-                                       collect_full(trace_contexts_p);
-                               }
-                               break;
-                       case collect_full_op:
-                               collect_full(trace_contexts_p);
-                               break;
-                       case collect_compact_op:
-                               collect_compact(trace_contexts_p);
-                               break;
-                       case collect_growing_heap_op:
-                               collect_growing_heap(requested_size,trace_contexts_p);
-                               break;
-                       default:
-                               critical_error("Bad GC op",current_gc->op);
-                               break;
-                       }
-
-                       break;
-               }
-               catch(const must_start_gc_again &)
-               {
-                       /* We come back here if the target generation is full. */
-                       start_gc_again();
-                       continue;
-               }
-       }
-
-       end_gc();
-
-       atomic::store(&current_gc_p, false);
-       delete current_gc;
-       current_gc = NULL;
-
-       /* Check the invariant again, just in case. */
-       FACTOR_ASSERT(!data->high_fragmentation_p());
+void factor_vm::gc(gc_op op, cell requested_size, bool trace_contexts_p) {
+  FACTOR_ASSERT(!gc_off);
+  FACTOR_ASSERT(!current_gc);
+
+  /* Important invariant: tenured space must have enough contiguous free
+     space to fit the entire contents of the aging space and nursery. This is
+     because when doing a full collection, objects from younger generations
+     are promoted before any unreachable tenured objects are freed. */
+  FACTOR_ASSERT(!data->high_fragmentation_p());
+
+  current_gc = new gc_state(op, this);
+  atomic::store(&current_gc_p, true);
+
+  /* Keep trying to GC higher and higher generations until we don't run
+     out of space in the target generation. */
+  for (;;) {
+    try {
+      if (gc_events)
+        current_gc->event->op = current_gc->op;
+
+      switch (current_gc->op) {
+        case collect_nursery_op:
+          collect_nursery();
+          break;
+        case collect_aging_op:
+          /* We end up here if the above fails. */
+          collect_aging();
+          if (data->high_fragmentation_p()) {
+            /* Change GC op so that if we fail again, we crash. */
+            set_current_gc_op(collect_full_op);
+            collect_full(trace_contexts_p);
+          }
+          break;
+        case collect_to_tenured_op:
+          /* We end up here if the above fails. */
+          collect_to_tenured();
+          if (data->high_fragmentation_p()) {
+            /* Change GC op so that if we fail again, we crash. */
+            set_current_gc_op(collect_full_op);
+            collect_full(trace_contexts_p);
+          }
+          break;
+        case collect_full_op:
+          collect_full(trace_contexts_p);
+          break;
+        case collect_compact_op:
+          collect_compact(trace_contexts_p);
+          break;
+        case collect_growing_heap_op:
+          collect_growing_heap(requested_size, trace_contexts_p);
+          break;
+        default:
+          critical_error("Bad GC op", current_gc->op);
+          break;
+      }
+
+      break;
+    }
+    catch (const must_start_gc_again&) {
+      /* We come back here if the target generation is full. */
+      start_gc_again();
+      continue;
+    }
+  }
+
+  end_gc();
+
+  atomic::store(&current_gc_p, false);
+  delete current_gc;
+  current_gc = NULL;
+
+  /* Check the invariant again, just in case. */
+  FACTOR_ASSERT(!data->high_fragmentation_p());
 }
 
 /* primitive_minor_gc() is invoked by inline GC checks, and it needs to fill in
-uninitialized stack locations before actually calling the GC. See the comment
-in compiler.cfg.stacks.uninitialized for details. */
+   uninitialized stack locations before actually calling the GC. See the
+   comment in compiler.cfg.stacks.uninitialized for details. */
 
 struct call_frame_scrubber {
-       factor_vm *parent;
-       context *ctx;
+  factor_vm* parent;
+  context* ctx;
 
-       explicit call_frame_scrubber(factor_vm *parent_, context *ctx_) :
-               parent(parent_), ctx(ctx_) {}
+  explicit call_frame_scrubber(factor_vm* parent_, context* ctx_)
+      : parent(parent_), ctx(ctx_) {}
 
-       void operator()(void *frame_top, cell frame_size, code_block *owner, void *addr)
-       {
-               cell return_address = owner->offset(addr);
+  void operator()(void* frame_top, cell frame_size, code_block* owner,
+                  void* addr) {
+    cell return_address = owner->offset(addr);
 
-               gc_info *info = owner->block_gc_info();
+    gc_info* info = owner->block_gc_info();
 
-               FACTOR_ASSERT(return_address < owner->size());
-               cell index = info->return_address_index(return_address);
-               if(index != (cell)-1)
-                       ctx->scrub_stacks(info,index);
-       }
+    FACTOR_ASSERT(return_address < owner->size());
+    cell index = info->return_address_index(return_address);
+    if (index != (cell) - 1)
+      ctx->scrub_stacks(info, index);
+  }
 };
 
-void factor_vm::scrub_context(context *ctx)
-{
-       call_frame_scrubber scrubber(this,ctx);
-       iterate_callstack(ctx,scrubber);
+void factor_vm::scrub_context(context* ctx) {
+  call_frame_scrubber scrubber(this, ctx);
+  iterate_callstack(ctx, scrubber);
 }
 
-void factor_vm::scrub_contexts()
-{
-       std::set<context *>::const_iterator begin = active_contexts.begin();
-       std::set<context *>::const_iterator end = active_contexts.end();
-       while(begin != end)
-       {
-               scrub_context(*begin);
-               begin++;
-       }
+void factor_vm::scrub_contexts() {
+  std::set<context*>::const_iterator begin = active_contexts.begin();
+  std::set<context*>::const_iterator end = active_contexts.end();
+  while (begin != end) {
+    scrub_context(*begin);
+    begin++;
+  }
 }
 
-void factor_vm::primitive_minor_gc()
-{
-       scrub_contexts();
+void factor_vm::primitive_minor_gc() {
+  scrub_contexts();
 
-       gc(collect_nursery_op,
-               0, /* requested size */
-               true /* trace contexts? */);
+  gc(collect_nursery_op, 0, /* requested size */
+     true /* trace contexts? */);
 }
 
-void factor_vm::primitive_full_gc()
-{
-       gc(collect_full_op,
-               0, /* requested size */
-               true /* trace contexts? */);
+void factor_vm::primitive_full_gc() {
+  gc(collect_full_op, 0, /* requested size */
+     true /* trace contexts? */);
 }
 
-void factor_vm::primitive_compact_gc()
-{
-       gc(collect_compact_op,
-               0, /* requested size */
-               true /* trace contexts? */);
+void factor_vm::primitive_compact_gc() {
+  gc(collect_compact_op, 0, /* requested size */
+     true /* trace contexts? */);
 }
 
 /*
@@ -295,66 +246,57 @@ void factor_vm::primitive_compact_gc()
  * fashion!
  */
 /* Allocates memory */
-object *factor_vm::allot_large_object(cell type, cell size)
-{
-       /* If tenured space does not have enough room, collect and compact */
-       cell requested_size = size + data->high_water_mark();
-       if(!data->tenured->can_allot_p(requested_size))
-       {
-               primitive_compact_gc();
-
-               /* If it still won't fit, grow the heap */
-               if(!data->tenured->can_allot_p(requested_size))
-               {
-                       gc(collect_growing_heap_op,
-                               size, /* requested size */
-                               true /* trace contexts? */);
-               }
-       }
-
-       object *obj = data->tenured->allot(size);
-
-       /* Allows initialization code to store old->new pointers
-       without hitting the write barrier in the common case of
-       a nursery allocation */
-       write_barrier(obj,size);
-
-       obj->initialize(type);
-       return obj;
+object* factor_vm::allot_large_object(cell type, cell size) {
+  /* If tenured space does not have enough room, collect and compact */
+  cell requested_size = size + data->high_water_mark();
+  if (!data->tenured->can_allot_p(requested_size)) {
+    primitive_compact_gc();
+
+    /* If it still won't fit, grow the heap */
+    if (!data->tenured->can_allot_p(requested_size)) {
+      gc(collect_growing_heap_op, size, /* requested size */
+         true /* trace contexts? */);
+    }
+  }
+
+  object* obj = data->tenured->allot(size);
+
+  /* Allows initialization code to store old->new pointers
+     without hitting the write barrier in the common case of
+     a nursery allocation */
+  write_barrier(obj, size);
+
+  obj->initialize(type);
+  return obj;
 }
 
-void factor_vm::primitive_enable_gc_events()
-{
-       gc_events = new std::vector<gc_event>();
+void factor_vm::primitive_enable_gc_events() {
+  gc_events = new std::vector<gc_event>();
 }
 
 /* Allocates memory */
-void factor_vm::primitive_disable_gc_events()
-{
-       if(gc_events)
-       {
-               growable_array result(this);
-
-               std::vector<gc_event> *gc_events = this->gc_events;
-               this->gc_events = NULL;
-
-               std::vector<gc_event>::const_iterator iter = gc_events->begin();
-               std::vector<gc_event>::const_iterator end = gc_events->end();
-
-               for(; iter != end; iter++)
-               {
-                       gc_event event = *iter;
-                       byte_array *obj = byte_array_from_value(&event);
-                       result.add(tag<byte_array>(obj));
-               }
-
-               result.trim();
-               ctx->push(result.elements.value());
-
-               delete this->gc_events;
-       }
-       else
-               ctx->push(false_object);
+void factor_vm::primitive_disable_gc_events() {
+  if (gc_events) {
+    growable_array result(this);
+
+    std::vector<gc_event>* gc_events = this->gc_events;
+    this->gc_events = NULL;
+
+    std::vector<gc_event>::const_iterator iter = gc_events->begin();
+    std::vector<gc_event>::const_iterator end = gc_events->end();
+
+    for (; iter != end; iter++) {
+      gc_event event = *iter;
+      byte_array* obj = byte_array_from_value(&event);
+      result.add(tag<byte_array>(obj));
+    }
+
+    result.trim();
+    ctx->push(result.elements.value());
+
+    delete this->gc_events;
+  } else
+    ctx->push(false_object);
 }
 
 }
index 76029d81ee851ab4b499eafc35822bd77c93d2e7..58b0a033054243037e232fd295335dbb1d6685ab 100644 (file)
--- a/vm/gc.hpp
+++ b/vm/gc.hpp
@@ -1,55 +1,54 @@
-namespace factor
-{
+namespace factor {
 
 enum gc_op {
-       collect_nursery_op,
-       collect_aging_op,
-       collect_to_tenured_op,
-       collect_full_op,
-       collect_compact_op,
-       collect_growing_heap_op
+  collect_nursery_op,
+  collect_aging_op,
+  collect_to_tenured_op,
+  collect_full_op,
+  collect_compact_op,
+  collect_growing_heap_op
 };
 
 struct gc_event {
-       gc_op op;
-       data_heap_room data_heap_before;
-       code_heap_room code_heap_before;
-       data_heap_room data_heap_after;
-       code_heap_room code_heap_after;
-       cell cards_scanned;
-       cell decks_scanned;
-       cell code_blocks_scanned;
-       u64 start_time;
-       cell total_time;
-       cell card_scan_time;
-       cell code_scan_time;
-       cell data_sweep_time;
-       cell code_sweep_time;
-       cell compaction_time;
-       u64 temp_time;
+  gc_op op;
+  data_heap_room data_heap_before;
+  code_heap_room code_heap_before;
+  data_heap_room data_heap_after;
+  code_heap_room code_heap_after;
+  cell cards_scanned;
+  cell decks_scanned;
+  cell code_blocks_scanned;
+  u64 start_time;
+  cell total_time;
+  cell card_scan_time;
+  cell code_scan_time;
+  cell data_sweep_time;
+  cell code_sweep_time;
+  cell compaction_time;
+  u64 temp_time;
 
-       gc_event(gc_op op_, factor_vm *parent);
-       void started_card_scan();
-       void ended_card_scan(cell cards_scanned_, cell decks_scanned_);
-       void started_code_scan();
-       void ended_code_scan(cell code_blocks_scanned_);
-       void started_data_sweep();
-       void ended_data_sweep();
-       void started_code_sweep();
-       void ended_code_sweep();
-       void started_compaction();
-       void ended_compaction();
-       void ended_gc(factor_vm *parent);
+  gc_event(gc_op op_, factor_vm* parent);
+  void started_card_scan();
+  void ended_card_scan(cell cards_scanned_, cell decks_scanned_);
+  void started_code_scan();
+  void ended_code_scan(cell code_blocks_scanned_);
+  void started_data_sweep();
+  void ended_data_sweep();
+  void started_code_sweep();
+  void ended_code_sweep();
+  void started_compaction();
+  void ended_compaction();
+  void ended_gc(factor_vm* parent);
 };
 
 struct gc_state {
-       gc_op op;
-       u64 start_time;
-       gc_event *event;
+  gc_op op;
+  u64 start_time;
+  gc_event* event;
 
-       explicit gc_state(gc_op op_, factor_vm *parent);
-       ~gc_state();
-       void start_again(gc_op op_, factor_vm *parent);
+  explicit gc_state(gc_op op_, factor_vm* parent);
+  ~gc_state();
+  void start_again(gc_op op_, factor_vm* parent);
 };
 
 }
index a693fc54557f39a373edb008c827e66a907bcb1b..332bbf63bd08855872967f4e978405d8014f44ae 100644 (file)
@@ -1,19 +1,16 @@
 #include "master.hpp"
 
-namespace factor
-{
+namespace factor {
 
-cell gc_info::return_address_index(cell return_address)
-{
-       u32 *return_address_array = return_addresses();
+cell gc_info::return_address_index(cell return_address) {
+  u32* return_address_array = return_addresses();
 
-       for(cell i = 0; i < return_address_count; i++)
-       {
-               if(return_address == return_address_array[i])
-                       return i;
-       }
+  for (cell i = 0; i < return_address_count; i++) {
+    if (return_address == return_address_array[i])
+      return i;
+  }
 
-       return (cell)-1;
+  return (cell) - 1;
 }
 
 }
index 9bff88b9b2882b67da956d3a983d975233a6886e..4dfac83ce5a6b35dfcde3f31d0386b0ef688efb8 100644 (file)
@@ -1,67 +1,48 @@
-namespace factor
-{
+namespace factor {
 
 struct gc_info {
-       u32 scrub_d_count;
-       u32 scrub_r_count;
-       u32 gc_root_count;
-       u32 derived_root_count;
-       u32 return_address_count;
+  u32 scrub_d_count;
+  u32 scrub_r_count;
+  u32 gc_root_count;
+  u32 derived_root_count;
+  u32 return_address_count;
 
-       cell callsite_bitmap_size()
-       {
-               return scrub_d_count + scrub_r_count + gc_root_count;
-       }
+  cell callsite_bitmap_size() {
+    return scrub_d_count + scrub_r_count + gc_root_count;
+  }
 
-       cell total_bitmap_size()
-       {
-               return return_address_count * callsite_bitmap_size();
-       }
+  cell total_bitmap_size() {
+    return return_address_count * callsite_bitmap_size();
+  }
 
-       cell total_bitmap_bytes()
-       {
-               return ((total_bitmap_size() + 7) / 8);
-       }
+  cell total_bitmap_bytes() { return ((total_bitmap_size() + 7) / 8); }
 
-       u32 *return_addresses()
-       {
-               return (u32 *)this - return_address_count;
-       }
+  u32* return_addresses() { return (u32*)this - return_address_count; }
 
-       u32 *base_pointer_map()
-       {
-               return return_addresses() - return_address_count * derived_root_count;
-       }
+  u32* base_pointer_map() {
+    return return_addresses() - return_address_count * derived_root_count;
+  }
 
-       u8 *gc_info_bitmap()
-       {
-               return (u8 *)base_pointer_map() - total_bitmap_bytes();
-       }
+  u8* gc_info_bitmap() {
+    return (u8*)base_pointer_map() - total_bitmap_bytes();
+  }
 
-       cell callsite_scrub_d(cell index)
-       {
-               return index * scrub_d_count;
-       }
+  cell callsite_scrub_d(cell index) { return index * scrub_d_count; }
 
-       cell callsite_scrub_r(cell index)
-       {
-               return return_address_count * scrub_d_count +
-                       index * scrub_r_count;
-       }
+  cell callsite_scrub_r(cell index) {
+    return return_address_count * scrub_d_count + index * scrub_r_count;
+  }
 
-       cell callsite_gc_roots(cell index)
-       {
-               return return_address_count * scrub_d_count
-                       + return_address_count * scrub_r_count
-                       + index * gc_root_count;
-       }
+  cell callsite_gc_roots(cell index) {
+    return return_address_count * scrub_d_count +
+           return_address_count * scrub_r_count + index * gc_root_count;
+  }
 
-       u32 lookup_base_pointer(cell index, cell derived_root)
-       {
-               return base_pointer_map()[index * derived_root_count + derived_root];
-       }
+  u32 lookup_base_pointer(cell index, cell derived_root) {
+    return base_pointer_map()[index * derived_root_count + derived_root];
+  }
 
-       cell return_address_index(cell return_address);
+  cell return_address_index(cell return_address);
 };
 
 }