10#include "gc/mmtk/mmtk.h"
12#include "ccan/list/list.h"
16#include <sys/sysctl.h>
25 size_t total_allocated_objects;
29 unsigned long long next_object_id;
35 struct ccan_list_head ractor_caches;
36 unsigned long live_ractor_cache_count;
38 pthread_mutex_t mutex;
40 pthread_cond_t cond_world_stopped;
41 pthread_cond_t cond_world_started;
42 size_t start_the_world_count;
44 struct rb_gc_vm_context vm_context;
48 struct ccan_list_node list_node;
50 MMTk_Mutator *mutator;
58 MMTK_FINAL_JOB_FINALIZE,
67 VALUE finalizer_array;
72#ifdef RB_THREAD_LOCAL_SPECIFIER
75# error We currently need language-supported TLS
83 rb_mmtk_gc_thread_tls = gc_thread_tls;
87rb_mmtk_is_mutator(
void)
93rb_mmtk_stop_the_world(
void)
98 if ((err = pthread_mutex_lock(&
objspace->mutex)) != 0) {
99 rb_bug(
"ERROR: cannot lock objspace->mutex: %s", strerror(err));
106 if ((err = pthread_mutex_unlock(&
objspace->mutex)) != 0) {
107 rb_bug(
"ERROR: cannot release objspace->mutex: %s", strerror(err));
112rb_mmtk_resume_mutators(
void)
117 if ((err = pthread_mutex_lock(&
objspace->mutex)) != 0) {
118 rb_bug(
"ERROR: cannot lock objspace->mutex: %s", strerror(err));
123 pthread_cond_broadcast(&
objspace->cond_world_started);
125 if ((err = pthread_mutex_unlock(&
objspace->mutex)) != 0) {
126 rb_bug(
"ERROR: cannot release objspace->mutex: %s", strerror(err));
135 size_t starting_gc_count =
objspace->gc_count;
136 int lock_lev = rb_gc_vm_lock();
138 if ((err = pthread_mutex_lock(&
objspace->mutex)) != 0) {
139 rb_bug(
"ERROR: cannot lock objspace->mutex: %s", strerror(err));
142 if (
objspace->gc_count == starting_gc_count) {
145 rb_gc_initialize_vm_context(&
objspace->vm_context);
147 mutator->gc_mutator_p =
true;
151 clock_gettime(CLOCK_MONOTONIC, &gc_start_time);
154 rb_gc_save_machine_context();
160 pthread_cond_broadcast(&
objspace->cond_world_stopped);
169 clock_gettime(CLOCK_MONOTONIC, &gc_end_time);
172 (gc_end_time.tv_sec - gc_start_time.tv_sec) * (1000 * 1000 * 1000) +
173 (gc_end_time.tv_nsec - gc_start_time.tv_nsec);
177 if ((err = pthread_mutex_unlock(&
objspace->mutex)) != 0) {
178 rb_bug(
"ERROR: cannot release objspace->mutex: %s", strerror(err));
180 rb_gc_vm_unlock(lock_lev);
184rb_mmtk_number_of_mutators(
void)
187 return objspace->live_ractor_cache_count;
191rb_mmtk_get_mutators(
void (*visit_mutator)(MMTk_Mutator *mutator,
void *data),
void *data)
198 ccan_list_for_each(&
objspace->ractor_caches, ractor_cache, list_node) {
199 visit_mutator(ractor_cache->mutator, data);
204rb_mmtk_scan_gc_roots(
void)
210pin_value(st_data_t key, st_data_t value, st_data_t data)
212 rb_gc_impl_mark_and_pin((
void *)data, (
VALUE)value);
218rb_mmtk_scan_objspace(
void)
222 if (
objspace->finalizer_table != NULL) {
226 st_foreach(
objspace->obj_to_id_tbl, gc_mark_tbl_no_pin_i, (st_data_t)
objspace);
229 while (job != NULL) {
231 case MMTK_FINAL_JOB_DFREE:
233 case MMTK_FINAL_JOB_FINALIZE:
234 rb_gc_impl_mark(
objspace, job->as.finalize.object_id);
235 rb_gc_impl_mark(
objspace, job->as.finalize.finalizer_array);
238 rb_bug(
"rb_mmtk_scan_objspace: unknown final job type %d", job->kind);
248 if (mutator->gc_mutator_p) {
251 rb_gc_worker_thread_set_vm_context(&
objspace->vm_context);
253 rb_gc_worker_thread_unset_vm_context(&
objspace->vm_context);
258rb_mmtk_scan_object_ruby_style(MMTk_ObjectReference
object)
260 rb_gc_mark_children(rb_gc_get_objspace(), (
VALUE)
object);
264rb_mmtk_call_gc_mark_children(MMTk_ObjectReference
object)
266 rb_gc_mark_children(rb_gc_get_objspace(), (
VALUE)
object);
270rb_mmtk_call_obj_free(MMTk_ObjectReference
object)
276 rb_gc_worker_thread_set_vm_context(&
objspace->vm_context);
278 rb_gc_worker_thread_unset_vm_context(&
objspace->vm_context);
285rb_mmtk_vm_live_bytes(
void)
294 RUBY_ASSERT(mmtk_is_reachable((MMTk_ObjectReference)table));
300 job->next =
objspace->finalizer_jobs;
301 job->kind = MMTK_FINAL_JOB_FINALIZE;
302 job->as.finalize.object_id = rb_obj_id((
VALUE)obj);
303 job->as.finalize.finalizer_array = table;
309rb_mmtk_update_finalizer_table_i(st_data_t key, st_data_t value, st_data_t data)
312 RUBY_ASSERT(mmtk_is_reachable((MMTk_ObjectReference)value));
317 if (!mmtk_is_reachable((MMTk_ObjectReference)key)) {
329rb_mmtk_update_finalizer_table(
void)
334 st_foreach(
objspace->finalizer_table, rb_mmtk_update_finalizer_table_i, (st_data_t)
objspace);
338rb_mmtk_update_table_i(
VALUE val,
void *data)
340 if (!mmtk_is_reachable((MMTk_ObjectReference)val)) {
348rb_mmtk_update_obj_id_tables_obj_to_id_i(st_data_t key, st_data_t val, st_data_t data)
352 if (!mmtk_is_reachable((MMTk_ObjectReference)key)) {
360rb_mmtk_update_obj_id_tables_id_to_obj_i(st_data_t key, st_data_t val, st_data_t data)
364 if (!mmtk_is_reachable((MMTk_ObjectReference)val)) {
372rb_mmtk_update_obj_id_tables(
void)
376 st_foreach(
objspace->obj_to_id_tbl, rb_mmtk_update_obj_id_tables_obj_to_id_i, 0);
377 st_foreach(
objspace->id_to_obj_tbl, rb_mmtk_update_obj_id_tables_id_to_obj_i, 0);
381rb_mmtk_global_tables_count(
void)
383 return RB_GC_VM_WEAK_TABLE_COUNT;
387rb_mmtk_update_global_tables(
int table)
391 rb_gc_vm_weak_table_foreach(rb_mmtk_update_table_i, NULL, NULL,
true, (
enum rb_gc_vm_weak_tables)table);
396 rb_mmtk_init_gc_worker_thread,
398 rb_mmtk_stop_the_world,
399 rb_mmtk_resume_mutators,
400 rb_mmtk_block_for_gc,
401 rb_mmtk_number_of_mutators,
402 rb_mmtk_get_mutators,
403 rb_mmtk_scan_gc_roots,
404 rb_mmtk_scan_objspace,
405 rb_mmtk_scan_roots_in_mutator_thread,
406 rb_mmtk_scan_object_ruby_style,
407 rb_mmtk_call_gc_mark_children,
408 rb_mmtk_call_obj_free,
409 rb_mmtk_vm_live_bytes,
410 rb_mmtk_update_global_tables,
411 rb_mmtk_global_tables_count,
412 rb_mmtk_update_finalizer_table,
413 rb_mmtk_update_obj_id_tables,
417#define RB_MMTK_HEAP_LIMIT_PERC 80
418#define RB_MMTK_DEFAULT_HEAP_MIN (1024 * 1024)
419#define RB_MMTK_DEFAULT_HEAP_MAX (rb_mmtk_system_physical_memory() / 100 * RB_MMTK_HEAP_LIMIT_PERC)
422 RB_MMTK_DYNAMIC_HEAP,
427rb_mmtk_builder_init(
void)
429 MMTk_Builder *builder = mmtk_builder_default();
434rb_gc_impl_objspace_alloc(
void)
436 MMTk_Builder *builder = rb_mmtk_builder_init();
437 mmtk_init_binding(builder, NULL, &ruby_upcalls, (MMTk_ObjectReference)
Qundef);
439 return calloc(1,
sizeof(
struct objspace));
443static void gc_run_finalizers(
void *data);
446rb_gc_impl_objspace_init(
void *objspace_ptr)
454 objspace->finalizer_table = st_init_numtable();
457 ccan_list_head_init(&
objspace->ractor_caches);
459 objspace->mutex = (pthread_mutex_t)PTHREAD_MUTEX_INITIALIZER;
460 objspace->cond_world_stopped = (pthread_cond_t)PTHREAD_COND_INITIALIZER;
461 objspace->cond_world_started = (pthread_cond_t)PTHREAD_COND_INITIALIZER;
465rb_gc_impl_objspace_free(
void *objspace_ptr)
471rb_gc_impl_ractor_cache_alloc(
void *objspace_ptr,
void *ractor)
474 if (
objspace->live_ractor_cache_count == 0) {
475 mmtk_initialize_collection(ractor);
477 objspace->live_ractor_cache_count++;
480 ccan_list_add(&
objspace->ractor_caches, &cache->list_node);
482 cache->mutator = mmtk_bind_mutator(cache);
488rb_gc_impl_ractor_cache_free(
void *objspace_ptr,
void *cache_ptr)
493 ccan_list_del(&cache->list_node);
496 objspace->live_ractor_cache_count--;
498 mmtk_destroy_mutator(cache->mutator);
501void rb_gc_impl_set_params(
void *objspace_ptr) { }
503static VALUE gc_verify_internal_consistency(
VALUE self) {
return Qnil; }
508 VALUE gc_constants = rb_hash_new();
510 rb_hash_aset(gc_constants,
ID2SYM(rb_intern(
"RVALUE_OVERHEAD")),
INT2NUM(0));
511 rb_hash_aset(gc_constants,
ID2SYM(rb_intern(
"RVARGC_MAX_ALLOCATE_SIZE")),
LONG2FIX(640));
513 rb_hash_aset(gc_constants,
ID2SYM(rb_intern(
"SIZE_POOL_COUNT")),
LONG2FIX(5));
515 rb_define_const(
rb_mGC,
"INTERNAL_CONSTANTS", gc_constants);
527static size_t heap_sizes[6] = {
528 40, 80, 160, 320, 640, 0
532rb_gc_impl_heap_sizes(
void *objspace_ptr)
538rb_mmtk_obj_free_iter_wrapper(
VALUE obj,
void *data)
543 rb_gc_obj_free_vm_weak_references(obj);
554rb_gc_impl_shutdown_free_objects(
void *objspace_ptr)
556 mmtk_set_gc_enabled(
false);
557 each_object(objspace_ptr, rb_mmtk_obj_free_iter_wrapper, objspace_ptr);
558 mmtk_set_gc_enabled(
true);
563rb_gc_impl_start(
void *objspace_ptr,
bool full_mark,
bool immediate_mark,
bool immediate_sweep,
bool compact)
565 bool enabled = mmtk_gc_enabled_p();
567 mmtk_set_gc_enabled(
true);
570 mmtk_handle_user_collection_request(rb_gc_get_ractor_newobj_cache());
573 mmtk_set_gc_enabled(
false);
578rb_gc_impl_during_gc_p(
void *objspace_ptr)
585rb_gc_impl_prepare_heap_i(MMTk_ObjectReference obj,
void *d)
587 rb_gc_prepare_heap_process_object((
VALUE)obj);
591rb_gc_impl_prepare_heap(
void *objspace_ptr)
593 mmtk_enumerate_objects(rb_gc_impl_prepare_heap_i, NULL);
597rb_gc_impl_gc_enable(
void *objspace_ptr)
599 mmtk_set_gc_enabled(
true);
603rb_gc_impl_gc_disable(
void *objspace_ptr,
bool finish_current_gc)
605 mmtk_set_gc_enabled(
false);
609rb_gc_impl_gc_enabled_p(
void *objspace_ptr)
611 return mmtk_gc_enabled_p();
615rb_gc_impl_stress_set(
void *objspace_ptr,
VALUE flag)
623rb_gc_impl_stress_get(
void *objspace_ptr)
631rb_gc_impl_config_get(
void *objspace_ptr)
633 VALUE hash = rb_hash_new();
638 size_t heap_min = mmtk_heap_min();
646rb_gc_impl_config_set(
void *objspace_ptr,
VALUE hash)
654rb_gc_impl_new_obj(
void *objspace_ptr,
void *cache_ptr,
VALUE klass,
VALUE flags,
VALUE v1,
VALUE v2,
VALUE v3,
bool wb_protected,
size_t alloc_size)
656#define MMTK_ALLOCATION_SEMANTICS_DEFAULT 0
660 if (alloc_size > 640) rb_bug(
"too big");
661 for (
int i = 0; i < 5; i++) {
662 if (alloc_size == heap_sizes[i])
break;
663 if (alloc_size < heap_sizes[i]) {
664 alloc_size = heap_sizes[i];
670 mmtk_handle_user_collection_request(ractor_cache);
673 VALUE *alloc_obj = mmtk_alloc(ractor_cache->mutator, alloc_size + 8, MMTk_MIN_OBJ_ALIGN, 0, MMTK_ALLOCATION_SEMANTICS_DEFAULT);
675 alloc_obj[-1] = alloc_size;
676 alloc_obj[0] = flags;
677 alloc_obj[1] = klass;
678 if (alloc_size > 16) alloc_obj[2] = v1;
679 if (alloc_size > 24) alloc_obj[3] = v2;
680 if (alloc_size > 32) alloc_obj[4] = v3;
682 mmtk_post_alloc(ractor_cache->mutator, (
void*)alloc_obj, alloc_size + 8, MMTK_ALLOCATION_SEMANTICS_DEFAULT);
685 mmtk_add_obj_free_candidate(alloc_obj);
687 objspace->total_allocated_objects++;
689 return (
VALUE)alloc_obj;
693rb_gc_impl_obj_slot_size(
VALUE obj)
695 return ((
VALUE *)obj)[-1];
699rb_gc_impl_heap_id_for_size(
void *objspace_ptr,
size_t size)
701 for (
int i = 0; i < 5; i++) {
702 if (size == heap_sizes[i])
return i;
703 if (size < heap_sizes[i])
return i;
706 rb_bug(
"size too big");
710rb_gc_impl_size_allocatable_p(
size_t size)
717rb_gc_impl_malloc(
void *objspace_ptr,
size_t size)
724rb_gc_impl_calloc(
void *objspace_ptr,
size_t size)
727 return calloc(1, size);
731rb_gc_impl_realloc(
void *objspace_ptr,
void *ptr,
size_t new_size,
size_t old_size)
734 return realloc(ptr, new_size);
738rb_gc_impl_free(
void *objspace_ptr,
void *ptr,
size_t old_size)
744void rb_gc_impl_adjust_memory_usage(
void *objspace_ptr, ssize_t diff) { }
748rb_gc_impl_mark(
void *objspace_ptr,
VALUE obj)
753 rb_mmtk_gc_thread_tls->gc_context,
754 (MMTk_ObjectReference)obj,
759rb_gc_impl_mark_and_move(
void *objspace_ptr,
VALUE *ptr)
764 rb_gc_impl_mark(objspace_ptr, *ptr);
768rb_gc_impl_mark_and_pin(
void *objspace_ptr,
VALUE obj)
773 rb_gc_impl_mark(objspace_ptr, obj);
777rb_gc_impl_mark_maybe(
void *objspace_ptr,
VALUE obj)
779 if (rb_gc_impl_pointer_to_heap_p(objspace_ptr, (
const void *)obj)) {
780 rb_gc_impl_mark_and_pin(objspace_ptr, obj);
785rb_gc_impl_mark_weak(
void *objspace_ptr,
VALUE *ptr)
787 mmtk_mark_weak((MMTk_ObjectReference *)ptr);
791rb_gc_impl_remove_weak(
void *objspace_ptr,
VALUE parent_obj,
VALUE *ptr)
793 mmtk_remove_weak((MMTk_ObjectReference *)ptr);
798rb_gc_impl_object_moved_p(
void *objspace_ptr,
VALUE obj)
800 rb_bug(
"unimplemented");
804rb_gc_impl_location(
void *objspace_ptr,
VALUE value)
806 rb_bug(
"unimplemented");
811rb_gc_impl_writebarrier(
void *objspace_ptr,
VALUE a,
VALUE b)
815 mmtk_object_reference_write_post(cache->mutator, (MMTk_ObjectReference)a);
819rb_gc_impl_writebarrier_unprotect(
void *objspace_ptr,
VALUE obj)
821 mmtk_register_wb_unprotected_object((MMTk_ObjectReference)obj);
825rb_gc_impl_writebarrier_remember(
void *objspace_ptr,
VALUE obj)
829 mmtk_object_reference_write_post(cache->mutator, (MMTk_ObjectReference)obj);
834each_objects_i(MMTk_ObjectReference obj,
void *d)
836 rb_darray(
VALUE) *objs = d;
838 rb_darray_append(objs, (
VALUE)obj);
844 rb_darray(
VALUE) objs;
845 rb_darray_make(&objs, 0);
847 mmtk_enumerate_objects(each_objects_i, &objs);
850 rb_darray_foreach(objs, i, obj_ptr) {
851 if (!mmtk_is_mmtk_object((MMTk_ObjectReference)*obj_ptr))
continue;
853 if (func(*obj_ptr, data) != 0) {
858 rb_darray_free(objs);
862 int (*func)(
void *,
void *, size_t,
void *);
867rb_gc_impl_each_objects_i(
VALUE obj,
void *d)
871 size_t slot_size = rb_gc_impl_obj_slot_size(obj);
873 return data->func((
void *)obj, (
void *)(obj + slot_size), slot_size, data->data);
877rb_gc_impl_each_objects(
void *objspace_ptr,
int (*func)(
void *,
void *,
size_t,
void *),
void *data)
884 each_object(objspace_ptr, rb_gc_impl_each_objects_i, &each_objects_data);
888 void (*func)(
VALUE,
void *);
893rb_gc_impl_each_object_i(
VALUE obj,
void *d)
897 data->func(obj, data->data);
903rb_gc_impl_each_object(
void *objspace_ptr,
void (*func)(
VALUE,
void *),
void *data)
910 each_object(objspace_ptr, rb_gc_impl_each_object_i, &each_object_data);
915gc_run_finalizers_get_final(
long i,
void *data)
923gc_run_finalizers(
void *data)
927 rb_gc_set_pending_interrupt();
929 while (
objspace->finalizer_jobs != NULL) {
931 objspace->finalizer_jobs = job->next;
934 case MMTK_FINAL_JOB_DFREE:
935 job->as.dfree.func(job->as.dfree.data);
937 case MMTK_FINAL_JOB_FINALIZE: {
938 VALUE object_id = job->as.finalize.object_id;
939 VALUE finalizer_array = job->as.finalize.finalizer_array;
941 rb_gc_run_obj_finalizer(
942 job->as.finalize.object_id,
944 gc_run_finalizers_get_final,
945 (
void *)finalizer_array
957 rb_gc_unset_pending_interrupt();
961rb_gc_impl_make_zombie(
void *objspace_ptr,
VALUE obj,
void (*dfree)(
void *),
void *data)
963 if (dfree == NULL)
return;
968 job->kind = MMTK_FINAL_JOB_DFREE;
969 job->as.dfree.func = dfree;
970 job->as.dfree.data = data;
974 job->next =
objspace->finalizer_jobs;
976 }
while (prev != job->next);
978 if (!ruby_free_at_exit_p()) {
984rb_gc_impl_define_finalizer(
void *objspace_ptr,
VALUE obj,
VALUE block)
992 if (st_lookup(
objspace->finalizer_table, obj, &data)) {
1000 for (i = 0; i <
len; i++) {
1008 rb_ary_push(table, block);
1013 st_add_direct(
objspace->finalizer_table, obj, table);
1020rb_gc_impl_undefine_finalizer(
void *objspace_ptr,
VALUE obj)
1024 st_data_t data = obj;
1025 st_delete(
objspace->finalizer_table, &data, 0);
1030rb_gc_impl_copy_finalizer(
void *objspace_ptr,
VALUE dest,
VALUE obj)
1038 if (RB_LIKELY(st_lookup(
objspace->finalizer_table, obj, &data))) {
1039 table = (
VALUE)data;
1040 st_insert(
objspace->finalizer_table, dest, table);
1044 rb_bug(
"rb_gc_copy_finalizer: FL_FINALIZE set but not found in finalizer_table: %s", rb_obj_info(obj));
1049move_finalizer_from_table_i(st_data_t key, st_data_t val, st_data_t arg)
1059rb_gc_impl_shutdown_call_finalizer(
void *objspace_ptr)
1063 while (
objspace->finalizer_table->num_entries) {
1064 st_foreach(
objspace->finalizer_table, move_finalizer_from_table_i, (st_data_t)
objspace);
1070 for (
size_t i = 0; i < registered_candidates.len; i++) {
1071 VALUE obj = (
VALUE)registered_candidates.ptr[i];
1073 if (rb_gc_shutdown_call_finalizer_p(obj)) {
1074 rb_gc_obj_free(objspace_ptr, obj);
1078 mmtk_free_raw_vec_of_obj_ref(registered_candidates);
1085object_id_cmp(st_data_t x, st_data_t y)
1088 return !rb_big_eql(x, y);
1096object_id_hash(st_data_t n)
1101#define OBJ_ID_INCREMENT (RUBY_IMMEDIATE_MASK + 1)
1102#define OBJ_ID_INITIAL (OBJ_ID_INCREMENT)
1104static const struct st_hash_type object_id_hash_type = {
1112 objspace->id_to_obj_tbl = st_init_table(&object_id_hash_type);
1113 objspace->obj_to_id_tbl = st_init_numtable();
1114 objspace->next_object_id = OBJ_ID_INITIAL;
1118rb_gc_impl_object_id(
void *objspace_ptr,
VALUE obj)
1122 unsigned int lev = rb_gc_vm_lock();
1125 if (st_lookup(
objspace->obj_to_id_tbl, (st_data_t)obj, &
id)) {
1132 objspace->next_object_id += OBJ_ID_INCREMENT;
1134 st_insert(
objspace->obj_to_id_tbl, (st_data_t)obj, (st_data_t)
id);
1135 st_insert(
objspace->id_to_obj_tbl, (st_data_t)
id, (st_data_t)obj);
1139 rb_gc_vm_unlock(lev);
1145rb_gc_impl_object_id_to_ref(
void *objspace_ptr,
VALUE object_id)
1150 if (st_lookup(
objspace->id_to_obj_tbl, object_id, &obj) &&
1151 !rb_gc_impl_garbage_object_p(
objspace, obj)) {
1166rb_gc_impl_before_fork(
void *objspace_ptr)
1172rb_gc_impl_after_fork(
void *objspace_ptr, rb_pid_t pid)
1174 mmtk_after_fork(rb_gc_get_ractor_newobj_cache());
1180rb_gc_impl_set_measure_total_time(
void *objspace_ptr,
VALUE flag)
1188rb_gc_impl_get_measure_total_time(
void *objspace_ptr)
1196rb_gc_impl_get_total_time(
void *objspace_ptr)
1204rb_gc_impl_gc_count(
void *objspace_ptr)
1212rb_gc_impl_latest_gc_info(
void *objspace_ptr,
VALUE hash_or_key)
1223 rb_bug(
"gc_info_decode: non-hash or symbol given");
1226#define SET(name, attr) \
1227 if (key == ID2SYM(rb_intern_const(#name))) \
1229 else if (hash != Qnil) \
1230 rb_hash_aset(hash, ID2SYM(rb_intern_const(#name)), (attr));
1248 gc_stat_sym_total_allocated_objects,
1249 gc_stat_sym_total_bytes,
1250 gc_stat_sym_used_bytes,
1251 gc_stat_sym_free_bytes,
1252 gc_stat_sym_starting_heap_address,
1253 gc_stat_sym_last_heap_address,
1257static VALUE gc_stat_symbols[gc_stat_sym_last];
1260setup_gc_stat_symbols(
void)
1262 if (gc_stat_symbols[0] == 0) {
1263#define S(s) gc_stat_symbols[gc_stat_sym_##s] = ID2SYM(rb_intern_const(#s))
1266 S(total_allocated_objects);
1270 S(starting_heap_address);
1271 S(last_heap_address);
1276rb_gc_impl_stat(
void *objspace_ptr,
VALUE hash_or_sym)
1281 setup_gc_stat_symbols();
1290 rb_bug(
"non-hash or symbol given");
1293#define SET(name, attr) \
1294 if (key == gc_stat_symbols[gc_stat_sym_##name]) \
1295 return SIZET2NUM(attr); \
1296 else if (hash != Qnil) \
1297 rb_hash_aset(hash, gc_stat_symbols[gc_stat_sym_##name], SIZET2NUM(attr));
1300 SET(time,
objspace->total_gc_time / (1000 * 1000));
1301 SET(total_allocated_objects,
objspace->total_allocated_objects);
1302 SET(total_bytes, mmtk_total_bytes());
1303 SET(used_bytes, mmtk_used_bytes());
1304 SET(free_bytes, mmtk_free_bytes());
1305 SET(starting_heap_address, (
size_t)mmtk_starting_heap_address());
1306 SET(last_heap_address, (
size_t)mmtk_last_heap_address());
1318rb_gc_impl_stat_heap(
void *objspace_ptr,
VALUE heap_name,
VALUE hash_or_sym)
1330#define RB_GC_OBJECT_METADATA_ENTRY_COUNT 0
1334rb_gc_impl_object_metadata(
void *objspace_ptr,
VALUE obj)
1336 object_metadata_entries[0].name = 0;
1337 object_metadata_entries[0].val = 0;
1339 return object_metadata_entries;
1343rb_gc_impl_pointer_to_heap_p(
void *objspace_ptr,
const void *ptr)
1345 if (ptr == NULL)
return false;
1346 if ((uintptr_t)ptr %
sizeof(
void*) != 0)
return false;
1347 return mmtk_is_mmtk_object((MMTk_Address)ptr);
1351rb_gc_impl_garbage_object_p(
void *objspace_ptr,
VALUE obj)
1356void rb_gc_impl_set_event_hook(
void *objspace_ptr,
const rb_event_flag_t event) { }
1359rb_gc_impl_copy_attributes(
void *objspace_ptr,
VALUE dest,
VALUE obj)
1361 if (mmtk_object_wb_unprotected_p((MMTk_ObjectReference)obj)) {
1362 rb_gc_impl_writebarrier_unprotect(objspace_ptr, dest);
1365 rb_gc_impl_copy_finalizer(objspace_ptr, dest, obj);
1371rb_gc_impl_active_gc_name(
void)
#define RUBY_ASSERT(...)
Asserts that the given expression is truthy if and only if RUBY_DEBUG is truthy.
#define RUBY_ATOMIC_PTR_CAS(var, oldval, newval)
Identical to RUBY_ATOMIC_CAS, except it expects its arguments are void*.
#define rb_define_singleton_method(klass, mid, func, arity)
Defines klass.mid.
unsigned int rb_postponed_job_handle_t
The type of a handle returned from rb_postponed_job_preregister and passed to rb_postponed_job_trigge...
void rb_postponed_job_trigger(rb_postponed_job_handle_t h)
Triggers a pre-registered job registered with rb_postponed_job_preregister, scheduling it for executi...
rb_postponed_job_handle_t rb_postponed_job_preregister(unsigned int flags, rb_postponed_job_func_t func, void *data)
Pre-registers a func in Ruby's postponed job preregistration table, returning an opaque handle which ...
#define RUBY_INTERNAL_EVENT_FREEOBJ
Object swept.
#define RUBY_INTERNAL_EVENT_GC_START
GC started.
uint32_t rb_event_flag_t
Represents event(s).
static VALUE RB_FL_TEST(VALUE obj, VALUE flags)
Tests if the given flag(s) are set or not.
static void RB_FL_UNSET(VALUE obj, VALUE flags)
Clears the given flag(s).
@ RUBY_FL_FINALIZE
This flag has something to do with finalisers.
#define xfree
Old name of ruby_xfree.
#define Qundef
Old name of RUBY_Qundef.
#define INT2FIX
Old name of RB_INT2FIX.
#define ID2SYM
Old name of RB_ID2SYM.
#define T_BIGNUM
Old name of RUBY_T_BIGNUM.
#define OBJ_FREEZE
Old name of RB_OBJ_FREEZE.
#define FL_SEEN_OBJ_ID
Old name of RUBY_FL_SEEN_OBJ_ID.
#define T_NONE
Old name of RUBY_T_NONE.
#define SIZET2NUM
Old name of RB_SIZE2NUM.
#define xmalloc
Old name of ruby_xmalloc.
#define LONG2FIX
Old name of RB_INT2FIX.
#define FL_FINALIZE
Old name of RUBY_FL_FINALIZE.
#define T_HASH
Old name of RUBY_T_HASH.
#define FL_SET
Old name of RB_FL_SET.
#define rb_ary_new3
Old name of rb_ary_new_from_args.
#define ULL2NUM
Old name of RB_ULL2NUM.
#define Qtrue
Old name of RUBY_Qtrue.
#define INT2NUM
Old name of RB_INT2NUM.
#define Qnil
Old name of RUBY_Qnil.
#define Qfalse
Old name of RUBY_Qfalse.
#define FIX2LONG
Old name of RB_FIX2LONG.
#define T_ARRAY
Old name of RUBY_T_ARRAY.
#define NIL_P
Old name of RB_NIL_P.
#define FL_TEST
Old name of RB_FL_TEST.
#define FL_UNSET
Old name of RB_FL_UNSET.
#define SYMBOL_P
Old name of RB_SYMBOL_P.
VALUE rb_eRangeError
RangeError exception.
VALUE rb_obj_hide(VALUE obj)
Make the object invisible from Ruby code.
VALUE rb_equal(VALUE lhs, VALUE rhs)
This function is an optimised version of calling #==.
VALUE rb_funcall(VALUE recv, ID mid, int n,...)
Calls a method.
#define rb_str_new_cstr(str)
Identical to rb_str_new, except it assumes the passed pointer is a pointer to a C string.
VALUE rb_f_notimplement(int argc, const VALUE *argv, VALUE obj, VALUE marker)
Raises rb_eNotImpError.
static ID rb_intern_const(const char *str)
This is a "tiny optimisation" over rb_intern().
int len
Length of the buffer.
#define RB_ULONG2NUM
Just another name of rb_ulong2num_inline.
#define RB_GC_GUARD(v)
Prevents premature destruction of local objects.
#define RARRAY_LEN
Just another name of rb_array_len.
#define RARRAY_AREF(a, i)
#define RBASIC(obj)
Convenient casting macro.
int ruby_native_thread_p(void)
Queries if the thread which calls this function is a ruby's thread.
static bool RB_SPECIAL_CONST_P(VALUE obj)
Checks if the given object is of enum ruby_special_consts.
#define RTEST
This is an old name of RB_TEST.
void * rust_closure
The pointer to the Rust-level closure object.
MMTk_ObjectClosureFunction c_function
The function to be called from C.
uintptr_t VALUE
Type that represents a Ruby object.
static enum ruby_value_type RB_BUILTIN_TYPE(VALUE obj)
Queries the type of the object.
static bool RB_TYPE_P(VALUE obj, enum ruby_value_type t)
Queries if the given object is of given type.