diff options
Diffstat (limited to 'variable.c')
| -rw-r--r-- | variable.c | 4859 |
1 files changed, 4226 insertions, 633 deletions
diff --git a/variable.c b/variable.c index c05853899e..ff8d24d78a 100644 --- a/variable.c +++ b/variable.c @@ -1,1070 +1,4663 @@ -/************************************************ +/********************************************************************** variable.c - $Author$ - $Date$ created at: Tue Apr 19 23:55:15 JST 1994 -************************************************/ + Copyright (C) 1993-2007 Yukihiro Matsumoto + Copyright (C) 2000 Network Applied Communication Laboratory, Inc. + Copyright (C) 2000 Information-technology Promotion Agency, Japan + +**********************************************************************/ + +#include "ruby/internal/config.h" +#include <stddef.h> +#include "ruby/internal/stdbool.h" +#include "ccan/list/list.h" +#include "constant.h" +#include "debug_counter.h" +#include "id.h" +#include "id_table.h" +#include "internal.h" +#include "internal/box.h" +#include "internal/class.h" +#include "internal/compilers.h" +#include "internal/error.h" +#include "internal/eval.h" +#include "internal/hash.h" +#include "internal/object.h" +#include "internal/gc.h" +#include "internal/re.h" +#include "internal/struct.h" +#include "internal/symbol.h" +#include "internal/thread.h" +#include "internal/variable.h" +#include "ruby/encoding.h" +#include "ruby/st.h" +#include "ruby/util.h" +#include "shape.h" +#include "symbol.h" +#include "variable.h" +#include "vm_core.h" +#include "ractor_core.h" +#include "vm_sync.h" + +RUBY_EXTERN rb_serial_t ruby_vm_global_cvar_state; +#define GET_GLOBAL_CVAR_STATE() (ruby_vm_global_cvar_state) + +typedef void rb_gvar_compact_t(void *var); + +static struct rb_id_table *rb_global_tbl; +static ID autoload; + +// This hash table maps file paths to loadable features. We use this to track +// autoload state until it's no longer needed. +// feature (file path) => struct autoload_data +static VALUE autoload_features; + +// This mutex is used to protect autoloading state. We use a global mutex which +// is held until a per-feature mutex can be created. This ensures there are no +// race conditions relating to autoload state. +static VALUE autoload_mutex; + +static void check_before_mod_set(VALUE, ID, VALUE, const char *); +static void setup_const_entry(rb_const_entry_t *, VALUE, VALUE, rb_const_flag_t); +static VALUE rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in); +static st_table *generic_fields_tbl_; + +typedef int rb_ivar_foreach_callback_func(ID key, VALUE val, st_data_t arg); +static void rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only); -#include "ruby.h" -#include "env.h" -#include "node.h" -#include "st.h" +void +Init_var_tables(void) +{ + rb_global_tbl = rb_id_table_create(0); + generic_fields_tbl_ = st_init_numtable(); + autoload = rb_intern_const("__autoload__"); -#ifdef USE_CWGUSI -char* strdup(char*); -#endif + autoload_mutex = rb_mutex_new(); + rb_obj_hide(autoload_mutex); + rb_vm_register_global_object(autoload_mutex); -static st_table *rb_global_tbl; -st_table *rb_class_tbl; -#define global_tbl rb_global_tbl -#define class_tbl rb_class_tbl + autoload_features = rb_ident_hash_new(); + rb_obj_hide(autoload_features); + rb_vm_register_global_object(autoload_features); +} -st_table* -new_idhash() +static inline bool +rb_namespace_p(VALUE obj) { - return st_init_numtable(); + if (RB_SPECIAL_CONST_P(obj)) return false; + switch (RB_BUILTIN_TYPE(obj)) { + case T_MODULE: case T_CLASS: return true; + default: break; + } + return false; } -void -Init_var_tables() +/** + * Returns +classpath+ of _klass_, if it is named, or +nil+ for + * anonymous +class+/+module+. A named +classpath+ may contain + * an anonymous component, but the last component is guaranteed + * to not be anonymous. <code>*permanent</code> is set to 1 + * if +classpath+ has no anonymous components. There is no builtin + * Ruby level APIs that can change a permanent +classpath+. + * + * YJIT needs this function to not allocate. + */ +static VALUE +classname(VALUE klass, bool *permanent) { - global_tbl = new_idhash(); - class_tbl = new_idhash(); + *permanent = false; + + VALUE classpath = RCLASS_CLASSPATH(klass); + if (classpath == 0) return Qnil; + + *permanent = RCLASS_PERMANENT_CLASSPATH_P(klass); + + return classpath; } -struct fc_result { - ID name; - VALUE klass; - VALUE path; - VALUE track; - struct fc_result *prev; -}; +VALUE +rb_mod_name0(VALUE klass, bool *permanent) +{ + return classname(klass, permanent); +} -static int -fc_i(key, value, res) - ID key; - VALUE value; - struct fc_result *res; +/* + * call-seq: + * mod.name -> string or nil + * + * Returns the name of the module <i>mod</i>. Returns +nil+ for anonymous modules. + */ + +VALUE +rb_mod_name(VALUE mod) { - VALUE path; - char *name; - - if (!rb_is_const_id(key)) return ST_CONTINUE; + // YJIT needs this function to not allocate. + bool permanent; + return classname(mod, &permanent); +} - name = rb_id2name(key); - if (res->path) { - path = str_dup(res->path); - str_cat(path, "::", 2); - str_cat(path, name, strlen(name)); +// Similar to logic in rb_mod_const_get(). +static bool +is_constant_path(VALUE name) +{ + const char *path = RSTRING_PTR(name); + const char *pend = RSTRING_END(name); + rb_encoding *enc = rb_enc_get(name); + + const char *p = path; + + if (p >= pend || !*p) { + return false; } - else { - path = str_new2(name); - } - if (value == res->klass) { - res->name = key; - res->path = path; - return ST_STOP; - } - if (obj_is_kind_of(value, cModule)) { - struct fc_result arg; - struct fc_result *list; - - - if (!RCLASS(value)->iv_tbl) return ST_CONTINUE; - - list = res; - while (list) { - if (list->track == value) return ST_CONTINUE; - list = list->prev; - } - - arg.name = 0; - arg.path = path; - arg.klass = res->klass; - arg.track = value; - arg.prev = res; - st_foreach(RCLASS(value)->iv_tbl, fc_i, &arg); - if (arg.name) { - res->name = arg.name; - res->path = arg.path; - return ST_STOP; - } + + while (p < pend) { + if (p + 2 <= pend && p[0] == ':' && p[1] == ':') { + p += 2; + } + + const char *pbeg = p; + while (p < pend && *p != ':') p++; + + if (pbeg == p) return false; + + if (rb_enc_symname_type(pbeg, p - pbeg, enc, 0) != ID_CONST) { + return false; + } } - return ST_CONTINUE; + + return true; } +struct sub_temporary_name_args { + VALUE names; + ID last; +}; + +static VALUE build_const_path(VALUE head, ID tail); +static void set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name); + static VALUE -find_class_path(klass) - VALUE klass; +set_sub_temporary_name_recursive(VALUE mod, VALUE data, int recursive) { - struct fc_result arg; + if (recursive) return Qfalse; - arg.name = 0; - arg.path = 0; - arg.klass = klass; - arg.track = cObject; - arg.prev = 0; - if (RCLASS(cObject)->iv_tbl) { - st_foreach(RCLASS(cObject)->iv_tbl, fc_i, &arg); - } - if (arg.name == 0) { - st_foreach(class_tbl, fc_i, &arg); - } - if (arg.name) { - rb_iv_set(klass, "__classpath__", arg.path); - return arg.path; + struct sub_temporary_name_args *args = (void *)data; + VALUE name = 0; + if (args->names) { + name = build_const_path(rb_ary_last(0, 0, args->names), args->last); } - return Qnil; + set_sub_temporary_name_foreach(mod, args, name); + return Qtrue; } static VALUE -classname(klass) - VALUE klass; +set_sub_temporary_name_topmost(VALUE mod, VALUE data, int recursive) { - VALUE path; + if (recursive) return Qfalse; - while (TYPE(klass) == T_ICLASS || FL_TEST(klass, FL_SINGLETON)) { - klass = (VALUE)RCLASS(klass)->super; + struct sub_temporary_name_args *args = (void *)data; + VALUE name = args->names; + if (name) { + args->names = rb_ary_hidden_new(0); } - path = rb_iv_get(klass, "__classpath__"); - if (NIL_P(path)) { - ID classid = rb_intern("__classid__"); + set_sub_temporary_name_foreach(mod, args, name); + return Qtrue; +} - path = rb_ivar_get(klass, classid); - if (!NIL_P(path)) { - path = str_new2(rb_id2name(FIX2INT(path))); - rb_ivar_set(klass, classid, path); - st_delete(RCLASS(klass)->iv_tbl, &classid, 0); - } - } - if (NIL_P(path)) { - path = find_class_path(klass); - if (NIL_P(path)) { - return 0; - } - return path; +static enum rb_id_table_iterator_result +set_sub_temporary_name_i(ID id, VALUE val, void *data) +{ + val = ((rb_const_entry_t *)val)->value; + if (rb_namespace_p(val) && !RCLASS_PERMANENT_CLASSPATH_P(val)) { + VALUE arg = (VALUE)data; + struct sub_temporary_name_args *args = data; + args->last = id; + rb_exec_recursive_paired(set_sub_temporary_name_recursive, val, arg, arg); } - if (TYPE(path) != T_STRING) Bug("class path is not set properly"); - return path; + return ID_TABLE_CONTINUE; } -VALUE -mod_name(mod) - VALUE mod; +static void +set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name) { - VALUE path = classname(mod); + RCLASS_WRITE_CLASSPATH(mod, name, FALSE); + struct rb_id_table *tbl = RCLASS_CONST_TBL(mod); + if (!tbl) return; + if (!name) { + rb_id_table_foreach(tbl, set_sub_temporary_name_i, args); + } + else { + long names_len = RARRAY_LEN(args->names); // paranoiac check? + rb_ary_push(args->names, name); + rb_id_table_foreach(tbl, set_sub_temporary_name_i, args); + rb_ary_set_len(args->names, names_len); + } +} - if (path) return str_dup(path); - return str_new(0,0); +static void +set_sub_temporary_name(VALUE mod, VALUE name) +{ + struct sub_temporary_name_args args = {name}; + VALUE arg = (VALUE)&args; + rb_exec_recursive_paired(set_sub_temporary_name_topmost, mod, arg, arg); } +/* + * call-seq: + * mod.set_temporary_name(string) -> self + * mod.set_temporary_name(nil) -> self + * + * Sets the temporary name of the module. This name is reflected in + * introspection of the module and the values that are related to it, such + * as instances, constants, and methods. + * + * The name should be +nil+ or a non-empty string that is not a valid constant + * path (to avoid confusing between permanent and temporary names). + * + * The method can be useful to distinguish dynamically generated classes and + * modules without assigning them to constants. + * + * If the module is given a permanent name by assigning it to a constant, + * the temporary name is discarded. A temporary name can't be assigned to + * modules that have a permanent name. + * + * If the given name is +nil+, the module becomes anonymous again. + * + * Example: + * + * m = Module.new # => #<Module:0x0000000102c68f38> + * m.name #=> nil + * + * m.set_temporary_name("fake_name") # => fake_name + * m.name #=> "fake_name" + * + * m.set_temporary_name(nil) # => #<Module:0x0000000102c68f38> + * m.name #=> nil + * + * c = Class.new + * c.set_temporary_name("MyClass(with description)") # => MyClass(with description) + * + * c.new # => #<MyClass(with description):0x0....> + * + * c::M = m + * c::M.name #=> "MyClass(with description)::M" + * + * # Assigning to a constant replaces the name with a permanent one + * C = c + * + * C.name #=> "C" + * C::M.name #=> "C::M" + * c.new # => #<C:0x0....> + */ + VALUE -rb_class_path(klass) - VALUE klass; +rb_mod_set_temporary_name(VALUE mod, VALUE name) { - VALUE path = classname(klass); + // We don't allow setting the name if the classpath is already permanent: + if (RCLASS_PERMANENT_CLASSPATH_P(mod)) { + rb_raise(rb_eRuntimeError, "can't change permanent name"); + } - if (path) return path; + if (NIL_P(name)) { + // Set the temporary classpath to NULL (anonymous): + RB_VM_LOCKING() { + set_sub_temporary_name(mod, 0); + } + } else { - char buf[256]; - char *s = "Class"; + // Ensure the name is a string: + StringValue(name); + + if (RSTRING_LEN(name) == 0) { + rb_raise(rb_eArgError, "empty class/module name"); + } - if (TYPE(klass) == T_MODULE) s = "Module"; - sprintf(buf, "#<%s 0x%x>", s, klass); - return str_new2(buf); + if (is_constant_path(name)) { + rb_raise(rb_eArgError, "the temporary name must not be a constant path to avoid confusion"); + } + + name = rb_str_new_frozen(name); + RB_OBJ_SET_SHAREABLE(name); + + // Set the temporary classpath to the given name: + RB_VM_LOCKING() { + set_sub_temporary_name(mod, name); + } } + + return mod; } -void -rb_set_class_path(klass, under, name) - VALUE klass, under; - char *name; +static VALUE +make_temporary_path(VALUE obj, VALUE klass) { - VALUE str; + VALUE path; + switch (klass) { + case Qnil: + path = rb_sprintf("#<Class:%p>", (void*)obj); + break; + case Qfalse: + path = rb_sprintf("#<Module:%p>", (void*)obj); + break; + default: + path = rb_sprintf("#<%"PRIsVALUE":%p>", klass, (void*)obj); + break; + } + OBJ_FREEZE(path); + return path; +} - if (under == cObject) { - str = str_new2(name); +typedef VALUE (*fallback_func)(VALUE obj, VALUE name); + +static VALUE +rb_tmp_class_path(VALUE klass, bool *permanent, fallback_func fallback) +{ + VALUE path = classname(klass, permanent); + + if (!NIL_P(path)) { + return path; } - else { - str = str_dup(rb_class_path(under)); - str_cat(str, "::", 2); - str_cat(str, name, strlen(name)); + + if (RB_TYPE_P(klass, T_MODULE)) { + if (rb_obj_class(klass) == rb_cModule) { + path = Qfalse; + } + else { + bool perm; + path = rb_tmp_class_path(RBASIC(klass)->klass, &perm, fallback); + } } - rb_iv_set(klass, "__classpath__", str); + + *permanent = false; + return fallback(klass, path); } VALUE -rb_path2class(path) - char *path; +rb_class_path(VALUE klass) { - if (path[0] == '#') { - ArgError("can't retrieve anonymous class %s", path); - } - return rb_eval_string(path); + bool permanent; + VALUE path = rb_tmp_class_path(klass, &permanent, make_temporary_path); + if (!NIL_P(path)) path = rb_str_dup(path); + return path; +} + +VALUE +rb_class_path_cached(VALUE klass) +{ + return rb_mod_name(klass); +} + +static VALUE +no_fallback(VALUE obj, VALUE name) +{ + return name; +} + +VALUE +rb_search_class_path(VALUE klass) +{ + bool permanent; + return rb_tmp_class_path(klass, &permanent, no_fallback); +} + +static VALUE +build_const_pathname(VALUE head, VALUE tail) +{ + VALUE path = rb_str_dup(head); + rb_str_cat2(path, "::"); + rb_str_append(path, tail); + return rb_fstring(path); +} + +static VALUE +build_const_path(VALUE head, ID tail) +{ + return build_const_pathname(head, rb_id2str(tail)); } void -rb_name_class(klass, id) - VALUE klass; - ID id; +rb_set_class_path_string(VALUE klass, VALUE under, VALUE name) { - if (cString) { - rb_iv_set(klass, "__classpath__", str_new2(rb_id2name(id))); + bool permanent = true; + + VALUE str; + if (under == rb_cObject) { + str = rb_str_new_frozen(name); } else { - rb_iv_set(klass, "__classid__", INT2FIX(id)); + str = rb_tmp_class_path(under, &permanent, make_temporary_path); + str = build_const_pathname(str, name); } + + RB_OBJ_SET_SHAREABLE(str); + RCLASS_SET_CLASSPATH(klass, str, permanent); } -static st_table *autoload_tbl = 0; +void +rb_set_class_path(VALUE klass, VALUE under, const char *name) +{ + VALUE str = rb_str_new2(name); + OBJ_FREEZE(str); + rb_set_class_path_string(klass, under, str); +} -static void -rb_autoload_id(id, filename) - ID id; - char *filename; +VALUE +rb_path_to_class(VALUE pathname) { - if (!rb_is_const_id(id)) { - NameError("autoload must be constant name", rb_id2name(id)); - } + rb_encoding *enc = rb_enc_get(pathname); + const char *pbeg, *pend, *p, *path = RSTRING_PTR(pathname); + ID id; + VALUE c = rb_cObject; - if (!autoload_tbl) { - autoload_tbl = new_idhash(); + if (!rb_enc_asciicompat(enc)) { + rb_raise(rb_eArgError, "invalid class path encoding (non ASCII)"); + } + pbeg = p = path; + pend = path + RSTRING_LEN(pathname); + if (path == pend || path[0] == '#') { + rb_raise(rb_eArgError, "can't retrieve anonymous class %"PRIsVALUE, + QUOTE(pathname)); } - st_insert(autoload_tbl, id, strdup(filename)); + while (p < pend) { + while (p < pend && *p != ':') p++; + id = rb_check_id_cstr(pbeg, p-pbeg, enc); + if (p < pend && p[0] == ':') { + if ((size_t)(pend - p) < 2 || p[1] != ':') goto undefined_class; + p += 2; + pbeg = p; + } + if (!id) { + goto undefined_class; + } + c = rb_const_search(c, id, TRUE, FALSE, FALSE, NULL); + if (UNDEF_P(c)) goto undefined_class; + if (!rb_namespace_p(c)) { + rb_raise(rb_eTypeError, "%"PRIsVALUE" does not refer to class/module", + pathname); + } + } + RB_GC_GUARD(pathname); + + return c; + + undefined_class: + rb_raise(rb_eArgError, "undefined class/module % "PRIsVALUE, + rb_str_subseq(pathname, 0, p-path)); + UNREACHABLE_RETURN(Qundef); } -void -rb_autoload(klass, filename) - char *klass, *filename; +VALUE +rb_path2class(const char *path) { - rb_autoload_id(rb_intern(klass), filename); + return rb_path_to_class(rb_str_new_cstr(path)); } VALUE -f_autoload(obj, klass, file) - VALUE obj, klass, file; +rb_class_name(VALUE klass) { - rb_autoload_id(rb_to_id(klass), STR2CSTR(file)); - return Qnil; + return rb_class_path(rb_class_real(klass)); } -char * -rb_class2name(klass) - VALUE klass; +const char * +rb_class2name(VALUE klass) { - return RSTRING(rb_class_path(klass))->ptr; + bool permanent; + VALUE path = rb_tmp_class_path(rb_class_real(klass), &permanent, make_temporary_path); + if (NIL_P(path)) return NULL; + return RSTRING_PTR(path); +} + +const char * +rb_obj_classname(VALUE obj) +{ + return rb_class2name(CLASS_OF(obj)); } struct trace_var { int removed; - void (*func)(); - void *data; + void (*func)(VALUE arg, VALUE val); + VALUE data; struct trace_var *next; }; -struct global_entry { - ID id; - void *data; - VALUE (*getter)(); - void (*setter)(); - void (*marker)(); +struct rb_global_variable { + int counter; int block_trace; + VALUE *data; + rb_gvar_getter_t *getter; + rb_gvar_setter_t *setter; + rb_gvar_marker_t *marker; + rb_gvar_compact_t *compactor; struct trace_var *trace; + bool box_ready; }; -static VALUE undef_getter(); -static void undef_setter(); -static void undef_marker(); +struct rb_global_entry { + struct rb_global_variable *var; + ID id; + bool ractor_local; +}; -static VALUE val_getter(); -static void val_setter(); -static void val_marker(); +static void +free_global_variable(struct rb_global_variable *var) +{ + RUBY_ASSERT(var->counter == 0); -static VALUE var_getter(); -static void var_setter(); -static void var_marker(); + struct trace_var *trace = var->trace; + while (trace) { + struct trace_var *next = trace->next; + xfree(trace); + trace = next; + } + xfree(var); +} -struct global_entry* -rb_global_entry(id) - ID id; +static enum rb_id_table_iterator_result +free_global_entry_i(VALUE val, void *arg) +{ + struct rb_global_entry *entry = (struct rb_global_entry *)val; + entry->var->counter--; + if (entry->var->counter == 0) { + free_global_variable(entry->var); + } + ruby_xfree(entry); + return ID_TABLE_DELETE; +} + +void +rb_free_rb_global_tbl(void) { - struct global_entry *entry; + rb_id_table_foreach_values(rb_global_tbl, free_global_entry_i, 0); + rb_id_table_free(rb_global_tbl); +} - if (!st_lookup(global_tbl, id, &entry)) { - entry = ALLOC(struct global_entry); - st_insert(global_tbl, id, entry); - entry->id = id; - entry->data = 0; - entry->getter = undef_getter; - entry->setter = undef_setter; - entry->marker = undef_marker; +void +rb_free_generic_fields_tbl_(void) +{ + st_free_table(generic_fields_tbl_); +} - entry->block_trace = 0; - entry->trace = 0; +static struct rb_global_entry* +rb_find_global_entry(ID id) +{ + struct rb_global_entry *entry; + VALUE data; + + RB_VM_LOCKING() { + if (!rb_id_table_lookup(rb_global_tbl, id, &data)) { + entry = NULL; + } + else { + entry = (struct rb_global_entry *)data; + RUBY_ASSERT(entry != NULL); + } + } + + if (UNLIKELY(!rb_ractor_main_p()) && (!entry || !entry->ractor_local)) { + rb_raise(rb_eRactorIsolationError, "can not access global variable %s from non-main Ractor", rb_id2name(id)); } + return entry; } -static VALUE -undef_getter(id) - ID id; +void +rb_gvar_ractor_local(const char *name) { - Warning("global variable `%s' not initialized", rb_id2name(id)); - return Qnil; + struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name)); + entry->ractor_local = true; +} + +void +rb_gvar_box_ready(const char *name) +{ + struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name)); + entry->var->box_ready = true; } static void -undef_setter(val, id, data, entry) - VALUE val; - ID id; - void *data; - struct global_entry *entry; +rb_gvar_undef_compactor(void *var) { - entry->getter = val_getter; - entry->setter = val_setter; - entry->marker = val_marker; +} - entry->data = (void*)val; +static struct rb_global_entry* +rb_global_entry(ID id) +{ + struct rb_global_entry *entry; + RB_VM_LOCKING() { + entry = rb_find_global_entry(id); + if (!entry) { + struct rb_global_variable *var; + entry = ALLOC(struct rb_global_entry); + var = ALLOC(struct rb_global_variable); + entry->id = id; + entry->var = var; + entry->ractor_local = false; + var->counter = 1; + var->data = 0; + var->getter = rb_gvar_undef_getter; + var->setter = rb_gvar_undef_setter; + var->marker = rb_gvar_undef_marker; + var->compactor = rb_gvar_undef_compactor; + + var->block_trace = 0; + var->trace = 0; + var->box_ready = false; + rb_id_table_insert(rb_global_tbl, id, (VALUE)entry); + } + } + return entry; +} + +VALUE +rb_gvar_undef_getter(ID id, VALUE *_) +{ + rb_warning("global variable '%"PRIsVALUE"' not initialized", QUOTE_ID(id)); + + return Qnil; } static void -undef_marker() +rb_gvar_val_compactor(void *_var) { + struct rb_global_variable *var = (struct rb_global_variable *)_var; + + VALUE obj = (VALUE)var->data; + + if (obj) { + VALUE new = rb_gc_location(obj); + if (new != obj) { + var->data = (void*)new; + } + } } -static VALUE -val_getter(id, val) - ID id; - VALUE val; +void +rb_gvar_undef_setter(VALUE val, ID id, VALUE *_) { - return val; + struct rb_global_variable *var = rb_global_entry(id)->var; + var->getter = rb_gvar_val_getter; + var->setter = rb_gvar_val_setter; + var->marker = rb_gvar_val_marker; + var->compactor = rb_gvar_val_compactor; + + var->data = (void*)val; } -static void -val_setter(val, id, data, entry) - VALUE val; - ID id; - void *data; - struct global_entry *entry; +void +rb_gvar_undef_marker(VALUE *var) { - entry->data = (void*)val; } -static void -val_marker(data) - void *data; +VALUE +rb_gvar_val_getter(ID id, VALUE *data) { - if (data) gc_mark_maybe(data); + return (VALUE)data; } -static VALUE -var_getter(id, var) - ID id; - VALUE *var; +void +rb_gvar_val_setter(VALUE val, ID id, VALUE *_) +{ + struct rb_global_variable *var = rb_global_entry(id)->var; + var->data = (void*)val; +} + +void +rb_gvar_val_marker(VALUE *var) +{ + VALUE data = (VALUE)var; + if (data) rb_gc_mark_movable(data); +} + +VALUE +rb_gvar_var_getter(ID id, VALUE *var) { - if (!var || !*var) return Qnil; + if (!var) return Qnil; return *var; } -static void -var_setter(val, id, var) - VALUE val; - ID id; - VALUE *var; +void +rb_gvar_var_setter(VALUE val, ID id, VALUE *data) { - *var = val; + *data = val; } -static void -var_marker(var) - VALUE **var; +void +rb_gvar_var_marker(VALUE *var) { - if (var) gc_mark_maybe(*var); + if (var) rb_gc_mark_maybe(*var); } -static void -readonly_setter(val, id, var) - VALUE val; - ID id; - void *var; +void +rb_gvar_readonly_setter(VALUE v, ID id, VALUE *_) { - NameError("Can't set variable %s", rb_id2name(id)); + rb_name_error(id, "%"PRIsVALUE" is a read-only variable", QUOTE_ID(id)); } -static int -mark_global_entry(key, entry) - ID key; - struct global_entry *entry; +static enum rb_id_table_iterator_result +mark_global_entry(VALUE v, void *ignored) { + struct rb_global_entry *entry = (struct rb_global_entry *)v; struct trace_var *trace; + struct rb_global_variable *var = entry->var; - (*entry->marker)(entry->data); - trace = entry->trace; + (*var->marker)(var->data); + trace = var->trace; while (trace) { - if (trace->data) gc_mark_maybe(trace->data); - trace = trace->next; + if (trace->data) rb_gc_mark_maybe(trace->data); + trace = trace->next; } - return ST_CONTINUE; + return ID_TABLE_CONTINUE; } +#define gc_mark_table(task) \ + if (rb_global_tbl) { rb_id_table_foreach_values(rb_global_tbl, task##_global_entry, 0); } + void -gc_mark_global_tbl() +rb_gc_mark_global_tbl(void) { - st_foreach(global_tbl, mark_global_entry, 0); + gc_mark_table(mark); +} + +static enum rb_id_table_iterator_result +update_global_entry(VALUE v, void *ignored) +{ + struct rb_global_entry *entry = (struct rb_global_entry *)v; + struct rb_global_variable *var = entry->var; + + (*var->compactor)(var); + return ID_TABLE_CONTINUE; +} + +void +rb_gc_update_global_tbl(void) +{ + gc_mark_table(update); } static ID -global_id(name) - char *name; +global_id(const char *name) { ID id; if (name[0] == '$') id = rb_intern(name); else { - char *buf = ALLOCA_N(char, strlen(name)+2); - buf[0] = '$'; - strcpy(buf+1, name); - id = rb_intern(buf); + size_t len = strlen(name); + VALUE vbuf = 0; + char *buf = ALLOCV_N(char, vbuf, len+1); + buf[0] = '$'; + memcpy(buf+1, name, len); + id = rb_intern2(buf, len+1); + ALLOCV_END(vbuf); + } + return id; +} + +static ID +find_global_id(const char *name) +{ + ID id; + size_t len = strlen(name); + + if (name[0] == '$') { + id = rb_check_id_cstr(name, len, NULL); + } + else { + VALUE vbuf = 0; + char *buf = ALLOCV_N(char, vbuf, len+1); + buf[0] = '$'; + memcpy(buf+1, name, len); + id = rb_check_id_cstr(buf, len+1, NULL); + ALLOCV_END(vbuf); } + return id; } void -rb_define_hooked_variable(name, var, getter, setter) - char *name; - VALUE *var; - VALUE (*getter)(); - void (*setter)(); +rb_define_hooked_variable( + const char *name, + VALUE *var, + rb_gvar_getter_t *getter, + rb_gvar_setter_t *setter) { - struct global_entry *entry; + volatile VALUE tmp = var ? *var : Qnil; ID id = global_id(name); + struct rb_global_variable *gvar = rb_global_entry(id)->var; + + gvar->data = (void*)var; + gvar->getter = getter ? (rb_gvar_getter_t *)getter : rb_gvar_var_getter; + gvar->setter = setter ? (rb_gvar_setter_t *)setter : rb_gvar_var_setter; + gvar->marker = rb_gvar_var_marker; - entry = rb_global_entry(id); - entry->data = (void*)var; - entry->getter = getter?getter:var_getter; - entry->setter = setter?setter:var_setter; - entry->marker = var_marker; + RB_GC_GUARD(tmp); } void -rb_define_variable(name, var) - char *name; - VALUE *var; +rb_define_variable(const char *name, VALUE *var) { rb_define_hooked_variable(name, var, 0, 0); } void -rb_define_readonly_variable(name, var) - char *name; - VALUE *var; +rb_define_readonly_variable(const char *name, const VALUE *var) { - rb_define_hooked_variable(name, var, 0, readonly_setter); + rb_define_hooked_variable(name, (VALUE *)var, 0, rb_gvar_readonly_setter); } void -rb_define_virtual_variable(name, getter, setter) - char *name; - VALUE (*getter)(); - void (*setter)(); +rb_define_virtual_variable( + const char *name, + rb_gvar_getter_t *getter, + rb_gvar_setter_t *setter) { - if (!getter) getter = val_getter; - if (!setter) setter = readonly_setter; + if (!getter) getter = rb_gvar_val_getter; + if (!setter) setter = rb_gvar_readonly_setter; rb_define_hooked_variable(name, 0, getter, setter); } static void -rb_trace_eval(cmd, val) - VALUE cmd, val; +rb_trace_eval(VALUE cmd, VALUE val) { - rb_eval_cmd(cmd, ary_new3(1, val)); + rb_eval_cmd_call_kw(cmd, 1, &val, RB_NO_KEYWORDS); } VALUE -f_trace_var(argc, argv) - int argc; - VALUE *argv; +rb_f_trace_var(int argc, const VALUE *argv) { VALUE var, cmd; - ID id; - struct global_entry *entry; + struct rb_global_entry *entry; struct trace_var *trace; if (rb_scan_args(argc, argv, "11", &var, &cmd) == 1) { - cmd = f_lambda(); + cmd = rb_block_proc(); } if (NIL_P(cmd)) { - return f_untrace_var(argc, argv); - } - id = rb_to_id(var); - if (!st_lookup(global_tbl, id, &entry)) { - NameError("undefined global variable %s", rb_id2name(id)); + return rb_f_untrace_var(argc, argv); } + entry = rb_global_entry(rb_to_id(var)); trace = ALLOC(struct trace_var); - trace->next = entry->trace; + trace->next = entry->var->trace; trace->func = rb_trace_eval; - trace->data = (void*)cmd; + trace->data = cmd; trace->removed = 0; - entry->trace = trace; + entry->var->trace = trace; return Qnil; } static void -remove_trace(entry) - struct global_entry *entry; +remove_trace(struct rb_global_variable *var) { - struct trace_var *trace = entry->trace; + struct trace_var *trace = var->trace; struct trace_var t; struct trace_var *next; t.next = trace; trace = &t; while (trace->next) { - next = trace->next; - if (next->removed) { - trace->next = next->next; - free(next); - } - trace = next; + next = trace->next; + if (next->removed) { + trace->next = next->next; + xfree(next); + } + else { + trace = next; + } } - entry->trace = t.next; + var->trace = t.next; } VALUE -f_untrace_var(argc, argv) - int argc; - VALUE *argv; +rb_f_untrace_var(int argc, const VALUE *argv) { VALUE var, cmd; ID id; - struct global_entry *entry; + struct rb_global_entry *entry; struct trace_var *trace; rb_scan_args(argc, argv, "11", &var, &cmd); - id = rb_to_id(var); - if (!st_lookup(global_tbl, id, &entry)) { - NameError("undefined global variable %s", rb_id2name(id)); + id = rb_check_id(&var); + if (!id) { + rb_name_error_str(var, "undefined global variable %"PRIsVALUE"", QUOTE(var)); + } + if ((entry = rb_find_global_entry(id)) == NULL) { + rb_name_error(id, "undefined global variable %"PRIsVALUE"", QUOTE_ID(id)); } - trace = entry->trace; + trace = entry->var->trace; if (NIL_P(cmd)) { - VALUE ary = ary_new(); + VALUE ary = rb_ary_new(); - while (trace) { - struct trace_var *next = trace->next; - ary_push(ary, (VALUE)trace->data); - trace->removed = 1; - trace = next; - } - entry->trace = 0; + while (trace) { + struct trace_var *next = trace->next; + rb_ary_push(ary, (VALUE)trace->data); + trace->removed = 1; + trace = next; + } - if (!entry->block_trace) remove_trace(entry); - return ary; + if (!entry->var->block_trace) remove_trace(entry->var); + return ary; } else { - while (trace) { - if (trace->data == (void*)cmd) { - trace->removed = 1; - if (!entry->block_trace) remove_trace(entry); - return ary_new3(1, cmd); - } - trace = trace->next; - } + while (trace) { + if (trace->data == cmd) { + trace->removed = 1; + if (!entry->var->block_trace) remove_trace(entry->var); + return rb_ary_new3(1, cmd); + } + trace = trace->next; + } } return Qnil; } -VALUE -rb_gvar_get(entry) - struct global_entry *entry; -{ - return (*entry->getter)(entry->id, entry->data, entry); -} struct trace_data { struct trace_var *trace; VALUE val; }; - + static VALUE -trace_ev(data) - struct trace_data *data; +trace_ev(VALUE v) { + struct trace_data *data = (void *)v; struct trace_var *trace = data->trace; while (trace) { - (*trace->func)(trace->data, data->val); - trace = trace->next; + (*trace->func)(trace->data, data->val); + trace = trace->next; } + + return Qnil; } static VALUE -trace_en(entry) - struct global_entry *entry; +trace_en(VALUE v) { - entry->block_trace = 0; - remove_trace(entry); + struct rb_global_variable *var = (void *)v; + var->block_trace = 0; + remove_trace(var); + return Qnil; /* not reached */ } -VALUE -rb_gvar_set(entry, val) - struct global_entry *entry; - VALUE val; +static VALUE +rb_gvar_set_entry(struct rb_global_entry *entry, VALUE val) { struct trace_data trace; + struct rb_global_variable *var = entry->var; - if (rb_safe_level() >= 4) { - Raise(eSecurityError, "cannot change global variable value"); - } - (*entry->setter)(val, entry->id, entry->data, entry); + (*var->setter)(val, entry->id, var->data); - if (entry->trace && !entry->block_trace) { - entry->block_trace = 1; - trace.trace = entry->trace; - trace.val = val; - rb_ensure(trace_ev, (VALUE)&trace, trace_en, (VALUE)entry); + if (var->trace && !var->block_trace) { + var->block_trace = 1; + trace.trace = var->trace; + trace.val = val; + rb_ensure(trace_ev, (VALUE)&trace, trace_en, (VALUE)var); } return val; } +#define USE_BOX_GVAR_TBL(ns,entry) \ + (BOX_USER_P(ns) && \ + (!entry || !entry->var->box_ready || entry->var->setter != rb_gvar_readonly_setter)) + VALUE -rb_gvar_set2(name, val) - char *name; - VALUE val; +rb_gvar_set(ID id, VALUE val) { - struct global_entry *entry; + VALUE retval; + struct rb_global_entry *entry; + const rb_box_t *box = rb_current_box(); + bool use_box_tbl = false; + + RB_VM_LOCKING() { + entry = rb_global_entry(id); + + if (USE_BOX_GVAR_TBL(box, entry)) { + use_box_tbl = true; + rb_hash_aset(box->gvar_tbl, rb_id2sym(entry->id), val); + retval = val; + // TODO: think about trace + } + } - entry = rb_global_entry(global_id(name)); - return rb_gvar_set(entry, val); + if (!use_box_tbl) { + retval = rb_gvar_set_entry(entry, val); + } + return retval; } VALUE -rb_gvar_defined(entry) - struct global_entry *entry; +rb_gv_set(const char *name, VALUE val) { - if (entry->getter == undef_getter) return FALSE; - return TRUE; + return rb_gvar_set(global_id(name), val); } -static int -gvar_i(key, entry, ary) - ID key; - struct global_entry *entry; - VALUE ary; +VALUE +rb_gvar_get(ID id) { - ary_push(ary, str_new2(rb_id2name(key))); - return ST_CONTINUE; + VALUE retval, gvars, key; + const rb_box_t *box = rb_current_box(); + bool use_box_tbl = false; + struct rb_global_entry *entry; + struct rb_global_variable *var; + // TODO: use lock-free rb_id_table when it's available for use (doesn't yet exist) + RB_VM_LOCKING() { + entry = rb_global_entry(id); + var = entry->var; + + if (USE_BOX_GVAR_TBL(box, entry)) { + use_box_tbl = true; + gvars = box->gvar_tbl; + key = rb_id2sym(entry->id); + if (RTEST(rb_hash_has_key(gvars, key))) { // this gvar is already cached + retval = rb_hash_aref(gvars, key); + } + else { + RB_VM_UNLOCK(); + { + retval = (*var->getter)(entry->id, var->data); + if (rb_obj_respond_to(retval, rb_intern("clone"), 1)) { + retval = rb_funcall(retval, rb_intern("clone"), 0); + } + } + RB_VM_LOCK(); + rb_hash_aset(gvars, key, retval); + } + } + } + if (!use_box_tbl) { + retval = (*var->getter)(entry->id, var->data); + } + return retval; } VALUE -f_global_variables() +rb_gv_get(const char *name) { - VALUE ary = ary_new(); - char buf[4]; - char *s = "&`'+123456789"; + ID id = find_global_id(name); - st_foreach(global_tbl, gvar_i, ary); - if (!NIL_P(backref_get())) { - while (*s) { - sprintf(buf, "$%c", *s++); - ary_push(ary, str_new2(buf)); - } + if (!id) { + rb_warning("global variable '%s' not initialized", name); + return Qnil; + } + + return rb_gvar_get(id); +} + +VALUE +rb_gvar_defined(ID id) +{ + struct rb_global_entry *entry = rb_global_entry(id); + return RBOOL(entry->var->getter != rb_gvar_undef_getter); +} + +rb_gvar_getter_t * +rb_gvar_getter_function_of(ID id) +{ + const struct rb_global_entry *entry = rb_global_entry(id); + return entry->var->getter; +} + +rb_gvar_setter_t * +rb_gvar_setter_function_of(ID id) +{ + const struct rb_global_entry *entry = rb_global_entry(id); + return entry->var->setter; +} + +static enum rb_id_table_iterator_result +gvar_i(ID key, VALUE val, void *a) +{ + VALUE ary = (VALUE)a; + rb_ary_push(ary, ID2SYM(key)); + return ID_TABLE_CONTINUE; +} + +VALUE +rb_f_global_variables(void) +{ + VALUE ary = rb_ary_new(); + VALUE sym, backref = rb_backref_get(); + + if (!rb_ractor_main_p()) { + rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors"); + } + /* gvar access (get/set) in boxes creates gvar entries globally */ + + rb_id_table_foreach(rb_global_tbl, gvar_i, (void *)ary); + if (!NIL_P(backref)) { + char buf[2]; + int i, nmatch = rb_match_count(backref); + buf[0] = '$'; + for (i = 1; i <= nmatch; ++i) { + if (!RTEST(rb_reg_nth_defined(i, backref))) continue; + if (i < 10) { + /* probably reused, make static ID */ + buf[1] = (char)(i + '0'); + sym = ID2SYM(rb_intern2(buf, 2)); + } + else { + /* dynamic symbol */ + sym = rb_str_intern(rb_sprintf("$%d", i)); + } + rb_ary_push(ary, sym); + } } return ary; } void -rb_alias_variable(name1, name2) - ID name1; - ID name2; +rb_alias_variable(ID name1, ID name2) +{ + struct rb_global_entry *entry1 = NULL, *entry2; + VALUE data1; + struct rb_id_table *gtbl = rb_global_tbl; + + if (!rb_ractor_main_p()) { + rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors"); + } + + RB_VM_LOCKING() { + entry2 = rb_global_entry(name2); + if (!rb_id_table_lookup(gtbl, name1, &data1)) { + entry1 = ZALLOC(struct rb_global_entry); + entry1->id = name1; + rb_id_table_insert(gtbl, name1, (VALUE)entry1); + } + else if ((entry1 = (struct rb_global_entry *)data1)->var != entry2->var) { + struct rb_global_variable *var = entry1->var; + if (var->block_trace) { + RB_VM_UNLOCK(); + rb_raise(rb_eRuntimeError, "can't alias in tracer"); + } + var->counter--; + if (var->counter == 0) { + free_global_variable(var); + } + } + if (entry1->var != entry2->var) { + entry2->var->counter++; + entry1->var = entry2->var; + } + } +} + +static void +IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(ID id) { - struct global_entry *entry1, *entry2; + if (UNLIKELY(!rb_ractor_main_p())) { + if (rb_is_instance_id(id)) { // check only normal ivars + rb_raise(rb_eRactorIsolationError, "can not set instance variables of classes/modules by non-main Ractors"); + } + } +} - entry1 = rb_global_entry(name1); - entry2 = rb_global_entry(name2); +static void +CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(VALUE klass, ID id) +{ + if (UNLIKELY(!rb_ractor_main_p())) { + rb_raise(rb_eRactorIsolationError, "can not access class variables from non-main Ractors (%"PRIsVALUE" from %"PRIsVALUE")", rb_id2str(id), klass); + } +} - entry1->data = entry2->data; - entry1->getter = entry2->getter; - entry1->setter = entry2->setter; - entry1->marker = entry2->marker; +static inline void +ivar_ractor_check(VALUE obj, ID id) +{ + if (LIKELY(rb_is_instance_id(id)) /* not internal ID */ && + !RB_OBJ_FROZEN_RAW(obj) && + UNLIKELY(!rb_ractor_main_p()) && + UNLIKELY(rb_ractor_shareable_p(obj))) { + + rb_raise(rb_eRactorIsolationError, "can not access instance variables of shareable objects from non-main Ractors"); + } +} + +static inline struct st_table * +generic_fields_tbl_no_ractor_check(void) +{ + ASSERT_vm_locking(); + + return generic_fields_tbl_; +} + +struct st_table * +rb_generic_fields_tbl_get(void) +{ + return generic_fields_tbl_; +} + +void +rb_mark_generic_ivar(VALUE obj) +{ + VALUE data; + // Bypass ASSERT_vm_locking() check because marking may happen concurrently with mmtk + if (st_lookup(generic_fields_tbl_, (st_data_t)obj, (st_data_t *)&data)) { + rb_gc_mark_movable(data); + } } VALUE -rb_ivar_get(obj, id) - VALUE obj; - ID id; +rb_obj_fields_generic_uncached(VALUE obj) { - VALUE val; + VALUE fields_obj = 0; + RB_VM_LOCKING() { + if (!st_lookup(generic_fields_tbl_, (st_data_t)obj, (st_data_t *)&fields_obj)) { + rb_bug("Object is missing entry in generic_fields_tbl"); + } + } + return fields_obj; +} - switch (TYPE(obj)) { +VALUE +rb_obj_fields(VALUE obj, ID field_name) +{ + RUBY_ASSERT(!RB_TYPE_P(obj, T_IMEMO)); + ivar_ractor_check(obj, field_name); + + VALUE fields_obj = 0; + if (rb_shape_obj_has_fields(obj)) { + switch (BUILTIN_TYPE(obj)) { + case T_DATA: + if (LIKELY(RTYPEDDATA_P(obj))) { + fields_obj = RTYPEDDATA(obj)->fields_obj; + break; + } + goto generic_fields; + case T_STRUCT: + if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) { + fields_obj = RSTRUCT_FIELDS_OBJ(obj); + break; + } + goto generic_fields; + default: + generic_fields: + { + rb_execution_context_t *ec = GET_EC(); + if (ec->gen_fields_cache.obj == obj && !UNDEF_P(ec->gen_fields_cache.fields_obj) && rb_imemo_fields_owner(ec->gen_fields_cache.fields_obj) == obj) { + fields_obj = ec->gen_fields_cache.fields_obj; + RUBY_ASSERT(fields_obj == rb_obj_fields_generic_uncached(obj)); + } + else { + fields_obj = rb_obj_fields_generic_uncached(obj); + ec->gen_fields_cache.fields_obj = fields_obj; + ec->gen_fields_cache.obj = obj; + } + } + } + } + return fields_obj; +} + +void +rb_free_generic_ivar(VALUE obj) +{ + if (rb_obj_gen_fields_p(obj)) { + st_data_t key = (st_data_t)obj, value; + switch (BUILTIN_TYPE(obj)) { + case T_DATA: + if (LIKELY(RTYPEDDATA_P(obj))) { + RB_OBJ_WRITE(obj, &RTYPEDDATA(obj)->fields_obj, 0); + break; + } + goto generic_fields; + case T_STRUCT: + if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) { + RSTRUCT_SET_FIELDS_OBJ(obj, 0); + break; + } + goto generic_fields; + default: + generic_fields: + { + // Other EC may have stale caches, so fields_obj should be + // invalidated and the GC will replace with Qundef + rb_execution_context_t *ec = GET_EC(); + if (ec->gen_fields_cache.obj == obj) { + ec->gen_fields_cache.obj = Qundef; + ec->gen_fields_cache.fields_obj = Qundef; + } + RB_VM_LOCKING() { + if (!st_delete(generic_fields_tbl_no_ractor_check(), &key, &value)) { + rb_bug("Object is missing entry in generic_fields_tbl"); + } + } + } + } + RBASIC_SET_SHAPE_ID(obj, ROOT_SHAPE_ID); + } +} + +static void +rb_obj_set_fields(VALUE obj, VALUE fields_obj, ID field_name, VALUE original_fields_obj) +{ + ivar_ractor_check(obj, field_name); + + if (!fields_obj) { + RUBY_ASSERT(original_fields_obj); + rb_free_generic_ivar(obj); + rb_imemo_fields_clear(original_fields_obj); + return; + } + + RUBY_ASSERT(IMEMO_TYPE_P(fields_obj, imemo_fields)); + RUBY_ASSERT(!original_fields_obj || IMEMO_TYPE_P(original_fields_obj, imemo_fields)); + + if (fields_obj != original_fields_obj) { + switch (BUILTIN_TYPE(obj)) { + case T_DATA: + if (LIKELY(RTYPEDDATA_P(obj))) { + RB_OBJ_WRITE(obj, &RTYPEDDATA(obj)->fields_obj, fields_obj); + break; + } + goto generic_fields; + case T_STRUCT: + if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) { + RSTRUCT_SET_FIELDS_OBJ(obj, fields_obj); + break; + } + goto generic_fields; + default: + generic_fields: + { + RB_VM_LOCKING() { + st_insert(generic_fields_tbl_, (st_data_t)obj, (st_data_t)fields_obj); + } + RB_OBJ_WRITTEN(obj, original_fields_obj, fields_obj); + + rb_execution_context_t *ec = GET_EC(); + if (ec->gen_fields_cache.fields_obj != fields_obj) { + ec->gen_fields_cache.obj = obj; + ec->gen_fields_cache.fields_obj = fields_obj; + } + } + } + + if (original_fields_obj) { + // Clear root shape to avoid triggering cleanup such as free_object_id. + rb_imemo_fields_clear(original_fields_obj); + } + } + + RBASIC_SET_SHAPE_ID(obj, RBASIC_SHAPE_ID(fields_obj)); +} + +void +rb_obj_replace_fields(VALUE obj, VALUE fields_obj) +{ + RB_VM_LOCKING() { + VALUE original_fields_obj = rb_obj_fields_no_ractor_check(obj); + rb_obj_set_fields(obj, fields_obj, 0, original_fields_obj); + } +} + +VALUE +rb_obj_field_get(VALUE obj, shape_id_t target_shape_id) +{ + RUBY_ASSERT(!SPECIAL_CONST_P(obj)); + RUBY_ASSERT(RSHAPE_TYPE_P(target_shape_id, SHAPE_IVAR) || RSHAPE_TYPE_P(target_shape_id, SHAPE_OBJ_ID)); + + VALUE fields_obj; + + switch (BUILTIN_TYPE(obj)) { + case T_CLASS: + case T_MODULE: + fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + break; case T_OBJECT: + fields_obj = obj; + break; + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + fields_obj = obj; + break; + default: + fields_obj = rb_obj_fields(obj, RSHAPE_EDGE_NAME(target_shape_id)); + break; + } + + if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) { + st_table *fields_hash = rb_imemo_fields_complex_tbl(fields_obj); + VALUE value = Qundef; + st_lookup(fields_hash, RSHAPE_EDGE_NAME(target_shape_id), &value); + RUBY_ASSERT(!UNDEF_P(value)); + return value; + } + + attr_index_t index = RSHAPE_INDEX(target_shape_id); + return rb_imemo_fields_ptr(fields_obj)[index]; +} + +VALUE +rb_ivar_lookup(VALUE obj, ID id, VALUE undef) +{ + if (SPECIAL_CONST_P(obj)) return undef; + + VALUE fields_obj; + + switch (BUILTIN_TYPE(obj)) { case T_CLASS: case T_MODULE: - if (ROBJECT(obj)->iv_tbl && st_lookup(ROBJECT(obj)->iv_tbl, id, &val)) - return val; - return Qnil; + { + VALUE val = rb_ivar_lookup(RCLASS_WRITABLE_FIELDS_OBJ(obj), id, undef); + if (val != undef && + rb_is_instance_id(id) && + UNLIKELY(!rb_ractor_main_p()) && + !rb_ractor_shareable_p(val)) { + rb_raise(rb_eRactorIsolationError, + "can not get unshareable values from instance variables of classes/modules from non-main Ractors (%"PRIsVALUE" from %"PRIsVALUE")", + rb_id2str(id), obj); + } + return val; + } + case T_IMEMO: + // Handled like T_OBJECT + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + fields_obj = obj; + break; + case T_OBJECT: + fields_obj = obj; + break; default: - TypeError("class %s can not have instance variables", - rb_class2name(CLASS_OF(obj))); - break; + fields_obj = rb_obj_fields(obj, id); + break; } - Warning("instance var %s not initialized", rb_id2name(id)); - return Qnil; + + if (!fields_obj) { + return undef; + } + + shape_id_t shape_id = RBASIC_SHAPE_ID(fields_obj); + + if (UNLIKELY(rb_shape_too_complex_p(shape_id))) { + st_table *iv_table = rb_imemo_fields_complex_tbl(fields_obj); + VALUE val; + if (rb_st_lookup(iv_table, (st_data_t)id, (st_data_t *)&val)) { + return val; + } + return undef; + } + + attr_index_t index = 0; + if (rb_shape_get_iv_index(shape_id, id, &index)) { + return rb_imemo_fields_ptr(fields_obj)[index]; + } + + return undef; } VALUE -rb_ivar_set(obj, id, val) - VALUE obj; - ID id; - VALUE val; +rb_ivar_get(VALUE obj, ID id) +{ + VALUE iv = rb_ivar_lookup(obj, id, Qnil); + RB_DEBUG_COUNTER_INC(ivar_get_base); + return iv; +} + +VALUE +rb_ivar_get_at(VALUE obj, attr_index_t index, ID id) { - if (rb_safe_level() >= 5) { - Raise(eSecurityError, "cannot change object status"); + RUBY_ASSERT(rb_is_instance_id(id)); + // Used by JITs, but never for T_OBJECT. + + switch (BUILTIN_TYPE(obj)) { + case T_OBJECT: + UNREACHABLE_RETURN(Qundef); + case T_CLASS: + case T_MODULE: + { + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + VALUE val = rb_imemo_fields_ptr(fields_obj)[index]; + + if (UNLIKELY(!rb_ractor_main_p()) && !rb_ractor_shareable_p(val)) { + rb_raise(rb_eRactorIsolationError, + "can not get unshareable values from instance variables of classes/modules from non-main Ractors"); + } + + return val; + } + default: + { + VALUE fields_obj = rb_obj_fields(obj, id); + return rb_imemo_fields_ptr(fields_obj)[index]; + } } - switch (TYPE(obj)) { +} + +VALUE +rb_ivar_get_at_no_ractor_check(VALUE obj, attr_index_t index) +{ + // Used by JITs, but never for T_OBJECT. + + VALUE fields_obj; + switch (BUILTIN_TYPE(obj)) { case T_OBJECT: + UNREACHABLE_RETURN(Qundef); case T_CLASS: case T_MODULE: - if (!ROBJECT(obj)->iv_tbl) ROBJECT(obj)->iv_tbl = new_idhash(); - st_insert(ROBJECT(obj)->iv_tbl, id, val); - break; + fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + break; default: - TypeError("class %s can not have instance variables", - rb_class2name(CLASS_OF(obj))); - break; + fields_obj = rb_obj_fields_no_ractor_check(obj); + break; } - return val; + return rb_imemo_fields_ptr(fields_obj)[index]; } VALUE -rb_ivar_defined(obj, id) - VALUE obj; - ID id; +rb_attr_get(VALUE obj, ID id) { - if (!rb_is_instance_id(id)) return FALSE; + return rb_ivar_lookup(obj, id, Qnil); +} - switch (TYPE(obj)) { +void rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table); +static VALUE imemo_fields_complex_from_obj(VALUE owner, VALUE source_fields_obj, shape_id_t shape_id); + +static shape_id_t +obj_transition_too_complex(VALUE obj, st_table *table) +{ + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + shape_id_t shape_id = rb_shape_transition_complex(obj); + + switch (BUILTIN_TYPE(obj)) { case T_OBJECT: + { + VALUE *old_fields = NULL; + if (FL_TEST_RAW(obj, ROBJECT_HEAP)) { + old_fields = ROBJECT_FIELDS(obj); + } + else { + FL_SET_RAW(obj, ROBJECT_HEAP); + } + RBASIC_SET_SHAPE_ID(obj, shape_id); + ROBJECT_SET_FIELDS_HASH(obj, table); + if (old_fields) { + xfree(old_fields); + } + } + break; + case T_CLASS: + case T_MODULE: + case T_IMEMO: + UNREACHABLE; + break; + default: + { + VALUE fields_obj = rb_imemo_fields_new_complex_tbl(obj, table, RB_OBJ_SHAREABLE_P(obj)); + RBASIC_SET_SHAPE_ID(fields_obj, shape_id); + rb_obj_replace_fields(obj, fields_obj); + } + } + + return shape_id; +} + +// Copy all object fields, including ivars and internal object_id, etc +static shape_id_t +rb_evict_fields_to_hash(VALUE obj) +{ + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + + st_table *table = st_init_numtable_with_size(RSHAPE_LEN(RBASIC_SHAPE_ID(obj))); + rb_obj_copy_fields_to_hash_table(obj, table); + shape_id_t new_shape_id = obj_transition_too_complex(obj, table); + + RUBY_ASSERT(rb_shape_obj_too_complex_p(obj)); + return new_shape_id; +} + +void +rb_evict_ivars_to_hash(VALUE obj) +{ + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + + st_table *table = st_init_numtable_with_size(rb_ivar_count(obj)); + + // Evacuate all previous values from shape into id_table + rb_obj_copy_ivs_to_hash_table(obj, table); + obj_transition_too_complex(obj, table); + + RUBY_ASSERT(rb_shape_obj_too_complex_p(obj)); +} + +static VALUE +rb_ivar_delete(VALUE obj, ID id, VALUE undef) +{ + rb_check_frozen(obj); + + VALUE val = undef; + VALUE fields_obj; + bool concurrent = false; + int type = BUILTIN_TYPE(obj); + + switch(type) { case T_CLASS: case T_MODULE: - if (ROBJECT(obj)->iv_tbl && st_lookup(ROBJECT(obj)->iv_tbl, id, 0)) - return TRUE; - break; + IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id); + + fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (rb_multi_ractor_p()) { + concurrent = true; + } + break; + case T_OBJECT: + fields_obj = obj; + break; + default: { + fields_obj = rb_obj_fields(obj, id); + break; + } + } + + if (!fields_obj) { + return undef; + } + + const VALUE original_fields_obj = fields_obj; + if (concurrent) { + fields_obj = rb_imemo_fields_clone(fields_obj); } - return FALSE; + + shape_id_t old_shape_id = RBASIC_SHAPE_ID(fields_obj); + shape_id_t removed_shape_id; + shape_id_t next_shape_id = rb_shape_transition_remove_ivar(fields_obj, id, &removed_shape_id); + + if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) { + if (UNLIKELY(!rb_shape_too_complex_p(old_shape_id))) { + if (type == T_OBJECT) { + rb_evict_fields_to_hash(obj); + } + else { + fields_obj = imemo_fields_complex_from_obj(obj, fields_obj, next_shape_id); + } + } + st_data_t key = id; + if (!st_delete(rb_imemo_fields_complex_tbl(fields_obj), &key, (st_data_t *)&val)) { + val = undef; + } + } + else { + if (next_shape_id == old_shape_id) { + return undef; + } + + RUBY_ASSERT(removed_shape_id != INVALID_SHAPE_ID); + RUBY_ASSERT(RSHAPE_LEN(next_shape_id) == RSHAPE_LEN(old_shape_id) - 1); + + VALUE *fields = rb_imemo_fields_ptr(fields_obj); + attr_index_t removed_index = RSHAPE_INDEX(removed_shape_id); + val = fields[removed_index]; + + attr_index_t new_fields_count = RSHAPE_LEN(next_shape_id); + if (new_fields_count) { + size_t trailing_fields = new_fields_count - removed_index; + + MEMMOVE(&fields[removed_index], &fields[removed_index + 1], VALUE, trailing_fields); + RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id); + + if (FL_TEST_RAW(fields_obj, OBJ_FIELD_HEAP) && rb_obj_embedded_size(new_fields_count) <= rb_gc_obj_slot_size(fields_obj)) { + // Re-embed objects when instances become small enough + // This is necessary because YJIT assumes that objects with the same shape + // have the same embeddedness for efficiency (avoid extra checks) + FL_UNSET_RAW(fields_obj, ROBJECT_HEAP); + MEMCPY(rb_imemo_fields_ptr(fields_obj), fields, VALUE, new_fields_count); + xfree(fields); + } + } + else { + fields_obj = 0; + rb_free_generic_ivar(obj); + } + } + + RBASIC_SET_SHAPE_ID(obj, next_shape_id); + if (fields_obj != original_fields_obj) { + switch (type) { + case T_OBJECT: + break; + case T_CLASS: + case T_MODULE: + RCLASS_WRITABLE_SET_FIELDS_OBJ(obj, fields_obj); + break; + default: + rb_obj_set_fields(obj, fields_obj, id, original_fields_obj); + break; + } + } + + return val; +} + +VALUE +rb_attr_delete(VALUE obj, ID id) +{ + return rb_ivar_delete(obj, id, Qnil); +} + +void +rb_obj_init_too_complex(VALUE obj, st_table *table) +{ + // This method is meant to be called on newly allocated object. + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + RUBY_ASSERT(rb_shape_canonical_p(RBASIC_SHAPE_ID(obj))); + RUBY_ASSERT(RSHAPE_LEN(RBASIC_SHAPE_ID(obj)) == 0); + + obj_transition_too_complex(obj, table); } static int -ivar_i(key, entry, ary) - ID key; - struct global_entry *entry; - VALUE ary; +imemo_fields_complex_from_obj_i(ID key, VALUE val, st_data_t arg) { - if (rb_is_instance_id(key)) { - ary_push(ary, str_new2(rb_id2name(key))); + VALUE fields = (VALUE)arg; + st_table *table = rb_imemo_fields_complex_tbl(fields); + + RUBY_ASSERT(!st_lookup(table, (st_data_t)key, NULL)); + st_add_direct(table, (st_data_t)key, (st_data_t)val); + RB_OBJ_WRITTEN(fields, Qundef, val); + + return ST_CONTINUE; +} + +static VALUE +imemo_fields_complex_from_obj(VALUE owner, VALUE source_fields_obj, shape_id_t shape_id) +{ + attr_index_t len = source_fields_obj ? RSHAPE_LEN(RBASIC_SHAPE_ID(source_fields_obj)) : 0; + VALUE fields_obj = rb_imemo_fields_new_complex(owner, len + 1, RB_OBJ_SHAREABLE_P(owner)); + + rb_field_foreach(source_fields_obj, imemo_fields_complex_from_obj_i, (st_data_t)fields_obj, false); + RBASIC_SET_SHAPE_ID(fields_obj, shape_id); + + return fields_obj; +} + +static VALUE +imemo_fields_copy_capa(VALUE owner, VALUE source_fields_obj, attr_index_t new_size) +{ + VALUE fields_obj = rb_imemo_fields_new(owner, new_size, RB_OBJ_SHAREABLE_P(owner)); + if (source_fields_obj) { + attr_index_t fields_count = RSHAPE_LEN(RBASIC_SHAPE_ID(source_fields_obj)); + VALUE *fields = rb_imemo_fields_ptr(fields_obj); + MEMCPY(fields, rb_imemo_fields_ptr(source_fields_obj), VALUE, fields_count); + RBASIC_SET_SHAPE_ID(fields_obj, RBASIC_SHAPE_ID(source_fields_obj)); + for (attr_index_t i = 0; i < fields_count; i++) { + RB_OBJ_WRITTEN(fields_obj, Qundef, fields[i]); + } + } + return fields_obj; +} + +static VALUE +imemo_fields_set(VALUE owner, VALUE fields_obj, shape_id_t target_shape_id, ID field_name, VALUE val, bool concurrent) +{ + const VALUE original_fields_obj = fields_obj; + shape_id_t current_shape_id = fields_obj ? RBASIC_SHAPE_ID(fields_obj) : ROOT_SHAPE_ID; + + if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) { + if (rb_shape_too_complex_p(current_shape_id)) { + if (concurrent) { + // In multi-ractor case, we must always work on a copy because + // even if the field already exist, inserting in a st_table may + // cause a rebuild. + fields_obj = rb_imemo_fields_clone(fields_obj); + } + } + else { + fields_obj = imemo_fields_complex_from_obj(owner, original_fields_obj, target_shape_id); + current_shape_id = target_shape_id; + } + + st_table *table = rb_imemo_fields_complex_tbl(fields_obj); + + RUBY_ASSERT(field_name); + st_insert(table, (st_data_t)field_name, (st_data_t)val); + RB_OBJ_WRITTEN(fields_obj, Qundef, val); + RBASIC_SET_SHAPE_ID(fields_obj, target_shape_id); } + else { + attr_index_t index = RSHAPE_INDEX(target_shape_id); + if (concurrent || index >= RSHAPE_CAPACITY(current_shape_id)) { + fields_obj = imemo_fields_copy_capa(owner, original_fields_obj, RSHAPE_CAPACITY(target_shape_id)); + } + + VALUE *table = rb_imemo_fields_ptr(fields_obj); + RB_OBJ_WRITE(fields_obj, &table[index], val); + + if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) { + RBASIC_SET_SHAPE_ID(fields_obj, target_shape_id); + } + } + + return fields_obj; +} + +static attr_index_t +generic_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val) +{ + if (!field_name) { + field_name = RSHAPE_EDGE_NAME(target_shape_id); + RUBY_ASSERT(field_name); + } + + const VALUE original_fields_obj = rb_obj_fields(obj, field_name); + VALUE fields_obj = imemo_fields_set(obj, original_fields_obj, target_shape_id, field_name, val, false); + + rb_obj_set_fields(obj, fields_obj, field_name, original_fields_obj); + return rb_shape_too_complex_p(target_shape_id) ? ATTR_INDEX_NOT_SET : RSHAPE_INDEX(target_shape_id); +} + +static shape_id_t +generic_shape_ivar(VALUE obj, ID id, bool *new_ivar_out) +{ + bool new_ivar = false; + shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj); + shape_id_t target_shape_id = current_shape_id; + + if (!rb_shape_too_complex_p(current_shape_id)) { + if (!rb_shape_find_ivar(current_shape_id, id, &target_shape_id)) { + if (RSHAPE_LEN(current_shape_id) >= SHAPE_MAX_FIELDS) { + rb_raise(rb_eArgError, "too many instance variables"); + } + + new_ivar = true; + target_shape_id = rb_shape_transition_add_ivar(obj, id); + } + } + + *new_ivar_out = new_ivar; + return target_shape_id; +} + +static attr_index_t +generic_ivar_set(VALUE obj, ID id, VALUE val) +{ + bool dontcare; + shape_id_t target_shape_id = generic_shape_ivar(obj, id, &dontcare); + return generic_field_set(obj, target_shape_id, id, val); +} + +void +rb_ensure_iv_list_size(VALUE obj, uint32_t current_len, uint32_t new_capacity) +{ + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + + if (FL_TEST_RAW(obj, ROBJECT_HEAP)) { + REALLOC_N(ROBJECT(obj)->as.heap.fields, VALUE, new_capacity); + } + else { + VALUE *ptr = ROBJECT_FIELDS(obj); + VALUE *newptr = ALLOC_N(VALUE, new_capacity); + MEMCPY(newptr, ptr, VALUE, current_len); + FL_SET_RAW(obj, ROBJECT_HEAP); + ROBJECT(obj)->as.heap.fields = newptr; + } +} + +static int +rb_obj_copy_ivs_to_hash_table_i(ID key, VALUE val, st_data_t arg) +{ + RUBY_ASSERT(!st_lookup((st_table *)arg, (st_data_t)key, NULL)); + + st_add_direct((st_table *)arg, (st_data_t)key, (st_data_t)val); return ST_CONTINUE; } +void +rb_obj_copy_ivs_to_hash_table(VALUE obj, st_table *table) +{ + rb_ivar_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table); +} + +void +rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table) +{ + rb_field_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table, false); +} + +static attr_index_t +obj_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val) +{ + shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj); + + if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) { + if (UNLIKELY(!rb_shape_too_complex_p(current_shape_id))) { + current_shape_id = rb_evict_fields_to_hash(obj); + } + + if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) { + RBASIC_SET_SHAPE_ID(obj, target_shape_id); + } + + if (!field_name) { + field_name = RSHAPE_EDGE_NAME(target_shape_id); + RUBY_ASSERT(field_name); + } + + st_insert(ROBJECT_FIELDS_HASH(obj), (st_data_t)field_name, (st_data_t)val); + RB_OBJ_WRITTEN(obj, Qundef, val); + + return ATTR_INDEX_NOT_SET; + } + else { + attr_index_t index = RSHAPE_INDEX(target_shape_id); + + if (index >= RSHAPE_LEN(current_shape_id)) { + if (UNLIKELY(index >= RSHAPE_CAPACITY(current_shape_id))) { + rb_ensure_iv_list_size(obj, RSHAPE_CAPACITY(current_shape_id), RSHAPE_CAPACITY(target_shape_id)); + } + RBASIC_SET_SHAPE_ID(obj, target_shape_id); + } + + RB_OBJ_WRITE(obj, &ROBJECT_FIELDS(obj)[index], val); + + return index; + } +} + +static attr_index_t +obj_ivar_set(VALUE obj, ID id, VALUE val) +{ + bool dontcare; + shape_id_t target_shape_id = generic_shape_ivar(obj, id, &dontcare); + return obj_field_set(obj, target_shape_id, id, val); +} + +/* Set the instance variable +val+ on object +obj+ at ivar name +id+. + * This function only works with T_OBJECT objects, so make sure + * +obj+ is of type T_OBJECT before using this function. + */ VALUE -obj_instance_variables(obj) - VALUE obj; +rb_vm_set_ivar_id(VALUE obj, ID id, VALUE val) { - VALUE ary; + rb_check_frozen(obj); + obj_ivar_set(obj, id, val); + return val; +} + +void rb_obj_freeze_inline(VALUE x) +{ + if (RB_FL_ABLE(x)) { + RB_FL_SET_RAW(x, RUBY_FL_FREEZE); + if (TYPE(x) == T_STRING) { + RB_FL_UNSET_RAW(x, FL_USER2 | FL_USER3); // STR_CHILLED + } + + RB_SET_SHAPE_ID(x, rb_shape_transition_frozen(x)); - switch (TYPE(obj)) { + if (RBASIC_CLASS(x)) { + rb_freeze_singleton_class(x); + } + } +} + +static attr_index_t class_ivar_set(VALUE obj, ID id, VALUE val, bool *new_ivar); + +static attr_index_t +ivar_set(VALUE obj, ID id, VALUE val) +{ + RB_DEBUG_COUNTER_INC(ivar_set_base); + + switch (BUILTIN_TYPE(obj)) { case T_OBJECT: + return obj_ivar_set(obj, id, val); case T_CLASS: case T_MODULE: - ary = ary_new(); - if (ROBJECT(obj)->iv_tbl) { - st_foreach(ROBJECT(obj)->iv_tbl, ivar_i, ary); - } - return ary; + { + IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id); + bool dontcare; + return class_ivar_set(obj, id, val, &dontcare); + } + default: + return generic_ivar_set(obj, id, val); } - return Qnil; } VALUE -obj_remove_instance_variable(obj, name) - VALUE obj, name; +rb_ivar_set(VALUE obj, ID id, VALUE val) +{ + rb_check_frozen(obj); + ivar_set(obj, id, val); + return val; +} + +attr_index_t +rb_ivar_set_index(VALUE obj, ID id, VALUE val) +{ + return ivar_set(obj, id, val); +} + +void +rb_ivar_set_internal(VALUE obj, ID id, VALUE val) +{ + // should be internal instance variable name (no @ prefix) + VM_ASSERT(!rb_is_instance_id(id)); + + ivar_set(obj, id, val); +} + +attr_index_t +rb_obj_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val) +{ + switch (BUILTIN_TYPE(obj)) { + case T_OBJECT: + return obj_field_set(obj, target_shape_id, field_name, val); + case T_CLASS: + case T_MODULE: + // The only field is object_id and T_CLASS handle it differently. + rb_bug("Unreachable"); + break; + default: + return generic_field_set(obj, target_shape_id, field_name, val); + } +} + +static VALUE +ivar_defined0(VALUE obj, ID id) +{ + attr_index_t index; + + if (rb_shape_obj_too_complex_p(obj)) { + VALUE idx; + st_table *table = NULL; + switch (BUILTIN_TYPE(obj)) { + case T_CLASS: + case T_MODULE: + rb_bug("Unreachable"); + break; + + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + table = rb_imemo_fields_complex_tbl(obj); + break; + + case T_OBJECT: + table = ROBJECT_FIELDS_HASH(obj); + break; + + default: { + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); // defined? doesn't require ractor checks + table = rb_imemo_fields_complex_tbl(fields_obj); + } + } + + if (!table || !rb_st_lookup(table, id, &idx)) { + return Qfalse; + } + + return Qtrue; + } + else { + return RBOOL(rb_shape_get_iv_index(RBASIC_SHAPE_ID(obj), id, &index)); + } +} + +VALUE +rb_ivar_defined(VALUE obj, ID id) +{ + if (SPECIAL_CONST_P(obj)) return Qfalse; + + VALUE defined = Qfalse; + switch (BUILTIN_TYPE(obj)) { + case T_CLASS: + case T_MODULE: + { + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (fields_obj) { + defined = ivar_defined0(fields_obj, id); + } + } + break; + default: + defined = ivar_defined0(obj, id); + break; + } + return defined; +} + +struct iv_itr_data { + VALUE obj; + struct gen_fields_tbl *fields_tbl; + st_data_t arg; + rb_ivar_foreach_callback_func *func; + VALUE *fields; + bool ivar_only; +}; + +static int +iterate_over_shapes_callback(shape_id_t shape_id, void *data) +{ + struct iv_itr_data *itr_data = data; + + if (itr_data->ivar_only && !RSHAPE_TYPE_P(shape_id, SHAPE_IVAR)) { + return ST_CONTINUE; + } + + VALUE *fields; + switch (BUILTIN_TYPE(itr_data->obj)) { + case T_OBJECT: + RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj)); + fields = ROBJECT_FIELDS(itr_data->obj); + break; + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(itr_data->obj, imemo_fields)); + RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj)); + + fields = rb_imemo_fields_ptr(itr_data->obj); + break; + default: + rb_bug("Unreachable"); + } + + VALUE val = fields[RSHAPE_INDEX(shape_id)]; + return itr_data->func(RSHAPE_EDGE_NAME(shape_id), val, itr_data->arg); +} + +/* + * Returns a flag to stop iterating depending on the result of +callback+. + */ +static void +iterate_over_shapes(shape_id_t shape_id, rb_ivar_foreach_callback_func *callback, struct iv_itr_data *itr_data) { - VALUE val = Qnil; - ID id = rb_to_id(name); + rb_shape_foreach_field(shape_id, iterate_over_shapes_callback, itr_data); +} + +static int +each_hash_iv(st_data_t id, st_data_t val, st_data_t data) +{ + struct iv_itr_data * itr_data = (struct iv_itr_data *)data; + rb_ivar_foreach_callback_func *callback = itr_data->func; + if (is_internal_id((ID)id)) { + return ST_CONTINUE; + } + return callback((ID)id, (VALUE)val, itr_data->arg); +} - if (!rb_is_instance_id(id)) { - NameError("`%s' is not an instance variable", rb_id2name(id)); +static void +obj_fields_each(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only) +{ + struct iv_itr_data itr_data = { + .obj = obj, + .arg = arg, + .func = func, + .ivar_only = ivar_only, + }; + + shape_id_t shape_id = RBASIC_SHAPE_ID(obj); + if (rb_shape_too_complex_p(shape_id)) { + rb_st_foreach(ROBJECT_FIELDS_HASH(obj), each_hash_iv, (st_data_t)&itr_data); } + else { + itr_data.fields = ROBJECT_FIELDS(obj); + iterate_over_shapes(shape_id, func, &itr_data); + } +} - switch (TYPE(obj)) { +static void +imemo_fields_each(VALUE fields_obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only) +{ + IMEMO_TYPE_P(fields_obj, imemo_fields); + + struct iv_itr_data itr_data = { + .obj = fields_obj, + .arg = arg, + .func = func, + .ivar_only = ivar_only, + }; + + shape_id_t shape_id = RBASIC_SHAPE_ID(fields_obj); + if (rb_shape_too_complex_p(shape_id)) { + rb_st_foreach(rb_imemo_fields_complex_tbl(fields_obj), each_hash_iv, (st_data_t)&itr_data); + } + else { + itr_data.fields = rb_imemo_fields_ptr(fields_obj); + iterate_over_shapes(shape_id, func, &itr_data); + } +} + +void +rb_copy_generic_ivar(VALUE dest, VALUE obj) +{ + VALUE new_fields_obj; + + rb_check_frozen(dest); + + if (!rb_obj_gen_fields_p(obj)) { + return; + } + + shape_id_t src_shape_id = rb_obj_shape_id(obj); + + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); + if (fields_obj) { + unsigned long src_num_ivs = rb_ivar_count(fields_obj); + if (!src_num_ivs) { + goto clear; + } + + if (rb_shape_too_complex_p(src_shape_id)) { + rb_shape_copy_complex_ivars(dest, obj, src_shape_id, rb_imemo_fields_complex_tbl(fields_obj)); + return; + } + + shape_id_t dest_shape_id = src_shape_id; + shape_id_t initial_shape_id = rb_obj_shape_id(dest); + + if (!rb_shape_canonical_p(src_shape_id)) { + RUBY_ASSERT(RSHAPE_TYPE_P(initial_shape_id, SHAPE_ROOT)); + + dest_shape_id = rb_shape_rebuild(initial_shape_id, src_shape_id); + if (UNLIKELY(rb_shape_too_complex_p(dest_shape_id))) { + st_table *table = rb_st_init_numtable_with_size(src_num_ivs); + rb_obj_copy_ivs_to_hash_table(obj, table); + rb_obj_init_too_complex(dest, table); + return; + } + } + + if (!RSHAPE_LEN(dest_shape_id)) { + RBASIC_SET_SHAPE_ID(dest, dest_shape_id); + return; + } + + new_fields_obj = rb_imemo_fields_new(dest, RSHAPE_CAPACITY(dest_shape_id), RB_OBJ_SHAREABLE_P(dest)); + VALUE *src_buf = rb_imemo_fields_ptr(fields_obj); + VALUE *dest_buf = rb_imemo_fields_ptr(new_fields_obj); + rb_shape_copy_fields(new_fields_obj, dest_buf, dest_shape_id, src_buf, src_shape_id); + RBASIC_SET_SHAPE_ID(new_fields_obj, dest_shape_id); + + rb_obj_replace_fields(dest, new_fields_obj); + } + return; + + clear: + rb_free_generic_ivar(dest); +} + +void +rb_replace_generic_ivar(VALUE clone, VALUE obj) +{ + RB_VM_LOCKING() { + st_data_t fields_tbl, obj_data = (st_data_t)obj; + if (st_delete(generic_fields_tbl_, &obj_data, &fields_tbl)) { + st_insert(generic_fields_tbl_, (st_data_t)clone, fields_tbl); + RB_OBJ_WRITTEN(clone, Qundef, fields_tbl); + } + else { + rb_bug("unreachable"); + } + } +} + +void +rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only) +{ + if (SPECIAL_CONST_P(obj)) return; + switch (BUILTIN_TYPE(obj)) { + case T_IMEMO: + if (IMEMO_TYPE_P(obj, imemo_fields)) { + imemo_fields_each(obj, func, arg, ivar_only); + } + break; case T_OBJECT: + obj_fields_each(obj, func, arg, ivar_only); + break; case T_CLASS: case T_MODULE: - if (ROBJECT(obj)->iv_tbl) { - st_delete(ROBJECT(obj)->iv_tbl, &id, &val); - } - break; + { + IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(0); + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (fields_obj) { + imemo_fields_each(fields_obj, func, arg, ivar_only); + } + } + break; default: - TypeError("object %s can not have instance variables", - rb_class2name(CLASS_OF(obj))); - break; + { + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); + if (fields_obj) { + imemo_fields_each(fields_obj, func, arg, ivar_only); + } + } + break; } - return val; } +void +rb_ivar_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg) +{ + rb_field_foreach(obj, func, arg, true); +} + +st_index_t +rb_ivar_count(VALUE obj) +{ + if (SPECIAL_CONST_P(obj)) return 0; + + st_index_t iv_count = 0; + switch (BUILTIN_TYPE(obj)) { + case T_OBJECT: + iv_count = ROBJECT_FIELDS_COUNT(obj); + break; + + case T_CLASS: + case T_MODULE: + { + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (!fields_obj) { + return 0; + } + if (rb_shape_obj_too_complex_p(fields_obj)) { + iv_count = rb_st_table_size(rb_imemo_fields_complex_tbl(fields_obj)); + } + else { + iv_count = RBASIC_FIELDS_COUNT(fields_obj); + } + } + break; + + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + + if (rb_shape_obj_too_complex_p(obj)) { + iv_count = rb_st_table_size(rb_imemo_fields_complex_tbl(obj)); + } + else { + iv_count = RBASIC_FIELDS_COUNT(obj); + } + break; + + default: + { + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); + if (fields_obj) { + if (rb_shape_obj_too_complex_p(fields_obj)) { + rb_st_table_size(rb_imemo_fields_complex_tbl(fields_obj)); + } + else { + iv_count = RBASIC_FIELDS_COUNT(obj); + } + } + } + break; + } + + if (rb_shape_obj_has_id(obj)) { + iv_count--; + } + + return iv_count; +} + +static int +ivar_i(ID key, VALUE v, st_data_t a) +{ + VALUE ary = (VALUE)a; + + if (rb_is_instance_id(key)) { + rb_ary_push(ary, ID2SYM(key)); + } + return ST_CONTINUE; +} + +/* + * call-seq: + * obj.instance_variables -> array + * + * Returns an array of instance variable names for the receiver. Note + * that simply defining an accessor does not create the corresponding + * instance variable. + * + * class Fred + * attr_accessor :a1 + * def initialize + * @iv = 3 + * end + * end + * Fred.new.instance_variables #=> [:@iv] + */ + VALUE -rb_const_get_at(klass, id) - VALUE klass; - ID id; +rb_obj_instance_variables(VALUE obj) { - VALUE value; + VALUE ary; - if (RCLASS(klass)->iv_tbl && st_lookup(RCLASS(klass)->iv_tbl, id, &value)) { - return value; + ary = rb_ary_new(); + rb_ivar_foreach(obj, ivar_i, ary); + return ary; +} + +#define rb_is_constant_id rb_is_const_id +#define rb_is_constant_name rb_is_const_name +#define id_for_var(obj, name, part, type) \ + id_for_var_message(obj, name, type, "'%1$s' is not allowed as "#part" "#type" variable name") +#define id_for_var_message(obj, name, type, message) \ + check_id_type(obj, &(name), rb_is_##type##_id, rb_is_##type##_name, message, strlen(message)) +static ID +check_id_type(VALUE obj, VALUE *pname, + int (*valid_id_p)(ID), int (*valid_name_p)(VALUE), + const char *message, size_t message_len) +{ + ID id = rb_check_id(pname); + VALUE name = *pname; + + if (id ? !valid_id_p(id) : !valid_name_p(name)) { + rb_name_err_raise_str(rb_fstring_new(message, message_len), + obj, name); } - if (klass == cObject) { - return rb_const_get(klass, id); + return id; +} + +/* + * call-seq: + * obj.remove_instance_variable(symbol) -> obj + * obj.remove_instance_variable(string) -> obj + * + * Removes the named instance variable from <i>obj</i>, returning that + * variable's value. The name can be passed as a symbol or as a string. + * + * class Dummy + * attr_reader :var + * def initialize + * @var = 99 + * end + * def remove + * remove_instance_variable(:@var) + * end + * end + * d = Dummy.new + * d.var #=> 99 + * d.remove #=> 99 + * d.var #=> nil + */ + +VALUE +rb_obj_remove_instance_variable(VALUE obj, VALUE name) +{ + const ID id = id_for_var(obj, name, an, instance); + + // Frozen check comes here because it's expected that we raise a + // NameError (from the id_for_var check) before we raise a FrozenError + rb_check_frozen(obj); + + if (id) { + VALUE val = rb_ivar_delete(obj, id, Qundef); + + if (!UNDEF_P(val)) return val; } - NameError("Uninitialized constant %s::%s", - RSTRING(rb_class_path(klass))->ptr, - rb_id2name(id)); - /* not reached */ + + rb_name_err_raise("instance variable %1$s not defined", + obj, name); + UNREACHABLE_RETURN(Qnil); } +NORETURN(static void uninitialized_constant(VALUE, VALUE)); +static void +uninitialized_constant(VALUE klass, VALUE name) +{ + if (klass && rb_class_real(klass) != rb_cObject) + rb_name_err_raise("uninitialized constant %2$s::%1$s", + klass, name); + else + rb_name_err_raise("uninitialized constant %1$s", + klass, name); +} VALUE -rb_const_get(klass, id) - VALUE klass; - ID id; +rb_const_missing(VALUE klass, VALUE name) +{ + VALUE value = rb_funcallv(klass, idConst_missing, 1, &name); + rb_vm_inc_const_missing_count(); + return value; +} + + +/* + * call-seq: + * mod.const_missing(sym) -> obj + * + * Invoked when a reference is made to an undefined constant in + * <i>mod</i>. It is passed a symbol for the undefined constant, and + * returns a value to be used for that constant. For example, consider: + * + * def Foo.const_missing(name) + * name # return the constant name as Symbol + * end + * + * Foo::UNDEFINED_CONST #=> :UNDEFINED_CONST: symbol returned + * + * As the example above shows, +const_missing+ is not required to create the + * missing constant in <i>mod</i>, though that is often a side-effect. The + * caller gets its return value when triggered. If the constant is also defined, + * further lookups won't hit +const_missing+ and will return the value stored in + * the constant as usual. Otherwise, +const_missing+ will be invoked again. + * + * In the next example, when a reference is made to an undefined constant, + * +const_missing+ attempts to load a file whose path is the lowercase version + * of the constant name (thus class <code>Fred</code> is assumed to be in file + * <code>fred.rb</code>). If defined as a side-effect of loading the file, the + * method returns the value stored in the constant. This implements an autoload + * feature similar to Kernel#autoload and Module#autoload, though it differs in + * important ways. + * + * def Object.const_missing(name) + * @looked_for ||= {} + * str_name = name.to_s + * raise "Constant not found: #{name}" if @looked_for[str_name] + * @looked_for[str_name] = 1 + * file = str_name.downcase + * require file + * const_get(name, false) + * end + * + */ + +VALUE +rb_mod_const_missing(VALUE klass, VALUE name) +{ + rb_execution_context_t *ec = GET_EC(); + VALUE ref = ec->private_const_reference; + rb_vm_pop_cfunc_frame(); + if (ref) { + ec->private_const_reference = 0; + rb_name_err_raise("private constant %2$s::%1$s referenced", ref, name); + } + uninitialized_constant(klass, name); + + UNREACHABLE_RETURN(Qnil); +} + +static void +autoload_table_mark(void *ptr) +{ + rb_mark_tbl_no_pin((st_table *)ptr); +} + +static void +autoload_table_free(void *ptr) +{ + st_free_table((st_table *)ptr); +} + +static size_t +autoload_table_memsize(const void *ptr) { + const st_table *tbl = ptr; + return st_memsize(tbl); +} + +static void +autoload_table_compact(void *ptr) +{ + rb_gc_ref_update_table_values_only((st_table *)ptr); +} + +static const rb_data_type_t autoload_table_type = { + "autoload_table", + {autoload_table_mark, autoload_table_free, autoload_table_memsize, autoload_table_compact,}, + 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED +}; + +#define check_autoload_table(av) \ + (struct st_table *)rb_check_typeddata((av), &autoload_table_type) + +static VALUE +autoload_data(VALUE mod, ID id) +{ + struct st_table *tbl; + st_data_t val; + + // If we are called with a non-origin ICLASS, fetch the autoload data from + // the original module. + if (RB_TYPE_P(mod, T_ICLASS)) { + if (RICLASS_IS_ORIGIN_P(mod)) { + return 0; + } + else { + mod = RBASIC(mod)->klass; + } + } + + RUBY_ASSERT(RB_TYPE_P(mod, T_CLASS) || RB_TYPE_P(mod, T_MODULE)); + + // Look up the instance variable table for `autoload`, then index into that table with the given constant name `id`. + + VALUE tbl_value = rb_ivar_lookup(mod, autoload, Qfalse); + if (!RTEST(tbl_value) || !(tbl = check_autoload_table(tbl_value)) || !st_lookup(tbl, (st_data_t)id, &val)) { + return 0; + } + + return (VALUE)val; +} + +// Every autoload constant has exactly one instance of autoload_const, stored in `autoload_features`. Since multiple autoload constants can refer to the same file, every `autoload_const` refers to a de-duplicated `autoload_data`. +struct autoload_const { + // The linked list node of all constants which are loaded by the related autoload feature. + struct ccan_list_node cnode; /* <=> autoload_data.constants */ + + // The shared "autoload_data" if multiple constants are defined from the same feature. + VALUE autoload_data_value; + + // The box object when the autoload is called in a user box + // Otherwise, Qnil means the root box + VALUE box_value; + + // The module we are loading a constant into. + VALUE module; + + // The name of the constant we are loading. + ID name; + + // The value of the constant (after it's loaded). VALUE value; - VALUE tmp; - tmp = klass; - while (tmp) { - if (RCLASS(tmp)->iv_tbl && st_lookup(RCLASS(tmp)->iv_tbl,id,&value)) { - return value; - } - tmp = RCLASS(tmp)->super; + // The constant entry flags which need to be re-applied after autoloading the feature. + rb_const_flag_t flag; + + // The source file and line number that defined this constant (different from feature path). + VALUE file; + int line; +}; + +// Each `autoload_data` uniquely represents a specific feature which can be loaded, and a list of constants which it is able to define. We use a mutex to coordinate multiple threads trying to load the same feature. +struct autoload_data { + // The feature path to require to load this constant. + VALUE feature; + + // The mutex which is protecting autoloading this feature. + VALUE mutex; + + // The process fork serial number since the autoload mutex will become invalid on fork. + rb_serial_t fork_gen; + + // The linked list of all constants that are going to be loaded by this autoload. + struct ccan_list_head constants; /* <=> autoload_const.cnode */ +}; + +static void +autoload_data_mark_and_move(void *ptr) +{ + struct autoload_data *p = ptr; + + rb_gc_mark_and_move(&p->feature); + rb_gc_mark_and_move(&p->mutex); +} + +static void +autoload_data_free(void *ptr) +{ + struct autoload_data *p = ptr; + + struct autoload_const *autoload_const, *next; + ccan_list_for_each_safe(&p->constants, autoload_const, next, cnode) { + ccan_list_del_init(&autoload_const->cnode); + } + + ruby_xfree(p); +} + +static size_t +autoload_data_memsize(const void *ptr) +{ + return sizeof(struct autoload_data); +} + +static const rb_data_type_t autoload_data_type = { + "autoload_data", + {autoload_data_mark_and_move, autoload_data_free, autoload_data_memsize, autoload_data_mark_and_move}, + 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED +}; + +static void +autoload_const_mark_and_move(void *ptr) +{ + struct autoload_const *ac = ptr; + + rb_gc_mark_and_move(&ac->module); + rb_gc_mark_and_move(&ac->autoload_data_value); + rb_gc_mark_and_move(&ac->value); + rb_gc_mark_and_move(&ac->file); + rb_gc_mark_and_move(&ac->box_value); +} + +static size_t +autoload_const_memsize(const void *ptr) +{ + return sizeof(struct autoload_const); +} + +static void +autoload_const_free(void *ptr) +{ + struct autoload_const *autoload_const = ptr; + + ccan_list_del(&autoload_const->cnode); + ruby_xfree(ptr); +} + +static const rb_data_type_t autoload_const_type = { + "autoload_const", + {autoload_const_mark_and_move, autoload_const_free, autoload_const_memsize, autoload_const_mark_and_move,}, + 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED +}; + +static struct autoload_data * +get_autoload_data(VALUE autoload_const_value, struct autoload_const **autoload_const_pointer) +{ + struct autoload_const *autoload_const = rb_check_typeddata(autoload_const_value, &autoload_const_type); + + VALUE autoload_data_value = autoload_const->autoload_data_value; + struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type); + + /* do not reach across stack for ->state after forking: */ + if (autoload_data && autoload_data->fork_gen != GET_VM()->fork_gen) { + RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil); + autoload_data->fork_gen = 0; } - if (BUILTIN_TYPE(klass) == T_MODULE) { - return rb_const_get(cObject, id); + + if (autoload_const_pointer) *autoload_const_pointer = autoload_const; + + return autoload_data; +} + +struct autoload_copy_table_data { + VALUE dst_tbl_value; + struct st_table *dst_tbl; + const rb_box_t *box; +}; + +static int +autoload_copy_table_for_box_i(st_data_t key, st_data_t value, st_data_t arg) +{ + struct autoload_const *autoload_const; + struct autoload_copy_table_data *data = (struct autoload_copy_table_data *)arg; + struct st_table *tbl = data->dst_tbl; + VALUE tbl_value = data->dst_tbl_value; + const rb_box_t *box = data->box; + + VALUE src_value = (VALUE)value; + struct autoload_const *src_const = rb_check_typeddata(src_value, &autoload_const_type); + // autoload_data can be shared between copies because the feature is equal between copies. + VALUE autoload_data_value = src_const->autoload_data_value; + struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type); + + VALUE new_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const); + RB_OBJ_WRITE(new_value, &autoload_const->box_value, rb_get_box_object((rb_box_t *)box)); + RB_OBJ_WRITE(new_value, &autoload_const->module, src_const->module); + autoload_const->name = src_const->name; + RB_OBJ_WRITE(new_value, &autoload_const->value, src_const->value); + autoload_const->flag = src_const->flag; + RB_OBJ_WRITE(new_value, &autoload_const->autoload_data_value, autoload_data_value); + ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode); + + st_insert(tbl, (st_data_t)autoload_const->name, (st_data_t)new_value); + RB_OBJ_WRITTEN(tbl_value, Qundef, new_value); + + return ST_CONTINUE; +} + +void +rb_autoload_copy_table_for_box(st_table *iv_ptr, const rb_box_t *box) +{ + struct st_table *src_tbl, *dst_tbl; + VALUE src_tbl_value, dst_tbl_value; + if (!rb_st_lookup(iv_ptr, (st_data_t)autoload, (st_data_t *)&src_tbl_value)) { + // the class has no autoload table yet. + return; } + if (!RTEST(src_tbl_value) || !(src_tbl = check_autoload_table(src_tbl_value))) { + // the __autoload__ ivar value isn't autoload table value. + return; + } + src_tbl = check_autoload_table(src_tbl_value); + + dst_tbl_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL); + RTYPEDDATA_DATA(dst_tbl_value) = dst_tbl = st_init_numtable(); + + struct autoload_copy_table_data data = { + .dst_tbl_value = dst_tbl_value, + .dst_tbl = dst_tbl, + .box = box, + }; + + st_foreach(src_tbl, autoload_copy_table_for_box_i, (st_data_t)&data); + st_insert(iv_ptr, (st_data_t)autoload, (st_data_t)dst_tbl_value); +} + +void +rb_autoload(VALUE module, ID name, const char *feature) +{ + if (!feature || !*feature) { + rb_raise(rb_eArgError, "empty feature name"); + } + + rb_autoload_str(module, name, rb_fstring_cstr(feature)); +} + +static void const_set(VALUE klass, ID id, VALUE val); +static void const_added(VALUE klass, ID const_name); + +struct autoload_arguments { + VALUE module; + ID name; + VALUE feature; + VALUE box_value; +}; + +static VALUE +autoload_feature_lookup_or_create(VALUE feature, struct autoload_data **autoload_data_pointer) +{ + RUBY_ASSERT_MUTEX_OWNED(autoload_mutex); + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); - /* pre-defined class */ - if (st_lookup(class_tbl, id, &value)) return value; + VALUE autoload_data_value = rb_hash_aref(autoload_features, feature); + struct autoload_data *autoload_data; - /* autoload */ - if (autoload_tbl && st_lookup(autoload_tbl, id, 0)) { - char *modname; - VALUE module; + if (NIL_P(autoload_data_value)) { + autoload_data_value = TypedData_Make_Struct(0, struct autoload_data, &autoload_data_type, autoload_data); + RB_OBJ_WRITE(autoload_data_value, &autoload_data->feature, feature); + RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil); + ccan_list_head_init(&autoload_data->constants); - st_delete(autoload_tbl, &id, &modname); - module = str_new2(modname); - free(modname); - f_require(0, module); - return rb_const_get(klass, id); + if (autoload_data_pointer) *autoload_data_pointer = autoload_data; + + rb_hash_aset(autoload_features, feature, autoload_data_value); + } + else if (autoload_data_pointer) { + *autoload_data_pointer = rb_check_typeddata(autoload_data_value, &autoload_data_type); } - /* Uninitialized constant */ - if (klass && klass != cObject) - NameError("Uninitialized constant %s::%s", - RSTRING(rb_class_path(klass))->ptr, - rb_id2name(id)); + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); + return autoload_data_value; +} + +static VALUE +autoload_table_lookup_or_create(VALUE module) +{ + VALUE autoload_table_value = rb_ivar_lookup(module, autoload, Qfalse); + if (RTEST(autoload_table_value)) { + return autoload_table_value; + } else { - NameError("Uninitialized constant %s",rb_id2name(id)); + autoload_table_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL); + rb_class_ivar_set(module, autoload, autoload_table_value); + RTYPEDDATA_DATA(autoload_table_value) = st_init_numtable(); + return autoload_table_value; + } +} + +static VALUE +autoload_synchronized(VALUE _arguments) +{ + struct autoload_arguments *arguments = (struct autoload_arguments *)_arguments; + + rb_const_entry_t *constant_entry = rb_const_lookup(arguments->module, arguments->name); + if (constant_entry && !UNDEF_P(constant_entry->value)) { + return Qfalse; + } + + // Reset any state associated with any previous constant: + const_set(arguments->module, arguments->name, Qundef); + + VALUE autoload_table_value = autoload_table_lookup_or_create(arguments->module); + struct st_table *autoload_table = check_autoload_table(autoload_table_value); + + // Ensure the string is uniqued since we use an identity lookup: + VALUE feature = rb_fstring(arguments->feature); + + struct autoload_data *autoload_data; + VALUE autoload_data_value = autoload_feature_lookup_or_create(feature, &autoload_data); + + { + struct autoload_const *autoload_const; + VALUE autoload_const_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const); + RB_OBJ_WRITE(autoload_const_value, &autoload_const->box_value, arguments->box_value); + RB_OBJ_WRITE(autoload_const_value, &autoload_const->module, arguments->module); + autoload_const->name = arguments->name; + autoload_const->value = Qundef; + autoload_const->flag = CONST_PUBLIC; + RB_OBJ_WRITE(autoload_const_value, &autoload_const->autoload_data_value, autoload_data_value); + ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode); + st_insert(autoload_table, (st_data_t)arguments->name, (st_data_t)autoload_const_value); + RB_OBJ_WRITTEN(autoload_table_value, Qundef, autoload_const_value); + } + + return Qtrue; +} + +void +rb_autoload_str(VALUE module, ID name, VALUE feature) +{ + const rb_box_t *box = rb_current_box(); + VALUE current_box_value = rb_get_box_object((rb_box_t *)box); + + if (!rb_is_const_id(name)) { + rb_raise(rb_eNameError, "autoload must be constant name: %"PRIsVALUE"", QUOTE_ID(name)); } - /* not reached */ + + Check_Type(feature, T_STRING); + if (!RSTRING_LEN(feature)) { + rb_raise(rb_eArgError, "empty feature name"); + } + + struct autoload_arguments arguments = { + .module = module, + .name = name, + .feature = feature, + .box_value = current_box_value, + }; + + VALUE result = rb_mutex_synchronize(autoload_mutex, autoload_synchronized, (VALUE)&arguments); + + if (result == Qtrue) { + const_added(module, name); + } +} + +static void +autoload_delete(VALUE module, ID name) +{ + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); + + st_data_t load = 0, key = name; + + RUBY_ASSERT(RB_TYPE_P(module, T_CLASS) || RB_TYPE_P(module, T_MODULE)); + + VALUE table_value = rb_ivar_lookup(module, autoload, Qfalse); + if (RTEST(table_value)) { + struct st_table *table = check_autoload_table(table_value); + + st_delete(table, &key, &load); + RB_OBJ_WRITTEN(table_value, load, Qundef); + + /* Qfalse can indicate already deleted */ + if (load != Qfalse) { + struct autoload_const *autoload_const; + struct autoload_data *autoload_data = get_autoload_data((VALUE)load, &autoload_const); + + VM_ASSERT(autoload_data); + VM_ASSERT(!ccan_list_empty(&autoload_data->constants)); + + /* + * we must delete here to avoid "already initialized" warnings + * with parallel autoload. Using list_del_init here so list_del + * works in autoload_const_free + */ + ccan_list_del_init(&autoload_const->cnode); + + if (ccan_list_empty(&autoload_data->constants)) { + rb_hash_delete(autoload_features, autoload_data->feature); + } + + // If the autoload table is empty, we can delete it. + if (table->num_entries == 0) { + rb_attr_delete(module, autoload); + } + } + } + + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); } static int -const_i(key, value, ary) - ID key; - VALUE value; - VALUE ary; +autoload_by_someone_else(struct autoload_data *ele) { - if (rb_is_const_id(key)) { - VALUE kval = str_new2(rb_id2name(key)); - if (!ary_includes(ary, kval)) { - ary_push(ary, kval); - } + return ele->mutex != Qnil && !rb_mutex_owned_p(ele->mutex); +} + +static VALUE +check_autoload_required(VALUE mod, ID id, const char **loadingpath) +{ + VALUE autoload_const_value = autoload_data(mod, id); + struct autoload_data *autoload_data; + const char *loading; + + if (!autoload_const_value || !(autoload_data = get_autoload_data(autoload_const_value, 0))) { + return 0; } - return ST_CONTINUE; + + VALUE feature = autoload_data->feature; + + /* + * if somebody else is autoloading, we MUST wait for them, since + * rb_provide_feature can provide a feature before autoload_const_set + * completes. We must wait until autoload_const_set finishes in + * the other thread. + */ + if (autoload_by_someone_else(autoload_data)) { + return autoload_const_value; + } + + loading = RSTRING_PTR(feature); + + if (!rb_feature_provided(loading, &loading)) { + return autoload_const_value; + } + + if (loadingpath && loading) { + *loadingpath = loading; + return autoload_const_value; + } + + return 0; } -VALUE -mod_remove_const(mod, name) - VALUE mod, name; +static struct autoload_const *autoloading_const_entry(VALUE mod, ID id); + +int +rb_autoloading_value(VALUE mod, ID id, VALUE* value, rb_const_flag_t *flag) { - ID id = rb_to_id(name); - VALUE val; + struct autoload_const *ac = autoloading_const_entry(mod, id); + if (!ac) return FALSE; - if (!rb_is_const_id(id)) { - NameError("`%s' is not constant", rb_id2name(id)); + if (value) { + *value = ac->value; } - if (RCLASS(mod)->iv_tbl && st_delete(ROBJECT(mod)->iv_tbl, &id, &val)) { - return val; + if (flag) { + *flag = ac->flag; } - if (rb_const_defined_at(mod, id)) { - NameError("cannot remove %s::%s", - rb_class2name(mod), rb_id2name(id)); + + return TRUE; +} + +static int +autoload_by_current(struct autoload_data *ele) +{ + return ele->mutex != Qnil && rb_mutex_owned_p(ele->mutex); +} + +// If there is an autoloading constant and it has been set by the current +// execution context, return it. This allows threads which are loading code to +// refer to their own autoloaded constants. +struct autoload_const * +autoloading_const_entry(VALUE mod, ID id) +{ + VALUE load = autoload_data(mod, id); + struct autoload_data *ele; + struct autoload_const *ac; + + // Find the autoloading state: + if (!load || !(ele = get_autoload_data(load, &ac))) { + // Couldn't be found: + return 0; } - NameError("constant %s::%s not defined", - rb_class2name(mod), rb_id2name(id)); + + // Check if it's being loaded by the current thread/fiber: + if (autoload_by_current(ele)) { + if (!UNDEF_P(ac->value)) { + return ac; + } + } + + return 0; } static int -autoload_i(key, name, ary) - ID key; - char *name; - VALUE ary; +autoload_defined_p(VALUE mod, ID id) { - VALUE kval = str_new2(rb_id2name(key)); - if (!ary_includes(ary, kval)) { - ary_push(ary, kval); + rb_const_entry_t *ce = rb_const_lookup(mod, id); + + // If there is no constant or the constant is not undefined (special marker for autoloading): + if (!ce || !UNDEF_P(ce->value)) { + // We are not autoloading: + return 0; } - return ST_CONTINUE; + + // Otherwise check if there is an autoload in flight right now: + return !rb_autoloading_value(mod, id, NULL, NULL); +} + +static void const_tbl_update(struct autoload_const *, int); + +struct autoload_load_arguments { + VALUE module; + ID name; + int flag; + + VALUE mutex; + + // The specific constant which triggered the autoload code to fire: + struct autoload_const *autoload_const; + + // The parent autoload data which is shared between multiple constants: + struct autoload_data *autoload_data; +}; + +static VALUE +autoload_const_set(struct autoload_const *ac) +{ + check_before_mod_set(ac->module, ac->name, ac->value, "constant"); + + RB_VM_LOCKING() { + const_tbl_update(ac, true); + } + + return 0; /* ignored */ +} + +static VALUE +autoload_load_needed(VALUE _arguments) +{ + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + const char *loading = 0, *src; + + if (!autoload_defined_p(arguments->module, arguments->name)) { + return Qfalse; + } + + VALUE autoload_const_value = check_autoload_required(arguments->module, arguments->name, &loading); + if (!autoload_const_value) { + return Qfalse; + } + + src = rb_sourcefile(); + if (src && loading && strcmp(src, loading) == 0) { + return Qfalse; + } + + struct autoload_const *autoload_const; + struct autoload_data *autoload_data; + if (!(autoload_data = get_autoload_data(autoload_const_value, &autoload_const))) { + return Qfalse; + } + + if (NIL_P(autoload_data->mutex)) { + RB_OBJ_WRITE(autoload_const->autoload_data_value, &autoload_data->mutex, rb_mutex_new()); + autoload_data->fork_gen = GET_VM()->fork_gen; + } + else if (rb_mutex_owned_p(autoload_data->mutex)) { + return Qfalse; + } + + arguments->mutex = autoload_data->mutex; + arguments->autoload_const = autoload_const; + + return autoload_const_value; +} + +static VALUE +autoload_apply_constants(VALUE _arguments) +{ + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); + + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + struct autoload_const *autoload_const = 0; // for ccan_container_off_var() + struct autoload_const *next; + + // We use safe iteration here because `autoload_const_set` will eventually invoke + // `autoload_delete` which will remove the constant from the linked list. In theory, once + // the `autoload_data->constants` linked list is empty, we can remove it. + + // Iterate over all constants and assign them: + ccan_list_for_each_safe(&arguments->autoload_data->constants, autoload_const, next, cnode) { + if (!UNDEF_P(autoload_const->value)) { + autoload_const_set(autoload_const); + } + } + + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); + + return Qtrue; +} + +static VALUE +autoload_feature_require(VALUE _arguments) +{ + VALUE receiver = rb_vm_top_self(); + + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + struct autoload_const *autoload_const = arguments->autoload_const; + VALUE autoload_box_value = autoload_const->box_value; + + // We save this for later use in autoload_apply_constants: + arguments->autoload_data = rb_check_typeddata(autoload_const->autoload_data_value, &autoload_data_type); + + if (rb_box_available() && BOX_OBJ_P(autoload_box_value)) + receiver = autoload_box_value; + + /* + * Clear the global cc cache table because the require method can be different from the current + * box's one and it may cause inconsistent cc-cme states. + * For example, the assertion below may fail in gccct_method_search(); + * VM_ASSERT(vm_cc_check_cme(cc, rb_callable_method_entry(klass, mid))) + */ + rb_gccct_clear_table(Qnil); + + VALUE result = rb_funcall(receiver, rb_intern("require"), 1, arguments->autoload_data->feature); + + if (RTEST(result)) { + return rb_mutex_synchronize(autoload_mutex, autoload_apply_constants, _arguments); + } + return result; +} + +static VALUE +autoload_try_load(VALUE _arguments) +{ + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + VALUE result = autoload_feature_require(_arguments); + + // After we loaded the feature, if the constant is not defined, we remove it completely: + rb_const_entry_t *ce = rb_const_lookup(arguments->module, arguments->name); + + if (!ce || UNDEF_P(ce->value)) { + result = Qfalse; + + rb_const_remove(arguments->module, arguments->name); + + if (arguments->module == rb_cObject) { + rb_warning( + "Expected %"PRIsVALUE" to define %"PRIsVALUE" but it didn't", + arguments->autoload_data->feature, + ID2SYM(arguments->name) + ); + } + else { + rb_warning( + "Expected %"PRIsVALUE" to define %"PRIsVALUE"::%"PRIsVALUE" but it didn't", + arguments->autoload_data->feature, + arguments->module, + ID2SYM(arguments->name) + ); + } + } + else { + // Otherwise, it was loaded, copy the flags from the autoload constant: + ce->flag |= arguments->flag; + } + + return result; } VALUE -mod_const_at(mod, ary) - VALUE mod, ary; +rb_autoload_load(VALUE module, ID name) { - if (RCLASS(mod)->iv_tbl) { - st_foreach(RCLASS(mod)->iv_tbl, const_i, ary); + rb_const_entry_t *ce = rb_const_lookup(module, name); + + // We bail out as early as possible without any synchronisation: + if (!ce || !UNDEF_P(ce->value)) { + return Qfalse; } - if ((VALUE)mod == cObject) { - st_foreach(class_tbl, const_i, ary); - if (autoload_tbl) { - st_foreach(autoload_tbl, autoload_i, ary); - } + + // At this point, we assume there might be autoloading, so fail if it's ractor: + if (UNLIKELY(!rb_ractor_main_p())) { + return rb_ractor_autoload_load(module, name); } - return ary; + + // This state is stored on the stack and is used during the autoload process. + struct autoload_load_arguments arguments = {.module = module, .name = name, .mutex = Qnil}; + + // Figure out whether we can autoload the named constant: + VALUE autoload_const_value = rb_mutex_synchronize(autoload_mutex, autoload_load_needed, (VALUE)&arguments); + + // This confirms whether autoloading is required or not: + if (autoload_const_value == Qfalse) return autoload_const_value; + + arguments.flag = ce->flag & (CONST_DEPRECATED | CONST_VISIBILITY_MASK); + + // Only one thread will enter here at a time: + VALUE result = rb_mutex_synchronize(arguments.mutex, autoload_try_load, (VALUE)&arguments); + + // If you don't guard this value, it's possible for the autoload constant to + // be freed by another thread which loads multiple constants, one of which + // resolves to the constant this thread is trying to load, so proteect this + // so that it is not freed until we are done with it in `autoload_try_load`: + RB_GC_GUARD(autoload_const_value); + + return result; } VALUE -mod_constants(mod) - VALUE mod; +rb_autoload_p(VALUE mod, ID id) { - return mod_const_at(mod, ary_new()); + return rb_autoload_at_p(mod, id, TRUE); } VALUE -mod_const_of(mod, ary) - VALUE mod; +rb_autoload_at_p(VALUE mod, ID id, int recur) +{ + VALUE load; + struct autoload_data *ele; + + while (!autoload_defined_p(mod, id)) { + if (!recur) return Qnil; + mod = RCLASS_SUPER(mod); + if (!mod) return Qnil; + } + load = check_autoload_required(mod, id, 0); + if (!load) return Qnil; + return (ele = get_autoload_data(load, 0)) ? ele->feature : Qnil; +} + +void +rb_const_warn_if_deprecated(const rb_const_entry_t *ce, VALUE klass, ID id) +{ + if (RB_CONST_DEPRECATED_P(ce) && + rb_warning_category_enabled_p(RB_WARN_CATEGORY_DEPRECATED)) { + if (klass == rb_cObject) { + rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant ::%"PRIsVALUE" is deprecated", QUOTE_ID(id)); + } + else { + rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant %"PRIsVALUE"::%"PRIsVALUE" is deprecated", + rb_class_name(klass), QUOTE_ID(id)); + } + } +} + +static VALUE +rb_const_get_0(VALUE klass, ID id, int exclude, int recurse, int visibility) +{ + VALUE found_in; + VALUE c = rb_const_search(klass, id, exclude, recurse, visibility, &found_in); + if (!UNDEF_P(c)) { + if (UNLIKELY(!rb_ractor_main_p())) { + if (!rb_ractor_shareable_p(c)) { + rb_raise(rb_eRactorIsolationError, "can not access non-shareable objects in constant %"PRIsVALUE"::%"PRIsVALUE" by non-main Ractor.", rb_class_path(found_in), rb_id2str(id)); + } + } + return c; + } + return rb_const_missing(klass, ID2SYM(id)); +} + +static VALUE +rb_const_search_from(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in) +{ + VALUE value, current; + bool first_iteration = true; + + for (current = klass; + RTEST(current); + current = RCLASS_SUPER(current), first_iteration = false) { + VALUE tmp; + VALUE am = 0; + rb_const_entry_t *ce; + + if (!first_iteration && RCLASS_ORIGIN(current) != current) { + // This item in the super chain has an origin iclass + // that comes later in the chain. Skip this item so + // prepended modules take precedence. + continue; + } + + // Do lookup in original class or module in case we are at an origin + // iclass in the chain. + tmp = current; + if (BUILTIN_TYPE(tmp) == T_ICLASS) tmp = RBASIC(tmp)->klass; + + // Do the lookup. Loop in case of autoload. + while ((ce = rb_const_lookup(tmp, id))) { + if (visibility && RB_CONST_PRIVATE_P(ce)) { + GET_EC()->private_const_reference = tmp; + return Qundef; + } + rb_const_warn_if_deprecated(ce, tmp, id); + value = ce->value; + if (UNDEF_P(value)) { + struct autoload_const *ac; + if (am == tmp) break; + am = tmp; + ac = autoloading_const_entry(tmp, id); + if (ac) { + if (found_in) { *found_in = tmp; } + return ac->value; + } + rb_autoload_load(tmp, id); + continue; + } + if (exclude && tmp == rb_cObject) { + goto not_found; + } + if (found_in) { *found_in = tmp; } + return value; + } + if (!recurse) break; + } + + not_found: + GET_EC()->private_const_reference = 0; + return Qundef; +} + +static VALUE +rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in) +{ + VALUE value; + + if (klass == rb_cObject) exclude = FALSE; + value = rb_const_search_from(klass, id, exclude, recurse, visibility, found_in); + if (!UNDEF_P(value)) return value; + if (exclude) return value; + if (BUILTIN_TYPE(klass) != T_MODULE) return value; + /* search global const too, if klass is a module */ + return rb_const_search_from(rb_cObject, id, FALSE, recurse, visibility, found_in); +} + +VALUE +rb_const_get_from(VALUE klass, ID id) +{ + return rb_const_get_0(klass, id, TRUE, TRUE, FALSE); +} + +VALUE +rb_const_get(VALUE klass, ID id) +{ + return rb_const_get_0(klass, id, FALSE, TRUE, FALSE); +} + +VALUE +rb_const_get_at(VALUE klass, ID id) +{ + return rb_const_get_0(klass, id, TRUE, FALSE, FALSE); +} + +VALUE +rb_public_const_get_from(VALUE klass, ID id) +{ + return rb_const_get_0(klass, id, TRUE, TRUE, TRUE); +} + +VALUE +rb_public_const_get_at(VALUE klass, ID id) +{ + return rb_const_get_0(klass, id, TRUE, FALSE, TRUE); +} + +NORETURN(static void undefined_constant(VALUE mod, VALUE name)); +static void +undefined_constant(VALUE mod, VALUE name) +{ + rb_name_err_raise("constant %2$s::%1$s not defined", + mod, name); +} + +static VALUE +rb_const_location_from(VALUE klass, ID id, int exclude, int recurse, int visibility) +{ + while (RTEST(klass)) { + rb_const_entry_t *ce; + + while ((ce = rb_const_lookup(klass, id))) { + if (visibility && RB_CONST_PRIVATE_P(ce)) { + return Qnil; + } + if (exclude && klass == rb_cObject) { + goto not_found; + } + + if (UNDEF_P(ce->value)) { // autoload + VALUE autoload_const_value = autoload_data(klass, id); + if (RTEST(autoload_const_value)) { + struct autoload_const *autoload_const; + struct autoload_data *autoload_data = get_autoload_data(autoload_const_value, &autoload_const); + + if (!UNDEF_P(autoload_const->value) && RTEST(rb_mutex_owned_p(autoload_data->mutex))) { + return rb_assoc_new(autoload_const->file, INT2NUM(autoload_const->line)); + } + } + } + + if (NIL_P(ce->file)) return rb_ary_new(); + return rb_assoc_new(ce->file, INT2NUM(ce->line)); + } + if (!recurse) break; + klass = RCLASS_SUPER(klass); + } + + not_found: + return Qnil; +} + +static VALUE +rb_const_location(VALUE klass, ID id, int exclude, int recurse, int visibility) +{ + VALUE loc; + + if (klass == rb_cObject) exclude = FALSE; + loc = rb_const_location_from(klass, id, exclude, recurse, visibility); + if (!NIL_P(loc)) return loc; + if (exclude) return loc; + if (BUILTIN_TYPE(klass) != T_MODULE) return loc; + /* search global const too, if klass is a module */ + return rb_const_location_from(rb_cObject, id, FALSE, recurse, visibility); +} + +VALUE +rb_const_source_location(VALUE klass, ID id) +{ + return rb_const_location(klass, id, FALSE, TRUE, FALSE); +} + +VALUE +rb_const_source_location_at(VALUE klass, ID id) +{ + return rb_const_location(klass, id, TRUE, FALSE, FALSE); +} + +/* + * call-seq: + * remove_const(sym) -> obj + * + * Removes the definition of the given constant, returning that + * constant's previous value. If that constant referred to + * a module, this will not change that module's name and can lead + * to confusion. + */ + +VALUE +rb_mod_remove_const(VALUE mod, VALUE name) +{ + const ID id = id_for_var(mod, name, a, constant); + + if (!id) { + undefined_constant(mod, name); + } + return rb_const_remove(mod, id); +} + +static rb_const_entry_t * const_lookup(struct rb_id_table *tbl, ID id); + +VALUE +rb_const_remove(VALUE mod, ID id) +{ + VALUE val; + rb_const_entry_t *ce; + + rb_check_frozen(mod); + + ce = rb_const_lookup(mod, id); + + if (!ce) { + if (rb_const_defined_at(mod, id)) { + rb_name_err_raise("cannot remove %2$s::%1$s", mod, ID2SYM(id)); + } + + undefined_constant(mod, ID2SYM(id)); + } + + VALUE writable_ce = 0; + if (rb_id_table_lookup(RCLASS_WRITABLE_CONST_TBL(mod), id, &writable_ce)) { + rb_id_table_delete(RCLASS_WRITABLE_CONST_TBL(mod), id); + if ((rb_const_entry_t *)writable_ce != ce) { + xfree((rb_const_entry_t *)writable_ce); + } + } + + rb_const_warn_if_deprecated(ce, mod, id); + rb_clear_constant_cache_for_id(id); + + val = ce->value; + + if (UNDEF_P(val)) { + autoload_delete(mod, id); + val = Qnil; + } + + if (ce != const_lookup(RCLASS_PRIME_CONST_TBL(mod), id)) { + ruby_xfree(ce); + } + // else - skip free'ing the ce because it still exists in the prime classext + + return val; +} + +static int +cv_i_update(st_data_t *k, st_data_t *v, st_data_t a, int existing) +{ + if (existing) return ST_STOP; + *v = a; + return ST_CONTINUE; +} + +static enum rb_id_table_iterator_result +sv_i(ID key, VALUE v, void *a) +{ + rb_const_entry_t *ce = (rb_const_entry_t *)v; + st_table *tbl = a; + + if (rb_is_const_id(key)) { + st_update(tbl, (st_data_t)key, cv_i_update, (st_data_t)ce); + } + return ID_TABLE_CONTINUE; +} + +static enum rb_id_table_iterator_result +rb_local_constants_i(ID const_name, VALUE const_value, void *ary) +{ + if (rb_is_const_id(const_name) && !RB_CONST_PRIVATE_P((rb_const_entry_t *)const_value)) { + rb_ary_push((VALUE)ary, ID2SYM(const_name)); + } + return ID_TABLE_CONTINUE; +} + +static VALUE +rb_local_constants(VALUE mod) +{ + struct rb_id_table *tbl = RCLASS_CONST_TBL(mod); VALUE ary; + + if (!tbl) return rb_ary_new2(0); + + RB_VM_LOCKING() { + ary = rb_ary_new2(rb_id_table_size(tbl)); + rb_id_table_foreach(tbl, rb_local_constants_i, (void *)ary); + } + + return ary; +} + +void* +rb_mod_const_at(VALUE mod, void *data) { - mod_const_at(mod, ary); + st_table *tbl = data; + if (!tbl) { + tbl = st_init_numtable(); + } + if (RCLASS_CONST_TBL(mod)) { + RB_VM_LOCKING() { + rb_id_table_foreach(RCLASS_CONST_TBL(mod), sv_i, tbl); + } + } + return tbl; +} + +void* +rb_mod_const_of(VALUE mod, void *data) +{ + VALUE tmp = mod; for (;;) { - mod = RCLASS(mod)->super; - if (!mod) break; - mod_const_at(mod, ary); + data = rb_mod_const_at(tmp, data); + tmp = RCLASS_SUPER(tmp); + if (!tmp) break; + if (tmp == rb_cObject && mod != rb_cObject) break; } + return data; +} + +static int +list_i(st_data_t key, st_data_t value, VALUE ary) +{ + ID sym = (ID)key; + rb_const_entry_t *ce = (rb_const_entry_t *)value; + if (RB_CONST_PUBLIC_P(ce)) rb_ary_push(ary, ID2SYM(sym)); + return ST_CONTINUE; +} + +VALUE +rb_const_list(void *data) +{ + st_table *tbl = data; + VALUE ary; + + if (!tbl) return rb_ary_new2(0); + ary = rb_ary_new2(tbl->num_entries); + st_foreach_safe(tbl, list_i, ary); + st_free_table(tbl); + return ary; } -int -rb_const_defined_at(klass, id) - VALUE klass; - ID id; +/* + * call-seq: + * mod.constants(inherit=true) -> array + * + * Returns an array of the names of the constants accessible in + * <i>mod</i>. This includes the names of constants in any included + * modules (example at start of section), unless the <i>inherit</i> + * parameter is set to <code>false</code>. + * + * The implementation makes no guarantees about the order in which the + * constants are yielded. + * + * IO.constants.include?(:SYNC) #=> true + * IO.constants(false).include?(:SYNC) #=> false + * + * Also see Module#const_defined?. + */ + +VALUE +rb_mod_constants(int argc, const VALUE *argv, VALUE mod) +{ + bool inherit = true; + + if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]); + + if (inherit) { + return rb_const_list(rb_mod_const_of(mod, 0)); + } + else { + return rb_local_constants(mod); + } +} + +static int +rb_const_defined_0(VALUE klass, ID id, int exclude, int recurse, int visibility) { - if (RCLASS(klass)->iv_tbl && st_lookup(RCLASS(klass)->iv_tbl, id, 0)) { - return TRUE; + VALUE tmp; + int mod_retry = 0; + rb_const_entry_t *ce; + + tmp = klass; + retry: + while (tmp) { + if ((ce = rb_const_lookup(tmp, id))) { + if (visibility && RB_CONST_PRIVATE_P(ce)) { + return (int)Qfalse; + } + if (UNDEF_P(ce->value) && !check_autoload_required(tmp, id, 0) && + !rb_autoloading_value(tmp, id, NULL, NULL)) + return (int)Qfalse; + + if (exclude && tmp == rb_cObject && klass != rb_cObject) { + return (int)Qfalse; + } + + return (int)Qtrue; + } + if (!recurse) break; + tmp = RCLASS_SUPER(tmp); } - if (klass == cObject) { - return rb_const_defined(klass, id); + if (!exclude && !mod_retry && BUILTIN_TYPE(klass) == T_MODULE) { + mod_retry = 1; + tmp = rb_cObject; + goto retry; } - return FALSE; + return (int)Qfalse; } int -rb_autoload_defined(id) - ID id; +rb_const_defined_from(VALUE klass, ID id) { - if (autoload_tbl && st_lookup(autoload_tbl, id, 0)) - return TRUE; - return FALSE; + return rb_const_defined_0(klass, id, TRUE, TRUE, FALSE); } int -rb_const_defined(klass, id) - VALUE klass; - ID id; +rb_const_defined(VALUE klass, ID id) { - VALUE tmp = klass; + return rb_const_defined_0(klass, id, FALSE, TRUE, FALSE); +} - while (tmp) { - if (RCLASS(tmp)->iv_tbl && st_lookup(RCLASS(tmp)->iv_tbl,id,0)) { - return TRUE; - } - tmp = RCLASS(tmp)->super; +int +rb_const_defined_at(VALUE klass, ID id) +{ + return rb_const_defined_0(klass, id, TRUE, FALSE, FALSE); +} + +int +rb_public_const_defined_from(VALUE klass, ID id) +{ + return rb_const_defined_0(klass, id, TRUE, TRUE, TRUE); +} + +static void +check_before_mod_set(VALUE klass, ID id, VALUE val, const char *dest) +{ + rb_check_frozen(klass); +} + +static void set_namespace_path(VALUE named_namespace, VALUE name); + +static enum rb_id_table_iterator_result +set_namespace_path_i(ID id, VALUE v, void *payload) +{ + rb_const_entry_t *ce = (rb_const_entry_t *)v; + VALUE value = ce->value; + VALUE parental_path = *((VALUE *) payload); + if (!rb_is_const_id(id) || !rb_namespace_p(value)) { + return ID_TABLE_CONTINUE; } - if (BUILTIN_TYPE(klass) == T_MODULE) { - return rb_const_defined(cObject, id); + + bool has_permanent_classpath; + classname(value, &has_permanent_classpath); + if (has_permanent_classpath) { + return ID_TABLE_CONTINUE; + } + set_namespace_path(value, build_const_path(parental_path, id)); + + if (!RCLASS_PERMANENT_CLASSPATH_P(value)) { + RCLASS_WRITE_CLASSPATH(value, 0, false); + } + + return ID_TABLE_CONTINUE; +} + +/* + * Assign permanent classpaths to all namespaces that are directly or indirectly + * nested under +named_namespace+. +named_namespace+ must have a permanent + * classpath. + */ +static void +set_namespace_path(VALUE named_namespace, VALUE namespace_path) +{ + struct rb_id_table *const_table = RCLASS_CONST_TBL(named_namespace); + RB_OBJ_SET_SHAREABLE(namespace_path); + + RB_VM_LOCKING() { + RCLASS_WRITE_CLASSPATH(named_namespace, namespace_path, true); + + if (const_table) { + rb_id_table_foreach(const_table, set_namespace_path_i, &namespace_path); + } + } +} + +static void +const_added(VALUE klass, ID const_name) +{ + if (GET_VM()->running) { + VALUE name = ID2SYM(const_name); + rb_funcallv(klass, idConst_added, 1, &name); + } +} + +static void +const_set(VALUE klass, ID id, VALUE val) +{ + rb_const_entry_t *ce; + + if (NIL_P(klass)) { + rb_raise(rb_eTypeError, "no class/module to define constant %"PRIsVALUE"", + QUOTE_ID(id)); + } + + if (!rb_ractor_main_p() && !rb_ractor_shareable_p(val)) { + rb_raise(rb_eRactorIsolationError, "can not set constants with non-shareable objects by non-main Ractors"); + } + + check_before_mod_set(klass, id, val, "constant"); + + RB_VM_LOCKING() { + struct rb_id_table *tbl = RCLASS_WRITABLE_CONST_TBL(klass); + if (!tbl) { + tbl = rb_id_table_create(0); + RCLASS_WRITE_CONST_TBL(klass, tbl, false); + rb_clear_constant_cache_for_id(id); + ce = ZALLOC(rb_const_entry_t); + rb_id_table_insert(tbl, id, (VALUE)ce); + setup_const_entry(ce, klass, val, CONST_PUBLIC); + } + else { + struct autoload_const ac = { + .module = klass, .name = id, + .value = val, .flag = CONST_PUBLIC, + /* fill the rest with 0 */ + }; + ac.file = rb_source_location(&ac.line); + const_tbl_update(&ac, false); + } + } + + /* + * Resolve and cache class name immediately to resolve ambiguity + * and avoid order-dependency on const_tbl + */ + if (rb_cObject && rb_namespace_p(val)) { + bool val_path_permanent; + VALUE val_path = classname(val, &val_path_permanent); + if (NIL_P(val_path) || !val_path_permanent) { + if (klass == rb_cObject) { + set_namespace_path(val, rb_id2str(id)); + } + else { + bool parental_path_permanent; + VALUE parental_path = classname(klass, &parental_path_permanent); + if (NIL_P(parental_path)) { + bool throwaway; + parental_path = rb_tmp_class_path(klass, &throwaway, make_temporary_path); + } + if (parental_path_permanent && !val_path_permanent) { + set_namespace_path(val, build_const_path(parental_path, id)); + } + else if (!parental_path_permanent && NIL_P(val_path)) { + VALUE path = build_const_path(parental_path, id); + RCLASS_SET_CLASSPATH(val, path, false); + } + } + } } - if (st_lookup(class_tbl, id, 0)) - return TRUE; - return rb_autoload_defined(id); } void -rb_const_set(klass, id, val) - VALUE klass; - ID id; - VALUE val; +rb_const_set(VALUE klass, ID id, VALUE val) { - if (!RCLASS(klass)->iv_tbl) { - RCLASS(klass)->iv_tbl = new_idhash(); + const_set(klass, id, val); + const_added(klass, id); +} + +static VALUE +autoload_const_value_for_named_constant(VALUE module, ID name, struct autoload_const **autoload_const_pointer) +{ + VALUE autoload_const_value = autoload_data(module, name); + if (!autoload_const_value) return Qfalse; + + struct autoload_data *autoload_data = get_autoload_data(autoload_const_value, autoload_const_pointer); + if (!autoload_data) return Qfalse; + + /* for autoloading thread, keep the defined value to autoloading storage */ + if (autoload_by_current(autoload_data)) { + return autoload_const_value; } - else if (st_lookup(RCLASS(klass)->iv_tbl, id, 0)) { - NameError("already initialized constant %s", rb_id2name(id)); + + return Qfalse; +} + +static void +const_tbl_update(struct autoload_const *ac, int autoload_force) +{ + VALUE value; + VALUE klass = ac->module; + VALUE val = ac->value; + ID id = ac->name; + struct rb_id_table *tbl = RCLASS_CONST_TBL(klass); + rb_const_flag_t visibility = ac->flag; + rb_const_entry_t *ce; + + if (rb_id_table_lookup(tbl, id, &value)) { + ce = (rb_const_entry_t *)value; + if (UNDEF_P(ce->value)) { + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); + VALUE file = ac->file; + int line = ac->line; + VALUE autoload_const_value = autoload_const_value_for_named_constant(klass, id, &ac); + + if (!autoload_force && autoload_const_value) { + rb_clear_constant_cache_for_id(id); + + RB_OBJ_WRITE(autoload_const_value, &ac->value, val); + RB_OBJ_WRITE(autoload_const_value, &ac->file, rb_source_location(&ac->line)); + } + else { + /* otherwise autoloaded constant, allow to override */ + autoload_delete(klass, id); + ce->flag = visibility; + RB_OBJ_WRITE(klass, &ce->value, val); + RB_OBJ_WRITE(klass, &ce->file, file); + ce->line = line; + } + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); + return; + } + else { + VALUE name = QUOTE_ID(id); + visibility = ce->flag; + if (klass == rb_cObject) + rb_warn("already initialized constant %"PRIsVALUE"", name); + else + rb_warn("already initialized constant %"PRIsVALUE"::%"PRIsVALUE"", + rb_class_name(klass), name); + if (!NIL_P(ce->file) && ce->line) { + rb_compile_warn(RSTRING_PTR(ce->file), ce->line, + "previous definition of %"PRIsVALUE" was here", name); + } + } + rb_clear_constant_cache_for_id(id); + setup_const_entry(ce, klass, val, visibility); } + else { + tbl = RCLASS_WRITABLE_CONST_TBL(klass); + rb_clear_constant_cache_for_id(id); - st_insert(RCLASS(klass)->iv_tbl, id, val); + ce = ZALLOC(rb_const_entry_t); + rb_id_table_insert(tbl, id, (VALUE)ce); + setup_const_entry(ce, klass, val, visibility); + } +} + +static void +setup_const_entry(rb_const_entry_t *ce, VALUE klass, VALUE val, + rb_const_flag_t visibility) +{ + ce->flag = visibility; + RB_OBJ_WRITE(klass, &ce->value, val); + RB_OBJ_WRITE(klass, &ce->file, rb_source_location(&ce->line)); } void -rb_define_const(klass, name, val) - VALUE klass; - char *name; - VALUE val; +rb_define_const(VALUE klass, const char *name, VALUE val) { ID id = rb_intern(name); + if (!rb_is_const_id(id)) { - NameError("wrong constant name %s", name); + rb_warn("rb_define_const: invalid name '%s' for constant", name); + } + if (!RB_SPECIAL_CONST_P(val)) { + rb_vm_register_global_object(val); } rb_const_set(klass, id, val); } void -rb_define_global_const(name, val) - char *name; - VALUE val; +rb_define_global_const(const char *name, VALUE val) { - rb_define_const(cObject, name, val); + rb_define_const(rb_cObject, name, val); } +static void +set_const_visibility(VALUE mod, int argc, const VALUE *argv, + rb_const_flag_t flag, rb_const_flag_t mask) +{ + int i; + rb_const_entry_t *ce; + ID id; + + rb_class_modify_check(mod); + if (argc == 0) { + rb_warning("%"PRIsVALUE" with no argument is just ignored", + QUOTE_ID(rb_frame_callee())); + return; + } + + for (i = 0; i < argc; i++) { + struct autoload_const *ac; + VALUE val = argv[i]; + id = rb_check_id(&val); + if (!id) { + undefined_constant(mod, val); + } + if ((ce = rb_const_lookup(mod, id))) { + ce->flag &= ~mask; + ce->flag |= flag; + if (UNDEF_P(ce->value)) { + if (autoload_const_value_for_named_constant(mod, id, &ac)) { + ac->flag &= ~mask; + ac->flag |= flag; + } + } + rb_clear_constant_cache_for_id(id); + } + else { + undefined_constant(mod, ID2SYM(id)); + } + } +} + +void +rb_deprecate_constant(VALUE mod, const char *name) +{ + rb_const_entry_t *ce; + ID id; + long len = strlen(name); + + rb_class_modify_check(mod); + if (!(id = rb_check_id_cstr(name, len, NULL))) { + undefined_constant(mod, rb_fstring_new(name, len)); + } + if (!(ce = rb_const_lookup(mod, id))) { + undefined_constant(mod, ID2SYM(id)); + } + ce->flag |= CONST_DEPRECATED; +} + +/* + * call-seq: + * mod.private_constant(symbol, ...) => mod + * + * Makes a list of existing constants private. + */ + VALUE -rb_iv_get(obj, name) - VALUE obj; - char *name; +rb_mod_private_constant(int argc, const VALUE *argv, VALUE obj) +{ + set_const_visibility(obj, argc, argv, CONST_PRIVATE, CONST_VISIBILITY_MASK); + return obj; +} + +/* + * call-seq: + * mod.public_constant(symbol, ...) => mod + * + * Makes a list of existing constants public. + */ + +VALUE +rb_mod_public_constant(int argc, const VALUE *argv, VALUE obj) +{ + set_const_visibility(obj, argc, argv, CONST_PUBLIC, CONST_VISIBILITY_MASK); + return obj; +} + +/* + * call-seq: + * mod.deprecate_constant(symbol, ...) => mod + * + * Makes a list of existing constants deprecated. Attempt + * to refer to them will produce a warning. + * + * module HTTP + * NotFound = Exception.new + * NOT_FOUND = NotFound # previous version of the library used this name + * + * deprecate_constant :NOT_FOUND + * end + * + * HTTP::NOT_FOUND + * # warning: constant HTTP::NOT_FOUND is deprecated + * + */ + +VALUE +rb_mod_deprecate_constant(int argc, const VALUE *argv, VALUE obj) +{ + set_const_visibility(obj, argc, argv, CONST_DEPRECATED, CONST_DEPRECATED); + return obj; +} + +static VALUE +original_module(VALUE c) +{ + if (RB_TYPE_P(c, T_ICLASS)) + return RBASIC(c)->klass; + return c; +} + +static int +cvar_lookup_at(VALUE klass, ID id, st_data_t *v) +{ + if (RB_TYPE_P(klass, T_ICLASS)) { + if (RICLASS_IS_ORIGIN_P(klass)) { + return 0; + } + else { + // check the original module + klass = RBASIC(klass)->klass; + } + } + + VALUE n = rb_ivar_lookup(klass, id, Qundef); + if (UNDEF_P(n)) return 0; + + if (v) *v = n; + return 1; +} + +static VALUE +cvar_front_klass(VALUE klass) +{ + if (RCLASS_SINGLETON_P(klass)) { + VALUE obj = RCLASS_ATTACHED_OBJECT(klass); + if (rb_namespace_p(obj)) { + return obj; + } + } + return RCLASS_SUPER(klass); +} + +static void +cvar_overtaken(VALUE front, VALUE target, ID id) +{ + if (front && target != front) { + if (original_module(front) != original_module(target)) { + rb_raise(rb_eRuntimeError, + "class variable % "PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"", + ID2SYM(id), rb_class_name(original_module(front)), + rb_class_name(original_module(target))); + } + if (BUILTIN_TYPE(front) == T_CLASS) { + rb_ivar_delete(front, id, Qundef); + } + } +} + +#define CVAR_FOREACH_ANCESTORS(klass, v, r) \ + for (klass = cvar_front_klass(klass); klass; klass = RCLASS_SUPER(klass)) { \ + if (cvar_lookup_at(klass, id, (v))) { \ + r; \ + } \ + } + +#define CVAR_LOOKUP(v,r) do {\ + CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(klass, id); \ + if (cvar_lookup_at(klass, id, (v))) {r;}\ + CVAR_FOREACH_ANCESTORS(klass, v, r);\ +} while(0) + +static VALUE +find_cvar(VALUE klass, VALUE * front, VALUE * target, ID id) +{ + VALUE v = Qundef; + CVAR_LOOKUP(&v, { + if (!*front) { + *front = klass; + } + *target = klass; + }); + + return v; +} + +static void +check_for_cvar_table(VALUE subclass, VALUE key) +{ + // Must not check ivar on ICLASS + if (!RB_TYPE_P(subclass, T_ICLASS) && RTEST(rb_ivar_defined(subclass, key))) { + RB_DEBUG_COUNTER_INC(cvar_class_invalidate); + ruby_vm_global_cvar_state++; + return; + } + + rb_class_foreach_subclass(subclass, check_for_cvar_table, key); +} + +void +rb_cvar_set(VALUE klass, ID id, VALUE val) +{ + VALUE tmp, front = 0, target = 0; + + tmp = klass; + CVAR_LOOKUP(0, {if (!front) front = klass; target = klass;}); + if (target) { + cvar_overtaken(front, target, id); + } + else { + target = tmp; + } + + if (RB_TYPE_P(target, T_ICLASS)) { + target = RBASIC(target)->klass; + } + check_before_mod_set(target, id, val, "class variable"); + + bool new_cvar = rb_class_ivar_set(target, id, val); + + struct rb_id_table *rb_cvc_tbl = RCLASS_WRITABLE_CVC_TBL(target); + + if (!rb_cvc_tbl) { + rb_cvc_tbl = rb_id_table_create(2); + RCLASS_WRITE_CVC_TBL(target, rb_cvc_tbl); + } + + struct rb_cvar_class_tbl_entry *ent; + VALUE ent_data; + + if (!rb_id_table_lookup(rb_cvc_tbl, id, &ent_data)) { + ent = ALLOC(struct rb_cvar_class_tbl_entry); + ent->class_value = target; + ent->global_cvar_state = GET_GLOBAL_CVAR_STATE(); + ent->cref = 0; + rb_id_table_insert(rb_cvc_tbl, id, (VALUE)ent); + RB_DEBUG_COUNTER_INC(cvar_inline_miss); + } + else { + ent = (void *)ent_data; + ent->global_cvar_state = GET_GLOBAL_CVAR_STATE(); + } + + // Break the cvar cache if this is a new class variable + // and target is a module or a subclass with the same + // cvar in this lookup. + if (new_cvar) { + if (RB_TYPE_P(target, T_CLASS)) { + if (RCLASS_SUBCLASSES_FIRST(target)) { + rb_class_foreach_subclass(target, check_for_cvar_table, id); + } + } + } +} + +VALUE +rb_cvar_find(VALUE klass, ID id, VALUE *front) +{ + VALUE target = 0; + VALUE value; + + value = find_cvar(klass, front, &target, id); + if (!target) { + rb_name_err_raise("uninitialized class variable %1$s in %2$s", + klass, ID2SYM(id)); + } + cvar_overtaken(*front, target, id); + return (VALUE)value; +} + +VALUE +rb_cvar_get(VALUE klass, ID id) +{ + VALUE front = 0; + return rb_cvar_find(klass, id, &front); +} + +VALUE +rb_cvar_defined(VALUE klass, ID id) +{ + if (!klass) return Qfalse; + CVAR_LOOKUP(0,return Qtrue); + return Qfalse; +} + +static ID +cv_intern(VALUE klass, const char *name) { ID id = rb_intern(name); + if (!rb_is_class_id(id)) { + rb_name_err_raise("wrong class variable name %1$s", + klass, rb_str_new_cstr(name)); + } + return id; +} +void +rb_cv_set(VALUE klass, const char *name, VALUE val) +{ + ID id = cv_intern(klass, name); + rb_cvar_set(klass, id, val); +} + +VALUE +rb_cv_get(VALUE klass, const char *name) +{ + ID id = cv_intern(klass, name); + return rb_cvar_get(klass, id); +} + +void +rb_define_class_variable(VALUE klass, const char *name, VALUE val) +{ + rb_cv_set(klass, name, val); +} + +static int +cv_i(ID key, VALUE v, st_data_t a) +{ + st_table *tbl = (st_table *)a; + + if (rb_is_class_id(key)) { + st_update(tbl, (st_data_t)key, cv_i_update, 0); + } + return ST_CONTINUE; +} + +static void* +mod_cvar_at(VALUE mod, void *data) +{ + st_table *tbl = data; + if (!tbl) { + tbl = st_init_numtable(); + } + mod = original_module(mod); + + rb_ivar_foreach(mod, cv_i, (st_data_t)tbl); + return tbl; +} + +static void* +mod_cvar_of(VALUE mod, void *data) +{ + VALUE tmp = mod; + if (RCLASS_SINGLETON_P(mod)) { + if (rb_namespace_p(RCLASS_ATTACHED_OBJECT(mod))) { + data = mod_cvar_at(tmp, data); + tmp = cvar_front_klass(tmp); + } + } + for (;;) { + data = mod_cvar_at(tmp, data); + tmp = RCLASS_SUPER(tmp); + if (!tmp) break; + } + return data; +} + +static int +cv_list_i(st_data_t key, st_data_t value, VALUE ary) +{ + ID sym = (ID)key; + rb_ary_push(ary, ID2SYM(sym)); + return ST_CONTINUE; +} + +static VALUE +cvar_list(void *data) +{ + st_table *tbl = data; + VALUE ary; + + if (!tbl) return rb_ary_new2(0); + ary = rb_ary_new2(tbl->num_entries); + st_foreach_safe(tbl, cv_list_i, ary); + st_free_table(tbl); + + return ary; +} + +/* + * call-seq: + * mod.class_variables(inherit=true) -> array + * + * Returns an array of the names of class variables in <i>mod</i>. + * This includes the names of class variables in any included + * modules, unless the <i>inherit</i> parameter is set to + * <code>false</code>. + * + * class One + * @@var1 = 1 + * end + * class Two < One + * @@var2 = 2 + * end + * One.class_variables #=> [:@@var1] + * Two.class_variables #=> [:@@var2, :@@var1] + * Two.class_variables(false) #=> [:@@var2] + */ + +VALUE +rb_mod_class_variables(int argc, const VALUE *argv, VALUE mod) +{ + bool inherit = true; + st_table *tbl; + + if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]); + if (inherit) { + tbl = mod_cvar_of(mod, 0); + } + else { + tbl = mod_cvar_at(mod, 0); + } + return cvar_list(tbl); +} + +/* + * call-seq: + * remove_class_variable(sym) -> obj + * + * Removes the named class variable from the receiver, returning that + * variable's value. + * + * class Example + * @@var = 99 + * puts remove_class_variable(:@@var) + * p(defined? @@var) + * end + * + * <em>produces:</em> + * + * 99 + * nil + */ + +VALUE +rb_mod_remove_cvar(VALUE mod, VALUE name) +{ + const ID id = id_for_var_message(mod, name, class, "wrong class variable name %1$s"); + st_data_t val; + + if (!id) { + goto not_defined; + } + rb_check_frozen(mod); + val = rb_ivar_delete(mod, id, Qundef); + if (!UNDEF_P(val)) { + return (VALUE)val; + } + if (rb_cvar_defined(mod, id)) { + rb_name_err_raise("cannot remove %1$s for %2$s", mod, ID2SYM(id)); + } + not_defined: + rb_name_err_raise("class variable %1$s not defined for %2$s", + mod, name); + UNREACHABLE_RETURN(Qundef); +} + +VALUE +rb_iv_get(VALUE obj, const char *name) +{ + ID id = rb_check_id_cstr(name, strlen(name), rb_usascii_encoding()); + + if (!id) { + return Qnil; + } return rb_ivar_get(obj, id); } VALUE -rb_iv_set(obj, name, val) - VALUE obj; - char *name; - VALUE val; +rb_iv_set(VALUE obj, const char *name, VALUE val) { ID id = rb_intern(name); return rb_ivar_set(obj, id, val); } + +static attr_index_t +class_fields_ivar_set(VALUE klass, VALUE fields_obj, ID id, VALUE val, bool concurrent, VALUE *new_fields_obj, bool *new_ivar_out) +{ + const VALUE original_fields_obj = fields_obj; + fields_obj = original_fields_obj ? original_fields_obj : rb_imemo_fields_new(klass, 1, true); + + shape_id_t current_shape_id = RBASIC_SHAPE_ID(fields_obj); + shape_id_t next_shape_id = current_shape_id; // for too_complex + if (UNLIKELY(rb_shape_too_complex_p(current_shape_id))) { + goto too_complex; + } + + bool new_ivar; + next_shape_id = generic_shape_ivar(fields_obj, id, &new_ivar); + + if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) { + fields_obj = imemo_fields_complex_from_obj(klass, fields_obj, next_shape_id); + goto too_complex; + } + + attr_index_t index = RSHAPE_INDEX(next_shape_id); + if (new_ivar) { + if (index >= RSHAPE_CAPACITY(current_shape_id)) { + // We allocate a new fields_obj even when concurrency isn't a concern + // so that we're embedded as long as possible. + fields_obj = imemo_fields_copy_capa(klass, fields_obj, RSHAPE_CAPACITY(next_shape_id)); + } + } + + VALUE *fields = rb_imemo_fields_ptr(fields_obj); + + if (concurrent && original_fields_obj == fields_obj) { + // In the concurrent case, if we're mutating the existing + // fields_obj, we must use an atomic write, because if we're + // adding a new field, the shape_id must be written after the field + // and if we're updating an existing field, we at least need a relaxed + // write to avoid reaping. + RB_OBJ_ATOMIC_WRITE(fields_obj, &fields[index], val); + } + else { + RB_OBJ_WRITE(fields_obj, &fields[index], val); + } + + if (new_ivar) { + RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id); + } + + *new_fields_obj = fields_obj; + *new_ivar_out = new_ivar; + return index; + +too_complex: + { + if (concurrent && fields_obj == original_fields_obj) { + // In multi-ractor case, we must always work on a copy because + // even if the field already exist, inserting in a st_table may + // cause a rebuild. + fields_obj = rb_imemo_fields_clone(fields_obj); + } + + st_table *table = rb_imemo_fields_complex_tbl(fields_obj); + new_ivar = !st_insert(table, (st_data_t)id, (st_data_t)val); + RB_OBJ_WRITTEN(fields_obj, Qundef, val); + + if (fields_obj != original_fields_obj) { + RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id); + } + } + + *new_fields_obj = fields_obj; + *new_ivar_out = new_ivar; + return ATTR_INDEX_NOT_SET; +} + +static attr_index_t +class_ivar_set(VALUE obj, ID id, VALUE val, bool *new_ivar) +{ + rb_class_ensure_writable(obj); + + const VALUE original_fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + VALUE new_fields_obj = 0; + + attr_index_t index = class_fields_ivar_set(obj, original_fields_obj, id, val, rb_multi_ractor_p(), &new_fields_obj, new_ivar); + + if (new_fields_obj != original_fields_obj) { + RCLASS_WRITABLE_SET_FIELDS_OBJ(obj, new_fields_obj); + } + + // TODO: What should we set as the T_CLASS shape_id? + // In most case we can replicate the single `fields_obj` shape + // but in namespaced case? Perhaps INVALID_SHAPE_ID? + RBASIC_SET_SHAPE_ID(obj, RBASIC_SHAPE_ID(new_fields_obj)); + return index; +} + +bool +rb_class_ivar_set(VALUE obj, ID id, VALUE val) +{ + RUBY_ASSERT(RB_TYPE_P(obj, T_CLASS) || RB_TYPE_P(obj, T_MODULE)); + rb_check_frozen(obj); + + bool new_ivar; + class_ivar_set(obj, id, val, &new_ivar); + return new_ivar; +} + +void +rb_fields_tbl_copy(VALUE dst, VALUE src) +{ + RUBY_ASSERT(rb_type(dst) == rb_type(src)); + RUBY_ASSERT(RB_TYPE_P(dst, T_CLASS) || RB_TYPE_P(dst, T_MODULE)); + RUBY_ASSERT(RSHAPE_TYPE_P(RBASIC_SHAPE_ID(dst), SHAPE_ROOT)); + + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(src); + if (fields_obj) { + RCLASS_WRITABLE_SET_FIELDS_OBJ(dst, rb_imemo_fields_clone(fields_obj)); + RBASIC_SET_SHAPE_ID(dst, RBASIC_SHAPE_ID(src)); + } +} + +static rb_const_entry_t * +const_lookup(struct rb_id_table *tbl, ID id) +{ + if (tbl) { + VALUE val; + bool r; + RB_VM_LOCKING() { + r = rb_id_table_lookup(tbl, id, &val); + } + + if (r) return (rb_const_entry_t *)val; + } + return NULL; +} + +rb_const_entry_t * +rb_const_lookup(VALUE klass, ID id) +{ + return const_lookup(RCLASS_CONST_TBL(klass), id); +} |
