diff options
Diffstat (limited to 'variable.c')
| -rw-r--r-- | variable.c | 4702 |
1 files changed, 3090 insertions, 1612 deletions
diff --git a/variable.c b/variable.c index bfb6fa2edd..ff8d24d78a 100644 --- a/variable.c +++ b/variable.c @@ -11,229 +11,325 @@ **********************************************************************/ -#include "internal.h" -#include "ruby/st.h" -#include "ruby/util.h" -#include "id_table.h" +#include "ruby/internal/config.h" +#include <stddef.h> +#include "ruby/internal/stdbool.h" +#include "ccan/list/list.h" #include "constant.h" +#include "debug_counter.h" #include "id.h" -#include "ccan/list/list.h" #include "id_table.h" -#include "debug_counter.h" +#include "internal.h" +#include "internal/box.h" +#include "internal/class.h" +#include "internal/compilers.h" +#include "internal/error.h" +#include "internal/eval.h" +#include "internal/hash.h" +#include "internal/object.h" +#include "internal/gc.h" +#include "internal/re.h" +#include "internal/struct.h" +#include "internal/symbol.h" +#include "internal/thread.h" +#include "internal/variable.h" +#include "ruby/encoding.h" +#include "ruby/st.h" +#include "ruby/util.h" +#include "shape.h" +#include "symbol.h" +#include "variable.h" #include "vm_core.h" +#include "ractor_core.h" +#include "vm_sync.h" -struct rb_id_table *rb_global_tbl; -static ID autoload, classpath, tmp_classpath, classid; +RUBY_EXTERN rb_serial_t ruby_vm_global_cvar_state; +#define GET_GLOBAL_CVAR_STATE() (ruby_vm_global_cvar_state) + +typedef void rb_gvar_compact_t(void *var); + +static struct rb_id_table *rb_global_tbl; +static ID autoload; + +// This hash table maps file paths to loadable features. We use this to track +// autoload state until it's no longer needed. +// feature (file path) => struct autoload_data +static VALUE autoload_features; + +// This mutex is used to protect autoloading state. We use a global mutex which +// is held until a per-feature mutex can be created. This ensures there are no +// race conditions relating to autoload state. +static VALUE autoload_mutex; static void check_before_mod_set(VALUE, ID, VALUE, const char *); static void setup_const_entry(rb_const_entry_t *, VALUE, VALUE, rb_const_flag_t); -static VALUE rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility); -static st_table *generic_iv_tbl; -static st_table *generic_iv_tbl_compat; - -/* per-object */ -struct gen_ivtbl { - uint32_t numiv; - VALUE ivptr[1]; /* flexible array */ -}; +static VALUE rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in); +static st_table *generic_fields_tbl_; -struct ivar_update { - union { - st_table *iv_index_tbl; - struct gen_ivtbl *ivtbl; - } u; - st_data_t index; - int iv_extended; -}; +typedef int rb_ivar_foreach_callback_func(ID key, VALUE val, st_data_t arg); +static void rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only); void Init_var_tables(void) { rb_global_tbl = rb_id_table_create(0); - generic_iv_tbl = st_init_numtable(); + generic_fields_tbl_ = st_init_numtable(); autoload = rb_intern_const("__autoload__"); - /* __classpath__: fully qualified class path */ - classpath = rb_intern_const("__classpath__"); - /* __tmp_classpath__: temporary class path which contains anonymous names */ - tmp_classpath = rb_intern_const("__tmp_classpath__"); - /* __classid__: name given to class/module under an anonymous namespace */ - classid = rb_intern_const("__classid__"); -} -struct fc_result { - ID name, preferred; - VALUE klass; - VALUE path; - VALUE track; - struct fc_result *prev; -}; + autoload_mutex = rb_mutex_new(); + rb_obj_hide(autoload_mutex); + rb_vm_register_global_object(autoload_mutex); -static VALUE -fc_path(struct fc_result *fc, ID name) -{ - VALUE path, tmp; - - path = rb_id2str(name); - while (fc) { - st_data_t n; - if (fc->track == rb_cObject) break; - if (RCLASS_IV_TBL(fc->track) && - st_lookup(RCLASS_IV_TBL(fc->track), (st_data_t)classpath, &n)) { - tmp = rb_str_dup((VALUE)n); - rb_str_cat2(tmp, "::"); - rb_str_append(tmp, path); - path = tmp; - break; - } - tmp = rb_str_dup(rb_id2str(fc->name)); - rb_str_cat2(tmp, "::"); - rb_str_append(tmp, path); - path = tmp; - fc = fc->prev; - } - OBJ_FREEZE(path); - return path; + autoload_features = rb_ident_hash_new(); + rb_obj_hide(autoload_features); + rb_vm_register_global_object(autoload_features); } -static enum rb_id_table_iterator_result -fc_i(ID key, VALUE v, void *a) +static inline bool +rb_namespace_p(VALUE obj) { - rb_const_entry_t *ce = (rb_const_entry_t *)v; - struct fc_result *res = a; - VALUE value = ce->value; - if (!rb_is_const_id(key)) return ID_TABLE_CONTINUE; - - if (value == res->klass && (!res->preferred || key == res->preferred)) { - res->path = fc_path(res, key); - return ID_TABLE_STOP; - } - if (RB_TYPE_P(value, T_MODULE) || RB_TYPE_P(value, T_CLASS)) { - if (!RCLASS_CONST_TBL(value)) return ID_TABLE_CONTINUE; - else { - struct fc_result arg; - struct fc_result *list; - - list = res; - while (list) { - if (list->track == value) return ID_TABLE_CONTINUE; - list = list->prev; - } - - arg.name = key; - arg.preferred = res->preferred; - arg.path = 0; - arg.klass = res->klass; - arg.track = value; - arg.prev = res; - rb_id_table_foreach(RCLASS_CONST_TBL(value), fc_i, &arg); - if (arg.path) { - res->path = arg.path; - return ID_TABLE_STOP; - } - } + if (RB_SPECIAL_CONST_P(obj)) return false; + switch (RB_BUILTIN_TYPE(obj)) { + case T_MODULE: case T_CLASS: return true; + default: break; } - return ID_TABLE_CONTINUE; + return false; } /** - * Traverse constant namespace and find +classpath+ for _klass_. If - * _preferred_ is not 0, choice the path whose base name is set to it. - * If +classpath+ is found, the hidden instance variable __classpath__ - * is set to the found path, and __tmp_classpath__ is removed. - * The path is frozen. + * Returns +classpath+ of _klass_, if it is named, or +nil+ for + * anonymous +class+/+module+. A named +classpath+ may contain + * an anonymous component, but the last component is guaranteed + * to not be anonymous. <code>*permanent</code> is set to 1 + * if +classpath+ has no anonymous components. There is no builtin + * Ruby level APIs that can change a permanent +classpath+. + * + * YJIT needs this function to not allocate. */ static VALUE -find_class_path(VALUE klass, ID preferred) +classname(VALUE klass, bool *permanent) +{ + *permanent = false; + + VALUE classpath = RCLASS_CLASSPATH(klass); + if (classpath == 0) return Qnil; + + *permanent = RCLASS_PERMANENT_CLASSPATH_P(klass); + + return classpath; +} + +VALUE +rb_mod_name0(VALUE klass, bool *permanent) +{ + return classname(klass, permanent); +} + +/* + * call-seq: + * mod.name -> string or nil + * + * Returns the name of the module <i>mod</i>. Returns +nil+ for anonymous modules. + */ + +VALUE +rb_mod_name(VALUE mod) { - struct fc_result arg; + // YJIT needs this function to not allocate. + bool permanent; + return classname(mod, &permanent); +} + +// Similar to logic in rb_mod_const_get(). +static bool +is_constant_path(VALUE name) +{ + const char *path = RSTRING_PTR(name); + const char *pend = RSTRING_END(name); + rb_encoding *enc = rb_enc_get(name); - arg.preferred = preferred; - arg.name = 0; - arg.path = 0; - arg.klass = klass; - arg.track = rb_cObject; - arg.prev = 0; - if (RCLASS_CONST_TBL(rb_cObject)) { - rb_id_table_foreach(RCLASS_CONST_TBL(rb_cObject), fc_i, &arg); + const char *p = path; + + if (p >= pend || !*p) { + return false; } - if (arg.path) { - st_data_t tmp = tmp_classpath; - if (!RCLASS_IV_TBL(klass)) { - RCLASS_IV_TBL(klass) = st_init_numtable(); - } - rb_class_ivar_set(klass, classpath, arg.path); - st_delete(RCLASS_IV_TBL(klass), &tmp, 0); - return arg.path; + while (p < pend) { + if (p + 2 <= pend && p[0] == ':' && p[1] == ':') { + p += 2; + } + + const char *pbeg = p; + while (p < pend && *p != ':') p++; + + if (pbeg == p) return false; + + if (rb_enc_symname_type(pbeg, p - pbeg, enc, 0) != ID_CONST) { + return false; + } } - return Qnil; + + return true; +} + +struct sub_temporary_name_args { + VALUE names; + ID last; +}; + +static VALUE build_const_path(VALUE head, ID tail); +static void set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name); + +static VALUE +set_sub_temporary_name_recursive(VALUE mod, VALUE data, int recursive) +{ + if (recursive) return Qfalse; + + struct sub_temporary_name_args *args = (void *)data; + VALUE name = 0; + if (args->names) { + name = build_const_path(rb_ary_last(0, 0, args->names), args->last); + } + set_sub_temporary_name_foreach(mod, args, name); + return Qtrue; } -/** - * Returns +classpath+ of _klass_, if it is named, or +nil+ for - * anonymous +class+/+module+. The last part of named +classpath+ is - * never anonymous, but anonymous +class+/+module+ names may be - * contained. If the path is "permanent", that means it has no - * anonymous names, <code>*permanent</code> is set to 1. - */ static VALUE -classname(VALUE klass, int *permanent) -{ - VALUE path = Qnil; - st_data_t n; - - if (!klass) klass = rb_cObject; - *permanent = 1; - if (RCLASS_IV_TBL(klass)) { - if (!st_lookup(RCLASS_IV_TBL(klass), (st_data_t)classpath, &n)) { - ID cid = 0; - if (st_lookup(RCLASS_IV_TBL(klass), (st_data_t)classid, &n)) { - VALUE cname = (VALUE)n; - cid = rb_check_id(&cname); - if (cid) path = find_class_path(klass, cid); - } - if (NIL_P(path)) { - path = find_class_path(klass, (ID)0); - } - if (NIL_P(path)) { - if (!cid) { - return Qnil; - } - if (!st_lookup(RCLASS_IV_TBL(klass), (st_data_t)tmp_classpath, &n)) { - path = rb_id2str(cid); - return path; - } - *permanent = 0; - path = (VALUE)n; - return path; - } - } - else { - path = (VALUE)n; - } - if (!RB_TYPE_P(path, T_STRING)) { - rb_bug("class path is not set properly"); - } - return path; - } - return find_class_path(klass, (ID)0); +set_sub_temporary_name_topmost(VALUE mod, VALUE data, int recursive) +{ + if (recursive) return Qfalse; + + struct sub_temporary_name_args *args = (void *)data; + VALUE name = args->names; + if (name) { + args->names = rb_ary_hidden_new(0); + } + set_sub_temporary_name_foreach(mod, args, name); + return Qtrue; +} + +static enum rb_id_table_iterator_result +set_sub_temporary_name_i(ID id, VALUE val, void *data) +{ + val = ((rb_const_entry_t *)val)->value; + if (rb_namespace_p(val) && !RCLASS_PERMANENT_CLASSPATH_P(val)) { + VALUE arg = (VALUE)data; + struct sub_temporary_name_args *args = data; + args->last = id; + rb_exec_recursive_paired(set_sub_temporary_name_recursive, val, arg, arg); + } + return ID_TABLE_CONTINUE; +} + +static void +set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name) +{ + RCLASS_WRITE_CLASSPATH(mod, name, FALSE); + struct rb_id_table *tbl = RCLASS_CONST_TBL(mod); + if (!tbl) return; + if (!name) { + rb_id_table_foreach(tbl, set_sub_temporary_name_i, args); + } + else { + long names_len = RARRAY_LEN(args->names); // paranoiac check? + rb_ary_push(args->names, name); + rb_id_table_foreach(tbl, set_sub_temporary_name_i, args); + rb_ary_set_len(args->names, names_len); + } +} + +static void +set_sub_temporary_name(VALUE mod, VALUE name) +{ + struct sub_temporary_name_args args = {name}; + VALUE arg = (VALUE)&args; + rb_exec_recursive_paired(set_sub_temporary_name_topmost, mod, arg, arg); } /* * call-seq: - * mod.name -> string + * mod.set_temporary_name(string) -> self + * mod.set_temporary_name(nil) -> self + * + * Sets the temporary name of the module. This name is reflected in + * introspection of the module and the values that are related to it, such + * as instances, constants, and methods. + * + * The name should be +nil+ or a non-empty string that is not a valid constant + * path (to avoid confusing between permanent and temporary names). + * + * The method can be useful to distinguish dynamically generated classes and + * modules without assigning them to constants. + * + * If the module is given a permanent name by assigning it to a constant, + * the temporary name is discarded. A temporary name can't be assigned to + * modules that have a permanent name. + * + * If the given name is +nil+, the module becomes anonymous again. + * + * Example: + * + * m = Module.new # => #<Module:0x0000000102c68f38> + * m.name #=> nil + * + * m.set_temporary_name("fake_name") # => fake_name + * m.name #=> "fake_name" + * + * m.set_temporary_name(nil) # => #<Module:0x0000000102c68f38> + * m.name #=> nil * - * Returns the name of the module <i>mod</i>. Returns nil for anonymous modules. + * c = Class.new + * c.set_temporary_name("MyClass(with description)") # => MyClass(with description) + * + * c.new # => #<MyClass(with description):0x0....> + * + * c::M = m + * c::M.name #=> "MyClass(with description)::M" + * + * # Assigning to a constant replaces the name with a permanent one + * C = c + * + * C.name #=> "C" + * C::M.name #=> "C::M" + * c.new # => #<C:0x0....> */ VALUE -rb_mod_name(VALUE mod) +rb_mod_set_temporary_name(VALUE mod, VALUE name) { - int permanent; - VALUE path = classname(mod, &permanent); + // We don't allow setting the name if the classpath is already permanent: + if (RCLASS_PERMANENT_CLASSPATH_P(mod)) { + rb_raise(rb_eRuntimeError, "can't change permanent name"); + } - if (!NIL_P(path)) return rb_str_dup(path); - return path; + if (NIL_P(name)) { + // Set the temporary classpath to NULL (anonymous): + RB_VM_LOCKING() { + set_sub_temporary_name(mod, 0); + } + } + else { + // Ensure the name is a string: + StringValue(name); + + if (RSTRING_LEN(name) == 0) { + rb_raise(rb_eArgError, "empty class/module name"); + } + + if (is_constant_path(name)) { + rb_raise(rb_eArgError, "the temporary name must not be a constant path to avoid confusion"); + } + + name = rb_str_new_frozen(name); + RB_OBJ_SET_SHAREABLE(name); + + // Set the temporary classpath to the given name: + RB_VM_LOCKING() { + set_sub_temporary_name(mod, name); + } + } + + return mod; } static VALUE @@ -242,150 +338,111 @@ make_temporary_path(VALUE obj, VALUE klass) VALUE path; switch (klass) { case Qnil: - path = rb_sprintf("#<Class:%p>", (void*)obj); - break; + path = rb_sprintf("#<Class:%p>", (void*)obj); + break; case Qfalse: - path = rb_sprintf("#<Module:%p>", (void*)obj); - break; + path = rb_sprintf("#<Module:%p>", (void*)obj); + break; default: - path = rb_sprintf("#<%"PRIsVALUE":%p>", klass, (void*)obj); - break; + path = rb_sprintf("#<%"PRIsVALUE":%p>", klass, (void*)obj); + break; } OBJ_FREEZE(path); return path; } -typedef VALUE (*path_cache_func)(VALUE obj, VALUE name); +typedef VALUE (*fallback_func)(VALUE obj, VALUE name); static VALUE -rb_tmp_class_path(VALUE klass, int *permanent, path_cache_func cache_path) +rb_tmp_class_path(VALUE klass, bool *permanent, fallback_func fallback) { VALUE path = classname(klass, permanent); - st_data_t n = (st_data_t)path; if (!NIL_P(path)) { - return path; - } - if (RCLASS_IV_TBL(klass) && st_lookup(RCLASS_IV_TBL(klass), - (st_data_t)tmp_classpath, &n)) { - *permanent = 0; - return (VALUE)n; + return path; } - else { - if (RB_TYPE_P(klass, T_MODULE)) { - if (rb_obj_class(klass) == rb_cModule) { - path = Qfalse; - } - else { - int perm; - path = rb_tmp_class_path(RBASIC(klass)->klass, &perm, cache_path); - } - } - *permanent = 0; - return cache_path(klass, path); + + if (RB_TYPE_P(klass, T_MODULE)) { + if (rb_obj_class(klass) == rb_cModule) { + path = Qfalse; + } + else { + bool perm; + path = rb_tmp_class_path(RBASIC(klass)->klass, &perm, fallback); + } } -} -static VALUE -ivar_cache(VALUE obj, VALUE name) -{ - return rb_ivar_set(obj, tmp_classpath, make_temporary_path(obj, name)); + *permanent = false; + return fallback(klass, path); } VALUE rb_class_path(VALUE klass) { - int permanent; - VALUE path = rb_tmp_class_path(klass, &permanent, ivar_cache); + bool permanent; + VALUE path = rb_tmp_class_path(klass, &permanent, make_temporary_path); if (!NIL_P(path)) path = rb_str_dup(path); return path; } -static VALUE -null_cache(VALUE obj, VALUE name) +VALUE +rb_class_path_cached(VALUE klass) { - return make_temporary_path(obj, name); + return rb_mod_name(klass); } -VALUE -rb_class_path_no_cache(VALUE klass) +static VALUE +no_fallback(VALUE obj, VALUE name) { - int permanent; - VALUE path = rb_tmp_class_path(klass, &permanent, null_cache); - if (!NIL_P(path)) path = rb_str_dup(path); - return path; + return name; } VALUE -rb_class_path_cached(VALUE klass) +rb_search_class_path(VALUE klass) { - st_table *ivtbl; - st_data_t n; - - if (!RCLASS_EXT(klass)) return Qnil; - if (!(ivtbl = RCLASS_IV_TBL(klass))) return Qnil; - if (st_lookup(ivtbl, (st_data_t)classpath, &n)) return (VALUE)n; - if (st_lookup(ivtbl, (st_data_t)tmp_classpath, &n)) return (VALUE)n; - return Qnil; + bool permanent; + return rb_tmp_class_path(klass, &permanent, no_fallback); } static VALUE -never_cache(VALUE obj, VALUE name) +build_const_pathname(VALUE head, VALUE tail) { - return name; + VALUE path = rb_str_dup(head); + rb_str_cat2(path, "::"); + rb_str_append(path, tail); + return rb_fstring(path); } -VALUE -rb_search_class_path(VALUE klass) +static VALUE +build_const_path(VALUE head, ID tail) { - int permanent; - return rb_tmp_class_path(klass, &permanent, never_cache); + return build_const_pathname(head, rb_id2str(tail)); } void rb_set_class_path_string(VALUE klass, VALUE under, VALUE name) { - VALUE str; - ID pathid = classpath; + bool permanent = true; + VALUE str; if (under == rb_cObject) { - str = rb_str_new_frozen(name); + str = rb_str_new_frozen(name); } else { - int permanent; - str = rb_str_dup(rb_tmp_class_path(under, &permanent, ivar_cache)); - rb_str_cat2(str, "::"); - rb_str_append(str, name); - OBJ_FREEZE(str); - if (!permanent) { - pathid = tmp_classpath; - rb_ivar_set(klass, classid, rb_str_intern(name)); - } + str = rb_tmp_class_path(under, &permanent, make_temporary_path); + str = build_const_pathname(str, name); } - rb_ivar_set(klass, pathid, str); + + RB_OBJ_SET_SHAREABLE(str); + RCLASS_SET_CLASSPATH(klass, str, permanent); } void rb_set_class_path(VALUE klass, VALUE under, const char *name) { - VALUE str; - ID pathid = classpath; - - if (under == rb_cObject) { - str = rb_str_new2(name); - } - else { - int permanent; - str = rb_str_dup(rb_tmp_class_path(under, &permanent, ivar_cache)); - rb_str_cat2(str, "::"); - rb_str_cat2(str, name); - if (!permanent) { - pathid = tmp_classpath; - rb_ivar_set(klass, classid, rb_str_intern(rb_str_new_cstr(name))); - } - } + VALUE str = rb_str_new2(name); OBJ_FREEZE(str); - rb_ivar_set(klass, pathid, str); + rb_set_class_path_string(klass, under, str); } VALUE @@ -397,37 +454,40 @@ rb_path_to_class(VALUE pathname) VALUE c = rb_cObject; if (!rb_enc_asciicompat(enc)) { - rb_raise(rb_eArgError, "invalid class path encoding (non ASCII)"); + rb_raise(rb_eArgError, "invalid class path encoding (non ASCII)"); } pbeg = p = path; pend = path + RSTRING_LEN(pathname); if (path == pend || path[0] == '#') { - rb_raise(rb_eArgError, "can't retrieve anonymous class %"PRIsVALUE, - QUOTE(pathname)); + rb_raise(rb_eArgError, "can't retrieve anonymous class %"PRIsVALUE, + QUOTE(pathname)); } while (p < pend) { - while (p < pend && *p != ':') p++; - id = rb_check_id_cstr(pbeg, p-pbeg, enc); - if (p < pend && p[0] == ':') { - if ((size_t)(pend - p) < 2 || p[1] != ':') goto undefined_class; - p += 2; - pbeg = p; - } - if (!id) { - undefined_class: - rb_raise(rb_eArgError, "undefined class/module % "PRIsVALUE, - rb_str_subseq(pathname, 0, p-path)); - } - c = rb_const_search(c, id, TRUE, FALSE, FALSE); - if (c == Qundef) goto undefined_class; - if (!RB_TYPE_P(c, T_MODULE) && !RB_TYPE_P(c, T_CLASS)) { - rb_raise(rb_eTypeError, "%"PRIsVALUE" does not refer to class/module", - pathname); - } + while (p < pend && *p != ':') p++; + id = rb_check_id_cstr(pbeg, p-pbeg, enc); + if (p < pend && p[0] == ':') { + if ((size_t)(pend - p) < 2 || p[1] != ':') goto undefined_class; + p += 2; + pbeg = p; + } + if (!id) { + goto undefined_class; + } + c = rb_const_search(c, id, TRUE, FALSE, FALSE, NULL); + if (UNDEF_P(c)) goto undefined_class; + if (!rb_namespace_p(c)) { + rb_raise(rb_eTypeError, "%"PRIsVALUE" does not refer to class/module", + pathname); + } } RB_GC_GUARD(pathname); return c; + + undefined_class: + rb_raise(rb_eArgError, "undefined class/module % "PRIsVALUE, + rb_str_subseq(pathname, 0, p-path)); + UNREACHABLE_RETURN(Qundef); } VALUE @@ -436,12 +496,6 @@ rb_path2class(const char *path) return rb_path_to_class(rb_str_new_cstr(path)); } -void -rb_name_class(VALUE klass, ID id) -{ - rb_ivar_set(klass, classid, ID2SYM(id)); -} - VALUE rb_class_name(VALUE klass) { @@ -451,8 +505,8 @@ rb_class_name(VALUE klass) const char * rb_class2name(VALUE klass) { - int permanent; - VALUE path = rb_tmp_class_path(rb_class_real(klass), &permanent, ivar_cache); + bool permanent; + VALUE path = rb_tmp_class_path(rb_class_real(klass), &permanent, make_temporary_path); if (NIL_P(path)) return NULL; return RSTRING_PTR(path); } @@ -473,55 +527,162 @@ struct trace_var { struct rb_global_variable { int counter; int block_trace; - void *data; + VALUE *data; rb_gvar_getter_t *getter; rb_gvar_setter_t *setter; rb_gvar_marker_t *marker; + rb_gvar_compact_t *compactor; struct trace_var *trace; + bool box_ready; }; -struct rb_global_entry* -rb_global_entry(ID id) +struct rb_global_entry { + struct rb_global_variable *var; + ID id; + bool ractor_local; +}; + +static void +free_global_variable(struct rb_global_variable *var) +{ + RUBY_ASSERT(var->counter == 0); + + struct trace_var *trace = var->trace; + while (trace) { + struct trace_var *next = trace->next; + xfree(trace); + trace = next; + } + xfree(var); +} + +static enum rb_id_table_iterator_result +free_global_entry_i(VALUE val, void *arg) +{ + struct rb_global_entry *entry = (struct rb_global_entry *)val; + entry->var->counter--; + if (entry->var->counter == 0) { + free_global_variable(entry->var); + } + ruby_xfree(entry); + return ID_TABLE_DELETE; +} + +void +rb_free_rb_global_tbl(void) +{ + rb_id_table_foreach_values(rb_global_tbl, free_global_entry_i, 0); + rb_id_table_free(rb_global_tbl); +} + +void +rb_free_generic_fields_tbl_(void) +{ + st_free_table(generic_fields_tbl_); +} + +static struct rb_global_entry* +rb_find_global_entry(ID id) { struct rb_global_entry *entry; VALUE data; - if (!rb_id_table_lookup(rb_global_tbl, id, &data)) { - struct rb_global_variable *var; - entry = ALLOC(struct rb_global_entry); - var = ALLOC(struct rb_global_variable); - entry->id = id; - entry->var = var; - var->counter = 1; - var->data = 0; - var->getter = rb_gvar_undef_getter; - var->setter = rb_gvar_undef_setter; - var->marker = rb_gvar_undef_marker; - - var->block_trace = 0; - var->trace = 0; - rb_id_table_insert(rb_global_tbl, id, (VALUE)entry); + RB_VM_LOCKING() { + if (!rb_id_table_lookup(rb_global_tbl, id, &data)) { + entry = NULL; + } + else { + entry = (struct rb_global_entry *)data; + RUBY_ASSERT(entry != NULL); + } } - else { - entry = (struct rb_global_entry *)data; + + if (UNLIKELY(!rb_ractor_main_p()) && (!entry || !entry->ractor_local)) { + rb_raise(rb_eRactorIsolationError, "can not access global variable %s from non-main Ractor", rb_id2name(id)); + } + + return entry; +} + +void +rb_gvar_ractor_local(const char *name) +{ + struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name)); + entry->ractor_local = true; +} + +void +rb_gvar_box_ready(const char *name) +{ + struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name)); + entry->var->box_ready = true; +} + +static void +rb_gvar_undef_compactor(void *var) +{ +} + +static struct rb_global_entry* +rb_global_entry(ID id) +{ + struct rb_global_entry *entry; + RB_VM_LOCKING() { + entry = rb_find_global_entry(id); + if (!entry) { + struct rb_global_variable *var; + entry = ALLOC(struct rb_global_entry); + var = ALLOC(struct rb_global_variable); + entry->id = id; + entry->var = var; + entry->ractor_local = false; + var->counter = 1; + var->data = 0; + var->getter = rb_gvar_undef_getter; + var->setter = rb_gvar_undef_setter; + var->marker = rb_gvar_undef_marker; + var->compactor = rb_gvar_undef_compactor; + + var->block_trace = 0; + var->trace = 0; + var->box_ready = false; + rb_id_table_insert(rb_global_tbl, id, (VALUE)entry); + } } return entry; } VALUE -rb_gvar_undef_getter(ID id, void *data, struct rb_global_variable *var) +rb_gvar_undef_getter(ID id, VALUE *_) { - rb_warning("global variable `%"PRIsVALUE"' not initialized", QUOTE_ID(id)); + rb_warning("global variable '%"PRIsVALUE"' not initialized", QUOTE_ID(id)); return Qnil; } +static void +rb_gvar_val_compactor(void *_var) +{ + struct rb_global_variable *var = (struct rb_global_variable *)_var; + + VALUE obj = (VALUE)var->data; + + if (obj) { + VALUE new = rb_gc_location(obj); + if (new != obj) { + var->data = (void*)new; + } + } +} + void -rb_gvar_undef_setter(VALUE val, ID id, void *d, struct rb_global_variable *var) +rb_gvar_undef_setter(VALUE val, ID id, VALUE *_) { + struct rb_global_variable *var = rb_global_entry(id)->var; var->getter = rb_gvar_val_getter; var->setter = rb_gvar_val_setter; var->marker = rb_gvar_val_marker; + var->compactor = rb_gvar_val_compactor; var->data = (void*)val; } @@ -532,14 +693,15 @@ rb_gvar_undef_marker(VALUE *var) } VALUE -rb_gvar_val_getter(ID id, void *data, struct rb_global_variable *var) +rb_gvar_val_getter(ID id, VALUE *data) { return (VALUE)data; } void -rb_gvar_val_setter(VALUE val, ID id, void *data, struct rb_global_variable *var) +rb_gvar_val_setter(VALUE val, ID id, VALUE *_) { + struct rb_global_variable *var = rb_global_entry(id)->var; var->data = (void*)val; } @@ -547,21 +709,20 @@ void rb_gvar_val_marker(VALUE *var) { VALUE data = (VALUE)var; - if (data) rb_gc_mark_maybe(data); + if (data) rb_gc_mark_movable(data); } VALUE -rb_gvar_var_getter(ID id, void *data, struct rb_global_variable *gvar) +rb_gvar_var_getter(ID id, VALUE *var) { - VALUE *var = data; if (!var) return Qnil; return *var; } void -rb_gvar_var_setter(VALUE val, ID id, void *data, struct rb_global_variable *g) +rb_gvar_var_setter(VALUE val, ID id, VALUE *data) { - *(VALUE *)data = val; + *data = val; } void @@ -571,7 +732,7 @@ rb_gvar_var_marker(VALUE *var) } void -rb_gvar_readonly_setter(VALUE v, ID id, void *d, struct rb_global_variable *g) +rb_gvar_readonly_setter(VALUE v, ID id, VALUE *_) { rb_name_error(id, "%"PRIsVALUE" is a read-only variable", QUOTE_ID(id)); } @@ -586,17 +747,35 @@ mark_global_entry(VALUE v, void *ignored) (*var->marker)(var->data); trace = var->trace; while (trace) { - if (trace->data) rb_gc_mark_maybe(trace->data); - trace = trace->next; + if (trace->data) rb_gc_mark_maybe(trace->data); + trace = trace->next; } return ID_TABLE_CONTINUE; } +#define gc_mark_table(task) \ + if (rb_global_tbl) { rb_id_table_foreach_values(rb_global_tbl, task##_global_entry, 0); } + void rb_gc_mark_global_tbl(void) { - if (rb_global_tbl) - rb_id_table_foreach_values(rb_global_tbl, mark_global_entry, 0); + gc_mark_table(mark); +} + +static enum rb_id_table_iterator_result +update_global_entry(VALUE v, void *ignored) +{ + struct rb_global_entry *entry = (struct rb_global_entry *)v; + struct rb_global_variable *var = entry->var; + + (*var->compactor)(var); + return ID_TABLE_CONTINUE; +} + +void +rb_gc_update_global_tbl(void) +{ + gc_mark_table(update); } static ID @@ -606,21 +785,44 @@ global_id(const char *name) if (name[0] == '$') id = rb_intern(name); else { - size_t len = strlen(name); - char *buf = ALLOCA_N(char, len+1); - buf[0] = '$'; - memcpy(buf+1, name, len); - id = rb_intern2(buf, len+1); + size_t len = strlen(name); + VALUE vbuf = 0; + char *buf = ALLOCV_N(char, vbuf, len+1); + buf[0] = '$'; + memcpy(buf+1, name, len); + id = rb_intern2(buf, len+1); + ALLOCV_END(vbuf); } return id; } +static ID +find_global_id(const char *name) +{ + ID id; + size_t len = strlen(name); + + if (name[0] == '$') { + id = rb_check_id_cstr(name, len, NULL); + } + else { + VALUE vbuf = 0; + char *buf = ALLOCV_N(char, vbuf, len+1); + buf[0] = '$'; + memcpy(buf+1, name, len); + id = rb_check_id_cstr(buf, len+1, NULL); + ALLOCV_END(vbuf); + } + + return id; +} + void rb_define_hooked_variable( const char *name, VALUE *var, - VALUE (*getter)(ANYARGS), - void (*setter)(ANYARGS)) + rb_gvar_getter_t *getter, + rb_gvar_setter_t *setter) { volatile VALUE tmp = var ? *var : Qnil; ID id = global_id(name); @@ -649,8 +851,8 @@ rb_define_readonly_variable(const char *name, const VALUE *var) void rb_define_virtual_variable( const char *name, - VALUE (*getter)(ANYARGS), - void (*setter)(ANYARGS)) + rb_gvar_getter_t *getter, + rb_gvar_setter_t *setter) { if (!getter) getter = rb_gvar_val_getter; if (!setter) setter = rb_gvar_readonly_setter; @@ -660,32 +862,9 @@ rb_define_virtual_variable( static void rb_trace_eval(VALUE cmd, VALUE val) { - rb_eval_cmd(cmd, rb_ary_new3(1, val), 0); + rb_eval_cmd_call_kw(cmd, 1, &val, RB_NO_KEYWORDS); } -/* - * call-seq: - * trace_var(symbol, cmd ) -> nil - * trace_var(symbol) {|val| block } -> nil - * - * Controls tracing of assignments to global variables. The parameter - * +symbol+ identifies the variable (as either a string name or a - * symbol identifier). _cmd_ (which may be a string or a - * +Proc+ object) or block is executed whenever the variable - * is assigned. The block or +Proc+ object receives the - * variable's new value as a parameter. Also see - * <code>Kernel::untrace_var</code>. - * - * trace_var :$_, proc {|v| puts "$_ is now '#{v}'" } - * $_ = "hello" - * $_ = ' there' - * - * <em>produces:</em> - * - * $_ is now 'hello' - * $_ is now ' there' - */ - VALUE rb_f_trace_var(int argc, const VALUE *argv) { @@ -694,15 +873,12 @@ rb_f_trace_var(int argc, const VALUE *argv) struct trace_var *trace; if (rb_scan_args(argc, argv, "11", &var, &cmd) == 1) { - cmd = rb_block_proc(); + cmd = rb_block_proc(); } if (NIL_P(cmd)) { - return rb_f_untrace_var(argc, argv); + return rb_f_untrace_var(argc, argv); } entry = rb_global_entry(rb_to_id(var)); - if (OBJ_TAINTED(cmd)) { - rb_raise(rb_eSecurityError, "Insecure: tainted variable trace"); - } trace = ALLOC(struct trace_var); trace->next = entry->var->trace; trace->func = rb_trace_eval; @@ -723,28 +899,18 @@ remove_trace(struct rb_global_variable *var) t.next = trace; trace = &t; while (trace->next) { - next = trace->next; - if (next->removed) { - trace->next = next->next; - xfree(next); - } - else { - trace = next; - } + next = trace->next; + if (next->removed) { + trace->next = next->next; + xfree(next); + } + else { + trace = next; + } } var->trace = t.next; } -/* - * call-seq: - * untrace_var(symbol [, cmd] ) -> array or nil - * - * Removes tracing for the specified command on the given global - * variable and returns +nil+. If no command is specified, - * removes all tracing for that variable and returns an array - * containing the commands actually removed. - */ - VALUE rb_f_untrace_var(int argc, const VALUE *argv) { @@ -752,117 +918,194 @@ rb_f_untrace_var(int argc, const VALUE *argv) ID id; struct rb_global_entry *entry; struct trace_var *trace; - VALUE data; rb_scan_args(argc, argv, "11", &var, &cmd); id = rb_check_id(&var); if (!id) { - rb_name_error_str(var, "undefined global variable %"PRIsVALUE"", QUOTE(var)); + rb_name_error_str(var, "undefined global variable %"PRIsVALUE"", QUOTE(var)); } - if (!rb_id_table_lookup(rb_global_tbl, id, &data)) { - rb_name_error(id, "undefined global variable %"PRIsVALUE"", QUOTE_ID(id)); + if ((entry = rb_find_global_entry(id)) == NULL) { + rb_name_error(id, "undefined global variable %"PRIsVALUE"", QUOTE_ID(id)); } - trace = (entry = (struct rb_global_entry *)data)->var->trace; + trace = entry->var->trace; if (NIL_P(cmd)) { - VALUE ary = rb_ary_new(); + VALUE ary = rb_ary_new(); - while (trace) { - struct trace_var *next = trace->next; - rb_ary_push(ary, (VALUE)trace->data); - trace->removed = 1; - trace = next; - } + while (trace) { + struct trace_var *next = trace->next; + rb_ary_push(ary, (VALUE)trace->data); + trace->removed = 1; + trace = next; + } - if (!entry->var->block_trace) remove_trace(entry->var); - return ary; + if (!entry->var->block_trace) remove_trace(entry->var); + return ary; } else { - while (trace) { - if (trace->data == cmd) { - trace->removed = 1; - if (!entry->var->block_trace) remove_trace(entry->var); - return rb_ary_new3(1, cmd); - } - trace = trace->next; - } + while (trace) { + if (trace->data == cmd) { + trace->removed = 1; + if (!entry->var->block_trace) remove_trace(entry->var); + return rb_ary_new3(1, cmd); + } + trace = trace->next; + } } return Qnil; } -VALUE -rb_gvar_get(struct rb_global_entry *entry) -{ - struct rb_global_variable *var = entry->var; - return (*var->getter)(entry->id, var->data, var); -} - struct trace_data { struct trace_var *trace; VALUE val; }; static VALUE -trace_ev(struct trace_data *data) +trace_ev(VALUE v) { + struct trace_data *data = (void *)v; struct trace_var *trace = data->trace; while (trace) { - (*trace->func)(trace->data, data->val); - trace = trace->next; + (*trace->func)(trace->data, data->val); + trace = trace->next; } return Qnil; } static VALUE -trace_en(struct rb_global_variable *var) +trace_en(VALUE v) { + struct rb_global_variable *var = (void *)v; var->block_trace = 0; remove_trace(var); return Qnil; /* not reached */ } -VALUE -rb_gvar_set(struct rb_global_entry *entry, VALUE val) +static VALUE +rb_gvar_set_entry(struct rb_global_entry *entry, VALUE val) { struct trace_data trace; struct rb_global_variable *var = entry->var; - (*var->setter)(val, entry->id, var->data, var); + (*var->setter)(val, entry->id, var->data); if (var->trace && !var->block_trace) { - var->block_trace = 1; - trace.trace = var->trace; - trace.val = val; - rb_ensure(trace_ev, (VALUE)&trace, trace_en, (VALUE)var); + var->block_trace = 1; + trace.trace = var->trace; + trace.val = val; + rb_ensure(trace_ev, (VALUE)&trace, trace_en, (VALUE)var); } return val; } +#define USE_BOX_GVAR_TBL(ns,entry) \ + (BOX_USER_P(ns) && \ + (!entry || !entry->var->box_ready || entry->var->setter != rb_gvar_readonly_setter)) + VALUE -rb_gv_set(const char *name, VALUE val) +rb_gvar_set(ID id, VALUE val) { + VALUE retval; struct rb_global_entry *entry; + const rb_box_t *box = rb_current_box(); + bool use_box_tbl = false; + + RB_VM_LOCKING() { + entry = rb_global_entry(id); + + if (USE_BOX_GVAR_TBL(box, entry)) { + use_box_tbl = true; + rb_hash_aset(box->gvar_tbl, rb_id2sym(entry->id), val); + retval = val; + // TODO: think about trace + } + } - entry = rb_global_entry(global_id(name)); - return rb_gvar_set(entry, val); + if (!use_box_tbl) { + retval = rb_gvar_set_entry(entry, val); + } + return retval; } VALUE -rb_gv_get(const char *name) +rb_gv_set(const char *name, VALUE val) { + return rb_gvar_set(global_id(name), val); +} + +VALUE +rb_gvar_get(ID id) +{ + VALUE retval, gvars, key; + const rb_box_t *box = rb_current_box(); + bool use_box_tbl = false; struct rb_global_entry *entry; + struct rb_global_variable *var; + // TODO: use lock-free rb_id_table when it's available for use (doesn't yet exist) + RB_VM_LOCKING() { + entry = rb_global_entry(id); + var = entry->var; + + if (USE_BOX_GVAR_TBL(box, entry)) { + use_box_tbl = true; + gvars = box->gvar_tbl; + key = rb_id2sym(entry->id); + if (RTEST(rb_hash_has_key(gvars, key))) { // this gvar is already cached + retval = rb_hash_aref(gvars, key); + } + else { + RB_VM_UNLOCK(); + { + retval = (*var->getter)(entry->id, var->data); + if (rb_obj_respond_to(retval, rb_intern("clone"), 1)) { + retval = rb_funcall(retval, rb_intern("clone"), 0); + } + } + RB_VM_LOCK(); + rb_hash_aset(gvars, key, retval); + } + } + } + if (!use_box_tbl) { + retval = (*var->getter)(entry->id, var->data); + } + return retval; +} - entry = rb_global_entry(global_id(name)); - return rb_gvar_get(entry); +VALUE +rb_gv_get(const char *name) +{ + ID id = find_global_id(name); + + if (!id) { + rb_warning("global variable '%s' not initialized", name); + return Qnil; + } + + return rb_gvar_get(id); } VALUE -rb_gvar_defined(struct rb_global_entry *entry) +rb_gvar_defined(ID id) { - if (entry->var->getter == rb_gvar_undef_getter) return Qfalse; - return Qtrue; + struct rb_global_entry *entry = rb_global_entry(id); + return RBOOL(entry->var->getter != rb_gvar_undef_getter); +} + +rb_gvar_getter_t * +rb_gvar_getter_function_of(ID id) +{ + const struct rb_global_entry *entry = rb_global_entry(id); + return entry->var->getter; +} + +rb_gvar_setter_t * +rb_gvar_setter_function_of(ID id) +{ + const struct rb_global_entry *entry = rb_global_entry(id); + return entry->var->setter; } static enum rb_id_table_iterator_result @@ -873,39 +1116,35 @@ gvar_i(ID key, VALUE val, void *a) return ID_TABLE_CONTINUE; } -/* - * call-seq: - * global_variables -> array - * - * Returns an array of the names of global variables. - * - * global_variables.grep /std/ #=> [:$stdin, :$stdout, :$stderr] - */ - VALUE rb_f_global_variables(void) { VALUE ary = rb_ary_new(); VALUE sym, backref = rb_backref_get(); + if (!rb_ractor_main_p()) { + rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors"); + } + /* gvar access (get/set) in boxes creates gvar entries globally */ + rb_id_table_foreach(rb_global_tbl, gvar_i, (void *)ary); if (!NIL_P(backref)) { - char buf[2]; - int i, nmatch = rb_match_count(backref); - buf[0] = '$'; - for (i = 1; i <= nmatch; ++i) { - if (!rb_match_nth_defined(i, backref)) continue; - if (i < 10) { - /* probably reused, make static ID */ - buf[1] = (char)(i + '0'); - sym = ID2SYM(rb_intern2(buf, 2)); - } - else { - /* dynamic symbol */ - sym = rb_str_intern(rb_sprintf("$%d", i)); - } - rb_ary_push(ary, sym); - } + char buf[2]; + int i, nmatch = rb_match_count(backref); + buf[0] = '$'; + for (i = 1; i <= nmatch; ++i) { + if (!RTEST(rb_reg_nth_defined(i, backref))) continue; + if (i < 10) { + /* probably reused, make static ID */ + buf[1] = (char)(i + '0'); + sym = ID2SYM(rb_intern2(buf, 2)); + } + else { + /* dynamic symbol */ + sym = rb_str_intern(rb_sprintf("$%d", i)); + } + rb_ary_push(ary, sym); + } } return ary; } @@ -913,313 +1152,398 @@ rb_f_global_variables(void) void rb_alias_variable(ID name1, ID name2) { - struct rb_global_entry *entry1, *entry2; + struct rb_global_entry *entry1 = NULL, *entry2; VALUE data1; + struct rb_id_table *gtbl = rb_global_tbl; - entry2 = rb_global_entry(name2); - if (!rb_id_table_lookup(rb_global_tbl, name1, &data1)) { - entry1 = ALLOC(struct rb_global_entry); - entry1->id = name1; - rb_id_table_insert(rb_global_tbl, name1, (VALUE)entry1); - } - else if ((entry1 = (struct rb_global_entry *)data1)->var != entry2->var) { - struct rb_global_variable *var = entry1->var; - if (var->block_trace) { - rb_raise(rb_eRuntimeError, "can't alias in tracer"); - } - var->counter--; - if (var->counter == 0) { - struct trace_var *trace = var->trace; - while (trace) { - struct trace_var *next = trace->next; - xfree(trace); - trace = next; - } - xfree(var); - } + if (!rb_ractor_main_p()) { + rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors"); } - else { - return; + + RB_VM_LOCKING() { + entry2 = rb_global_entry(name2); + if (!rb_id_table_lookup(gtbl, name1, &data1)) { + entry1 = ZALLOC(struct rb_global_entry); + entry1->id = name1; + rb_id_table_insert(gtbl, name1, (VALUE)entry1); + } + else if ((entry1 = (struct rb_global_entry *)data1)->var != entry2->var) { + struct rb_global_variable *var = entry1->var; + if (var->block_trace) { + RB_VM_UNLOCK(); + rb_raise(rb_eRuntimeError, "can't alias in tracer"); + } + var->counter--; + if (var->counter == 0) { + free_global_variable(var); + } + } + if (entry1->var != entry2->var) { + entry2->var->counter++; + entry1->var = entry2->var; + } } - entry2->var->counter++; - entry1->var = entry2->var; } -static int -gen_ivtbl_get(VALUE obj, struct gen_ivtbl **ivtbl) +static void +IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(ID id) { - st_data_t data; - - if (st_lookup(generic_iv_tbl, (st_data_t)obj, &data)) { - *ivtbl = (struct gen_ivtbl *)data; - return 1; + if (UNLIKELY(!rb_ractor_main_p())) { + if (rb_is_instance_id(id)) { // check only normal ivars + rb_raise(rb_eRactorIsolationError, "can not set instance variables of classes/modules by non-main Ractors"); + } } - return 0; } -static VALUE -generic_ivar_delete(VALUE obj, ID id, VALUE undef) +static void +CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(VALUE klass, ID id) { - struct gen_ivtbl *ivtbl; - - if (gen_ivtbl_get(obj, &ivtbl)) { - st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj)); - st_data_t index; - - if (st_lookup(iv_index_tbl, (st_data_t)id, &index)) { - if (index < ivtbl->numiv) { - VALUE ret = ivtbl->ivptr[index]; - - ivtbl->ivptr[index] = Qundef; - return ret == Qundef ? undef : ret; - } - } + if (UNLIKELY(!rb_ractor_main_p())) { + rb_raise(rb_eRactorIsolationError, "can not access class variables from non-main Ractors (%"PRIsVALUE" from %"PRIsVALUE")", rb_id2str(id), klass); } - return undef; } -static VALUE -generic_ivar_get(VALUE obj, ID id, VALUE undef) +static inline void +ivar_ractor_check(VALUE obj, ID id) { - struct gen_ivtbl *ivtbl; - - if (gen_ivtbl_get(obj, &ivtbl)) { - st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj)); - st_data_t index; + if (LIKELY(rb_is_instance_id(id)) /* not internal ID */ && + !RB_OBJ_FROZEN_RAW(obj) && + UNLIKELY(!rb_ractor_main_p()) && + UNLIKELY(rb_ractor_shareable_p(obj))) { - if (st_lookup(iv_index_tbl, (st_data_t)id, &index)) { - if (index < ivtbl->numiv) { - VALUE ret = ivtbl->ivptr[index]; - - return ret == Qundef ? undef : ret; - } - } + rb_raise(rb_eRactorIsolationError, "can not access instance variables of shareable objects from non-main Ractors"); } - return undef; } -static size_t -gen_ivtbl_bytes(size_t n) +static inline struct st_table * +generic_fields_tbl_no_ractor_check(void) { - return sizeof(struct gen_ivtbl) + n * sizeof(VALUE) - sizeof(VALUE); + ASSERT_vm_locking(); + + return generic_fields_tbl_; } -static struct gen_ivtbl * -gen_ivtbl_resize(struct gen_ivtbl *old, uint32_t n) +struct st_table * +rb_generic_fields_tbl_get(void) { - uint32_t len = old ? old->numiv : 0; - struct gen_ivtbl *ivtbl = xrealloc(old, gen_ivtbl_bytes(n)); - - ivtbl->numiv = n; - for (; len < n; len++) { - ivtbl->ivptr[len] = Qundef; - } - - return ivtbl; + return generic_fields_tbl_; } -#if 0 -static struct gen_ivtbl * -gen_ivtbl_dup(const struct gen_ivtbl *orig) +void +rb_mark_generic_ivar(VALUE obj) { - size_t s = gen_ivtbl_bytes(orig->numiv); - struct gen_ivtbl *ivtbl = xmalloc(s); - - memcpy(ivtbl, orig, s); - - return ivtbl; + VALUE data; + // Bypass ASSERT_vm_locking() check because marking may happen concurrently with mmtk + if (st_lookup(generic_fields_tbl_, (st_data_t)obj, (st_data_t *)&data)) { + rb_gc_mark_movable(data); + } } -#endif -static uint32_t -iv_index_tbl_newsize(struct ivar_update *ivup) +VALUE +rb_obj_fields_generic_uncached(VALUE obj) { - uint32_t index = (uint32_t)ivup->index; /* should not overflow */ - uint32_t newsize = (index+1) + (index+1)/4; /* (index+1)*1.25 */ + VALUE fields_obj = 0; + RB_VM_LOCKING() { + if (!st_lookup(generic_fields_tbl_, (st_data_t)obj, (st_data_t *)&fields_obj)) { + rb_bug("Object is missing entry in generic_fields_tbl"); + } + } + return fields_obj; +} - if (!ivup->iv_extended && - ivup->u.iv_index_tbl->num_entries < (st_index_t)newsize) { - newsize = (uint32_t)ivup->u.iv_index_tbl->num_entries; +VALUE +rb_obj_fields(VALUE obj, ID field_name) +{ + RUBY_ASSERT(!RB_TYPE_P(obj, T_IMEMO)); + ivar_ractor_check(obj, field_name); + + VALUE fields_obj = 0; + if (rb_shape_obj_has_fields(obj)) { + switch (BUILTIN_TYPE(obj)) { + case T_DATA: + if (LIKELY(RTYPEDDATA_P(obj))) { + fields_obj = RTYPEDDATA(obj)->fields_obj; + break; + } + goto generic_fields; + case T_STRUCT: + if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) { + fields_obj = RSTRUCT_FIELDS_OBJ(obj); + break; + } + goto generic_fields; + default: + generic_fields: + { + rb_execution_context_t *ec = GET_EC(); + if (ec->gen_fields_cache.obj == obj && !UNDEF_P(ec->gen_fields_cache.fields_obj) && rb_imemo_fields_owner(ec->gen_fields_cache.fields_obj) == obj) { + fields_obj = ec->gen_fields_cache.fields_obj; + RUBY_ASSERT(fields_obj == rb_obj_fields_generic_uncached(obj)); + } + else { + fields_obj = rb_obj_fields_generic_uncached(obj); + ec->gen_fields_cache.fields_obj = fields_obj; + ec->gen_fields_cache.obj = obj; + } + } + } } - return newsize; + return fields_obj; } -static int -generic_ivar_update(st_data_t *k, st_data_t *v, st_data_t u, int existing) +void +rb_free_generic_ivar(VALUE obj) { - VALUE obj = (VALUE)*k; - struct ivar_update *ivup = (struct ivar_update *)u; - uint32_t newsize; - int ret = ST_CONTINUE; - struct gen_ivtbl *ivtbl; - - if (existing) { - ivtbl = (struct gen_ivtbl *)*v; - if (ivup->index >= ivtbl->numiv) { - goto resize; - } - ret = ST_STOP; - } - else { - FL_SET(obj, FL_EXIVAR); - ivtbl = 0; -resize: - newsize = iv_index_tbl_newsize(ivup); - ivtbl = gen_ivtbl_resize(ivtbl, newsize); - *v = (st_data_t)ivtbl; + if (rb_obj_gen_fields_p(obj)) { + st_data_t key = (st_data_t)obj, value; + switch (BUILTIN_TYPE(obj)) { + case T_DATA: + if (LIKELY(RTYPEDDATA_P(obj))) { + RB_OBJ_WRITE(obj, &RTYPEDDATA(obj)->fields_obj, 0); + break; + } + goto generic_fields; + case T_STRUCT: + if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) { + RSTRUCT_SET_FIELDS_OBJ(obj, 0); + break; + } + goto generic_fields; + default: + generic_fields: + { + // Other EC may have stale caches, so fields_obj should be + // invalidated and the GC will replace with Qundef + rb_execution_context_t *ec = GET_EC(); + if (ec->gen_fields_cache.obj == obj) { + ec->gen_fields_cache.obj = Qundef; + ec->gen_fields_cache.fields_obj = Qundef; + } + RB_VM_LOCKING() { + if (!st_delete(generic_fields_tbl_no_ractor_check(), &key, &value)) { + rb_bug("Object is missing entry in generic_fields_tbl"); + } + } + } + } + RBASIC_SET_SHAPE_ID(obj, ROOT_SHAPE_ID); } - ivup->u.ivtbl = ivtbl; - return ret; } -static VALUE -generic_ivar_defined(VALUE obj, ID id) +static void +rb_obj_set_fields(VALUE obj, VALUE fields_obj, ID field_name, VALUE original_fields_obj) { - struct gen_ivtbl *ivtbl; - st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj)); - st_data_t index; + ivar_ractor_check(obj, field_name); - if (!iv_index_tbl) return Qfalse; - if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) return Qfalse; - if (!gen_ivtbl_get(obj, &ivtbl)) return Qfalse; - - if ((index < ivtbl->numiv) && (ivtbl->ivptr[index] != Qundef)) - return Qtrue; + if (!fields_obj) { + RUBY_ASSERT(original_fields_obj); + rb_free_generic_ivar(obj); + rb_imemo_fields_clear(original_fields_obj); + return; + } - return Qfalse; -} + RUBY_ASSERT(IMEMO_TYPE_P(fields_obj, imemo_fields)); + RUBY_ASSERT(!original_fields_obj || IMEMO_TYPE_P(original_fields_obj, imemo_fields)); -static int -generic_ivar_remove(VALUE obj, ID id, VALUE *valp) -{ - struct gen_ivtbl *ivtbl; - st_data_t key = (st_data_t)id; - st_data_t index; - st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj)); + if (fields_obj != original_fields_obj) { + switch (BUILTIN_TYPE(obj)) { + case T_DATA: + if (LIKELY(RTYPEDDATA_P(obj))) { + RB_OBJ_WRITE(obj, &RTYPEDDATA(obj)->fields_obj, fields_obj); + break; + } + goto generic_fields; + case T_STRUCT: + if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) { + RSTRUCT_SET_FIELDS_OBJ(obj, fields_obj); + break; + } + goto generic_fields; + default: + generic_fields: + { + RB_VM_LOCKING() { + st_insert(generic_fields_tbl_, (st_data_t)obj, (st_data_t)fields_obj); + } + RB_OBJ_WRITTEN(obj, original_fields_obj, fields_obj); - if (!iv_index_tbl) return 0; - if (!st_lookup(iv_index_tbl, key, &index)) return 0; - if (!gen_ivtbl_get(obj, &ivtbl)) return 0; + rb_execution_context_t *ec = GET_EC(); + if (ec->gen_fields_cache.fields_obj != fields_obj) { + ec->gen_fields_cache.obj = obj; + ec->gen_fields_cache.fields_obj = fields_obj; + } + } + } - if (index < ivtbl->numiv) { - if (ivtbl->ivptr[index] != Qundef) { - *valp = ivtbl->ivptr[index]; - ivtbl->ivptr[index] = Qundef; - return 1; - } + if (original_fields_obj) { + // Clear root shape to avoid triggering cleanup such as free_object_id. + rb_imemo_fields_clear(original_fields_obj); + } } - return 0; -} - -static void -gen_ivtbl_mark(const struct gen_ivtbl *ivtbl) -{ - uint32_t i; - for (i = 0; i < ivtbl->numiv; i++) { - rb_gc_mark(ivtbl->ivptr[i]); - } + RBASIC_SET_SHAPE_ID(obj, RBASIC_SHAPE_ID(fields_obj)); } void -rb_mark_generic_ivar(VALUE obj) +rb_obj_replace_fields(VALUE obj, VALUE fields_obj) { - struct gen_ivtbl *ivtbl; - - if (gen_ivtbl_get(obj, &ivtbl)) { - gen_ivtbl_mark(ivtbl); + RB_VM_LOCKING() { + VALUE original_fields_obj = rb_obj_fields_no_ractor_check(obj); + rb_obj_set_fields(obj, fields_obj, 0, original_fields_obj); } } -void -rb_free_generic_ivar(VALUE obj) +VALUE +rb_obj_field_get(VALUE obj, shape_id_t target_shape_id) { - st_data_t key = (st_data_t)obj; - struct gen_ivtbl *ivtbl; + RUBY_ASSERT(!SPECIAL_CONST_P(obj)); + RUBY_ASSERT(RSHAPE_TYPE_P(target_shape_id, SHAPE_IVAR) || RSHAPE_TYPE_P(target_shape_id, SHAPE_OBJ_ID)); - if (st_delete(generic_iv_tbl, &key, (st_data_t *)&ivtbl)) - xfree(ivtbl); + VALUE fields_obj; - if (generic_iv_tbl_compat) { - st_table *tbl; + switch (BUILTIN_TYPE(obj)) { + case T_CLASS: + case T_MODULE: + fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + break; + case T_OBJECT: + fields_obj = obj; + break; + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + fields_obj = obj; + break; + default: + fields_obj = rb_obj_fields(obj, RSHAPE_EDGE_NAME(target_shape_id)); + break; + } - if (st_delete(generic_iv_tbl_compat, &key, (st_data_t *)&tbl)) - st_free_table(tbl); + if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) { + st_table *fields_hash = rb_imemo_fields_complex_tbl(fields_obj); + VALUE value = Qundef; + st_lookup(fields_hash, RSHAPE_EDGE_NAME(target_shape_id), &value); + RUBY_ASSERT(!UNDEF_P(value)); + return value; } + + attr_index_t index = RSHAPE_INDEX(target_shape_id); + return rb_imemo_fields_ptr(fields_obj)[index]; } -RUBY_FUNC_EXPORTED size_t -rb_generic_ivar_memsize(VALUE obj) +VALUE +rb_ivar_lookup(VALUE obj, ID id, VALUE undef) { - struct gen_ivtbl *ivtbl; + if (SPECIAL_CONST_P(obj)) return undef; - if (gen_ivtbl_get(obj, &ivtbl)) - return gen_ivtbl_bytes(ivtbl->numiv); - return 0; -} + VALUE fields_obj; -static size_t -gen_ivtbl_count(const struct gen_ivtbl *ivtbl) -{ - uint32_t i; - size_t n = 0; + switch (BUILTIN_TYPE(obj)) { + case T_CLASS: + case T_MODULE: + { + VALUE val = rb_ivar_lookup(RCLASS_WRITABLE_FIELDS_OBJ(obj), id, undef); + if (val != undef && + rb_is_instance_id(id) && + UNLIKELY(!rb_ractor_main_p()) && + !rb_ractor_shareable_p(val)) { + rb_raise(rb_eRactorIsolationError, + "can not get unshareable values from instance variables of classes/modules from non-main Ractors (%"PRIsVALUE" from %"PRIsVALUE")", + rb_id2str(id), obj); + } + return val; + } + case T_IMEMO: + // Handled like T_OBJECT + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + fields_obj = obj; + break; + case T_OBJECT: + fields_obj = obj; + break; + default: + fields_obj = rb_obj_fields(obj, id); + break; + } + + if (!fields_obj) { + return undef; + } + + shape_id_t shape_id = RBASIC_SHAPE_ID(fields_obj); - for (i = 0; i < ivtbl->numiv; i++) { - if (ivtbl->ivptr[i] != Qundef) { - n++; - } + if (UNLIKELY(rb_shape_too_complex_p(shape_id))) { + st_table *iv_table = rb_imemo_fields_complex_tbl(fields_obj); + VALUE val; + if (rb_st_lookup(iv_table, (st_data_t)id, (st_data_t *)&val)) { + return val; + } + return undef; } - return n; + attr_index_t index = 0; + if (rb_shape_get_iv_index(shape_id, id, &index)) { + return rb_imemo_fields_ptr(fields_obj)[index]; + } + + return undef; } VALUE -rb_ivar_lookup(VALUE obj, ID id, VALUE undef) +rb_ivar_get(VALUE obj, ID id) +{ + VALUE iv = rb_ivar_lookup(obj, id, Qnil); + RB_DEBUG_COUNTER_INC(ivar_get_base); + return iv; +} + +VALUE +rb_ivar_get_at(VALUE obj, attr_index_t index, ID id) { - VALUE val, *ptr; - struct st_table *iv_index_tbl; - uint32_t len; - st_data_t index; + RUBY_ASSERT(rb_is_instance_id(id)); + // Used by JITs, but never for T_OBJECT. - if (SPECIAL_CONST_P(obj)) return undef; switch (BUILTIN_TYPE(obj)) { case T_OBJECT: - len = ROBJECT_NUMIV(obj); - ptr = ROBJECT_IVPTR(obj); - iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj); - if (!iv_index_tbl) break; - if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break; - if (len <= index) break; - val = ptr[index]; - if (val != Qundef) - return val; - break; + UNREACHABLE_RETURN(Qundef); case T_CLASS: case T_MODULE: - if (RCLASS_IV_TBL(obj) && - st_lookup(RCLASS_IV_TBL(obj), (st_data_t)id, &index)) - return (VALUE)index; - break; + { + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + VALUE val = rb_imemo_fields_ptr(fields_obj)[index]; + + if (UNLIKELY(!rb_ractor_main_p()) && !rb_ractor_shareable_p(val)) { + rb_raise(rb_eRactorIsolationError, + "can not get unshareable values from instance variables of classes/modules from non-main Ractors"); + } + + return val; + } default: - if (FL_TEST(obj, FL_EXIVAR)) - return generic_ivar_get(obj, id, undef); - break; + { + VALUE fields_obj = rb_obj_fields(obj, id); + return rb_imemo_fields_ptr(fields_obj)[index]; + } } - return undef; } VALUE -rb_ivar_get(VALUE obj, ID id) +rb_ivar_get_at_no_ractor_check(VALUE obj, attr_index_t index) { - VALUE iv = rb_ivar_lookup(obj, id, Qundef); - RB_DEBUG_COUNTER_INC(ivar_get_base); + // Used by JITs, but never for T_OBJECT. - if (iv == Qundef) { - if (RTEST(ruby_verbose)) - rb_warning("instance variable %"PRIsVALUE" not initialized", QUOTE_ID(id)); - iv = Qnil; + VALUE fields_obj; + switch (BUILTIN_TYPE(obj)) { + case T_OBJECT: + UNREACHABLE_RETURN(Qundef); + case T_CLASS: + case T_MODULE: + fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + break; + default: + fields_obj = rb_obj_fields_no_ractor_check(obj); + break; } - return iv; + return rb_imemo_fields_ptr(fields_obj)[index]; } VALUE @@ -1228,40 +1552,182 @@ rb_attr_get(VALUE obj, ID id) return rb_ivar_lookup(obj, id, Qnil); } -static VALUE -rb_ivar_delete(VALUE obj, ID id, VALUE undef) +void rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table); +static VALUE imemo_fields_complex_from_obj(VALUE owner, VALUE source_fields_obj, shape_id_t shape_id); + +static shape_id_t +obj_transition_too_complex(VALUE obj, st_table *table) { - VALUE val, *ptr; - struct st_table *iv_index_tbl; - uint32_t len; - st_data_t index; + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + shape_id_t shape_id = rb_shape_transition_complex(obj); - rb_check_frozen(obj); switch (BUILTIN_TYPE(obj)) { case T_OBJECT: - len = ROBJECT_NUMIV(obj); - ptr = ROBJECT_IVPTR(obj); - iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj); - if (!iv_index_tbl) break; - if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break; - if (len <= index) break; - val = ptr[index]; - ptr[index] = Qundef; - if (val != Qundef) - return val; - break; + { + VALUE *old_fields = NULL; + if (FL_TEST_RAW(obj, ROBJECT_HEAP)) { + old_fields = ROBJECT_FIELDS(obj); + } + else { + FL_SET_RAW(obj, ROBJECT_HEAP); + } + RBASIC_SET_SHAPE_ID(obj, shape_id); + ROBJECT_SET_FIELDS_HASH(obj, table); + if (old_fields) { + xfree(old_fields); + } + } + break; case T_CLASS: case T_MODULE: - if (RCLASS_IV_TBL(obj) && - st_delete(RCLASS_IV_TBL(obj), (st_data_t *)&id, &index)) - return (VALUE)index; - break; + case T_IMEMO: + UNREACHABLE; + break; default: - if (FL_TEST(obj, FL_EXIVAR)) - return generic_ivar_delete(obj, id, undef); - break; + { + VALUE fields_obj = rb_imemo_fields_new_complex_tbl(obj, table, RB_OBJ_SHAREABLE_P(obj)); + RBASIC_SET_SHAPE_ID(fields_obj, shape_id); + rb_obj_replace_fields(obj, fields_obj); + } } - return undef; + + return shape_id; +} + +// Copy all object fields, including ivars and internal object_id, etc +static shape_id_t +rb_evict_fields_to_hash(VALUE obj) +{ + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + + st_table *table = st_init_numtable_with_size(RSHAPE_LEN(RBASIC_SHAPE_ID(obj))); + rb_obj_copy_fields_to_hash_table(obj, table); + shape_id_t new_shape_id = obj_transition_too_complex(obj, table); + + RUBY_ASSERT(rb_shape_obj_too_complex_p(obj)); + return new_shape_id; +} + +void +rb_evict_ivars_to_hash(VALUE obj) +{ + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + + st_table *table = st_init_numtable_with_size(rb_ivar_count(obj)); + + // Evacuate all previous values from shape into id_table + rb_obj_copy_ivs_to_hash_table(obj, table); + obj_transition_too_complex(obj, table); + + RUBY_ASSERT(rb_shape_obj_too_complex_p(obj)); +} + +static VALUE +rb_ivar_delete(VALUE obj, ID id, VALUE undef) +{ + rb_check_frozen(obj); + + VALUE val = undef; + VALUE fields_obj; + bool concurrent = false; + int type = BUILTIN_TYPE(obj); + + switch(type) { + case T_CLASS: + case T_MODULE: + IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id); + + fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (rb_multi_ractor_p()) { + concurrent = true; + } + break; + case T_OBJECT: + fields_obj = obj; + break; + default: { + fields_obj = rb_obj_fields(obj, id); + break; + } + } + + if (!fields_obj) { + return undef; + } + + const VALUE original_fields_obj = fields_obj; + if (concurrent) { + fields_obj = rb_imemo_fields_clone(fields_obj); + } + + shape_id_t old_shape_id = RBASIC_SHAPE_ID(fields_obj); + shape_id_t removed_shape_id; + shape_id_t next_shape_id = rb_shape_transition_remove_ivar(fields_obj, id, &removed_shape_id); + + if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) { + if (UNLIKELY(!rb_shape_too_complex_p(old_shape_id))) { + if (type == T_OBJECT) { + rb_evict_fields_to_hash(obj); + } + else { + fields_obj = imemo_fields_complex_from_obj(obj, fields_obj, next_shape_id); + } + } + st_data_t key = id; + if (!st_delete(rb_imemo_fields_complex_tbl(fields_obj), &key, (st_data_t *)&val)) { + val = undef; + } + } + else { + if (next_shape_id == old_shape_id) { + return undef; + } + + RUBY_ASSERT(removed_shape_id != INVALID_SHAPE_ID); + RUBY_ASSERT(RSHAPE_LEN(next_shape_id) == RSHAPE_LEN(old_shape_id) - 1); + + VALUE *fields = rb_imemo_fields_ptr(fields_obj); + attr_index_t removed_index = RSHAPE_INDEX(removed_shape_id); + val = fields[removed_index]; + + attr_index_t new_fields_count = RSHAPE_LEN(next_shape_id); + if (new_fields_count) { + size_t trailing_fields = new_fields_count - removed_index; + + MEMMOVE(&fields[removed_index], &fields[removed_index + 1], VALUE, trailing_fields); + RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id); + + if (FL_TEST_RAW(fields_obj, OBJ_FIELD_HEAP) && rb_obj_embedded_size(new_fields_count) <= rb_gc_obj_slot_size(fields_obj)) { + // Re-embed objects when instances become small enough + // This is necessary because YJIT assumes that objects with the same shape + // have the same embeddedness for efficiency (avoid extra checks) + FL_UNSET_RAW(fields_obj, ROBJECT_HEAP); + MEMCPY(rb_imemo_fields_ptr(fields_obj), fields, VALUE, new_fields_count); + xfree(fields); + } + } + else { + fields_obj = 0; + rb_free_generic_ivar(obj); + } + } + + RBASIC_SET_SHAPE_ID(obj, next_shape_id); + if (fields_obj != original_fields_obj) { + switch (type) { + case T_OBJECT: + break; + case T_CLASS: + case T_MODULE: + RCLASS_WRITABLE_SET_FIELDS_OBJ(obj, fields_obj); + break; + default: + rb_obj_set_fields(obj, fields_obj, id, original_fields_obj); + break; + } + } + + return val; } VALUE @@ -1270,347 +1736,666 @@ rb_attr_delete(VALUE obj, ID id) return rb_ivar_delete(obj, id, Qnil); } -static st_table * -iv_index_tbl_make(VALUE obj) +void +rb_obj_init_too_complex(VALUE obj, st_table *table) +{ + // This method is meant to be called on newly allocated object. + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); + RUBY_ASSERT(rb_shape_canonical_p(RBASIC_SHAPE_ID(obj))); + RUBY_ASSERT(RSHAPE_LEN(RBASIC_SHAPE_ID(obj)) == 0); + + obj_transition_too_complex(obj, table); +} + +static int +imemo_fields_complex_from_obj_i(ID key, VALUE val, st_data_t arg) { - VALUE klass = rb_obj_class(obj); - st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(klass); + VALUE fields = (VALUE)arg; + st_table *table = rb_imemo_fields_complex_tbl(fields); + + RUBY_ASSERT(!st_lookup(table, (st_data_t)key, NULL)); + st_add_direct(table, (st_data_t)key, (st_data_t)val); + RB_OBJ_WRITTEN(fields, Qundef, val); - if (!iv_index_tbl) { - iv_index_tbl = RCLASS_IV_INDEX_TBL(klass) = st_init_numtable(); + return ST_CONTINUE; +} + +static VALUE +imemo_fields_complex_from_obj(VALUE owner, VALUE source_fields_obj, shape_id_t shape_id) +{ + attr_index_t len = source_fields_obj ? RSHAPE_LEN(RBASIC_SHAPE_ID(source_fields_obj)) : 0; + VALUE fields_obj = rb_imemo_fields_new_complex(owner, len + 1, RB_OBJ_SHAREABLE_P(owner)); + + rb_field_foreach(source_fields_obj, imemo_fields_complex_from_obj_i, (st_data_t)fields_obj, false); + RBASIC_SET_SHAPE_ID(fields_obj, shape_id); + + return fields_obj; +} + +static VALUE +imemo_fields_copy_capa(VALUE owner, VALUE source_fields_obj, attr_index_t new_size) +{ + VALUE fields_obj = rb_imemo_fields_new(owner, new_size, RB_OBJ_SHAREABLE_P(owner)); + if (source_fields_obj) { + attr_index_t fields_count = RSHAPE_LEN(RBASIC_SHAPE_ID(source_fields_obj)); + VALUE *fields = rb_imemo_fields_ptr(fields_obj); + MEMCPY(fields, rb_imemo_fields_ptr(source_fields_obj), VALUE, fields_count); + RBASIC_SET_SHAPE_ID(fields_obj, RBASIC_SHAPE_ID(source_fields_obj)); + for (attr_index_t i = 0; i < fields_count; i++) { + RB_OBJ_WRITTEN(fields_obj, Qundef, fields[i]); + } + } + return fields_obj; +} + +static VALUE +imemo_fields_set(VALUE owner, VALUE fields_obj, shape_id_t target_shape_id, ID field_name, VALUE val, bool concurrent) +{ + const VALUE original_fields_obj = fields_obj; + shape_id_t current_shape_id = fields_obj ? RBASIC_SHAPE_ID(fields_obj) : ROOT_SHAPE_ID; + + if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) { + if (rb_shape_too_complex_p(current_shape_id)) { + if (concurrent) { + // In multi-ractor case, we must always work on a copy because + // even if the field already exist, inserting in a st_table may + // cause a rebuild. + fields_obj = rb_imemo_fields_clone(fields_obj); + } + } + else { + fields_obj = imemo_fields_complex_from_obj(owner, original_fields_obj, target_shape_id); + current_shape_id = target_shape_id; + } + + st_table *table = rb_imemo_fields_complex_tbl(fields_obj); + + RUBY_ASSERT(field_name); + st_insert(table, (st_data_t)field_name, (st_data_t)val); + RB_OBJ_WRITTEN(fields_obj, Qundef, val); + RBASIC_SET_SHAPE_ID(fields_obj, target_shape_id); } + else { + attr_index_t index = RSHAPE_INDEX(target_shape_id); + if (concurrent || index >= RSHAPE_CAPACITY(current_shape_id)) { + fields_obj = imemo_fields_copy_capa(owner, original_fields_obj, RSHAPE_CAPACITY(target_shape_id)); + } - return iv_index_tbl; + VALUE *table = rb_imemo_fields_ptr(fields_obj); + RB_OBJ_WRITE(fields_obj, &table[index], val); + + if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) { + RBASIC_SET_SHAPE_ID(fields_obj, target_shape_id); + } + } + + return fields_obj; } -static void -iv_index_tbl_extend(struct ivar_update *ivup, ID id) +static attr_index_t +generic_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val) { - if (st_lookup(ivup->u.iv_index_tbl, (st_data_t)id, &ivup->index)) { - return; + if (!field_name) { + field_name = RSHAPE_EDGE_NAME(target_shape_id); + RUBY_ASSERT(field_name); } - if (ivup->u.iv_index_tbl->num_entries >= INT_MAX) { - rb_raise(rb_eArgError, "too many instance variables"); + + const VALUE original_fields_obj = rb_obj_fields(obj, field_name); + VALUE fields_obj = imemo_fields_set(obj, original_fields_obj, target_shape_id, field_name, val, false); + + rb_obj_set_fields(obj, fields_obj, field_name, original_fields_obj); + return rb_shape_too_complex_p(target_shape_id) ? ATTR_INDEX_NOT_SET : RSHAPE_INDEX(target_shape_id); +} + +static shape_id_t +generic_shape_ivar(VALUE obj, ID id, bool *new_ivar_out) +{ + bool new_ivar = false; + shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj); + shape_id_t target_shape_id = current_shape_id; + + if (!rb_shape_too_complex_p(current_shape_id)) { + if (!rb_shape_find_ivar(current_shape_id, id, &target_shape_id)) { + if (RSHAPE_LEN(current_shape_id) >= SHAPE_MAX_FIELDS) { + rb_raise(rb_eArgError, "too many instance variables"); + } + + new_ivar = true; + target_shape_id = rb_shape_transition_add_ivar(obj, id); + } } - ivup->index = (st_data_t)ivup->u.iv_index_tbl->num_entries; - st_add_direct(ivup->u.iv_index_tbl, (st_data_t)id, ivup->index); - ivup->iv_extended = 1; + + *new_ivar_out = new_ivar; + return target_shape_id; } -static void +static attr_index_t generic_ivar_set(VALUE obj, ID id, VALUE val) { - struct ivar_update ivup; + bool dontcare; + shape_id_t target_shape_id = generic_shape_ivar(obj, id, &dontcare); + return generic_field_set(obj, target_shape_id, id, val); +} - ivup.iv_extended = 0; - ivup.u.iv_index_tbl = iv_index_tbl_make(obj); - iv_index_tbl_extend(&ivup, id); - st_update(generic_iv_tbl, (st_data_t)obj, generic_ivar_update, - (st_data_t)&ivup); +void +rb_ensure_iv_list_size(VALUE obj, uint32_t current_len, uint32_t new_capacity) +{ + RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj)); - ivup.u.ivtbl->ivptr[ivup.index] = val; + if (FL_TEST_RAW(obj, ROBJECT_HEAP)) { + REALLOC_N(ROBJECT(obj)->as.heap.fields, VALUE, new_capacity); + } + else { + VALUE *ptr = ROBJECT_FIELDS(obj); + VALUE *newptr = ALLOC_N(VALUE, new_capacity); + MEMCPY(newptr, ptr, VALUE, current_len); + FL_SET_RAW(obj, ROBJECT_HEAP); + ROBJECT(obj)->as.heap.fields = newptr; + } +} + +static int +rb_obj_copy_ivs_to_hash_table_i(ID key, VALUE val, st_data_t arg) +{ + RUBY_ASSERT(!st_lookup((st_table *)arg, (st_data_t)key, NULL)); + + st_add_direct((st_table *)arg, (st_data_t)key, (st_data_t)val); + return ST_CONTINUE; +} + +void +rb_obj_copy_ivs_to_hash_table(VALUE obj, st_table *table) +{ + rb_ivar_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table); +} - RB_OBJ_WRITTEN(obj, Qundef, val); +void +rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table) +{ + rb_field_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table, false); } +static attr_index_t +obj_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val) +{ + shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj); + + if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) { + if (UNLIKELY(!rb_shape_too_complex_p(current_shape_id))) { + current_shape_id = rb_evict_fields_to_hash(obj); + } + + if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) { + RBASIC_SET_SHAPE_ID(obj, target_shape_id); + } + + if (!field_name) { + field_name = RSHAPE_EDGE_NAME(target_shape_id); + RUBY_ASSERT(field_name); + } + + st_insert(ROBJECT_FIELDS_HASH(obj), (st_data_t)field_name, (st_data_t)val); + RB_OBJ_WRITTEN(obj, Qundef, val); + + return ATTR_INDEX_NOT_SET; + } + else { + attr_index_t index = RSHAPE_INDEX(target_shape_id); + + if (index >= RSHAPE_LEN(current_shape_id)) { + if (UNLIKELY(index >= RSHAPE_CAPACITY(current_shape_id))) { + rb_ensure_iv_list_size(obj, RSHAPE_CAPACITY(current_shape_id), RSHAPE_CAPACITY(target_shape_id)); + } + RBASIC_SET_SHAPE_ID(obj, target_shape_id); + } + + RB_OBJ_WRITE(obj, &ROBJECT_FIELDS(obj)[index], val); + + return index; + } +} + +static attr_index_t +obj_ivar_set(VALUE obj, ID id, VALUE val) +{ + bool dontcare; + shape_id_t target_shape_id = generic_shape_ivar(obj, id, &dontcare); + return obj_field_set(obj, target_shape_id, id, val); +} + +/* Set the instance variable +val+ on object +obj+ at ivar name +id+. + * This function only works with T_OBJECT objects, so make sure + * +obj+ is of type T_OBJECT before using this function. + */ VALUE -rb_ivar_set(VALUE obj, ID id, VALUE val) +rb_vm_set_ivar_id(VALUE obj, ID id, VALUE val) +{ + rb_check_frozen(obj); + obj_ivar_set(obj, id, val); + return val; +} + +void rb_obj_freeze_inline(VALUE x) { - struct ivar_update ivup; - uint32_t i, len; + if (RB_FL_ABLE(x)) { + RB_FL_SET_RAW(x, RUBY_FL_FREEZE); + if (TYPE(x) == T_STRING) { + RB_FL_UNSET_RAW(x, FL_USER2 | FL_USER3); // STR_CHILLED + } + + RB_SET_SHAPE_ID(x, rb_shape_transition_frozen(x)); + + if (RBASIC_CLASS(x)) { + rb_freeze_singleton_class(x); + } + } +} + +static attr_index_t class_ivar_set(VALUE obj, ID id, VALUE val, bool *new_ivar); +static attr_index_t +ivar_set(VALUE obj, ID id, VALUE val) +{ RB_DEBUG_COUNTER_INC(ivar_set_base); + switch (BUILTIN_TYPE(obj)) { + case T_OBJECT: + return obj_ivar_set(obj, id, val); + case T_CLASS: + case T_MODULE: + { + IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id); + bool dontcare; + return class_ivar_set(obj, id, val, &dontcare); + } + default: + return generic_ivar_set(obj, id, val); + } +} + +VALUE +rb_ivar_set(VALUE obj, ID id, VALUE val) +{ rb_check_frozen(obj); + ivar_set(obj, id, val); + return val; +} + +attr_index_t +rb_ivar_set_index(VALUE obj, ID id, VALUE val) +{ + return ivar_set(obj, id, val); +} + +void +rb_ivar_set_internal(VALUE obj, ID id, VALUE val) +{ + // should be internal instance variable name (no @ prefix) + VM_ASSERT(!rb_is_instance_id(id)); + ivar_set(obj, id, val); +} + +attr_index_t +rb_obj_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val) +{ switch (BUILTIN_TYPE(obj)) { case T_OBJECT: - ivup.iv_extended = 0; - ivup.u.iv_index_tbl = iv_index_tbl_make(obj); - iv_index_tbl_extend(&ivup, id); - len = ROBJECT_NUMIV(obj); - if (len <= ivup.index) { - VALUE *ptr = ROBJECT_IVPTR(obj); - if (ivup.index < ROBJECT_EMBED_LEN_MAX) { - RBASIC(obj)->flags |= ROBJECT_EMBED; - ptr = ROBJECT(obj)->as.ary; - for (i = 0; i < ROBJECT_EMBED_LEN_MAX; i++) { - ptr[i] = Qundef; - } - } - else { - VALUE *newptr; - uint32_t newsize = iv_index_tbl_newsize(&ivup); - - if (RBASIC(obj)->flags & ROBJECT_EMBED) { - newptr = ALLOC_N(VALUE, newsize); - MEMCPY(newptr, ptr, VALUE, len); - RBASIC(obj)->flags &= ~ROBJECT_EMBED; - ROBJECT(obj)->as.heap.ivptr = newptr; - } - else { - REALLOC_N(ROBJECT(obj)->as.heap.ivptr, VALUE, newsize); - newptr = ROBJECT(obj)->as.heap.ivptr; - } - for (; len < newsize; len++) - newptr[len] = Qundef; - ROBJECT(obj)->as.heap.numiv = newsize; - ROBJECT(obj)->as.heap.iv_index_tbl = ivup.u.iv_index_tbl; - } - } - RB_OBJ_WRITE(obj, &ROBJECT_IVPTR(obj)[ivup.index], val); - break; + return obj_field_set(obj, target_shape_id, field_name, val); case T_CLASS: case T_MODULE: - if (!RCLASS_IV_TBL(obj)) RCLASS_IV_TBL(obj) = st_init_numtable(); - rb_class_ivar_set(obj, id, val); + // The only field is object_id and T_CLASS handle it differently. + rb_bug("Unreachable"); break; default: - generic_ivar_set(obj, id, val); - break; + return generic_field_set(obj, target_shape_id, field_name, val); + } +} + +static VALUE +ivar_defined0(VALUE obj, ID id) +{ + attr_index_t index; + + if (rb_shape_obj_too_complex_p(obj)) { + VALUE idx; + st_table *table = NULL; + switch (BUILTIN_TYPE(obj)) { + case T_CLASS: + case T_MODULE: + rb_bug("Unreachable"); + break; + + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + table = rb_imemo_fields_complex_tbl(obj); + break; + + case T_OBJECT: + table = ROBJECT_FIELDS_HASH(obj); + break; + + default: { + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); // defined? doesn't require ractor checks + table = rb_imemo_fields_complex_tbl(fields_obj); + } + } + + if (!table || !rb_st_lookup(table, id, &idx)) { + return Qfalse; + } + + return Qtrue; + } + else { + return RBOOL(rb_shape_get_iv_index(RBASIC_SHAPE_ID(obj), id, &index)); } - return val; } VALUE rb_ivar_defined(VALUE obj, ID id) { - VALUE val; - struct st_table *iv_index_tbl; - st_data_t index; - if (SPECIAL_CONST_P(obj)) return Qfalse; + + VALUE defined = Qfalse; switch (BUILTIN_TYPE(obj)) { - case T_OBJECT: - iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj); - if (!iv_index_tbl) break; - if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break; - if (ROBJECT_NUMIV(obj) <= index) break; - val = ROBJECT_IVPTR(obj)[index]; - if (val != Qundef) - return Qtrue; - break; case T_CLASS: case T_MODULE: - if (RCLASS_IV_TBL(obj) && st_lookup(RCLASS_IV_TBL(obj), (st_data_t)id, 0)) - return Qtrue; - break; + { + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (fields_obj) { + defined = ivar_defined0(fields_obj, id); + } + } + break; default: - if (FL_TEST(obj, FL_EXIVAR)) - return generic_ivar_defined(obj, id); - break; + defined = ivar_defined0(obj, id); + break; } - return Qfalse; + return defined; } -struct obj_ivar_tag { +struct iv_itr_data { VALUE obj; - int (*func)(ID key, VALUE val, st_data_t arg); + struct gen_fields_tbl *fields_tbl; st_data_t arg; + rb_ivar_foreach_callback_func *func; + VALUE *fields; + bool ivar_only; }; static int -obj_ivar_i(st_data_t key, st_data_t index, st_data_t arg) +iterate_over_shapes_callback(shape_id_t shape_id, void *data) { - struct obj_ivar_tag *data = (struct obj_ivar_tag *)arg; - if (index < ROBJECT_NUMIV(data->obj)) { - VALUE val = ROBJECT_IVPTR(data->obj)[index]; - if (val != Qundef) { - return (data->func)((ID)key, val, data->arg); - } - } - return ST_CONTINUE; -} + struct iv_itr_data *itr_data = data; -static void -obj_ivar_each(VALUE obj, int (*func)(ANYARGS), st_data_t arg) -{ - st_table *tbl; - struct obj_ivar_tag data; + if (itr_data->ivar_only && !RSHAPE_TYPE_P(shape_id, SHAPE_IVAR)) { + return ST_CONTINUE; + } - tbl = ROBJECT_IV_INDEX_TBL(obj); - if (!tbl) - return; + VALUE *fields; + switch (BUILTIN_TYPE(itr_data->obj)) { + case T_OBJECT: + RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj)); + fields = ROBJECT_FIELDS(itr_data->obj); + break; + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(itr_data->obj, imemo_fields)); + RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj)); - data.obj = obj; - data.func = (int (*)(ID key, VALUE val, st_data_t arg))func; - data.arg = arg; + fields = rb_imemo_fields_ptr(itr_data->obj); + break; + default: + rb_bug("Unreachable"); + } - st_foreach_safe(tbl, obj_ivar_i, (st_data_t)&data); + VALUE val = fields[RSHAPE_INDEX(shape_id)]; + return itr_data->func(RSHAPE_EDGE_NAME(shape_id), val, itr_data->arg); } -struct gen_ivar_tag { - struct gen_ivtbl *ivtbl; - int (*func)(ID key, VALUE val, st_data_t arg); - st_data_t arg; -}; +/* + * Returns a flag to stop iterating depending on the result of +callback+. + */ +static void +iterate_over_shapes(shape_id_t shape_id, rb_ivar_foreach_callback_func *callback, struct iv_itr_data *itr_data) +{ + rb_shape_foreach_field(shape_id, iterate_over_shapes_callback, itr_data); +} static int -gen_ivar_each_i(st_data_t key, st_data_t index, st_data_t data) +each_hash_iv(st_data_t id, st_data_t val, st_data_t data) { - struct gen_ivar_tag *arg = (struct gen_ivar_tag *)data; - - if (index < arg->ivtbl->numiv) { - VALUE val = arg->ivtbl->ivptr[index]; - if (val != Qundef) { - return (arg->func)((ID)key, val, arg->arg); - } + struct iv_itr_data * itr_data = (struct iv_itr_data *)data; + rb_ivar_foreach_callback_func *callback = itr_data->func; + if (is_internal_id((ID)id)) { + return ST_CONTINUE; } - return ST_CONTINUE; + return callback((ID)id, (VALUE)val, itr_data->arg); } static void -gen_ivar_each(VALUE obj, int (*func)(ANYARGS), st_data_t arg) +obj_fields_each(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only) { - struct gen_ivar_tag data; - st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj)); - - if (!iv_index_tbl) return; - if (!gen_ivtbl_get(obj, &data.ivtbl)) return; + struct iv_itr_data itr_data = { + .obj = obj, + .arg = arg, + .func = func, + .ivar_only = ivar_only, + }; - data.func = (int (*)(ID key, VALUE val, st_data_t arg))func; - data.arg = arg; - - st_foreach_safe(iv_index_tbl, gen_ivar_each_i, (st_data_t)&data); + shape_id_t shape_id = RBASIC_SHAPE_ID(obj); + if (rb_shape_too_complex_p(shape_id)) { + rb_st_foreach(ROBJECT_FIELDS_HASH(obj), each_hash_iv, (st_data_t)&itr_data); + } + else { + itr_data.fields = ROBJECT_FIELDS(obj); + iterate_over_shapes(shape_id, func, &itr_data); + } } -struct givar_copy { - VALUE obj; - st_table *iv_index_tbl; - struct gen_ivtbl *ivtbl; -}; - -static int -gen_ivar_copy(ID id, VALUE val, st_data_t arg) +static void +imemo_fields_each(VALUE fields_obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only) { - struct givar_copy *c = (struct givar_copy *)arg; - struct ivar_update ivup; - - ivup.iv_extended = 0; - ivup.u.iv_index_tbl = c->iv_index_tbl; - iv_index_tbl_extend(&ivup, id); - if (ivup.index >= c->ivtbl->numiv) { - uint32_t newsize = iv_index_tbl_newsize(&ivup); - c->ivtbl = gen_ivtbl_resize(c->ivtbl, newsize); - } - c->ivtbl->ivptr[ivup.index] = val; + IMEMO_TYPE_P(fields_obj, imemo_fields); - RB_OBJ_WRITTEN(c->obj, Qundef, val); + struct iv_itr_data itr_data = { + .obj = fields_obj, + .arg = arg, + .func = func, + .ivar_only = ivar_only, + }; - return ST_CONTINUE; + shape_id_t shape_id = RBASIC_SHAPE_ID(fields_obj); + if (rb_shape_too_complex_p(shape_id)) { + rb_st_foreach(rb_imemo_fields_complex_tbl(fields_obj), each_hash_iv, (st_data_t)&itr_data); + } + else { + itr_data.fields = rb_imemo_fields_ptr(fields_obj); + iterate_over_shapes(shape_id, func, &itr_data); + } } void -rb_copy_generic_ivar(VALUE clone, VALUE obj) +rb_copy_generic_ivar(VALUE dest, VALUE obj) { - struct gen_ivtbl *ivtbl; + VALUE new_fields_obj; - rb_check_frozen(clone); + rb_check_frozen(dest); - if (!FL_TEST(obj, FL_EXIVAR)) { - clear: - if (FL_TEST(clone, FL_EXIVAR)) { - rb_free_generic_ivar(clone); - FL_UNSET(clone, FL_EXIVAR); - } + if (!rb_obj_gen_fields_p(obj)) { return; } - if (gen_ivtbl_get(obj, &ivtbl)) { - struct givar_copy c; - uint32_t i; - if (gen_ivtbl_count(ivtbl) == 0) - goto clear; + shape_id_t src_shape_id = rb_obj_shape_id(obj); + + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); + if (fields_obj) { + unsigned long src_num_ivs = rb_ivar_count(fields_obj); + if (!src_num_ivs) { + goto clear; + } + + if (rb_shape_too_complex_p(src_shape_id)) { + rb_shape_copy_complex_ivars(dest, obj, src_shape_id, rb_imemo_fields_complex_tbl(fields_obj)); + return; + } + + shape_id_t dest_shape_id = src_shape_id; + shape_id_t initial_shape_id = rb_obj_shape_id(dest); - if (gen_ivtbl_get(clone, &c.ivtbl)) { - for (i = 0; i < c.ivtbl->numiv; i++) - c.ivtbl->ivptr[i] = Qundef; - } - else { - c.ivtbl = gen_ivtbl_resize(0, ivtbl->numiv); - FL_SET(clone, FL_EXIVAR); - } + if (!rb_shape_canonical_p(src_shape_id)) { + RUBY_ASSERT(RSHAPE_TYPE_P(initial_shape_id, SHAPE_ROOT)); + + dest_shape_id = rb_shape_rebuild(initial_shape_id, src_shape_id); + if (UNLIKELY(rb_shape_too_complex_p(dest_shape_id))) { + st_table *table = rb_st_init_numtable_with_size(src_num_ivs); + rb_obj_copy_ivs_to_hash_table(obj, table); + rb_obj_init_too_complex(dest, table); + return; + } + } + + if (!RSHAPE_LEN(dest_shape_id)) { + RBASIC_SET_SHAPE_ID(dest, dest_shape_id); + return; + } - c.iv_index_tbl = iv_index_tbl_make(clone); - c.obj = clone; - gen_ivar_each(obj, gen_ivar_copy, (st_data_t)&c); - /* - * c.ivtbl may change in gen_ivar_copy due to realloc, - * no need to free - */ - st_insert(generic_iv_tbl, (st_data_t)clone, (st_data_t)c.ivtbl); + new_fields_obj = rb_imemo_fields_new(dest, RSHAPE_CAPACITY(dest_shape_id), RB_OBJ_SHAREABLE_P(dest)); + VALUE *src_buf = rb_imemo_fields_ptr(fields_obj); + VALUE *dest_buf = rb_imemo_fields_ptr(new_fields_obj); + rb_shape_copy_fields(new_fields_obj, dest_buf, dest_shape_id, src_buf, src_shape_id); + RBASIC_SET_SHAPE_ID(new_fields_obj, dest_shape_id); + + rb_obj_replace_fields(dest, new_fields_obj); + } + return; + + clear: + rb_free_generic_ivar(dest); +} + +void +rb_replace_generic_ivar(VALUE clone, VALUE obj) +{ + RB_VM_LOCKING() { + st_data_t fields_tbl, obj_data = (st_data_t)obj; + if (st_delete(generic_fields_tbl_, &obj_data, &fields_tbl)) { + st_insert(generic_fields_tbl_, (st_data_t)clone, fields_tbl); + RB_OBJ_WRITTEN(clone, Qundef, fields_tbl); + } + else { + rb_bug("unreachable"); + } } } void -rb_ivar_foreach(VALUE obj, int (*func)(ANYARGS), st_data_t arg) +rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only) { if (SPECIAL_CONST_P(obj)) return; switch (BUILTIN_TYPE(obj)) { + case T_IMEMO: + if (IMEMO_TYPE_P(obj, imemo_fields)) { + imemo_fields_each(obj, func, arg, ivar_only); + } + break; case T_OBJECT: - obj_ivar_each(obj, func, arg); - break; + obj_fields_each(obj, func, arg, ivar_only); + break; case T_CLASS: case T_MODULE: - if (RCLASS_IV_TBL(obj)) { - st_foreach_safe(RCLASS_IV_TBL(obj), func, arg); - } - break; + { + IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(0); + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (fields_obj) { + imemo_fields_each(fields_obj, func, arg, ivar_only); + } + } + break; default: - if (FL_TEST(obj, FL_EXIVAR)) { - gen_ivar_each(obj, func, arg); - } - break; + { + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); + if (fields_obj) { + imemo_fields_each(fields_obj, func, arg, ivar_only); + } + } + break; } } +void +rb_ivar_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg) +{ + rb_field_foreach(obj, func, arg, true); +} + st_index_t rb_ivar_count(VALUE obj) { - st_table *tbl; - if (SPECIAL_CONST_P(obj)) return 0; + st_index_t iv_count = 0; switch (BUILTIN_TYPE(obj)) { case T_OBJECT: - if ((tbl = ROBJECT_IV_INDEX_TBL(obj)) != 0) { - st_index_t i, count, num = ROBJECT_NUMIV(obj); - const VALUE *const ivptr = ROBJECT_IVPTR(obj); - for (i = count = 0; i < num; ++i) { - if (ivptr[i] != Qundef) { - count++; - } - } - return count; - } - break; + iv_count = ROBJECT_FIELDS_COUNT(obj); + break; + case T_CLASS: case T_MODULE: - if ((tbl = RCLASS_IV_TBL(obj)) != 0) { - return tbl->num_entries; - } - break; + { + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + if (!fields_obj) { + return 0; + } + if (rb_shape_obj_too_complex_p(fields_obj)) { + iv_count = rb_st_table_size(rb_imemo_fields_complex_tbl(fields_obj)); + } + else { + iv_count = RBASIC_FIELDS_COUNT(fields_obj); + } + } + break; + + case T_IMEMO: + RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields)); + + if (rb_shape_obj_too_complex_p(obj)) { + iv_count = rb_st_table_size(rb_imemo_fields_complex_tbl(obj)); + } + else { + iv_count = RBASIC_FIELDS_COUNT(obj); + } + break; + default: - if (FL_TEST(obj, FL_EXIVAR)) { - struct gen_ivtbl *ivtbl; + { + VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); + if (fields_obj) { + if (rb_shape_obj_too_complex_p(fields_obj)) { + rb_st_table_size(rb_imemo_fields_complex_tbl(fields_obj)); + } + else { + iv_count = RBASIC_FIELDS_COUNT(obj); + } + } + } + break; + } - if (gen_ivtbl_get(obj, &ivtbl)) { - return gen_ivtbl_count(ivtbl); - } - } - break; + if (rb_shape_obj_has_id(obj)) { + iv_count--; } - return 0; + + return iv_count; } static int -ivar_i(st_data_t k, st_data_t v, st_data_t a) +ivar_i(ID key, VALUE v, st_data_t a) { - ID key = (ID)k; VALUE ary = (VALUE)a; if (rb_is_instance_id(key)) { - rb_ary_push(ary, ID2SYM(key)); + rb_ary_push(ary, ID2SYM(key)); } return ST_CONTINUE; } @@ -1645,20 +2430,20 @@ rb_obj_instance_variables(VALUE obj) #define rb_is_constant_id rb_is_const_id #define rb_is_constant_name rb_is_const_name #define id_for_var(obj, name, part, type) \ - id_for_var_message(obj, name, type, "`%1$s' is not allowed as "#part" "#type" variable name") + id_for_var_message(obj, name, type, "'%1$s' is not allowed as "#part" "#type" variable name") #define id_for_var_message(obj, name, type, message) \ check_id_type(obj, &(name), rb_is_##type##_id, rb_is_##type##_name, message, strlen(message)) static ID check_id_type(VALUE obj, VALUE *pname, - int (*valid_id_p)(ID), int (*valid_name_p)(VALUE), - const char *message, size_t message_len) + int (*valid_id_p)(ID), int (*valid_name_p)(VALUE), + const char *message, size_t message_len) { ID id = rb_check_id(pname); VALUE name = *pname; if (id ? !valid_id_p(id) : !valid_name_p(name)) { - rb_name_err_raise_str(rb_fstring_new(message, message_len), - obj, name); + rb_name_err_raise_str(rb_fstring_new(message, message_len), + obj, name); } return id; } @@ -1666,9 +2451,10 @@ check_id_type(VALUE obj, VALUE *pname, /* * call-seq: * obj.remove_instance_variable(symbol) -> obj + * obj.remove_instance_variable(string) -> obj * * Removes the named instance variable from <i>obj</i>, returning that - * variable's value. + * variable's value. The name can be passed as a symbol or as a string. * * class Dummy * attr_reader :var @@ -1688,49 +2474,21 @@ check_id_type(VALUE obj, VALUE *pname, VALUE rb_obj_remove_instance_variable(VALUE obj, VALUE name) { - VALUE val = Qnil; const ID id = id_for_var(obj, name, an, instance); - st_data_t n, v; - struct st_table *iv_index_tbl; - st_data_t index; + // Frozen check comes here because it's expected that we raise a + // NameError (from the id_for_var check) before we raise a FrozenError rb_check_frozen(obj); - if (!id) { - goto not_defined; - } - switch (BUILTIN_TYPE(obj)) { - case T_OBJECT: - iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj); - if (!iv_index_tbl) break; - if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break; - if (ROBJECT_NUMIV(obj) <= index) break; - val = ROBJECT_IVPTR(obj)[index]; - if (val != Qundef) { - ROBJECT_IVPTR(obj)[index] = Qundef; - return val; - } - break; - case T_CLASS: - case T_MODULE: - n = id; - if (RCLASS_IV_TBL(obj) && st_delete(RCLASS_IV_TBL(obj), &n, &v)) { - return (VALUE)v; - } - break; - default: - if (FL_TEST(obj, FL_EXIVAR)) { - if (generic_ivar_remove(obj, id, &val)) { - return val; - } - } - break; + if (id) { + VALUE val = rb_ivar_delete(obj, id, Qundef); + + if (!UNDEF_P(val)) return val; } - not_defined: rb_name_err_raise("instance variable %1$s not defined", - obj, name); - UNREACHABLE; + obj, name); + UNREACHABLE_RETURN(Qnil); } NORETURN(static void uninitialized_constant(VALUE, VALUE)); @@ -1738,17 +2496,17 @@ static void uninitialized_constant(VALUE klass, VALUE name) { if (klass && rb_class_real(klass) != rb_cObject) - rb_name_err_raise("uninitialized constant %2$s::%1$s", - klass, name); + rb_name_err_raise("uninitialized constant %2$s::%1$s", + klass, name); else - rb_name_err_raise("uninitialized constant %1$s", - klass, name); + rb_name_err_raise("uninitialized constant %1$s", + klass, name); } VALUE rb_const_missing(VALUE klass, VALUE name) { - VALUE value = rb_funcallv(klass, rb_intern("const_missing"), 1, &name); + VALUE value = rb_funcallv(klass, idConst_missing, 1, &name); rb_vm_inc_const_missing_count(); return value; } @@ -1760,8 +2518,7 @@ rb_const_missing(VALUE klass, VALUE name) * * Invoked when a reference is made to an undefined constant in * <i>mod</i>. It is passed a symbol for the undefined constant, and - * returns a value to be used for that constant. The - * following code is an example of the same: + * returns a value to be used for that constant. For example, consider: * * def Foo.const_missing(name) * name # return the constant name as Symbol @@ -1769,23 +2526,28 @@ rb_const_missing(VALUE klass, VALUE name) * * Foo::UNDEFINED_CONST #=> :UNDEFINED_CONST: symbol returned * - * In the next example when a reference is made to an undefined constant, - * it attempts to load a file whose name is the lowercase version of the - * constant (thus class <code>Fred</code> is assumed to be in file - * <code>fred.rb</code>). If found, it returns the loaded class. It - * therefore implements an autoload feature similar to Kernel#autoload and - * Module#autoload. + * As the example above shows, +const_missing+ is not required to create the + * missing constant in <i>mod</i>, though that is often a side-effect. The + * caller gets its return value when triggered. If the constant is also defined, + * further lookups won't hit +const_missing+ and will return the value stored in + * the constant as usual. Otherwise, +const_missing+ will be invoked again. + * + * In the next example, when a reference is made to an undefined constant, + * +const_missing+ attempts to load a file whose path is the lowercase version + * of the constant name (thus class <code>Fred</code> is assumed to be in file + * <code>fred.rb</code>). If defined as a side-effect of loading the file, the + * method returns the value stored in the constant. This implements an autoload + * feature similar to Kernel#autoload and Module#autoload, though it differs in + * important ways. * * def Object.const_missing(name) * @looked_for ||= {} * str_name = name.to_s - * raise "Class not found: #{name}" if @looked_for[str_name] + * raise "Constant not found: #{name}" if @looked_for[str_name] * @looked_for[str_name] = 1 * file = str_name.downcase * require file - * klass = const_get(name) - * return klass if klass - * raise "Class not found: #{name}" + * const_get(name, false) * end * */ @@ -1793,39 +2555,51 @@ rb_const_missing(VALUE klass, VALUE name) VALUE rb_mod_const_missing(VALUE klass, VALUE name) { + rb_execution_context_t *ec = GET_EC(); + VALUE ref = ec->private_const_reference; rb_vm_pop_cfunc_frame(); + if (ref) { + ec->private_const_reference = 0; + rb_name_err_raise("private constant %2$s::%1$s referenced", ref, name); + } uninitialized_constant(klass, name); - UNREACHABLE; + UNREACHABLE_RETURN(Qnil); } static void -autoload_mark(void *ptr) +autoload_table_mark(void *ptr) { - rb_mark_tbl((st_table *)ptr); + rb_mark_tbl_no_pin((st_table *)ptr); } static void -autoload_free(void *ptr) +autoload_table_free(void *ptr) { st_free_table((st_table *)ptr); } static size_t -autoload_memsize(const void *ptr) +autoload_table_memsize(const void *ptr) { const st_table *tbl = ptr; return st_memsize(tbl); } -static const rb_data_type_t autoload_data_type = { - "autoload", - {autoload_mark, autoload_free, autoload_memsize,}, - 0, 0, RUBY_TYPED_FREE_IMMEDIATELY +static void +autoload_table_compact(void *ptr) +{ + rb_gc_ref_update_table_values_only((st_table *)ptr); +} + +static const rb_data_type_t autoload_table_type = { + "autoload_table", + {autoload_table_mark, autoload_table_free, autoload_table_memsize, autoload_table_compact,}, + 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED }; #define check_autoload_table(av) \ - (struct st_table *)rb_check_typeddata((av), &autoload_data_type) + (struct st_table *)rb_check_typeddata((av), &autoload_table_type) static VALUE autoload_data(VALUE mod, ID id) @@ -1833,219 +2607,489 @@ autoload_data(VALUE mod, ID id) struct st_table *tbl; st_data_t val; - if (!st_lookup(RCLASS_IV_TBL(mod), autoload, &val) || - !(tbl = check_autoload_table((VALUE)val)) || - !st_lookup(tbl, (st_data_t)id, &val)) { - return 0; + // If we are called with a non-origin ICLASS, fetch the autoload data from + // the original module. + if (RB_TYPE_P(mod, T_ICLASS)) { + if (RICLASS_IS_ORIGIN_P(mod)) { + return 0; + } + else { + mod = RBASIC(mod)->klass; + } } + + RUBY_ASSERT(RB_TYPE_P(mod, T_CLASS) || RB_TYPE_P(mod, T_MODULE)); + + // Look up the instance variable table for `autoload`, then index into that table with the given constant name `id`. + + VALUE tbl_value = rb_ivar_lookup(mod, autoload, Qfalse); + if (!RTEST(tbl_value) || !(tbl = check_autoload_table(tbl_value)) || !st_lookup(tbl, (st_data_t)id, &val)) { + return 0; + } + return (VALUE)val; } -/* always on stack, no need to mark */ -struct autoload_state { - struct autoload_data_i *ele; - VALUE mod; - VALUE result; - ID id; - VALUE thread; - union { - struct list_node node; - struct list_head head; - } waitq; +// Every autoload constant has exactly one instance of autoload_const, stored in `autoload_features`. Since multiple autoload constants can refer to the same file, every `autoload_const` refers to a de-duplicated `autoload_data`. +struct autoload_const { + // The linked list node of all constants which are loaded by the related autoload feature. + struct ccan_list_node cnode; /* <=> autoload_data.constants */ + + // The shared "autoload_data" if multiple constants are defined from the same feature. + VALUE autoload_data_value; + + // The box object when the autoload is called in a user box + // Otherwise, Qnil means the root box + VALUE box_value; + + // The module we are loading a constant into. + VALUE module; + + // The name of the constant we are loading. + ID name; + + // The value of the constant (after it's loaded). + VALUE value; + + // The constant entry flags which need to be re-applied after autoloading the feature. + rb_const_flag_t flag; + + // The source file and line number that defined this constant (different from feature path). + VALUE file; + int line; }; -struct autoload_data_i { +// Each `autoload_data` uniquely represents a specific feature which can be loaded, and a list of constants which it is able to define. We use a mutex to coordinate multiple threads trying to load the same feature. +struct autoload_data { + // The feature path to require to load this constant. VALUE feature; - int safe_level; - rb_const_flag_t flag; - VALUE value; - struct autoload_state *state; /* points to on-stack struct */ + + // The mutex which is protecting autoloading this feature. + VALUE mutex; + + // The process fork serial number since the autoload mutex will become invalid on fork. rb_serial_t fork_gen; + + // The linked list of all constants that are going to be loaded by this autoload. + struct ccan_list_head constants; /* <=> autoload_const.cnode */ }; static void -autoload_i_mark(void *ptr) +autoload_data_mark_and_move(void *ptr) { - struct autoload_data_i *p = ptr; - rb_gc_mark(p->feature); - rb_gc_mark(p->value); + struct autoload_data *p = ptr; + + rb_gc_mark_and_move(&p->feature); + rb_gc_mark_and_move(&p->mutex); +} + +static void +autoload_data_free(void *ptr) +{ + struct autoload_data *p = ptr; + + struct autoload_const *autoload_const, *next; + ccan_list_for_each_safe(&p->constants, autoload_const, next, cnode) { + ccan_list_del_init(&autoload_const->cnode); + } + + ruby_xfree(p); +} + +static size_t +autoload_data_memsize(const void *ptr) +{ + return sizeof(struct autoload_data); +} + +static const rb_data_type_t autoload_data_type = { + "autoload_data", + {autoload_data_mark_and_move, autoload_data_free, autoload_data_memsize, autoload_data_mark_and_move}, + 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED +}; + +static void +autoload_const_mark_and_move(void *ptr) +{ + struct autoload_const *ac = ptr; + + rb_gc_mark_and_move(&ac->module); + rb_gc_mark_and_move(&ac->autoload_data_value); + rb_gc_mark_and_move(&ac->value); + rb_gc_mark_and_move(&ac->file); + rb_gc_mark_and_move(&ac->box_value); } static size_t -autoload_i_memsize(const void *ptr) +autoload_const_memsize(const void *ptr) +{ + return sizeof(struct autoload_const); +} + +static void +autoload_const_free(void *ptr) { - return sizeof(struct autoload_data_i); + struct autoload_const *autoload_const = ptr; + + ccan_list_del(&autoload_const->cnode); + ruby_xfree(ptr); } -static const rb_data_type_t autoload_data_i_type = { - "autoload_i", - {autoload_i_mark, RUBY_TYPED_DEFAULT_FREE, autoload_i_memsize,}, - 0, 0, RUBY_TYPED_FREE_IMMEDIATELY +static const rb_data_type_t autoload_const_type = { + "autoload_const", + {autoload_const_mark_and_move, autoload_const_free, autoload_const_memsize, autoload_const_mark_and_move,}, + 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED }; -static struct autoload_data_i * -get_autoload_data(VALUE av) +static struct autoload_data * +get_autoload_data(VALUE autoload_const_value, struct autoload_const **autoload_const_pointer) { - struct autoload_data_i *ele = rb_check_typeddata(av, &autoload_data_i_type); + struct autoload_const *autoload_const = rb_check_typeddata(autoload_const_value, &autoload_const_type); + + VALUE autoload_data_value = autoload_const->autoload_data_value; + struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type); /* do not reach across stack for ->state after forking: */ - if (ele && ele->state && ele->fork_gen != GET_VM()->fork_gen) { - ele->state = 0; - ele->fork_gen = 0; + if (autoload_data && autoload_data->fork_gen != GET_VM()->fork_gen) { + RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil); + autoload_data->fork_gen = 0; } - return ele; + + if (autoload_const_pointer) *autoload_const_pointer = autoload_const; + + return autoload_data; +} + +struct autoload_copy_table_data { + VALUE dst_tbl_value; + struct st_table *dst_tbl; + const rb_box_t *box; +}; + +static int +autoload_copy_table_for_box_i(st_data_t key, st_data_t value, st_data_t arg) +{ + struct autoload_const *autoload_const; + struct autoload_copy_table_data *data = (struct autoload_copy_table_data *)arg; + struct st_table *tbl = data->dst_tbl; + VALUE tbl_value = data->dst_tbl_value; + const rb_box_t *box = data->box; + + VALUE src_value = (VALUE)value; + struct autoload_const *src_const = rb_check_typeddata(src_value, &autoload_const_type); + // autoload_data can be shared between copies because the feature is equal between copies. + VALUE autoload_data_value = src_const->autoload_data_value; + struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type); + + VALUE new_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const); + RB_OBJ_WRITE(new_value, &autoload_const->box_value, rb_get_box_object((rb_box_t *)box)); + RB_OBJ_WRITE(new_value, &autoload_const->module, src_const->module); + autoload_const->name = src_const->name; + RB_OBJ_WRITE(new_value, &autoload_const->value, src_const->value); + autoload_const->flag = src_const->flag; + RB_OBJ_WRITE(new_value, &autoload_const->autoload_data_value, autoload_data_value); + ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode); + + st_insert(tbl, (st_data_t)autoload_const->name, (st_data_t)new_value); + RB_OBJ_WRITTEN(tbl_value, Qundef, new_value); + + return ST_CONTINUE; } -RUBY_FUNC_EXPORTED void -rb_autoload(VALUE mod, ID id, const char *file) +void +rb_autoload_copy_table_for_box(st_table *iv_ptr, const rb_box_t *box) { - if (!file || !*file) { - rb_raise(rb_eArgError, "empty file name"); + struct st_table *src_tbl, *dst_tbl; + VALUE src_tbl_value, dst_tbl_value; + if (!rb_st_lookup(iv_ptr, (st_data_t)autoload, (st_data_t *)&src_tbl_value)) { + // the class has no autoload table yet. + return; } - rb_autoload_str(mod, id, rb_fstring_cstr(file)); + if (!RTEST(src_tbl_value) || !(src_tbl = check_autoload_table(src_tbl_value))) { + // the __autoload__ ivar value isn't autoload table value. + return; + } + src_tbl = check_autoload_table(src_tbl_value); + + dst_tbl_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL); + RTYPEDDATA_DATA(dst_tbl_value) = dst_tbl = st_init_numtable(); + + struct autoload_copy_table_data data = { + .dst_tbl_value = dst_tbl_value, + .dst_tbl = dst_tbl, + .box = box, + }; + + st_foreach(src_tbl, autoload_copy_table_for_box_i, (st_data_t)&data); + st_insert(iv_ptr, (st_data_t)autoload, (st_data_t)dst_tbl_value); } void -rb_autoload_str(VALUE mod, ID id, VALUE file) +rb_autoload(VALUE module, ID name, const char *feature) { - st_data_t av; - VALUE ad; - struct st_table *tbl; - struct autoload_data_i *ele; - rb_const_entry_t *ce; - - if (!rb_is_const_id(id)) { - rb_raise(rb_eNameError, "autoload must be constant name: %"PRIsVALUE"", - QUOTE_ID(id)); + if (!feature || !*feature) { + rb_raise(rb_eArgError, "empty feature name"); } - Check_Type(file, T_STRING); - if (!RSTRING_LEN(file)) { - rb_raise(rb_eArgError, "empty file name"); - } + rb_autoload_str(module, name, rb_fstring_cstr(feature)); +} - ce = rb_const_lookup(mod, id); - if (ce && ce->value != Qundef) { - return; +static void const_set(VALUE klass, ID id, VALUE val); +static void const_added(VALUE klass, ID const_name); + +struct autoload_arguments { + VALUE module; + ID name; + VALUE feature; + VALUE box_value; +}; + +static VALUE +autoload_feature_lookup_or_create(VALUE feature, struct autoload_data **autoload_data_pointer) +{ + RUBY_ASSERT_MUTEX_OWNED(autoload_mutex); + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); + + VALUE autoload_data_value = rb_hash_aref(autoload_features, feature); + struct autoload_data *autoload_data; + + if (NIL_P(autoload_data_value)) { + autoload_data_value = TypedData_Make_Struct(0, struct autoload_data, &autoload_data_type, autoload_data); + RB_OBJ_WRITE(autoload_data_value, &autoload_data->feature, feature); + RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil); + ccan_list_head_init(&autoload_data->constants); + + if (autoload_data_pointer) *autoload_data_pointer = autoload_data; + + rb_hash_aset(autoload_features, feature, autoload_data_value); + } + else if (autoload_data_pointer) { + *autoload_data_pointer = rb_check_typeddata(autoload_data_value, &autoload_data_type); } - rb_const_set(mod, id, Qundef); - tbl = RCLASS_IV_TBL(mod); - if (tbl && st_lookup(tbl, (st_data_t)autoload, &av)) { - tbl = check_autoload_table((VALUE)av); + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); + return autoload_data_value; +} + +static VALUE +autoload_table_lookup_or_create(VALUE module) +{ + VALUE autoload_table_value = rb_ivar_lookup(module, autoload, Qfalse); + if (RTEST(autoload_table_value)) { + return autoload_table_value; } else { - if (!tbl) tbl = RCLASS_IV_TBL(mod) = st_init_numtable(); - av = (st_data_t)TypedData_Wrap_Struct(0, &autoload_data_type, 0); - st_add_direct(tbl, (st_data_t)autoload, av); - RB_OBJ_WRITTEN(mod, Qnil, av); - DATA_PTR(av) = tbl = st_init_numtable(); + autoload_table_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL); + rb_class_ivar_set(module, autoload, autoload_table_value); + RTYPEDDATA_DATA(autoload_table_value) = st_init_numtable(); + return autoload_table_value; } +} - ad = TypedData_Make_Struct(0, struct autoload_data_i, &autoload_data_i_type, ele); - if (OBJ_TAINTED(file)) { - file = rb_str_dup(file); - FL_UNSET(file, FL_TAINT); +static VALUE +autoload_synchronized(VALUE _arguments) +{ + struct autoload_arguments *arguments = (struct autoload_arguments *)_arguments; + + rb_const_entry_t *constant_entry = rb_const_lookup(arguments->module, arguments->name); + if (constant_entry && !UNDEF_P(constant_entry->value)) { + return Qfalse; } - ele->feature = rb_fstring(file); - ele->safe_level = rb_safe_level(); - ele->value = Qundef; - ele->state = 0; - ele->flag = CONST_PUBLIC; - st_insert(tbl, (st_data_t)id, (st_data_t)ad); + + // Reset any state associated with any previous constant: + const_set(arguments->module, arguments->name, Qundef); + + VALUE autoload_table_value = autoload_table_lookup_or_create(arguments->module); + struct st_table *autoload_table = check_autoload_table(autoload_table_value); + + // Ensure the string is uniqued since we use an identity lookup: + VALUE feature = rb_fstring(arguments->feature); + + struct autoload_data *autoload_data; + VALUE autoload_data_value = autoload_feature_lookup_or_create(feature, &autoload_data); + + { + struct autoload_const *autoload_const; + VALUE autoload_const_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const); + RB_OBJ_WRITE(autoload_const_value, &autoload_const->box_value, arguments->box_value); + RB_OBJ_WRITE(autoload_const_value, &autoload_const->module, arguments->module); + autoload_const->name = arguments->name; + autoload_const->value = Qundef; + autoload_const->flag = CONST_PUBLIC; + RB_OBJ_WRITE(autoload_const_value, &autoload_const->autoload_data_value, autoload_data_value); + ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode); + st_insert(autoload_table, (st_data_t)arguments->name, (st_data_t)autoload_const_value); + RB_OBJ_WRITTEN(autoload_table_value, Qundef, autoload_const_value); + } + + return Qtrue; } -static void -autoload_delete(VALUE mod, ID id) +void +rb_autoload_str(VALUE module, ID name, VALUE feature) { - st_data_t val, load = 0, n = id; + const rb_box_t *box = rb_current_box(); + VALUE current_box_value = rb_get_box_object((rb_box_t *)box); + + if (!rb_is_const_id(name)) { + rb_raise(rb_eNameError, "autoload must be constant name: %"PRIsVALUE"", QUOTE_ID(name)); + } + + Check_Type(feature, T_STRING); + if (!RSTRING_LEN(feature)) { + rb_raise(rb_eArgError, "empty feature name"); + } - if (st_lookup(RCLASS_IV_TBL(mod), (st_data_t)autoload, &val)) { - struct st_table *tbl = check_autoload_table((VALUE)val); + struct autoload_arguments arguments = { + .module = module, + .name = name, + .feature = feature, + .box_value = current_box_value, + }; - st_delete(tbl, &n, &load); + VALUE result = rb_mutex_synchronize(autoload_mutex, autoload_synchronized, (VALUE)&arguments); - if (tbl->num_entries == 0) { - n = autoload; - st_delete(RCLASS_IV_TBL(mod), &n, &val); - } + if (result == Qtrue) { + const_added(module, name); } } -static VALUE -autoload_provided(VALUE arg) +static void +autoload_delete(VALUE module, ID name) { - const char **p = (const char **)arg; - return rb_feature_provided(*p, p); + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); + + st_data_t load = 0, key = name; + + RUBY_ASSERT(RB_TYPE_P(module, T_CLASS) || RB_TYPE_P(module, T_MODULE)); + + VALUE table_value = rb_ivar_lookup(module, autoload, Qfalse); + if (RTEST(table_value)) { + struct st_table *table = check_autoload_table(table_value); + + st_delete(table, &key, &load); + RB_OBJ_WRITTEN(table_value, load, Qundef); + + /* Qfalse can indicate already deleted */ + if (load != Qfalse) { + struct autoload_const *autoload_const; + struct autoload_data *autoload_data = get_autoload_data((VALUE)load, &autoload_const); + + VM_ASSERT(autoload_data); + VM_ASSERT(!ccan_list_empty(&autoload_data->constants)); + + /* + * we must delete here to avoid "already initialized" warnings + * with parallel autoload. Using list_del_init here so list_del + * works in autoload_const_free + */ + ccan_list_del_init(&autoload_const->cnode); + + if (ccan_list_empty(&autoload_data->constants)) { + rb_hash_delete(autoload_features, autoload_data->feature); + } + + // If the autoload table is empty, we can delete it. + if (table->num_entries == 0) { + rb_attr_delete(module, autoload); + } + } + } + + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); } -static VALUE -reset_safe(VALUE safe) +static int +autoload_by_someone_else(struct autoload_data *ele) { - rb_set_safe_level_force((int)safe); - return safe; + return ele->mutex != Qnil && !rb_mutex_owned_p(ele->mutex); } static VALUE check_autoload_required(VALUE mod, ID id, const char **loadingpath) { - VALUE file, load; - struct autoload_data_i *ele; + VALUE autoload_const_value = autoload_data(mod, id); + struct autoload_data *autoload_data; const char *loading; - int safe; - if (!(load = autoload_data(mod, id)) || !(ele = get_autoload_data(load))) { - return 0; - } - file = ele->feature; - Check_Type(file, T_STRING); - if (!RSTRING_LEN(file) || !*RSTRING_PTR(file)) { - rb_raise(rb_eArgError, "empty file name"); + if (!autoload_const_value || !(autoload_data = get_autoload_data(autoload_const_value, 0))) { + return 0; } + VALUE feature = autoload_data->feature; + /* * if somebody else is autoloading, we MUST wait for them, since * rb_provide_feature can provide a feature before autoload_const_set * completes. We must wait until autoload_const_set finishes in * the other thread. */ - if (ele->state && ele->state->thread != rb_thread_current()) { - return load; + if (autoload_by_someone_else(autoload_data)) { + return autoload_const_value; } - loading = RSTRING_PTR(file); - safe = rb_safe_level(); - rb_set_safe_level_force(0); - if (!rb_ensure(autoload_provided, (VALUE)&loading, reset_safe, (VALUE)safe)) { - return load; + loading = RSTRING_PTR(feature); + + if (!rb_feature_provided(loading, &loading)) { + return autoload_const_value; } + if (loadingpath && loading) { - *loadingpath = loading; - return load; + *loadingpath = loading; + return autoload_const_value; } + return 0; } +static struct autoload_const *autoloading_const_entry(VALUE mod, ID id); + int rb_autoloading_value(VALUE mod, ID id, VALUE* value, rb_const_flag_t *flag) { - VALUE load; - struct autoload_data_i *ele; - - if (!(load = autoload_data(mod, id)) || !(ele = get_autoload_data(load))) { - return 0; - } - if (ele->state && ele->state->thread == rb_thread_current()) { - if (ele->value != Qundef) { - if (value) { - *value = ele->value; - } - if (flag) { - *flag = ele->flag; - } - return 1; - } + struct autoload_const *ac = autoloading_const_entry(mod, id); + if (!ac) return FALSE; + + if (value) { + *value = ac->value; + } + + if (flag) { + *flag = ac->flag; + } + + return TRUE; +} + +static int +autoload_by_current(struct autoload_data *ele) +{ + return ele->mutex != Qnil && rb_mutex_owned_p(ele->mutex); +} + +// If there is an autoloading constant and it has been set by the current +// execution context, return it. This allows threads which are loading code to +// refer to their own autoloaded constants. +struct autoload_const * +autoloading_const_entry(VALUE mod, ID id) +{ + VALUE load = autoload_data(mod, id); + struct autoload_data *ele; + struct autoload_const *ac; + + // Find the autoloading state: + if (!load || !(ele = get_autoload_data(load, &ac))) { + // Couldn't be found: + return 0; } + + // Check if it's being loaded by the current thread/fiber: + if (autoload_by_current(ele)) { + if (!UNDEF_P(ac->value)) { + return ac; + } + } + return 0; } @@ -2054,258 +3098,346 @@ autoload_defined_p(VALUE mod, ID id) { rb_const_entry_t *ce = rb_const_lookup(mod, id); - if (!ce || ce->value != Qundef) { - return 0; + // If there is no constant or the constant is not undefined (special marker for autoloading): + if (!ce || !UNDEF_P(ce->value)) { + // We are not autoloading: + return 0; } + + // Otherwise check if there is an autoload in flight right now: return !rb_autoloading_value(mod, id, NULL, NULL); } -struct autoload_const_set_args { - VALUE mod; - ID id; - VALUE value; - rb_const_flag_t flag; -}; +static void const_tbl_update(struct autoload_const *, int); -static void const_tbl_update(struct autoload_const_set_args *); +struct autoload_load_arguments { + VALUE module; + ID name; + int flag; -static VALUE -autoload_const_set(VALUE arg) -{ - struct autoload_const_set_args* args = (struct autoload_const_set_args *)arg; - VALUE klass = args->mod; - ID id = args->id; - check_before_mod_set(klass, id, args->value, "constant"); - const_tbl_update(args); - return 0; /* ignored */ -} + VALUE mutex; + + // The specific constant which triggered the autoload code to fire: + struct autoload_const *autoload_const; + + // The parent autoload data which is shared between multiple constants: + struct autoload_data *autoload_data; +}; static VALUE -autoload_require(VALUE arg) +autoload_const_set(struct autoload_const *ac) { - struct autoload_state *state = (struct autoload_state *)arg; + check_before_mod_set(ac->module, ac->name, ac->value, "constant"); - /* this may release GVL and switch threads: */ - state->result = rb_funcall(rb_vm_top_self(), rb_intern("require"), 1, - state->ele->feature); + RB_VM_LOCKING() { + const_tbl_update(ac, true); + } - return state->result; + return 0; /* ignored */ } static VALUE -autoload_reset(VALUE arg) +autoload_load_needed(VALUE _arguments) { - struct autoload_state *state = (struct autoload_state *)arg; - int need_wakeups = 0; + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + const char *loading = 0, *src; + + if (!autoload_defined_p(arguments->module, arguments->name)) { + return Qfalse; + } - if (state->ele->state == state) { - need_wakeups = 1; - state->ele->state = 0; - state->ele->fork_gen = 0; + VALUE autoload_const_value = check_autoload_required(arguments->module, arguments->name, &loading); + if (!autoload_const_value) { + return Qfalse; } - /* At the last, move a value defined in autoload to constant table */ - if (RTEST(state->result) && state->ele->value != Qundef) { - int safe_backup; - struct autoload_const_set_args args; + src = rb_sourcefile(); + if (src && loading && strcmp(src, loading) == 0) { + return Qfalse; + } - args.mod = state->mod; - args.id = state->id; - args.value = state->ele->value; - args.flag = state->ele->flag; - safe_backup = rb_safe_level(); - rb_set_safe_level_force(state->ele->safe_level); - rb_ensure(autoload_const_set, (VALUE)&args, - reset_safe, (VALUE)safe_backup); + struct autoload_const *autoload_const; + struct autoload_data *autoload_data; + if (!(autoload_data = get_autoload_data(autoload_const_value, &autoload_const))) { + return Qfalse; } - /* wakeup any waiters we had */ - if (need_wakeups) { - struct autoload_state *cur = 0, *nxt; + if (NIL_P(autoload_data->mutex)) { + RB_OBJ_WRITE(autoload_const->autoload_data_value, &autoload_data->mutex, rb_mutex_new()); + autoload_data->fork_gen = GET_VM()->fork_gen; + } + else if (rb_mutex_owned_p(autoload_data->mutex)) { + return Qfalse; + } - list_for_each_safe(&state->waitq.head, cur, nxt, waitq.node) { - VALUE th = cur->thread; + arguments->mutex = autoload_data->mutex; + arguments->autoload_const = autoload_const; - cur->thread = Qfalse; - list_del_init(&cur->waitq.node); /* idempotent */ + return autoload_const_value; +} - /* - * cur is stored on the stack of cur->waiting_th, - * do not touch cur after waking up waiting_th - */ - rb_thread_wakeup_alive(th); - } +static VALUE +autoload_apply_constants(VALUE _arguments) +{ + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); + + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + struct autoload_const *autoload_const = 0; // for ccan_container_off_var() + struct autoload_const *next; + + // We use safe iteration here because `autoload_const_set` will eventually invoke + // `autoload_delete` which will remove the constant from the linked list. In theory, once + // the `autoload_data->constants` linked list is empty, we can remove it. + + // Iterate over all constants and assign them: + ccan_list_for_each_safe(&arguments->autoload_data->constants, autoload_const, next, cnode) { + if (!UNDEF_P(autoload_const->value)) { + autoload_const_set(autoload_const); + } } - return 0; /* ignored */ + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); + + return Qtrue; } static VALUE -autoload_sleep(VALUE arg) +autoload_feature_require(VALUE _arguments) { - struct autoload_state *state = (struct autoload_state *)arg; + VALUE receiver = rb_vm_top_self(); + + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + struct autoload_const *autoload_const = arguments->autoload_const; + VALUE autoload_box_value = autoload_const->box_value; + + // We save this for later use in autoload_apply_constants: + arguments->autoload_data = rb_check_typeddata(autoload_const->autoload_data_value, &autoload_data_type); + + if (rb_box_available() && BOX_OBJ_P(autoload_box_value)) + receiver = autoload_box_value; /* - * autoload_reset in other thread will resume us and remove us - * from the waitq list + * Clear the global cc cache table because the require method can be different from the current + * box's one and it may cause inconsistent cc-cme states. + * For example, the assertion below may fail in gccct_method_search(); + * VM_ASSERT(vm_cc_check_cme(cc, rb_callable_method_entry(klass, mid))) */ - do { - rb_thread_sleep_deadly(); - } while (state->thread != Qfalse); + rb_gccct_clear_table(Qnil); - return Qfalse; + VALUE result = rb_funcall(receiver, rb_intern("require"), 1, arguments->autoload_data->feature); + + if (RTEST(result)) { + return rb_mutex_synchronize(autoload_mutex, autoload_apply_constants, _arguments); + } + return result; } static VALUE -autoload_sleep_done(VALUE arg) +autoload_try_load(VALUE _arguments) { - struct autoload_state *state = (struct autoload_state *)arg; + struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments; + + VALUE result = autoload_feature_require(_arguments); + + // After we loaded the feature, if the constant is not defined, we remove it completely: + rb_const_entry_t *ce = rb_const_lookup(arguments->module, arguments->name); + + if (!ce || UNDEF_P(ce->value)) { + result = Qfalse; + + rb_const_remove(arguments->module, arguments->name); - if (state->thread != Qfalse && rb_thread_to_be_killed(state->thread)) { - list_del(&state->waitq.node); /* idempotent after list_del_init */ + if (arguments->module == rb_cObject) { + rb_warning( + "Expected %"PRIsVALUE" to define %"PRIsVALUE" but it didn't", + arguments->autoload_data->feature, + ID2SYM(arguments->name) + ); + } + else { + rb_warning( + "Expected %"PRIsVALUE" to define %"PRIsVALUE"::%"PRIsVALUE" but it didn't", + arguments->autoload_data->feature, + arguments->module, + ID2SYM(arguments->name) + ); + } + } + else { + // Otherwise, it was loaded, copy the flags from the autoload constant: + ce->flag |= arguments->flag; } - return Qfalse; + return result; } VALUE -rb_autoload_load(VALUE mod, ID id) +rb_autoload_load(VALUE module, ID name) { - VALUE load, result; - const char *loading = 0, *src; - struct autoload_data_i *ele; - struct autoload_state state; + rb_const_entry_t *ce = rb_const_lookup(module, name); - if (!autoload_defined_p(mod, id)) return Qfalse; - load = check_autoload_required(mod, id, &loading); - if (!load) return Qfalse; - src = rb_sourcefile(); - if (src && loading && strcmp(src, loading) == 0) return Qfalse; + // We bail out as early as possible without any synchronisation: + if (!ce || !UNDEF_P(ce->value)) { + return Qfalse; + } - /* set ele->state for a marker of autoloading thread */ - if (!(ele = get_autoload_data(load))) { - return Qfalse; + // At this point, we assume there might be autoloading, so fail if it's ractor: + if (UNLIKELY(!rb_ractor_main_p())) { + return rb_ractor_autoload_load(module, name); } - state.ele = ele; - state.mod = mod; - state.id = id; - state.thread = rb_thread_current(); - if (!ele->state) { - ele->state = &state; - ele->fork_gen = GET_VM()->fork_gen; + // This state is stored on the stack and is used during the autoload process. + struct autoload_load_arguments arguments = {.module = module, .name = name, .mutex = Qnil}; - /* - * autoload_reset will wake up any threads added to this - * iff the GVL is released during autoload_require - */ - list_head_init(&state.waitq.head); - } - else if (state.thread == ele->state->thread) { - return Qfalse; - } - else { - list_add_tail(&ele->state->waitq.head, &state.waitq.node); + // Figure out whether we can autoload the named constant: + VALUE autoload_const_value = rb_mutex_synchronize(autoload_mutex, autoload_load_needed, (VALUE)&arguments); - rb_ensure(autoload_sleep, (VALUE)&state, - autoload_sleep_done, (VALUE)&state); - } + // This confirms whether autoloading is required or not: + if (autoload_const_value == Qfalse) return autoload_const_value; + + arguments.flag = ce->flag & (CONST_DEPRECATED | CONST_VISIBILITY_MASK); + + // Only one thread will enter here at a time: + VALUE result = rb_mutex_synchronize(arguments.mutex, autoload_try_load, (VALUE)&arguments); - /* autoload_data_i can be deleted by another thread while require */ - result = rb_ensure(autoload_require, (VALUE)&state, - autoload_reset, (VALUE)&state); + // If you don't guard this value, it's possible for the autoload constant to + // be freed by another thread which loads multiple constants, one of which + // resolves to the constant this thread is trying to load, so proteect this + // so that it is not freed until we are done with it in `autoload_try_load`: + RB_GC_GUARD(autoload_const_value); - RB_GC_GUARD(load); return result; } VALUE rb_autoload_p(VALUE mod, ID id) { + return rb_autoload_at_p(mod, id, TRUE); +} + +VALUE +rb_autoload_at_p(VALUE mod, ID id, int recur) +{ VALUE load; - struct autoload_data_i *ele; + struct autoload_data *ele; while (!autoload_defined_p(mod, id)) { - mod = RCLASS_SUPER(mod); - if (!mod) return Qnil; + if (!recur) return Qnil; + mod = RCLASS_SUPER(mod); + if (!mod) return Qnil; } load = check_autoload_required(mod, id, 0); if (!load) return Qnil; - return (ele = get_autoload_data(load)) ? ele->feature : Qnil; + return (ele = get_autoload_data(load, 0)) ? ele->feature : Qnil; } void rb_const_warn_if_deprecated(const rb_const_entry_t *ce, VALUE klass, ID id) { - if (RB_CONST_DEPRECATED_P(ce)) { - if (klass == rb_cObject) { - rb_warn("constant ::%"PRIsVALUE" is deprecated", QUOTE_ID(id)); - } - else { - rb_warn("constant %"PRIsVALUE"::%"PRIsVALUE" is deprecated", - rb_class_name(klass), QUOTE_ID(id)); - } + if (RB_CONST_DEPRECATED_P(ce) && + rb_warning_category_enabled_p(RB_WARN_CATEGORY_DEPRECATED)) { + if (klass == rb_cObject) { + rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant ::%"PRIsVALUE" is deprecated", QUOTE_ID(id)); + } + else { + rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant %"PRIsVALUE"::%"PRIsVALUE" is deprecated", + rb_class_name(klass), QUOTE_ID(id)); + } } } static VALUE rb_const_get_0(VALUE klass, ID id, int exclude, int recurse, int visibility) { - VALUE c = rb_const_search(klass, id, exclude, recurse, visibility); - if (c != Qundef) return c; + VALUE found_in; + VALUE c = rb_const_search(klass, id, exclude, recurse, visibility, &found_in); + if (!UNDEF_P(c)) { + if (UNLIKELY(!rb_ractor_main_p())) { + if (!rb_ractor_shareable_p(c)) { + rb_raise(rb_eRactorIsolationError, "can not access non-shareable objects in constant %"PRIsVALUE"::%"PRIsVALUE" by non-main Ractor.", rb_class_path(found_in), rb_id2str(id)); + } + } + return c; + } return rb_const_missing(klass, ID2SYM(id)); } static VALUE -rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility) -{ - VALUE value, tmp, av; - rb_const_flag_t flag; - int mod_retry = 0; +rb_const_search_from(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in) +{ + VALUE value, current; + bool first_iteration = true; + + for (current = klass; + RTEST(current); + current = RCLASS_SUPER(current), first_iteration = false) { + VALUE tmp; + VALUE am = 0; + rb_const_entry_t *ce; + + if (!first_iteration && RCLASS_ORIGIN(current) != current) { + // This item in the super chain has an origin iclass + // that comes later in the chain. Skip this item so + // prepended modules take precedence. + continue; + } - tmp = klass; - retry: - while (RTEST(tmp)) { - VALUE am = 0; - rb_const_entry_t *ce; - - while ((ce = rb_const_lookup(tmp, id))) { - if (visibility && RB_CONST_PRIVATE_P(ce)) { - if (BUILTIN_TYPE(tmp) == T_ICLASS) tmp = RBASIC(tmp)->klass; - rb_name_err_raise("private constant %2$s::%1$s referenced", - tmp, ID2SYM(id)); - } - rb_const_warn_if_deprecated(ce, tmp, id); - value = ce->value; - if (value == Qundef) { - if (am == tmp) break; - am = tmp; - if (rb_autoloading_value(tmp, id, &av, &flag)) return av; - rb_autoload_load(tmp, id); - continue; - } - if (exclude && tmp == rb_cObject && klass != rb_cObject) { -#if 0 - rb_warn("toplevel constant %"PRIsVALUE" referenced by %"PRIsVALUE"::%"PRIsVALUE"", - QUOTE_ID(id), rb_class_name(klass), QUOTE_ID(id)); -#else - return Qundef; -#endif - } - return value; - } - if (!recurse) break; - tmp = RCLASS_SUPER(tmp); - } - if (!exclude && !mod_retry && BUILTIN_TYPE(klass) == T_MODULE) { - mod_retry = 1; - tmp = rb_cObject; - goto retry; + // Do lookup in original class or module in case we are at an origin + // iclass in the chain. + tmp = current; + if (BUILTIN_TYPE(tmp) == T_ICLASS) tmp = RBASIC(tmp)->klass; + + // Do the lookup. Loop in case of autoload. + while ((ce = rb_const_lookup(tmp, id))) { + if (visibility && RB_CONST_PRIVATE_P(ce)) { + GET_EC()->private_const_reference = tmp; + return Qundef; + } + rb_const_warn_if_deprecated(ce, tmp, id); + value = ce->value; + if (UNDEF_P(value)) { + struct autoload_const *ac; + if (am == tmp) break; + am = tmp; + ac = autoloading_const_entry(tmp, id); + if (ac) { + if (found_in) { *found_in = tmp; } + return ac->value; + } + rb_autoload_load(tmp, id); + continue; + } + if (exclude && tmp == rb_cObject) { + goto not_found; + } + if (found_in) { *found_in = tmp; } + return value; + } + if (!recurse) break; } + not_found: + GET_EC()->private_const_reference = 0; return Qundef; } +static VALUE +rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in) +{ + VALUE value; + + if (klass == rb_cObject) exclude = FALSE; + value = rb_const_search_from(klass, id, exclude, recurse, visibility, found_in); + if (!UNDEF_P(value)) return value; + if (exclude) return value; + if (BUILTIN_TYPE(klass) != T_MODULE) return value; + /* search global const too, if klass is a module */ + return rb_const_search_from(rb_cObject, id, FALSE, recurse, visibility, found_in); +} + VALUE rb_const_get_from(VALUE klass, ID id) { @@ -2331,15 +3463,80 @@ rb_public_const_get_from(VALUE klass, ID id) } VALUE -rb_public_const_get(VALUE klass, ID id) +rb_public_const_get_at(VALUE klass, ID id) +{ + return rb_const_get_0(klass, id, TRUE, FALSE, TRUE); +} + +NORETURN(static void undefined_constant(VALUE mod, VALUE name)); +static void +undefined_constant(VALUE mod, VALUE name) { - return rb_const_get_0(klass, id, FALSE, TRUE, TRUE); + rb_name_err_raise("constant %2$s::%1$s not defined", + mod, name); +} + +static VALUE +rb_const_location_from(VALUE klass, ID id, int exclude, int recurse, int visibility) +{ + while (RTEST(klass)) { + rb_const_entry_t *ce; + + while ((ce = rb_const_lookup(klass, id))) { + if (visibility && RB_CONST_PRIVATE_P(ce)) { + return Qnil; + } + if (exclude && klass == rb_cObject) { + goto not_found; + } + + if (UNDEF_P(ce->value)) { // autoload + VALUE autoload_const_value = autoload_data(klass, id); + if (RTEST(autoload_const_value)) { + struct autoload_const *autoload_const; + struct autoload_data *autoload_data = get_autoload_data(autoload_const_value, &autoload_const); + + if (!UNDEF_P(autoload_const->value) && RTEST(rb_mutex_owned_p(autoload_data->mutex))) { + return rb_assoc_new(autoload_const->file, INT2NUM(autoload_const->line)); + } + } + } + + if (NIL_P(ce->file)) return rb_ary_new(); + return rb_assoc_new(ce->file, INT2NUM(ce->line)); + } + if (!recurse) break; + klass = RCLASS_SUPER(klass); + } + + not_found: + return Qnil; +} + +static VALUE +rb_const_location(VALUE klass, ID id, int exclude, int recurse, int visibility) +{ + VALUE loc; + + if (klass == rb_cObject) exclude = FALSE; + loc = rb_const_location_from(klass, id, exclude, recurse, visibility); + if (!NIL_P(loc)) return loc; + if (exclude) return loc; + if (BUILTIN_TYPE(klass) != T_MODULE) return loc; + /* search global const too, if klass is a module */ + return rb_const_location_from(rb_cObject, id, FALSE, recurse, visibility); } VALUE -rb_public_const_get_at(VALUE klass, ID id) +rb_const_source_location(VALUE klass, ID id) { - return rb_const_get_0(klass, id, TRUE, FALSE, TRUE); + return rb_const_location(klass, id, FALSE, TRUE, FALSE); +} + +VALUE +rb_const_source_location_at(VALUE klass, ID id) +{ + return rb_const_location(klass, id, TRUE, FALSE, FALSE); } /* @@ -2358,12 +3555,13 @@ rb_mod_remove_const(VALUE mod, VALUE name) const ID id = id_for_var(mod, name, a, constant); if (!id) { - rb_name_err_raise("constant %2$s::%1$s not defined", - mod, name); + undefined_constant(mod, name); } return rb_const_remove(mod, id); } +static rb_const_entry_t * const_lookup(struct rb_id_table *tbl, ID id); + VALUE rb_const_remove(VALUE mod, ID id) { @@ -2371,24 +3569,40 @@ rb_const_remove(VALUE mod, ID id) rb_const_entry_t *ce; rb_check_frozen(mod); + ce = rb_const_lookup(mod, id); - if (!ce || !rb_id_table_delete(RCLASS_CONST_TBL(mod), id)) { - if (rb_const_defined_at(mod, id)) { - rb_name_err_raise("cannot remove %2$s::%1$s", - mod, ID2SYM(id)); - } - rb_name_err_raise("constant %2$s::%1$s not defined", - mod, ID2SYM(id)); + + if (!ce) { + if (rb_const_defined_at(mod, id)) { + rb_name_err_raise("cannot remove %2$s::%1$s", mod, ID2SYM(id)); + } + + undefined_constant(mod, ID2SYM(id)); } - rb_clear_constant_cache(); + VALUE writable_ce = 0; + if (rb_id_table_lookup(RCLASS_WRITABLE_CONST_TBL(mod), id, &writable_ce)) { + rb_id_table_delete(RCLASS_WRITABLE_CONST_TBL(mod), id); + if ((rb_const_entry_t *)writable_ce != ce) { + xfree((rb_const_entry_t *)writable_ce); + } + } + + rb_const_warn_if_deprecated(ce, mod, id); + rb_clear_constant_cache_for_id(id); val = ce->value; - if (val == Qundef) { - autoload_delete(mod, id); - val = Qnil; + + if (UNDEF_P(val)) { + autoload_delete(mod, id); + val = Qnil; + } + + if (ce != const_lookup(RCLASS_PRIME_CONST_TBL(mod), id)) { + ruby_xfree(ce); } - xfree(ce); + // else - skip free'ing the ce because it still exists in the prime classext + return val; } @@ -2407,7 +3621,7 @@ sv_i(ID key, VALUE v, void *a) st_table *tbl = a; if (rb_is_const_id(key)) { - st_update(tbl, (st_data_t)key, cv_i_update, (st_data_t)ce); + st_update(tbl, (st_data_t)key, cv_i_update, (st_data_t)ce); } return ID_TABLE_CONTINUE; } @@ -2416,7 +3630,7 @@ static enum rb_id_table_iterator_result rb_local_constants_i(ID const_name, VALUE const_value, void *ary) { if (rb_is_const_id(const_name) && !RB_CONST_PRIVATE_P((rb_const_entry_t *)const_value)) { - rb_ary_push((VALUE)ary, ID2SYM(const_name)); + rb_ary_push((VALUE)ary, ID2SYM(const_name)); } return ID_TABLE_CONTINUE; } @@ -2429,8 +3643,11 @@ rb_local_constants(VALUE mod) if (!tbl) return rb_ary_new2(0); - ary = rb_ary_new2(rb_id_table_size(tbl)); - rb_id_table_foreach(tbl, rb_local_constants_i, (void *)ary); + RB_VM_LOCKING() { + ary = rb_ary_new2(rb_id_table_size(tbl)); + rb_id_table_foreach(tbl, rb_local_constants_i, (void *)ary); + } + return ary; } @@ -2439,10 +3656,12 @@ rb_mod_const_at(VALUE mod, void *data) { st_table *tbl = data; if (!tbl) { - tbl = st_init_numtable(); + tbl = st_init_numtable(); } if (RCLASS_CONST_TBL(mod)) { - rb_id_table_foreach(RCLASS_CONST_TBL(mod), sv_i, tbl); + RB_VM_LOCKING() { + rb_id_table_foreach(RCLASS_CONST_TBL(mod), sv_i, tbl); + } } return tbl; } @@ -2452,10 +3671,10 @@ rb_mod_const_of(VALUE mod, void *data) { VALUE tmp = mod; for (;;) { - data = rb_mod_const_at(tmp, data); - tmp = RCLASS_SUPER(tmp); - if (!tmp) break; - if (tmp == rb_cObject && mod != rb_cObject) break; + data = rb_mod_const_at(tmp, data); + tmp = RCLASS_SUPER(tmp); + if (!tmp) break; + if (tmp == rb_cObject && mod != rb_cObject) break; } return data; } @@ -2498,26 +3717,21 @@ rb_const_list(void *data) * IO.constants.include?(:SYNC) #=> true * IO.constants(false).include?(:SYNC) #=> false * - * Also see <code>Module::const_defined?</code>. + * Also see Module#const_defined?. */ VALUE rb_mod_constants(int argc, const VALUE *argv, VALUE mod) { - VALUE inherit; + bool inherit = true; - if (argc == 0) { - inherit = Qtrue; - } - else { - rb_scan_args(argc, argv, "01", &inherit); - } + if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]); - if (RTEST(inherit)) { - return rb_const_list(rb_mod_const_of(mod, 0)); + if (inherit) { + return rb_const_list(rb_mod_const_of(mod, 0)); } else { - return rb_local_constants(mod); + return rb_local_constants(mod); } } @@ -2531,27 +3745,27 @@ rb_const_defined_0(VALUE klass, ID id, int exclude, int recurse, int visibility) tmp = klass; retry: while (tmp) { - if ((ce = rb_const_lookup(tmp, id))) { - if (visibility && RB_CONST_PRIVATE_P(ce)) { - return (int)Qfalse; - } - if (ce->value == Qundef && !check_autoload_required(tmp, id, 0) && - !rb_autoloading_value(tmp, id, NULL, NULL)) - return (int)Qfalse; - - if (exclude && tmp == rb_cObject && klass != rb_cObject) { - return (int)Qfalse; - } - - return (int)Qtrue; - } - if (!recurse) break; - tmp = RCLASS_SUPER(tmp); + if ((ce = rb_const_lookup(tmp, id))) { + if (visibility && RB_CONST_PRIVATE_P(ce)) { + return (int)Qfalse; + } + if (UNDEF_P(ce->value) && !check_autoload_required(tmp, id, 0) && + !rb_autoloading_value(tmp, id, NULL, NULL)) + return (int)Qfalse; + + if (exclude && tmp == rb_cObject && klass != rb_cObject) { + return (int)Qfalse; + } + + return (int)Qtrue; + } + if (!recurse) break; + tmp = RCLASS_SUPER(tmp); } if (!exclude && !mod_retry && BUILTIN_TYPE(klass) == T_MODULE) { - mod_retry = 1; - tmp = rb_cObject; - goto retry; + mod_retry = 1; + tmp = rb_cObject; + goto retry; } return (int)Qfalse; } @@ -2580,148 +3794,223 @@ rb_public_const_defined_from(VALUE klass, ID id) return rb_const_defined_0(klass, id, TRUE, TRUE, TRUE); } -int -rb_public_const_defined(VALUE klass, ID id) +static void +check_before_mod_set(VALUE klass, ID id, VALUE val, const char *dest) { - return rb_const_defined_0(klass, id, FALSE, TRUE, TRUE); + rb_check_frozen(klass); } -int -rb_public_const_defined_at(VALUE klass, ID id) +static void set_namespace_path(VALUE named_namespace, VALUE name); + +static enum rb_id_table_iterator_result +set_namespace_path_i(ID id, VALUE v, void *payload) { - return rb_const_defined_0(klass, id, TRUE, FALSE, TRUE); + rb_const_entry_t *ce = (rb_const_entry_t *)v; + VALUE value = ce->value; + VALUE parental_path = *((VALUE *) payload); + if (!rb_is_const_id(id) || !rb_namespace_p(value)) { + return ID_TABLE_CONTINUE; + } + + bool has_permanent_classpath; + classname(value, &has_permanent_classpath); + if (has_permanent_classpath) { + return ID_TABLE_CONTINUE; + } + set_namespace_path(value, build_const_path(parental_path, id)); + + if (!RCLASS_PERMANENT_CLASSPATH_P(value)) { + RCLASS_WRITE_CLASSPATH(value, 0, false); + } + + return ID_TABLE_CONTINUE; } +/* + * Assign permanent classpaths to all namespaces that are directly or indirectly + * nested under +named_namespace+. +named_namespace+ must have a permanent + * classpath. + */ static void -check_before_mod_set(VALUE klass, ID id, VALUE val, const char *dest) +set_namespace_path(VALUE named_namespace, VALUE namespace_path) { - rb_check_frozen(klass); + struct rb_id_table *const_table = RCLASS_CONST_TBL(named_namespace); + RB_OBJ_SET_SHAREABLE(namespace_path); + + RB_VM_LOCKING() { + RCLASS_WRITE_CLASSPATH(named_namespace, namespace_path, true); + + if (const_table) { + rb_id_table_foreach(const_table, set_namespace_path_i, &namespace_path); + } + } } -void -rb_const_set(VALUE klass, ID id, VALUE val) +static void +const_added(VALUE klass, ID const_name) +{ + if (GET_VM()->running) { + VALUE name = ID2SYM(const_name); + rb_funcallv(klass, idConst_added, 1, &name); + } +} + +static void +const_set(VALUE klass, ID id, VALUE val) { rb_const_entry_t *ce; - struct rb_id_table *tbl = RCLASS_CONST_TBL(klass); if (NIL_P(klass)) { - rb_raise(rb_eTypeError, "no class/module to define constant %"PRIsVALUE"", - QUOTE_ID(id)); + rb_raise(rb_eTypeError, "no class/module to define constant %"PRIsVALUE"", + QUOTE_ID(id)); } - check_before_mod_set(klass, id, val, "constant"); - if (!tbl) { - RCLASS_CONST_TBL(klass) = tbl = rb_id_table_create(0); - rb_clear_constant_cache(); - ce = ZALLOC(rb_const_entry_t); - rb_id_table_insert(tbl, id, (VALUE)ce); - setup_const_entry(ce, klass, val, CONST_PUBLIC); + if (!rb_ractor_main_p() && !rb_ractor_shareable_p(val)) { + rb_raise(rb_eRactorIsolationError, "can not set constants with non-shareable objects by non-main Ractors"); } - else { - struct autoload_const_set_args args; - args.mod = klass; - args.id = id; - args.value = val; - args.flag = CONST_PUBLIC; - const_tbl_update(&args); + + check_before_mod_set(klass, id, val, "constant"); + + RB_VM_LOCKING() { + struct rb_id_table *tbl = RCLASS_WRITABLE_CONST_TBL(klass); + if (!tbl) { + tbl = rb_id_table_create(0); + RCLASS_WRITE_CONST_TBL(klass, tbl, false); + rb_clear_constant_cache_for_id(id); + ce = ZALLOC(rb_const_entry_t); + rb_id_table_insert(tbl, id, (VALUE)ce); + setup_const_entry(ce, klass, val, CONST_PUBLIC); + } + else { + struct autoload_const ac = { + .module = klass, .name = id, + .value = val, .flag = CONST_PUBLIC, + /* fill the rest with 0 */ + }; + ac.file = rb_source_location(&ac.line); + const_tbl_update(&ac, false); + } } + /* * Resolve and cache class name immediately to resolve ambiguity * and avoid order-dependency on const_tbl */ - if (rb_cObject && (RB_TYPE_P(val, T_MODULE) || RB_TYPE_P(val, T_CLASS))) { - if (NIL_P(rb_class_path_cached(val))) { - if (klass == rb_cObject) { - rb_ivar_set(val, classpath, rb_id2str(id)); - rb_name_class(val, id); - } - else { - VALUE path; - ID pathid; - st_data_t n; - st_table *ivtbl = RCLASS_IV_TBL(klass); - if (ivtbl && - (st_lookup(ivtbl, (st_data_t)(pathid = classpath), &n) || - st_lookup(ivtbl, (st_data_t)(pathid = tmp_classpath), &n))) { - path = rb_str_dup((VALUE)n); - rb_str_append(rb_str_cat2(path, "::"), rb_id2str(id)); - OBJ_FREEZE(path); - rb_ivar_set(val, pathid, path); - rb_name_class(val, id); - } - } - } - } -} - -static struct autoload_data_i * -current_autoload_data(VALUE mod, ID id) -{ - struct autoload_data_i *ele; - VALUE load = autoload_data(mod, id); - if (!load) return 0; - ele = get_autoload_data(load); - if (!ele) return 0; + if (rb_cObject && rb_namespace_p(val)) { + bool val_path_permanent; + VALUE val_path = classname(val, &val_path_permanent); + if (NIL_P(val_path) || !val_path_permanent) { + if (klass == rb_cObject) { + set_namespace_path(val, rb_id2str(id)); + } + else { + bool parental_path_permanent; + VALUE parental_path = classname(klass, &parental_path_permanent); + if (NIL_P(parental_path)) { + bool throwaway; + parental_path = rb_tmp_class_path(klass, &throwaway, make_temporary_path); + } + if (parental_path_permanent && !val_path_permanent) { + set_namespace_path(val, build_const_path(parental_path, id)); + } + else if (!parental_path_permanent && NIL_P(val_path)) { + VALUE path = build_const_path(parental_path, id); + RCLASS_SET_CLASSPATH(val, path, false); + } + } + } + } +} + +void +rb_const_set(VALUE klass, ID id, VALUE val) +{ + const_set(klass, id, val); + const_added(klass, id); +} + +static VALUE +autoload_const_value_for_named_constant(VALUE module, ID name, struct autoload_const **autoload_const_pointer) +{ + VALUE autoload_const_value = autoload_data(module, name); + if (!autoload_const_value) return Qfalse; + + struct autoload_data *autoload_data = get_autoload_data(autoload_const_value, autoload_const_pointer); + if (!autoload_data) return Qfalse; + /* for autoloading thread, keep the defined value to autoloading storage */ - if (ele->state && (ele->state->thread == rb_thread_current())) { - return ele; + if (autoload_by_current(autoload_data)) { + return autoload_const_value; } - return 0; + + return Qfalse; } static void -const_tbl_update(struct autoload_const_set_args *args) +const_tbl_update(struct autoload_const *ac, int autoload_force) { VALUE value; - VALUE klass = args->mod; - VALUE val = args->value; - ID id = args->id; + VALUE klass = ac->module; + VALUE val = ac->value; + ID id = ac->name; struct rb_id_table *tbl = RCLASS_CONST_TBL(klass); - rb_const_flag_t visibility = args->flag; + rb_const_flag_t visibility = ac->flag; rb_const_entry_t *ce; if (rb_id_table_lookup(tbl, id, &value)) { - ce = (rb_const_entry_t *)value; - if (ce->value == Qundef) { - struct autoload_data_i *ele = current_autoload_data(klass, id); - - if (ele) { - rb_clear_constant_cache(); - - ele->value = val; /* autoload_i is non-WB-protected */ - return; - } - /* otherwise, allow to override */ - autoload_delete(klass, id); - } - else { - VALUE name = QUOTE_ID(id); - visibility = ce->flag; - if (klass == rb_cObject) - rb_warn("already initialized constant %"PRIsVALUE"", name); - else - rb_warn("already initialized constant %"PRIsVALUE"::%"PRIsVALUE"", - rb_class_name(klass), name); - if (!NIL_P(ce->file) && ce->line) { - rb_compile_warn(RSTRING_PTR(ce->file), ce->line, - "previous definition of %"PRIsVALUE" was here", name); - } - } - rb_clear_constant_cache(); - setup_const_entry(ce, klass, val, visibility); + ce = (rb_const_entry_t *)value; + if (UNDEF_P(ce->value)) { + RUBY_ASSERT_CRITICAL_SECTION_ENTER(); + VALUE file = ac->file; + int line = ac->line; + VALUE autoload_const_value = autoload_const_value_for_named_constant(klass, id, &ac); + + if (!autoload_force && autoload_const_value) { + rb_clear_constant_cache_for_id(id); + + RB_OBJ_WRITE(autoload_const_value, &ac->value, val); + RB_OBJ_WRITE(autoload_const_value, &ac->file, rb_source_location(&ac->line)); + } + else { + /* otherwise autoloaded constant, allow to override */ + autoload_delete(klass, id); + ce->flag = visibility; + RB_OBJ_WRITE(klass, &ce->value, val); + RB_OBJ_WRITE(klass, &ce->file, file); + ce->line = line; + } + RUBY_ASSERT_CRITICAL_SECTION_LEAVE(); + return; + } + else { + VALUE name = QUOTE_ID(id); + visibility = ce->flag; + if (klass == rb_cObject) + rb_warn("already initialized constant %"PRIsVALUE"", name); + else + rb_warn("already initialized constant %"PRIsVALUE"::%"PRIsVALUE"", + rb_class_name(klass), name); + if (!NIL_P(ce->file) && ce->line) { + rb_compile_warn(RSTRING_PTR(ce->file), ce->line, + "previous definition of %"PRIsVALUE" was here", name); + } + } + rb_clear_constant_cache_for_id(id); + setup_const_entry(ce, klass, val, visibility); } else { - rb_clear_constant_cache(); + tbl = RCLASS_WRITABLE_CONST_TBL(klass); + rb_clear_constant_cache_for_id(id); - ce = ZALLOC(rb_const_entry_t); - rb_id_table_insert(tbl, id, (VALUE)ce); - setup_const_entry(ce, klass, val, visibility); + ce = ZALLOC(rb_const_entry_t); + rb_id_table_insert(tbl, id, (VALUE)ce); + setup_const_entry(ce, klass, val, visibility); } } static void setup_const_entry(rb_const_entry_t *ce, VALUE klass, VALUE val, - rb_const_flag_t visibility) + rb_const_flag_t visibility) { ce->flag = visibility; RB_OBJ_WRITE(klass, &ce->value, val); @@ -2734,7 +4023,10 @@ rb_define_const(VALUE klass, const char *name, VALUE val) ID id = rb_intern(name); if (!rb_is_const_id(id)) { - rb_warn("rb_define_const: invalid name `%s' for constant", name); + rb_warn("rb_define_const: invalid name '%s' for constant", name); + } + if (!RB_SPECIAL_CONST_P(val)) { + rb_vm_register_global_object(val); } rb_const_set(klass, id, val); } @@ -2747,50 +4039,41 @@ rb_define_global_const(const char *name, VALUE val) static void set_const_visibility(VALUE mod, int argc, const VALUE *argv, - rb_const_flag_t flag, rb_const_flag_t mask) + rb_const_flag_t flag, rb_const_flag_t mask) { int i; rb_const_entry_t *ce; ID id; - rb_frozen_class_p(mod); + rb_class_modify_check(mod); if (argc == 0) { - rb_warning("%"PRIsVALUE" with no argument is just ignored", - QUOTE_ID(rb_frame_callee())); - return; + rb_warning("%"PRIsVALUE" with no argument is just ignored", + QUOTE_ID(rb_frame_callee())); + return; } for (i = 0; i < argc; i++) { - VALUE val = argv[i]; - id = rb_check_id(&val); - if (!id) { - if (i > 0) { - rb_clear_constant_cache(); - } - - rb_name_err_raise("constant %2$s::%1$s not defined", - mod, val); - } - if ((ce = rb_const_lookup(mod, id))) { - ce->flag &= ~mask; - ce->flag |= flag; - if (ce->value == Qundef) { - struct autoload_data_i *ele = current_autoload_data(mod, id); - if (ele) { - ele->flag &= ~mask; - ele->flag |= flag; - } - } - } - else { - if (i > 0) { - rb_clear_constant_cache(); - } - rb_name_err_raise("constant %2$s::%1$s not defined", - mod, ID2SYM(id)); - } - } - rb_clear_constant_cache(); + struct autoload_const *ac; + VALUE val = argv[i]; + id = rb_check_id(&val); + if (!id) { + undefined_constant(mod, val); + } + if ((ce = rb_const_lookup(mod, id))) { + ce->flag &= ~mask; + ce->flag |= flag; + if (UNDEF_P(ce->value)) { + if (autoload_const_value_for_named_constant(mod, id, &ac)) { + ac->flag &= ~mask; + ac->flag |= flag; + } + } + rb_clear_constant_cache_for_id(id); + } + else { + undefined_constant(mod, ID2SYM(id)); + } + } } void @@ -2800,11 +4083,12 @@ rb_deprecate_constant(VALUE mod, const char *name) ID id; long len = strlen(name); - rb_frozen_class_p(mod); - if (!(id = rb_check_id_cstr(name, len, NULL)) || - !(ce = rb_const_lookup(mod, id))) { - rb_name_err_raise("constant %2$s::%1$s not defined", - mod, rb_fstring_new(name, len)); + rb_class_modify_check(mod); + if (!(id = rb_check_id_cstr(name, len, NULL))) { + undefined_constant(mod, rb_fstring_new(name, len)); + } + if (!(ce = rb_const_lookup(mod, id))) { + undefined_constant(mod, ID2SYM(id)); } ce->flag |= CONST_DEPRECATED; } @@ -2841,7 +4125,19 @@ rb_mod_public_constant(int argc, const VALUE *argv, VALUE obj) * call-seq: * mod.deprecate_constant(symbol, ...) => mod * - * Makes a list of existing constants deprecated. + * Makes a list of existing constants deprecated. Attempt + * to refer to them will produce a warning. + * + * module HTTP + * NotFound = Exception.new + * NOT_FOUND = NotFound # previous version of the library used this name + * + * deprecate_constant :NOT_FOUND + * end + * + * HTTP::NOT_FOUND + * # warning: constant HTTP::NOT_FOUND is deprecated + * */ VALUE @@ -2855,41 +4151,98 @@ static VALUE original_module(VALUE c) { if (RB_TYPE_P(c, T_ICLASS)) - return RBASIC(c)->klass; + return RBASIC(c)->klass; return c; } static int cvar_lookup_at(VALUE klass, ID id, st_data_t *v) { - if (!RCLASS_IV_TBL(klass)) return 0; - return st_lookup(RCLASS_IV_TBL(klass), (st_data_t)id, v); + if (RB_TYPE_P(klass, T_ICLASS)) { + if (RICLASS_IS_ORIGIN_P(klass)) { + return 0; + } + else { + // check the original module + klass = RBASIC(klass)->klass; + } + } + + VALUE n = rb_ivar_lookup(klass, id, Qundef); + if (UNDEF_P(n)) return 0; + + if (v) *v = n; + return 1; } static VALUE cvar_front_klass(VALUE klass) { - if (FL_TEST(klass, FL_SINGLETON)) { - VALUE obj = rb_ivar_get(klass, id__attached__); - if (RB_TYPE_P(obj, T_MODULE) || RB_TYPE_P(obj, T_CLASS)) { - return obj; - } + if (RCLASS_SINGLETON_P(klass)) { + VALUE obj = RCLASS_ATTACHED_OBJECT(klass); + if (rb_namespace_p(obj)) { + return obj; + } } return RCLASS_SUPER(klass); } +static void +cvar_overtaken(VALUE front, VALUE target, ID id) +{ + if (front && target != front) { + if (original_module(front) != original_module(target)) { + rb_raise(rb_eRuntimeError, + "class variable % "PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"", + ID2SYM(id), rb_class_name(original_module(front)), + rb_class_name(original_module(target))); + } + if (BUILTIN_TYPE(front) == T_CLASS) { + rb_ivar_delete(front, id, Qundef); + } + } +} + #define CVAR_FOREACH_ANCESTORS(klass, v, r) \ for (klass = cvar_front_klass(klass); klass; klass = RCLASS_SUPER(klass)) { \ - if (cvar_lookup_at(klass, id, (v))) { \ - r; \ - } \ + if (cvar_lookup_at(klass, id, (v))) { \ + r; \ + } \ } #define CVAR_LOOKUP(v,r) do {\ + CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(klass, id); \ if (cvar_lookup_at(klass, id, (v))) {r;}\ CVAR_FOREACH_ANCESTORS(klass, v, r);\ } while(0) +static VALUE +find_cvar(VALUE klass, VALUE * front, VALUE * target, ID id) +{ + VALUE v = Qundef; + CVAR_LOOKUP(&v, { + if (!*front) { + *front = klass; + } + *target = klass; + }); + + return v; +} + +static void +check_for_cvar_table(VALUE subclass, VALUE key) +{ + // Must not check ivar on ICLASS + if (!RB_TYPE_P(subclass, T_ICLASS) && RTEST(rb_ivar_defined(subclass, key))) { + RB_DEBUG_COUNTER_INC(cvar_class_invalidate); + ruby_vm_global_cvar_state++; + return; + } + + rb_class_foreach_subclass(subclass, check_for_cvar_table, key); +} + void rb_cvar_set(VALUE klass, ID id, VALUE val) { @@ -2898,59 +4251,77 @@ rb_cvar_set(VALUE klass, ID id, VALUE val) tmp = klass; CVAR_LOOKUP(0, {if (!front) front = klass; target = klass;}); if (target) { - if (front && target != front) { - st_data_t did = id; - - if (RTEST(ruby_verbose)) { - rb_warning("class variable %"PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"", - QUOTE_ID(id), rb_class_name(original_module(front)), - rb_class_name(original_module(target))); - } - if (BUILTIN_TYPE(front) == T_CLASS) { - st_delete(RCLASS_IV_TBL(front),&did,0); - } - } + cvar_overtaken(front, target, id); } else { - target = tmp; + target = tmp; } + if (RB_TYPE_P(target, T_ICLASS)) { + target = RBASIC(target)->klass; + } check_before_mod_set(target, id, val, "class variable"); - if (!RCLASS_IV_TBL(target)) { - RCLASS_IV_TBL(target) = st_init_numtable(); + + bool new_cvar = rb_class_ivar_set(target, id, val); + + struct rb_id_table *rb_cvc_tbl = RCLASS_WRITABLE_CVC_TBL(target); + + if (!rb_cvc_tbl) { + rb_cvc_tbl = rb_id_table_create(2); + RCLASS_WRITE_CVC_TBL(target, rb_cvc_tbl); + } + + struct rb_cvar_class_tbl_entry *ent; + VALUE ent_data; + + if (!rb_id_table_lookup(rb_cvc_tbl, id, &ent_data)) { + ent = ALLOC(struct rb_cvar_class_tbl_entry); + ent->class_value = target; + ent->global_cvar_state = GET_GLOBAL_CVAR_STATE(); + ent->cref = 0; + rb_id_table_insert(rb_cvc_tbl, id, (VALUE)ent); + RB_DEBUG_COUNTER_INC(cvar_inline_miss); + } + else { + ent = (void *)ent_data; + ent->global_cvar_state = GET_GLOBAL_CVAR_STATE(); } - rb_class_ivar_set(target, id, val); + // Break the cvar cache if this is a new class variable + // and target is a module or a subclass with the same + // cvar in this lookup. + if (new_cvar) { + if (RB_TYPE_P(target, T_CLASS)) { + if (RCLASS_SUBCLASSES_FIRST(target)) { + rb_class_foreach_subclass(target, check_for_cvar_table, id); + } + } + } } VALUE -rb_cvar_get(VALUE klass, ID id) +rb_cvar_find(VALUE klass, ID id, VALUE *front) { - VALUE tmp, front = 0, target = 0; - st_data_t value; + VALUE target = 0; + VALUE value; - tmp = klass; - CVAR_LOOKUP(&value, {if (!front) front = klass; target = klass;}); + value = find_cvar(klass, front, &target, id); if (!target) { - rb_name_err_raise("uninitialized class variable %1$s in %2$s", - tmp, ID2SYM(id)); - } - if (front && target != front) { - st_data_t did = id; - - if (RTEST(ruby_verbose)) { - rb_warning("class variable %"PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"", - QUOTE_ID(id), rb_class_name(original_module(front)), - rb_class_name(original_module(target))); - } - if (BUILTIN_TYPE(front) == T_CLASS) { - st_delete(RCLASS_IV_TBL(front),&did,0); - } + rb_name_err_raise("uninitialized class variable %1$s in %2$s", + klass, ID2SYM(id)); } + cvar_overtaken(*front, target, id); return (VALUE)value; } VALUE +rb_cvar_get(VALUE klass, ID id) +{ + VALUE front = 0; + return rb_cvar_find(klass, id, &front); +} + +VALUE rb_cvar_defined(VALUE klass, ID id) { if (!klass) return Qfalse; @@ -2963,8 +4334,8 @@ cv_intern(VALUE klass, const char *name) { ID id = rb_intern(name); if (!rb_is_class_id(id)) { - rb_name_err_raise("wrong class variable name %1$s", - klass, rb_str_new_cstr(name)); + rb_name_err_raise("wrong class variable name %1$s", + klass, rb_str_new_cstr(name)); } return id; } @@ -2986,18 +4357,16 @@ rb_cv_get(VALUE klass, const char *name) void rb_define_class_variable(VALUE klass, const char *name, VALUE val) { - ID id = cv_intern(klass, name); - rb_cvar_set(klass, id, val); + rb_cv_set(klass, name, val); } static int -cv_i(st_data_t k, st_data_t v, st_data_t a) +cv_i(ID key, VALUE v, st_data_t a) { - ID key = (ID)k; st_table *tbl = (st_table *)a; if (rb_is_class_id(key)) { - st_update(tbl, (st_data_t)key, cv_i_update, 0); + st_update(tbl, (st_data_t)key, cv_i_update, 0); } return ST_CONTINUE; } @@ -3007,11 +4376,11 @@ mod_cvar_at(VALUE mod, void *data) { st_table *tbl = data; if (!tbl) { - tbl = st_init_numtable(); - } - if (RCLASS_IV_TBL(mod)) { - st_foreach_safe(RCLASS_IV_TBL(mod), cv_i, (st_data_t)tbl); + tbl = st_init_numtable(); } + mod = original_module(mod); + + rb_ivar_foreach(mod, cv_i, (st_data_t)tbl); return tbl; } @@ -3019,10 +4388,16 @@ static void* mod_cvar_of(VALUE mod, void *data) { VALUE tmp = mod; + if (RCLASS_SINGLETON_P(mod)) { + if (rb_namespace_p(RCLASS_ATTACHED_OBJECT(mod))) { + data = mod_cvar_at(tmp, data); + tmp = cvar_front_klass(tmp); + } + } for (;;) { - data = mod_cvar_at(tmp, data); - tmp = RCLASS_SUPER(tmp); - if (!tmp) break; + data = mod_cvar_at(tmp, data); + tmp = RCLASS_SUPER(tmp); + if (!tmp) break; } return data; } @@ -3072,20 +4447,15 @@ cvar_list(void *data) VALUE rb_mod_class_variables(int argc, const VALUE *argv, VALUE mod) { - VALUE inherit; + bool inherit = true; st_table *tbl; - if (argc == 0) { - inherit = Qtrue; + if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]); + if (inherit) { + tbl = mod_cvar_of(mod, 0); } else { - rb_scan_args(argc, argv, "01", &inherit); - } - if (RTEST(inherit)) { - tbl = mod_cvar_of(mod, 0); - } - else { - tbl = mod_cvar_at(mod, 0); + tbl = mod_cvar_at(mod, 0); } return cvar_list(tbl); } @@ -3094,13 +4464,12 @@ rb_mod_class_variables(int argc, const VALUE *argv, VALUE mod) * call-seq: * remove_class_variable(sym) -> obj * - * Removes the definition of the <i>sym</i>, returning that - * constant's value. + * Removes the named class variable from the receiver, returning that + * variable's value. * - * class Dummy + * class Example * @@var = 99 - * puts @@var - * remove_class_variable(:@@var) + * puts remove_class_variable(:@@var) * p(defined? @@var) * end * @@ -3114,28 +4483,33 @@ VALUE rb_mod_remove_cvar(VALUE mod, VALUE name) { const ID id = id_for_var_message(mod, name, class, "wrong class variable name %1$s"); - st_data_t val, n = id; + st_data_t val; if (!id) { - not_defined: - rb_name_err_raise("class variable %1$s not defined for %2$s", - mod, name); + goto not_defined; } rb_check_frozen(mod); - if (RCLASS_IV_TBL(mod) && st_delete(RCLASS_IV_TBL(mod), &n, &val)) { - return (VALUE)val; + val = rb_ivar_delete(mod, id, Qundef); + if (!UNDEF_P(val)) { + return (VALUE)val; } if (rb_cvar_defined(mod, id)) { - rb_name_err_raise("cannot remove %1$s for %2$s", mod, ID2SYM(id)); + rb_name_err_raise("cannot remove %1$s for %2$s", mod, ID2SYM(id)); } - goto not_defined; + not_defined: + rb_name_err_raise("class variable %1$s not defined for %2$s", + mod, name); + UNREACHABLE_RETURN(Qundef); } VALUE rb_iv_get(VALUE obj, const char *name) { - ID id = rb_intern(name); + ID id = rb_check_id_cstr(name, strlen(name), rb_usascii_encoding()); + if (!id) { + return Qnil; + } return rb_ivar_get(obj, id); } @@ -3147,39 +4521,143 @@ rb_iv_set(VALUE obj, const char *name, VALUE val) return rb_ivar_set(obj, id, val); } -/* tbl = xx(obj); tbl[key] = value; */ -int -rb_class_ivar_set(VALUE obj, ID key, VALUE value) +static attr_index_t +class_fields_ivar_set(VALUE klass, VALUE fields_obj, ID id, VALUE val, bool concurrent, VALUE *new_fields_obj, bool *new_ivar_out) { - st_table *tbl = RCLASS_IV_TBL(obj); - int result = st_insert(tbl, (st_data_t)key, (st_data_t)value); - RB_OBJ_WRITTEN(obj, Qundef, value); - return result; + const VALUE original_fields_obj = fields_obj; + fields_obj = original_fields_obj ? original_fields_obj : rb_imemo_fields_new(klass, 1, true); + + shape_id_t current_shape_id = RBASIC_SHAPE_ID(fields_obj); + shape_id_t next_shape_id = current_shape_id; // for too_complex + if (UNLIKELY(rb_shape_too_complex_p(current_shape_id))) { + goto too_complex; + } + + bool new_ivar; + next_shape_id = generic_shape_ivar(fields_obj, id, &new_ivar); + + if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) { + fields_obj = imemo_fields_complex_from_obj(klass, fields_obj, next_shape_id); + goto too_complex; + } + + attr_index_t index = RSHAPE_INDEX(next_shape_id); + if (new_ivar) { + if (index >= RSHAPE_CAPACITY(current_shape_id)) { + // We allocate a new fields_obj even when concurrency isn't a concern + // so that we're embedded as long as possible. + fields_obj = imemo_fields_copy_capa(klass, fields_obj, RSHAPE_CAPACITY(next_shape_id)); + } + } + + VALUE *fields = rb_imemo_fields_ptr(fields_obj); + + if (concurrent && original_fields_obj == fields_obj) { + // In the concurrent case, if we're mutating the existing + // fields_obj, we must use an atomic write, because if we're + // adding a new field, the shape_id must be written after the field + // and if we're updating an existing field, we at least need a relaxed + // write to avoid reaping. + RB_OBJ_ATOMIC_WRITE(fields_obj, &fields[index], val); + } + else { + RB_OBJ_WRITE(fields_obj, &fields[index], val); + } + + if (new_ivar) { + RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id); + } + + *new_fields_obj = fields_obj; + *new_ivar_out = new_ivar; + return index; + +too_complex: + { + if (concurrent && fields_obj == original_fields_obj) { + // In multi-ractor case, we must always work on a copy because + // even if the field already exist, inserting in a st_table may + // cause a rebuild. + fields_obj = rb_imemo_fields_clone(fields_obj); + } + + st_table *table = rb_imemo_fields_complex_tbl(fields_obj); + new_ivar = !st_insert(table, (st_data_t)id, (st_data_t)val); + RB_OBJ_WRITTEN(fields_obj, Qundef, val); + + if (fields_obj != original_fields_obj) { + RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id); + } + } + + *new_fields_obj = fields_obj; + *new_ivar_out = new_ivar; + return ATTR_INDEX_NOT_SET; } -static int -tbl_copy_i(st_data_t key, st_data_t value, st_data_t data) +static attr_index_t +class_ivar_set(VALUE obj, ID id, VALUE val, bool *new_ivar) { - RB_OBJ_WRITTEN((VALUE)data, Qundef, (VALUE)value); - return ST_CONTINUE; + rb_class_ensure_writable(obj); + + const VALUE original_fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj); + VALUE new_fields_obj = 0; + + attr_index_t index = class_fields_ivar_set(obj, original_fields_obj, id, val, rb_multi_ractor_p(), &new_fields_obj, new_ivar); + + if (new_fields_obj != original_fields_obj) { + RCLASS_WRITABLE_SET_FIELDS_OBJ(obj, new_fields_obj); + } + + // TODO: What should we set as the T_CLASS shape_id? + // In most case we can replicate the single `fields_obj` shape + // but in namespaced case? Perhaps INVALID_SHAPE_ID? + RBASIC_SET_SHAPE_ID(obj, RBASIC_SHAPE_ID(new_fields_obj)); + return index; } -st_table * -rb_st_copy(VALUE obj, struct st_table *orig_tbl) +bool +rb_class_ivar_set(VALUE obj, ID id, VALUE val) { - st_table *new_tbl = st_copy(orig_tbl); - st_foreach(new_tbl, tbl_copy_i, (st_data_t)obj); - return new_tbl; + RUBY_ASSERT(RB_TYPE_P(obj, T_CLASS) || RB_TYPE_P(obj, T_MODULE)); + rb_check_frozen(obj); + + bool new_ivar; + class_ivar_set(obj, id, val, &new_ivar); + return new_ivar; } -rb_const_entry_t * -rb_const_lookup(VALUE klass, ID id) +void +rb_fields_tbl_copy(VALUE dst, VALUE src) { - struct rb_id_table *tbl = RCLASS_CONST_TBL(klass); - VALUE val; + RUBY_ASSERT(rb_type(dst) == rb_type(src)); + RUBY_ASSERT(RB_TYPE_P(dst, T_CLASS) || RB_TYPE_P(dst, T_MODULE)); + RUBY_ASSERT(RSHAPE_TYPE_P(RBASIC_SHAPE_ID(dst), SHAPE_ROOT)); - if (tbl && rb_id_table_lookup(tbl, id, &val)) { - return (rb_const_entry_t *)val; + VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(src); + if (fields_obj) { + RCLASS_WRITABLE_SET_FIELDS_OBJ(dst, rb_imemo_fields_clone(fields_obj)); + RBASIC_SET_SHAPE_ID(dst, RBASIC_SHAPE_ID(src)); } - return 0; +} + +static rb_const_entry_t * +const_lookup(struct rb_id_table *tbl, ID id) +{ + if (tbl) { + VALUE val; + bool r; + RB_VM_LOCKING() { + r = rb_id_table_lookup(tbl, id, &val); + } + + if (r) return (rb_const_entry_t *)val; + } + return NULL; +} + +rb_const_entry_t * +rb_const_lookup(VALUE klass, ID id) +{ + return const_lookup(RCLASS_CONST_TBL(klass), id); } |
