summaryrefslogtreecommitdiff
path: root/variable.c
diff options
context:
space:
mode:
Diffstat (limited to 'variable.c')
-rw-r--r--variable.c4702
1 files changed, 3090 insertions, 1612 deletions
diff --git a/variable.c b/variable.c
index bfb6fa2edd..ff8d24d78a 100644
--- a/variable.c
+++ b/variable.c
@@ -11,229 +11,325 @@
**********************************************************************/
-#include "internal.h"
-#include "ruby/st.h"
-#include "ruby/util.h"
-#include "id_table.h"
+#include "ruby/internal/config.h"
+#include <stddef.h>
+#include "ruby/internal/stdbool.h"
+#include "ccan/list/list.h"
#include "constant.h"
+#include "debug_counter.h"
#include "id.h"
-#include "ccan/list/list.h"
#include "id_table.h"
-#include "debug_counter.h"
+#include "internal.h"
+#include "internal/box.h"
+#include "internal/class.h"
+#include "internal/compilers.h"
+#include "internal/error.h"
+#include "internal/eval.h"
+#include "internal/hash.h"
+#include "internal/object.h"
+#include "internal/gc.h"
+#include "internal/re.h"
+#include "internal/struct.h"
+#include "internal/symbol.h"
+#include "internal/thread.h"
+#include "internal/variable.h"
+#include "ruby/encoding.h"
+#include "ruby/st.h"
+#include "ruby/util.h"
+#include "shape.h"
+#include "symbol.h"
+#include "variable.h"
#include "vm_core.h"
+#include "ractor_core.h"
+#include "vm_sync.h"
-struct rb_id_table *rb_global_tbl;
-static ID autoload, classpath, tmp_classpath, classid;
+RUBY_EXTERN rb_serial_t ruby_vm_global_cvar_state;
+#define GET_GLOBAL_CVAR_STATE() (ruby_vm_global_cvar_state)
+
+typedef void rb_gvar_compact_t(void *var);
+
+static struct rb_id_table *rb_global_tbl;
+static ID autoload;
+
+// This hash table maps file paths to loadable features. We use this to track
+// autoload state until it's no longer needed.
+// feature (file path) => struct autoload_data
+static VALUE autoload_features;
+
+// This mutex is used to protect autoloading state. We use a global mutex which
+// is held until a per-feature mutex can be created. This ensures there are no
+// race conditions relating to autoload state.
+static VALUE autoload_mutex;
static void check_before_mod_set(VALUE, ID, VALUE, const char *);
static void setup_const_entry(rb_const_entry_t *, VALUE, VALUE, rb_const_flag_t);
-static VALUE rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility);
-static st_table *generic_iv_tbl;
-static st_table *generic_iv_tbl_compat;
-
-/* per-object */
-struct gen_ivtbl {
- uint32_t numiv;
- VALUE ivptr[1]; /* flexible array */
-};
+static VALUE rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in);
+static st_table *generic_fields_tbl_;
-struct ivar_update {
- union {
- st_table *iv_index_tbl;
- struct gen_ivtbl *ivtbl;
- } u;
- st_data_t index;
- int iv_extended;
-};
+typedef int rb_ivar_foreach_callback_func(ID key, VALUE val, st_data_t arg);
+static void rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only);
void
Init_var_tables(void)
{
rb_global_tbl = rb_id_table_create(0);
- generic_iv_tbl = st_init_numtable();
+ generic_fields_tbl_ = st_init_numtable();
autoload = rb_intern_const("__autoload__");
- /* __classpath__: fully qualified class path */
- classpath = rb_intern_const("__classpath__");
- /* __tmp_classpath__: temporary class path which contains anonymous names */
- tmp_classpath = rb_intern_const("__tmp_classpath__");
- /* __classid__: name given to class/module under an anonymous namespace */
- classid = rb_intern_const("__classid__");
-}
-struct fc_result {
- ID name, preferred;
- VALUE klass;
- VALUE path;
- VALUE track;
- struct fc_result *prev;
-};
+ autoload_mutex = rb_mutex_new();
+ rb_obj_hide(autoload_mutex);
+ rb_vm_register_global_object(autoload_mutex);
-static VALUE
-fc_path(struct fc_result *fc, ID name)
-{
- VALUE path, tmp;
-
- path = rb_id2str(name);
- while (fc) {
- st_data_t n;
- if (fc->track == rb_cObject) break;
- if (RCLASS_IV_TBL(fc->track) &&
- st_lookup(RCLASS_IV_TBL(fc->track), (st_data_t)classpath, &n)) {
- tmp = rb_str_dup((VALUE)n);
- rb_str_cat2(tmp, "::");
- rb_str_append(tmp, path);
- path = tmp;
- break;
- }
- tmp = rb_str_dup(rb_id2str(fc->name));
- rb_str_cat2(tmp, "::");
- rb_str_append(tmp, path);
- path = tmp;
- fc = fc->prev;
- }
- OBJ_FREEZE(path);
- return path;
+ autoload_features = rb_ident_hash_new();
+ rb_obj_hide(autoload_features);
+ rb_vm_register_global_object(autoload_features);
}
-static enum rb_id_table_iterator_result
-fc_i(ID key, VALUE v, void *a)
+static inline bool
+rb_namespace_p(VALUE obj)
{
- rb_const_entry_t *ce = (rb_const_entry_t *)v;
- struct fc_result *res = a;
- VALUE value = ce->value;
- if (!rb_is_const_id(key)) return ID_TABLE_CONTINUE;
-
- if (value == res->klass && (!res->preferred || key == res->preferred)) {
- res->path = fc_path(res, key);
- return ID_TABLE_STOP;
- }
- if (RB_TYPE_P(value, T_MODULE) || RB_TYPE_P(value, T_CLASS)) {
- if (!RCLASS_CONST_TBL(value)) return ID_TABLE_CONTINUE;
- else {
- struct fc_result arg;
- struct fc_result *list;
-
- list = res;
- while (list) {
- if (list->track == value) return ID_TABLE_CONTINUE;
- list = list->prev;
- }
-
- arg.name = key;
- arg.preferred = res->preferred;
- arg.path = 0;
- arg.klass = res->klass;
- arg.track = value;
- arg.prev = res;
- rb_id_table_foreach(RCLASS_CONST_TBL(value), fc_i, &arg);
- if (arg.path) {
- res->path = arg.path;
- return ID_TABLE_STOP;
- }
- }
+ if (RB_SPECIAL_CONST_P(obj)) return false;
+ switch (RB_BUILTIN_TYPE(obj)) {
+ case T_MODULE: case T_CLASS: return true;
+ default: break;
}
- return ID_TABLE_CONTINUE;
+ return false;
}
/**
- * Traverse constant namespace and find +classpath+ for _klass_. If
- * _preferred_ is not 0, choice the path whose base name is set to it.
- * If +classpath+ is found, the hidden instance variable __classpath__
- * is set to the found path, and __tmp_classpath__ is removed.
- * The path is frozen.
+ * Returns +classpath+ of _klass_, if it is named, or +nil+ for
+ * anonymous +class+/+module+. A named +classpath+ may contain
+ * an anonymous component, but the last component is guaranteed
+ * to not be anonymous. <code>*permanent</code> is set to 1
+ * if +classpath+ has no anonymous components. There is no builtin
+ * Ruby level APIs that can change a permanent +classpath+.
+ *
+ * YJIT needs this function to not allocate.
*/
static VALUE
-find_class_path(VALUE klass, ID preferred)
+classname(VALUE klass, bool *permanent)
+{
+ *permanent = false;
+
+ VALUE classpath = RCLASS_CLASSPATH(klass);
+ if (classpath == 0) return Qnil;
+
+ *permanent = RCLASS_PERMANENT_CLASSPATH_P(klass);
+
+ return classpath;
+}
+
+VALUE
+rb_mod_name0(VALUE klass, bool *permanent)
+{
+ return classname(klass, permanent);
+}
+
+/*
+ * call-seq:
+ * mod.name -> string or nil
+ *
+ * Returns the name of the module <i>mod</i>. Returns +nil+ for anonymous modules.
+ */
+
+VALUE
+rb_mod_name(VALUE mod)
{
- struct fc_result arg;
+ // YJIT needs this function to not allocate.
+ bool permanent;
+ return classname(mod, &permanent);
+}
+
+// Similar to logic in rb_mod_const_get().
+static bool
+is_constant_path(VALUE name)
+{
+ const char *path = RSTRING_PTR(name);
+ const char *pend = RSTRING_END(name);
+ rb_encoding *enc = rb_enc_get(name);
- arg.preferred = preferred;
- arg.name = 0;
- arg.path = 0;
- arg.klass = klass;
- arg.track = rb_cObject;
- arg.prev = 0;
- if (RCLASS_CONST_TBL(rb_cObject)) {
- rb_id_table_foreach(RCLASS_CONST_TBL(rb_cObject), fc_i, &arg);
+ const char *p = path;
+
+ if (p >= pend || !*p) {
+ return false;
}
- if (arg.path) {
- st_data_t tmp = tmp_classpath;
- if (!RCLASS_IV_TBL(klass)) {
- RCLASS_IV_TBL(klass) = st_init_numtable();
- }
- rb_class_ivar_set(klass, classpath, arg.path);
- st_delete(RCLASS_IV_TBL(klass), &tmp, 0);
- return arg.path;
+ while (p < pend) {
+ if (p + 2 <= pend && p[0] == ':' && p[1] == ':') {
+ p += 2;
+ }
+
+ const char *pbeg = p;
+ while (p < pend && *p != ':') p++;
+
+ if (pbeg == p) return false;
+
+ if (rb_enc_symname_type(pbeg, p - pbeg, enc, 0) != ID_CONST) {
+ return false;
+ }
}
- return Qnil;
+
+ return true;
+}
+
+struct sub_temporary_name_args {
+ VALUE names;
+ ID last;
+};
+
+static VALUE build_const_path(VALUE head, ID tail);
+static void set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name);
+
+static VALUE
+set_sub_temporary_name_recursive(VALUE mod, VALUE data, int recursive)
+{
+ if (recursive) return Qfalse;
+
+ struct sub_temporary_name_args *args = (void *)data;
+ VALUE name = 0;
+ if (args->names) {
+ name = build_const_path(rb_ary_last(0, 0, args->names), args->last);
+ }
+ set_sub_temporary_name_foreach(mod, args, name);
+ return Qtrue;
}
-/**
- * Returns +classpath+ of _klass_, if it is named, or +nil+ for
- * anonymous +class+/+module+. The last part of named +classpath+ is
- * never anonymous, but anonymous +class+/+module+ names may be
- * contained. If the path is "permanent", that means it has no
- * anonymous names, <code>*permanent</code> is set to 1.
- */
static VALUE
-classname(VALUE klass, int *permanent)
-{
- VALUE path = Qnil;
- st_data_t n;
-
- if (!klass) klass = rb_cObject;
- *permanent = 1;
- if (RCLASS_IV_TBL(klass)) {
- if (!st_lookup(RCLASS_IV_TBL(klass), (st_data_t)classpath, &n)) {
- ID cid = 0;
- if (st_lookup(RCLASS_IV_TBL(klass), (st_data_t)classid, &n)) {
- VALUE cname = (VALUE)n;
- cid = rb_check_id(&cname);
- if (cid) path = find_class_path(klass, cid);
- }
- if (NIL_P(path)) {
- path = find_class_path(klass, (ID)0);
- }
- if (NIL_P(path)) {
- if (!cid) {
- return Qnil;
- }
- if (!st_lookup(RCLASS_IV_TBL(klass), (st_data_t)tmp_classpath, &n)) {
- path = rb_id2str(cid);
- return path;
- }
- *permanent = 0;
- path = (VALUE)n;
- return path;
- }
- }
- else {
- path = (VALUE)n;
- }
- if (!RB_TYPE_P(path, T_STRING)) {
- rb_bug("class path is not set properly");
- }
- return path;
- }
- return find_class_path(klass, (ID)0);
+set_sub_temporary_name_topmost(VALUE mod, VALUE data, int recursive)
+{
+ if (recursive) return Qfalse;
+
+ struct sub_temporary_name_args *args = (void *)data;
+ VALUE name = args->names;
+ if (name) {
+ args->names = rb_ary_hidden_new(0);
+ }
+ set_sub_temporary_name_foreach(mod, args, name);
+ return Qtrue;
+}
+
+static enum rb_id_table_iterator_result
+set_sub_temporary_name_i(ID id, VALUE val, void *data)
+{
+ val = ((rb_const_entry_t *)val)->value;
+ if (rb_namespace_p(val) && !RCLASS_PERMANENT_CLASSPATH_P(val)) {
+ VALUE arg = (VALUE)data;
+ struct sub_temporary_name_args *args = data;
+ args->last = id;
+ rb_exec_recursive_paired(set_sub_temporary_name_recursive, val, arg, arg);
+ }
+ return ID_TABLE_CONTINUE;
+}
+
+static void
+set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name)
+{
+ RCLASS_WRITE_CLASSPATH(mod, name, FALSE);
+ struct rb_id_table *tbl = RCLASS_CONST_TBL(mod);
+ if (!tbl) return;
+ if (!name) {
+ rb_id_table_foreach(tbl, set_sub_temporary_name_i, args);
+ }
+ else {
+ long names_len = RARRAY_LEN(args->names); // paranoiac check?
+ rb_ary_push(args->names, name);
+ rb_id_table_foreach(tbl, set_sub_temporary_name_i, args);
+ rb_ary_set_len(args->names, names_len);
+ }
+}
+
+static void
+set_sub_temporary_name(VALUE mod, VALUE name)
+{
+ struct sub_temporary_name_args args = {name};
+ VALUE arg = (VALUE)&args;
+ rb_exec_recursive_paired(set_sub_temporary_name_topmost, mod, arg, arg);
}
/*
* call-seq:
- * mod.name -> string
+ * mod.set_temporary_name(string) -> self
+ * mod.set_temporary_name(nil) -> self
+ *
+ * Sets the temporary name of the module. This name is reflected in
+ * introspection of the module and the values that are related to it, such
+ * as instances, constants, and methods.
+ *
+ * The name should be +nil+ or a non-empty string that is not a valid constant
+ * path (to avoid confusing between permanent and temporary names).
+ *
+ * The method can be useful to distinguish dynamically generated classes and
+ * modules without assigning them to constants.
+ *
+ * If the module is given a permanent name by assigning it to a constant,
+ * the temporary name is discarded. A temporary name can't be assigned to
+ * modules that have a permanent name.
+ *
+ * If the given name is +nil+, the module becomes anonymous again.
+ *
+ * Example:
+ *
+ * m = Module.new # => #<Module:0x0000000102c68f38>
+ * m.name #=> nil
+ *
+ * m.set_temporary_name("fake_name") # => fake_name
+ * m.name #=> "fake_name"
+ *
+ * m.set_temporary_name(nil) # => #<Module:0x0000000102c68f38>
+ * m.name #=> nil
*
- * Returns the name of the module <i>mod</i>. Returns nil for anonymous modules.
+ * c = Class.new
+ * c.set_temporary_name("MyClass(with description)") # => MyClass(with description)
+ *
+ * c.new # => #<MyClass(with description):0x0....>
+ *
+ * c::M = m
+ * c::M.name #=> "MyClass(with description)::M"
+ *
+ * # Assigning to a constant replaces the name with a permanent one
+ * C = c
+ *
+ * C.name #=> "C"
+ * C::M.name #=> "C::M"
+ * c.new # => #<C:0x0....>
*/
VALUE
-rb_mod_name(VALUE mod)
+rb_mod_set_temporary_name(VALUE mod, VALUE name)
{
- int permanent;
- VALUE path = classname(mod, &permanent);
+ // We don't allow setting the name if the classpath is already permanent:
+ if (RCLASS_PERMANENT_CLASSPATH_P(mod)) {
+ rb_raise(rb_eRuntimeError, "can't change permanent name");
+ }
- if (!NIL_P(path)) return rb_str_dup(path);
- return path;
+ if (NIL_P(name)) {
+ // Set the temporary classpath to NULL (anonymous):
+ RB_VM_LOCKING() {
+ set_sub_temporary_name(mod, 0);
+ }
+ }
+ else {
+ // Ensure the name is a string:
+ StringValue(name);
+
+ if (RSTRING_LEN(name) == 0) {
+ rb_raise(rb_eArgError, "empty class/module name");
+ }
+
+ if (is_constant_path(name)) {
+ rb_raise(rb_eArgError, "the temporary name must not be a constant path to avoid confusion");
+ }
+
+ name = rb_str_new_frozen(name);
+ RB_OBJ_SET_SHAREABLE(name);
+
+ // Set the temporary classpath to the given name:
+ RB_VM_LOCKING() {
+ set_sub_temporary_name(mod, name);
+ }
+ }
+
+ return mod;
}
static VALUE
@@ -242,150 +338,111 @@ make_temporary_path(VALUE obj, VALUE klass)
VALUE path;
switch (klass) {
case Qnil:
- path = rb_sprintf("#<Class:%p>", (void*)obj);
- break;
+ path = rb_sprintf("#<Class:%p>", (void*)obj);
+ break;
case Qfalse:
- path = rb_sprintf("#<Module:%p>", (void*)obj);
- break;
+ path = rb_sprintf("#<Module:%p>", (void*)obj);
+ break;
default:
- path = rb_sprintf("#<%"PRIsVALUE":%p>", klass, (void*)obj);
- break;
+ path = rb_sprintf("#<%"PRIsVALUE":%p>", klass, (void*)obj);
+ break;
}
OBJ_FREEZE(path);
return path;
}
-typedef VALUE (*path_cache_func)(VALUE obj, VALUE name);
+typedef VALUE (*fallback_func)(VALUE obj, VALUE name);
static VALUE
-rb_tmp_class_path(VALUE klass, int *permanent, path_cache_func cache_path)
+rb_tmp_class_path(VALUE klass, bool *permanent, fallback_func fallback)
{
VALUE path = classname(klass, permanent);
- st_data_t n = (st_data_t)path;
if (!NIL_P(path)) {
- return path;
- }
- if (RCLASS_IV_TBL(klass) && st_lookup(RCLASS_IV_TBL(klass),
- (st_data_t)tmp_classpath, &n)) {
- *permanent = 0;
- return (VALUE)n;
+ return path;
}
- else {
- if (RB_TYPE_P(klass, T_MODULE)) {
- if (rb_obj_class(klass) == rb_cModule) {
- path = Qfalse;
- }
- else {
- int perm;
- path = rb_tmp_class_path(RBASIC(klass)->klass, &perm, cache_path);
- }
- }
- *permanent = 0;
- return cache_path(klass, path);
+
+ if (RB_TYPE_P(klass, T_MODULE)) {
+ if (rb_obj_class(klass) == rb_cModule) {
+ path = Qfalse;
+ }
+ else {
+ bool perm;
+ path = rb_tmp_class_path(RBASIC(klass)->klass, &perm, fallback);
+ }
}
-}
-static VALUE
-ivar_cache(VALUE obj, VALUE name)
-{
- return rb_ivar_set(obj, tmp_classpath, make_temporary_path(obj, name));
+ *permanent = false;
+ return fallback(klass, path);
}
VALUE
rb_class_path(VALUE klass)
{
- int permanent;
- VALUE path = rb_tmp_class_path(klass, &permanent, ivar_cache);
+ bool permanent;
+ VALUE path = rb_tmp_class_path(klass, &permanent, make_temporary_path);
if (!NIL_P(path)) path = rb_str_dup(path);
return path;
}
-static VALUE
-null_cache(VALUE obj, VALUE name)
+VALUE
+rb_class_path_cached(VALUE klass)
{
- return make_temporary_path(obj, name);
+ return rb_mod_name(klass);
}
-VALUE
-rb_class_path_no_cache(VALUE klass)
+static VALUE
+no_fallback(VALUE obj, VALUE name)
{
- int permanent;
- VALUE path = rb_tmp_class_path(klass, &permanent, null_cache);
- if (!NIL_P(path)) path = rb_str_dup(path);
- return path;
+ return name;
}
VALUE
-rb_class_path_cached(VALUE klass)
+rb_search_class_path(VALUE klass)
{
- st_table *ivtbl;
- st_data_t n;
-
- if (!RCLASS_EXT(klass)) return Qnil;
- if (!(ivtbl = RCLASS_IV_TBL(klass))) return Qnil;
- if (st_lookup(ivtbl, (st_data_t)classpath, &n)) return (VALUE)n;
- if (st_lookup(ivtbl, (st_data_t)tmp_classpath, &n)) return (VALUE)n;
- return Qnil;
+ bool permanent;
+ return rb_tmp_class_path(klass, &permanent, no_fallback);
}
static VALUE
-never_cache(VALUE obj, VALUE name)
+build_const_pathname(VALUE head, VALUE tail)
{
- return name;
+ VALUE path = rb_str_dup(head);
+ rb_str_cat2(path, "::");
+ rb_str_append(path, tail);
+ return rb_fstring(path);
}
-VALUE
-rb_search_class_path(VALUE klass)
+static VALUE
+build_const_path(VALUE head, ID tail)
{
- int permanent;
- return rb_tmp_class_path(klass, &permanent, never_cache);
+ return build_const_pathname(head, rb_id2str(tail));
}
void
rb_set_class_path_string(VALUE klass, VALUE under, VALUE name)
{
- VALUE str;
- ID pathid = classpath;
+ bool permanent = true;
+ VALUE str;
if (under == rb_cObject) {
- str = rb_str_new_frozen(name);
+ str = rb_str_new_frozen(name);
}
else {
- int permanent;
- str = rb_str_dup(rb_tmp_class_path(under, &permanent, ivar_cache));
- rb_str_cat2(str, "::");
- rb_str_append(str, name);
- OBJ_FREEZE(str);
- if (!permanent) {
- pathid = tmp_classpath;
- rb_ivar_set(klass, classid, rb_str_intern(name));
- }
+ str = rb_tmp_class_path(under, &permanent, make_temporary_path);
+ str = build_const_pathname(str, name);
}
- rb_ivar_set(klass, pathid, str);
+
+ RB_OBJ_SET_SHAREABLE(str);
+ RCLASS_SET_CLASSPATH(klass, str, permanent);
}
void
rb_set_class_path(VALUE klass, VALUE under, const char *name)
{
- VALUE str;
- ID pathid = classpath;
-
- if (under == rb_cObject) {
- str = rb_str_new2(name);
- }
- else {
- int permanent;
- str = rb_str_dup(rb_tmp_class_path(under, &permanent, ivar_cache));
- rb_str_cat2(str, "::");
- rb_str_cat2(str, name);
- if (!permanent) {
- pathid = tmp_classpath;
- rb_ivar_set(klass, classid, rb_str_intern(rb_str_new_cstr(name)));
- }
- }
+ VALUE str = rb_str_new2(name);
OBJ_FREEZE(str);
- rb_ivar_set(klass, pathid, str);
+ rb_set_class_path_string(klass, under, str);
}
VALUE
@@ -397,37 +454,40 @@ rb_path_to_class(VALUE pathname)
VALUE c = rb_cObject;
if (!rb_enc_asciicompat(enc)) {
- rb_raise(rb_eArgError, "invalid class path encoding (non ASCII)");
+ rb_raise(rb_eArgError, "invalid class path encoding (non ASCII)");
}
pbeg = p = path;
pend = path + RSTRING_LEN(pathname);
if (path == pend || path[0] == '#') {
- rb_raise(rb_eArgError, "can't retrieve anonymous class %"PRIsVALUE,
- QUOTE(pathname));
+ rb_raise(rb_eArgError, "can't retrieve anonymous class %"PRIsVALUE,
+ QUOTE(pathname));
}
while (p < pend) {
- while (p < pend && *p != ':') p++;
- id = rb_check_id_cstr(pbeg, p-pbeg, enc);
- if (p < pend && p[0] == ':') {
- if ((size_t)(pend - p) < 2 || p[1] != ':') goto undefined_class;
- p += 2;
- pbeg = p;
- }
- if (!id) {
- undefined_class:
- rb_raise(rb_eArgError, "undefined class/module % "PRIsVALUE,
- rb_str_subseq(pathname, 0, p-path));
- }
- c = rb_const_search(c, id, TRUE, FALSE, FALSE);
- if (c == Qundef) goto undefined_class;
- if (!RB_TYPE_P(c, T_MODULE) && !RB_TYPE_P(c, T_CLASS)) {
- rb_raise(rb_eTypeError, "%"PRIsVALUE" does not refer to class/module",
- pathname);
- }
+ while (p < pend && *p != ':') p++;
+ id = rb_check_id_cstr(pbeg, p-pbeg, enc);
+ if (p < pend && p[0] == ':') {
+ if ((size_t)(pend - p) < 2 || p[1] != ':') goto undefined_class;
+ p += 2;
+ pbeg = p;
+ }
+ if (!id) {
+ goto undefined_class;
+ }
+ c = rb_const_search(c, id, TRUE, FALSE, FALSE, NULL);
+ if (UNDEF_P(c)) goto undefined_class;
+ if (!rb_namespace_p(c)) {
+ rb_raise(rb_eTypeError, "%"PRIsVALUE" does not refer to class/module",
+ pathname);
+ }
}
RB_GC_GUARD(pathname);
return c;
+
+ undefined_class:
+ rb_raise(rb_eArgError, "undefined class/module % "PRIsVALUE,
+ rb_str_subseq(pathname, 0, p-path));
+ UNREACHABLE_RETURN(Qundef);
}
VALUE
@@ -436,12 +496,6 @@ rb_path2class(const char *path)
return rb_path_to_class(rb_str_new_cstr(path));
}
-void
-rb_name_class(VALUE klass, ID id)
-{
- rb_ivar_set(klass, classid, ID2SYM(id));
-}
-
VALUE
rb_class_name(VALUE klass)
{
@@ -451,8 +505,8 @@ rb_class_name(VALUE klass)
const char *
rb_class2name(VALUE klass)
{
- int permanent;
- VALUE path = rb_tmp_class_path(rb_class_real(klass), &permanent, ivar_cache);
+ bool permanent;
+ VALUE path = rb_tmp_class_path(rb_class_real(klass), &permanent, make_temporary_path);
if (NIL_P(path)) return NULL;
return RSTRING_PTR(path);
}
@@ -473,55 +527,162 @@ struct trace_var {
struct rb_global_variable {
int counter;
int block_trace;
- void *data;
+ VALUE *data;
rb_gvar_getter_t *getter;
rb_gvar_setter_t *setter;
rb_gvar_marker_t *marker;
+ rb_gvar_compact_t *compactor;
struct trace_var *trace;
+ bool box_ready;
};
-struct rb_global_entry*
-rb_global_entry(ID id)
+struct rb_global_entry {
+ struct rb_global_variable *var;
+ ID id;
+ bool ractor_local;
+};
+
+static void
+free_global_variable(struct rb_global_variable *var)
+{
+ RUBY_ASSERT(var->counter == 0);
+
+ struct trace_var *trace = var->trace;
+ while (trace) {
+ struct trace_var *next = trace->next;
+ xfree(trace);
+ trace = next;
+ }
+ xfree(var);
+}
+
+static enum rb_id_table_iterator_result
+free_global_entry_i(VALUE val, void *arg)
+{
+ struct rb_global_entry *entry = (struct rb_global_entry *)val;
+ entry->var->counter--;
+ if (entry->var->counter == 0) {
+ free_global_variable(entry->var);
+ }
+ ruby_xfree(entry);
+ return ID_TABLE_DELETE;
+}
+
+void
+rb_free_rb_global_tbl(void)
+{
+ rb_id_table_foreach_values(rb_global_tbl, free_global_entry_i, 0);
+ rb_id_table_free(rb_global_tbl);
+}
+
+void
+rb_free_generic_fields_tbl_(void)
+{
+ st_free_table(generic_fields_tbl_);
+}
+
+static struct rb_global_entry*
+rb_find_global_entry(ID id)
{
struct rb_global_entry *entry;
VALUE data;
- if (!rb_id_table_lookup(rb_global_tbl, id, &data)) {
- struct rb_global_variable *var;
- entry = ALLOC(struct rb_global_entry);
- var = ALLOC(struct rb_global_variable);
- entry->id = id;
- entry->var = var;
- var->counter = 1;
- var->data = 0;
- var->getter = rb_gvar_undef_getter;
- var->setter = rb_gvar_undef_setter;
- var->marker = rb_gvar_undef_marker;
-
- var->block_trace = 0;
- var->trace = 0;
- rb_id_table_insert(rb_global_tbl, id, (VALUE)entry);
+ RB_VM_LOCKING() {
+ if (!rb_id_table_lookup(rb_global_tbl, id, &data)) {
+ entry = NULL;
+ }
+ else {
+ entry = (struct rb_global_entry *)data;
+ RUBY_ASSERT(entry != NULL);
+ }
}
- else {
- entry = (struct rb_global_entry *)data;
+
+ if (UNLIKELY(!rb_ractor_main_p()) && (!entry || !entry->ractor_local)) {
+ rb_raise(rb_eRactorIsolationError, "can not access global variable %s from non-main Ractor", rb_id2name(id));
+ }
+
+ return entry;
+}
+
+void
+rb_gvar_ractor_local(const char *name)
+{
+ struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name));
+ entry->ractor_local = true;
+}
+
+void
+rb_gvar_box_ready(const char *name)
+{
+ struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name));
+ entry->var->box_ready = true;
+}
+
+static void
+rb_gvar_undef_compactor(void *var)
+{
+}
+
+static struct rb_global_entry*
+rb_global_entry(ID id)
+{
+ struct rb_global_entry *entry;
+ RB_VM_LOCKING() {
+ entry = rb_find_global_entry(id);
+ if (!entry) {
+ struct rb_global_variable *var;
+ entry = ALLOC(struct rb_global_entry);
+ var = ALLOC(struct rb_global_variable);
+ entry->id = id;
+ entry->var = var;
+ entry->ractor_local = false;
+ var->counter = 1;
+ var->data = 0;
+ var->getter = rb_gvar_undef_getter;
+ var->setter = rb_gvar_undef_setter;
+ var->marker = rb_gvar_undef_marker;
+ var->compactor = rb_gvar_undef_compactor;
+
+ var->block_trace = 0;
+ var->trace = 0;
+ var->box_ready = false;
+ rb_id_table_insert(rb_global_tbl, id, (VALUE)entry);
+ }
}
return entry;
}
VALUE
-rb_gvar_undef_getter(ID id, void *data, struct rb_global_variable *var)
+rb_gvar_undef_getter(ID id, VALUE *_)
{
- rb_warning("global variable `%"PRIsVALUE"' not initialized", QUOTE_ID(id));
+ rb_warning("global variable '%"PRIsVALUE"' not initialized", QUOTE_ID(id));
return Qnil;
}
+static void
+rb_gvar_val_compactor(void *_var)
+{
+ struct rb_global_variable *var = (struct rb_global_variable *)_var;
+
+ VALUE obj = (VALUE)var->data;
+
+ if (obj) {
+ VALUE new = rb_gc_location(obj);
+ if (new != obj) {
+ var->data = (void*)new;
+ }
+ }
+}
+
void
-rb_gvar_undef_setter(VALUE val, ID id, void *d, struct rb_global_variable *var)
+rb_gvar_undef_setter(VALUE val, ID id, VALUE *_)
{
+ struct rb_global_variable *var = rb_global_entry(id)->var;
var->getter = rb_gvar_val_getter;
var->setter = rb_gvar_val_setter;
var->marker = rb_gvar_val_marker;
+ var->compactor = rb_gvar_val_compactor;
var->data = (void*)val;
}
@@ -532,14 +693,15 @@ rb_gvar_undef_marker(VALUE *var)
}
VALUE
-rb_gvar_val_getter(ID id, void *data, struct rb_global_variable *var)
+rb_gvar_val_getter(ID id, VALUE *data)
{
return (VALUE)data;
}
void
-rb_gvar_val_setter(VALUE val, ID id, void *data, struct rb_global_variable *var)
+rb_gvar_val_setter(VALUE val, ID id, VALUE *_)
{
+ struct rb_global_variable *var = rb_global_entry(id)->var;
var->data = (void*)val;
}
@@ -547,21 +709,20 @@ void
rb_gvar_val_marker(VALUE *var)
{
VALUE data = (VALUE)var;
- if (data) rb_gc_mark_maybe(data);
+ if (data) rb_gc_mark_movable(data);
}
VALUE
-rb_gvar_var_getter(ID id, void *data, struct rb_global_variable *gvar)
+rb_gvar_var_getter(ID id, VALUE *var)
{
- VALUE *var = data;
if (!var) return Qnil;
return *var;
}
void
-rb_gvar_var_setter(VALUE val, ID id, void *data, struct rb_global_variable *g)
+rb_gvar_var_setter(VALUE val, ID id, VALUE *data)
{
- *(VALUE *)data = val;
+ *data = val;
}
void
@@ -571,7 +732,7 @@ rb_gvar_var_marker(VALUE *var)
}
void
-rb_gvar_readonly_setter(VALUE v, ID id, void *d, struct rb_global_variable *g)
+rb_gvar_readonly_setter(VALUE v, ID id, VALUE *_)
{
rb_name_error(id, "%"PRIsVALUE" is a read-only variable", QUOTE_ID(id));
}
@@ -586,17 +747,35 @@ mark_global_entry(VALUE v, void *ignored)
(*var->marker)(var->data);
trace = var->trace;
while (trace) {
- if (trace->data) rb_gc_mark_maybe(trace->data);
- trace = trace->next;
+ if (trace->data) rb_gc_mark_maybe(trace->data);
+ trace = trace->next;
}
return ID_TABLE_CONTINUE;
}
+#define gc_mark_table(task) \
+ if (rb_global_tbl) { rb_id_table_foreach_values(rb_global_tbl, task##_global_entry, 0); }
+
void
rb_gc_mark_global_tbl(void)
{
- if (rb_global_tbl)
- rb_id_table_foreach_values(rb_global_tbl, mark_global_entry, 0);
+ gc_mark_table(mark);
+}
+
+static enum rb_id_table_iterator_result
+update_global_entry(VALUE v, void *ignored)
+{
+ struct rb_global_entry *entry = (struct rb_global_entry *)v;
+ struct rb_global_variable *var = entry->var;
+
+ (*var->compactor)(var);
+ return ID_TABLE_CONTINUE;
+}
+
+void
+rb_gc_update_global_tbl(void)
+{
+ gc_mark_table(update);
}
static ID
@@ -606,21 +785,44 @@ global_id(const char *name)
if (name[0] == '$') id = rb_intern(name);
else {
- size_t len = strlen(name);
- char *buf = ALLOCA_N(char, len+1);
- buf[0] = '$';
- memcpy(buf+1, name, len);
- id = rb_intern2(buf, len+1);
+ size_t len = strlen(name);
+ VALUE vbuf = 0;
+ char *buf = ALLOCV_N(char, vbuf, len+1);
+ buf[0] = '$';
+ memcpy(buf+1, name, len);
+ id = rb_intern2(buf, len+1);
+ ALLOCV_END(vbuf);
}
return id;
}
+static ID
+find_global_id(const char *name)
+{
+ ID id;
+ size_t len = strlen(name);
+
+ if (name[0] == '$') {
+ id = rb_check_id_cstr(name, len, NULL);
+ }
+ else {
+ VALUE vbuf = 0;
+ char *buf = ALLOCV_N(char, vbuf, len+1);
+ buf[0] = '$';
+ memcpy(buf+1, name, len);
+ id = rb_check_id_cstr(buf, len+1, NULL);
+ ALLOCV_END(vbuf);
+ }
+
+ return id;
+}
+
void
rb_define_hooked_variable(
const char *name,
VALUE *var,
- VALUE (*getter)(ANYARGS),
- void (*setter)(ANYARGS))
+ rb_gvar_getter_t *getter,
+ rb_gvar_setter_t *setter)
{
volatile VALUE tmp = var ? *var : Qnil;
ID id = global_id(name);
@@ -649,8 +851,8 @@ rb_define_readonly_variable(const char *name, const VALUE *var)
void
rb_define_virtual_variable(
const char *name,
- VALUE (*getter)(ANYARGS),
- void (*setter)(ANYARGS))
+ rb_gvar_getter_t *getter,
+ rb_gvar_setter_t *setter)
{
if (!getter) getter = rb_gvar_val_getter;
if (!setter) setter = rb_gvar_readonly_setter;
@@ -660,32 +862,9 @@ rb_define_virtual_variable(
static void
rb_trace_eval(VALUE cmd, VALUE val)
{
- rb_eval_cmd(cmd, rb_ary_new3(1, val), 0);
+ rb_eval_cmd_call_kw(cmd, 1, &val, RB_NO_KEYWORDS);
}
-/*
- * call-seq:
- * trace_var(symbol, cmd ) -> nil
- * trace_var(symbol) {|val| block } -> nil
- *
- * Controls tracing of assignments to global variables. The parameter
- * +symbol+ identifies the variable (as either a string name or a
- * symbol identifier). _cmd_ (which may be a string or a
- * +Proc+ object) or block is executed whenever the variable
- * is assigned. The block or +Proc+ object receives the
- * variable's new value as a parameter. Also see
- * <code>Kernel::untrace_var</code>.
- *
- * trace_var :$_, proc {|v| puts "$_ is now '#{v}'" }
- * $_ = "hello"
- * $_ = ' there'
- *
- * <em>produces:</em>
- *
- * $_ is now 'hello'
- * $_ is now ' there'
- */
-
VALUE
rb_f_trace_var(int argc, const VALUE *argv)
{
@@ -694,15 +873,12 @@ rb_f_trace_var(int argc, const VALUE *argv)
struct trace_var *trace;
if (rb_scan_args(argc, argv, "11", &var, &cmd) == 1) {
- cmd = rb_block_proc();
+ cmd = rb_block_proc();
}
if (NIL_P(cmd)) {
- return rb_f_untrace_var(argc, argv);
+ return rb_f_untrace_var(argc, argv);
}
entry = rb_global_entry(rb_to_id(var));
- if (OBJ_TAINTED(cmd)) {
- rb_raise(rb_eSecurityError, "Insecure: tainted variable trace");
- }
trace = ALLOC(struct trace_var);
trace->next = entry->var->trace;
trace->func = rb_trace_eval;
@@ -723,28 +899,18 @@ remove_trace(struct rb_global_variable *var)
t.next = trace;
trace = &t;
while (trace->next) {
- next = trace->next;
- if (next->removed) {
- trace->next = next->next;
- xfree(next);
- }
- else {
- trace = next;
- }
+ next = trace->next;
+ if (next->removed) {
+ trace->next = next->next;
+ xfree(next);
+ }
+ else {
+ trace = next;
+ }
}
var->trace = t.next;
}
-/*
- * call-seq:
- * untrace_var(symbol [, cmd] ) -> array or nil
- *
- * Removes tracing for the specified command on the given global
- * variable and returns +nil+. If no command is specified,
- * removes all tracing for that variable and returns an array
- * containing the commands actually removed.
- */
-
VALUE
rb_f_untrace_var(int argc, const VALUE *argv)
{
@@ -752,117 +918,194 @@ rb_f_untrace_var(int argc, const VALUE *argv)
ID id;
struct rb_global_entry *entry;
struct trace_var *trace;
- VALUE data;
rb_scan_args(argc, argv, "11", &var, &cmd);
id = rb_check_id(&var);
if (!id) {
- rb_name_error_str(var, "undefined global variable %"PRIsVALUE"", QUOTE(var));
+ rb_name_error_str(var, "undefined global variable %"PRIsVALUE"", QUOTE(var));
}
- if (!rb_id_table_lookup(rb_global_tbl, id, &data)) {
- rb_name_error(id, "undefined global variable %"PRIsVALUE"", QUOTE_ID(id));
+ if ((entry = rb_find_global_entry(id)) == NULL) {
+ rb_name_error(id, "undefined global variable %"PRIsVALUE"", QUOTE_ID(id));
}
- trace = (entry = (struct rb_global_entry *)data)->var->trace;
+ trace = entry->var->trace;
if (NIL_P(cmd)) {
- VALUE ary = rb_ary_new();
+ VALUE ary = rb_ary_new();
- while (trace) {
- struct trace_var *next = trace->next;
- rb_ary_push(ary, (VALUE)trace->data);
- trace->removed = 1;
- trace = next;
- }
+ while (trace) {
+ struct trace_var *next = trace->next;
+ rb_ary_push(ary, (VALUE)trace->data);
+ trace->removed = 1;
+ trace = next;
+ }
- if (!entry->var->block_trace) remove_trace(entry->var);
- return ary;
+ if (!entry->var->block_trace) remove_trace(entry->var);
+ return ary;
}
else {
- while (trace) {
- if (trace->data == cmd) {
- trace->removed = 1;
- if (!entry->var->block_trace) remove_trace(entry->var);
- return rb_ary_new3(1, cmd);
- }
- trace = trace->next;
- }
+ while (trace) {
+ if (trace->data == cmd) {
+ trace->removed = 1;
+ if (!entry->var->block_trace) remove_trace(entry->var);
+ return rb_ary_new3(1, cmd);
+ }
+ trace = trace->next;
+ }
}
return Qnil;
}
-VALUE
-rb_gvar_get(struct rb_global_entry *entry)
-{
- struct rb_global_variable *var = entry->var;
- return (*var->getter)(entry->id, var->data, var);
-}
-
struct trace_data {
struct trace_var *trace;
VALUE val;
};
static VALUE
-trace_ev(struct trace_data *data)
+trace_ev(VALUE v)
{
+ struct trace_data *data = (void *)v;
struct trace_var *trace = data->trace;
while (trace) {
- (*trace->func)(trace->data, data->val);
- trace = trace->next;
+ (*trace->func)(trace->data, data->val);
+ trace = trace->next;
}
return Qnil;
}
static VALUE
-trace_en(struct rb_global_variable *var)
+trace_en(VALUE v)
{
+ struct rb_global_variable *var = (void *)v;
var->block_trace = 0;
remove_trace(var);
return Qnil; /* not reached */
}
-VALUE
-rb_gvar_set(struct rb_global_entry *entry, VALUE val)
+static VALUE
+rb_gvar_set_entry(struct rb_global_entry *entry, VALUE val)
{
struct trace_data trace;
struct rb_global_variable *var = entry->var;
- (*var->setter)(val, entry->id, var->data, var);
+ (*var->setter)(val, entry->id, var->data);
if (var->trace && !var->block_trace) {
- var->block_trace = 1;
- trace.trace = var->trace;
- trace.val = val;
- rb_ensure(trace_ev, (VALUE)&trace, trace_en, (VALUE)var);
+ var->block_trace = 1;
+ trace.trace = var->trace;
+ trace.val = val;
+ rb_ensure(trace_ev, (VALUE)&trace, trace_en, (VALUE)var);
}
return val;
}
+#define USE_BOX_GVAR_TBL(ns,entry) \
+ (BOX_USER_P(ns) && \
+ (!entry || !entry->var->box_ready || entry->var->setter != rb_gvar_readonly_setter))
+
VALUE
-rb_gv_set(const char *name, VALUE val)
+rb_gvar_set(ID id, VALUE val)
{
+ VALUE retval;
struct rb_global_entry *entry;
+ const rb_box_t *box = rb_current_box();
+ bool use_box_tbl = false;
+
+ RB_VM_LOCKING() {
+ entry = rb_global_entry(id);
+
+ if (USE_BOX_GVAR_TBL(box, entry)) {
+ use_box_tbl = true;
+ rb_hash_aset(box->gvar_tbl, rb_id2sym(entry->id), val);
+ retval = val;
+ // TODO: think about trace
+ }
+ }
- entry = rb_global_entry(global_id(name));
- return rb_gvar_set(entry, val);
+ if (!use_box_tbl) {
+ retval = rb_gvar_set_entry(entry, val);
+ }
+ return retval;
}
VALUE
-rb_gv_get(const char *name)
+rb_gv_set(const char *name, VALUE val)
{
+ return rb_gvar_set(global_id(name), val);
+}
+
+VALUE
+rb_gvar_get(ID id)
+{
+ VALUE retval, gvars, key;
+ const rb_box_t *box = rb_current_box();
+ bool use_box_tbl = false;
struct rb_global_entry *entry;
+ struct rb_global_variable *var;
+ // TODO: use lock-free rb_id_table when it's available for use (doesn't yet exist)
+ RB_VM_LOCKING() {
+ entry = rb_global_entry(id);
+ var = entry->var;
+
+ if (USE_BOX_GVAR_TBL(box, entry)) {
+ use_box_tbl = true;
+ gvars = box->gvar_tbl;
+ key = rb_id2sym(entry->id);
+ if (RTEST(rb_hash_has_key(gvars, key))) { // this gvar is already cached
+ retval = rb_hash_aref(gvars, key);
+ }
+ else {
+ RB_VM_UNLOCK();
+ {
+ retval = (*var->getter)(entry->id, var->data);
+ if (rb_obj_respond_to(retval, rb_intern("clone"), 1)) {
+ retval = rb_funcall(retval, rb_intern("clone"), 0);
+ }
+ }
+ RB_VM_LOCK();
+ rb_hash_aset(gvars, key, retval);
+ }
+ }
+ }
+ if (!use_box_tbl) {
+ retval = (*var->getter)(entry->id, var->data);
+ }
+ return retval;
+}
- entry = rb_global_entry(global_id(name));
- return rb_gvar_get(entry);
+VALUE
+rb_gv_get(const char *name)
+{
+ ID id = find_global_id(name);
+
+ if (!id) {
+ rb_warning("global variable '%s' not initialized", name);
+ return Qnil;
+ }
+
+ return rb_gvar_get(id);
}
VALUE
-rb_gvar_defined(struct rb_global_entry *entry)
+rb_gvar_defined(ID id)
{
- if (entry->var->getter == rb_gvar_undef_getter) return Qfalse;
- return Qtrue;
+ struct rb_global_entry *entry = rb_global_entry(id);
+ return RBOOL(entry->var->getter != rb_gvar_undef_getter);
+}
+
+rb_gvar_getter_t *
+rb_gvar_getter_function_of(ID id)
+{
+ const struct rb_global_entry *entry = rb_global_entry(id);
+ return entry->var->getter;
+}
+
+rb_gvar_setter_t *
+rb_gvar_setter_function_of(ID id)
+{
+ const struct rb_global_entry *entry = rb_global_entry(id);
+ return entry->var->setter;
}
static enum rb_id_table_iterator_result
@@ -873,39 +1116,35 @@ gvar_i(ID key, VALUE val, void *a)
return ID_TABLE_CONTINUE;
}
-/*
- * call-seq:
- * global_variables -> array
- *
- * Returns an array of the names of global variables.
- *
- * global_variables.grep /std/ #=> [:$stdin, :$stdout, :$stderr]
- */
-
VALUE
rb_f_global_variables(void)
{
VALUE ary = rb_ary_new();
VALUE sym, backref = rb_backref_get();
+ if (!rb_ractor_main_p()) {
+ rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors");
+ }
+ /* gvar access (get/set) in boxes creates gvar entries globally */
+
rb_id_table_foreach(rb_global_tbl, gvar_i, (void *)ary);
if (!NIL_P(backref)) {
- char buf[2];
- int i, nmatch = rb_match_count(backref);
- buf[0] = '$';
- for (i = 1; i <= nmatch; ++i) {
- if (!rb_match_nth_defined(i, backref)) continue;
- if (i < 10) {
- /* probably reused, make static ID */
- buf[1] = (char)(i + '0');
- sym = ID2SYM(rb_intern2(buf, 2));
- }
- else {
- /* dynamic symbol */
- sym = rb_str_intern(rb_sprintf("$%d", i));
- }
- rb_ary_push(ary, sym);
- }
+ char buf[2];
+ int i, nmatch = rb_match_count(backref);
+ buf[0] = '$';
+ for (i = 1; i <= nmatch; ++i) {
+ if (!RTEST(rb_reg_nth_defined(i, backref))) continue;
+ if (i < 10) {
+ /* probably reused, make static ID */
+ buf[1] = (char)(i + '0');
+ sym = ID2SYM(rb_intern2(buf, 2));
+ }
+ else {
+ /* dynamic symbol */
+ sym = rb_str_intern(rb_sprintf("$%d", i));
+ }
+ rb_ary_push(ary, sym);
+ }
}
return ary;
}
@@ -913,313 +1152,398 @@ rb_f_global_variables(void)
void
rb_alias_variable(ID name1, ID name2)
{
- struct rb_global_entry *entry1, *entry2;
+ struct rb_global_entry *entry1 = NULL, *entry2;
VALUE data1;
+ struct rb_id_table *gtbl = rb_global_tbl;
- entry2 = rb_global_entry(name2);
- if (!rb_id_table_lookup(rb_global_tbl, name1, &data1)) {
- entry1 = ALLOC(struct rb_global_entry);
- entry1->id = name1;
- rb_id_table_insert(rb_global_tbl, name1, (VALUE)entry1);
- }
- else if ((entry1 = (struct rb_global_entry *)data1)->var != entry2->var) {
- struct rb_global_variable *var = entry1->var;
- if (var->block_trace) {
- rb_raise(rb_eRuntimeError, "can't alias in tracer");
- }
- var->counter--;
- if (var->counter == 0) {
- struct trace_var *trace = var->trace;
- while (trace) {
- struct trace_var *next = trace->next;
- xfree(trace);
- trace = next;
- }
- xfree(var);
- }
+ if (!rb_ractor_main_p()) {
+ rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors");
}
- else {
- return;
+
+ RB_VM_LOCKING() {
+ entry2 = rb_global_entry(name2);
+ if (!rb_id_table_lookup(gtbl, name1, &data1)) {
+ entry1 = ZALLOC(struct rb_global_entry);
+ entry1->id = name1;
+ rb_id_table_insert(gtbl, name1, (VALUE)entry1);
+ }
+ else if ((entry1 = (struct rb_global_entry *)data1)->var != entry2->var) {
+ struct rb_global_variable *var = entry1->var;
+ if (var->block_trace) {
+ RB_VM_UNLOCK();
+ rb_raise(rb_eRuntimeError, "can't alias in tracer");
+ }
+ var->counter--;
+ if (var->counter == 0) {
+ free_global_variable(var);
+ }
+ }
+ if (entry1->var != entry2->var) {
+ entry2->var->counter++;
+ entry1->var = entry2->var;
+ }
}
- entry2->var->counter++;
- entry1->var = entry2->var;
}
-static int
-gen_ivtbl_get(VALUE obj, struct gen_ivtbl **ivtbl)
+static void
+IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(ID id)
{
- st_data_t data;
-
- if (st_lookup(generic_iv_tbl, (st_data_t)obj, &data)) {
- *ivtbl = (struct gen_ivtbl *)data;
- return 1;
+ if (UNLIKELY(!rb_ractor_main_p())) {
+ if (rb_is_instance_id(id)) { // check only normal ivars
+ rb_raise(rb_eRactorIsolationError, "can not set instance variables of classes/modules by non-main Ractors");
+ }
}
- return 0;
}
-static VALUE
-generic_ivar_delete(VALUE obj, ID id, VALUE undef)
+static void
+CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(VALUE klass, ID id)
{
- struct gen_ivtbl *ivtbl;
-
- if (gen_ivtbl_get(obj, &ivtbl)) {
- st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj));
- st_data_t index;
-
- if (st_lookup(iv_index_tbl, (st_data_t)id, &index)) {
- if (index < ivtbl->numiv) {
- VALUE ret = ivtbl->ivptr[index];
-
- ivtbl->ivptr[index] = Qundef;
- return ret == Qundef ? undef : ret;
- }
- }
+ if (UNLIKELY(!rb_ractor_main_p())) {
+ rb_raise(rb_eRactorIsolationError, "can not access class variables from non-main Ractors (%"PRIsVALUE" from %"PRIsVALUE")", rb_id2str(id), klass);
}
- return undef;
}
-static VALUE
-generic_ivar_get(VALUE obj, ID id, VALUE undef)
+static inline void
+ivar_ractor_check(VALUE obj, ID id)
{
- struct gen_ivtbl *ivtbl;
-
- if (gen_ivtbl_get(obj, &ivtbl)) {
- st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj));
- st_data_t index;
+ if (LIKELY(rb_is_instance_id(id)) /* not internal ID */ &&
+ !RB_OBJ_FROZEN_RAW(obj) &&
+ UNLIKELY(!rb_ractor_main_p()) &&
+ UNLIKELY(rb_ractor_shareable_p(obj))) {
- if (st_lookup(iv_index_tbl, (st_data_t)id, &index)) {
- if (index < ivtbl->numiv) {
- VALUE ret = ivtbl->ivptr[index];
-
- return ret == Qundef ? undef : ret;
- }
- }
+ rb_raise(rb_eRactorIsolationError, "can not access instance variables of shareable objects from non-main Ractors");
}
- return undef;
}
-static size_t
-gen_ivtbl_bytes(size_t n)
+static inline struct st_table *
+generic_fields_tbl_no_ractor_check(void)
{
- return sizeof(struct gen_ivtbl) + n * sizeof(VALUE) - sizeof(VALUE);
+ ASSERT_vm_locking();
+
+ return generic_fields_tbl_;
}
-static struct gen_ivtbl *
-gen_ivtbl_resize(struct gen_ivtbl *old, uint32_t n)
+struct st_table *
+rb_generic_fields_tbl_get(void)
{
- uint32_t len = old ? old->numiv : 0;
- struct gen_ivtbl *ivtbl = xrealloc(old, gen_ivtbl_bytes(n));
-
- ivtbl->numiv = n;
- for (; len < n; len++) {
- ivtbl->ivptr[len] = Qundef;
- }
-
- return ivtbl;
+ return generic_fields_tbl_;
}
-#if 0
-static struct gen_ivtbl *
-gen_ivtbl_dup(const struct gen_ivtbl *orig)
+void
+rb_mark_generic_ivar(VALUE obj)
{
- size_t s = gen_ivtbl_bytes(orig->numiv);
- struct gen_ivtbl *ivtbl = xmalloc(s);
-
- memcpy(ivtbl, orig, s);
-
- return ivtbl;
+ VALUE data;
+ // Bypass ASSERT_vm_locking() check because marking may happen concurrently with mmtk
+ if (st_lookup(generic_fields_tbl_, (st_data_t)obj, (st_data_t *)&data)) {
+ rb_gc_mark_movable(data);
+ }
}
-#endif
-static uint32_t
-iv_index_tbl_newsize(struct ivar_update *ivup)
+VALUE
+rb_obj_fields_generic_uncached(VALUE obj)
{
- uint32_t index = (uint32_t)ivup->index; /* should not overflow */
- uint32_t newsize = (index+1) + (index+1)/4; /* (index+1)*1.25 */
+ VALUE fields_obj = 0;
+ RB_VM_LOCKING() {
+ if (!st_lookup(generic_fields_tbl_, (st_data_t)obj, (st_data_t *)&fields_obj)) {
+ rb_bug("Object is missing entry in generic_fields_tbl");
+ }
+ }
+ return fields_obj;
+}
- if (!ivup->iv_extended &&
- ivup->u.iv_index_tbl->num_entries < (st_index_t)newsize) {
- newsize = (uint32_t)ivup->u.iv_index_tbl->num_entries;
+VALUE
+rb_obj_fields(VALUE obj, ID field_name)
+{
+ RUBY_ASSERT(!RB_TYPE_P(obj, T_IMEMO));
+ ivar_ractor_check(obj, field_name);
+
+ VALUE fields_obj = 0;
+ if (rb_shape_obj_has_fields(obj)) {
+ switch (BUILTIN_TYPE(obj)) {
+ case T_DATA:
+ if (LIKELY(RTYPEDDATA_P(obj))) {
+ fields_obj = RTYPEDDATA(obj)->fields_obj;
+ break;
+ }
+ goto generic_fields;
+ case T_STRUCT:
+ if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) {
+ fields_obj = RSTRUCT_FIELDS_OBJ(obj);
+ break;
+ }
+ goto generic_fields;
+ default:
+ generic_fields:
+ {
+ rb_execution_context_t *ec = GET_EC();
+ if (ec->gen_fields_cache.obj == obj && !UNDEF_P(ec->gen_fields_cache.fields_obj) && rb_imemo_fields_owner(ec->gen_fields_cache.fields_obj) == obj) {
+ fields_obj = ec->gen_fields_cache.fields_obj;
+ RUBY_ASSERT(fields_obj == rb_obj_fields_generic_uncached(obj));
+ }
+ else {
+ fields_obj = rb_obj_fields_generic_uncached(obj);
+ ec->gen_fields_cache.fields_obj = fields_obj;
+ ec->gen_fields_cache.obj = obj;
+ }
+ }
+ }
}
- return newsize;
+ return fields_obj;
}
-static int
-generic_ivar_update(st_data_t *k, st_data_t *v, st_data_t u, int existing)
+void
+rb_free_generic_ivar(VALUE obj)
{
- VALUE obj = (VALUE)*k;
- struct ivar_update *ivup = (struct ivar_update *)u;
- uint32_t newsize;
- int ret = ST_CONTINUE;
- struct gen_ivtbl *ivtbl;
-
- if (existing) {
- ivtbl = (struct gen_ivtbl *)*v;
- if (ivup->index >= ivtbl->numiv) {
- goto resize;
- }
- ret = ST_STOP;
- }
- else {
- FL_SET(obj, FL_EXIVAR);
- ivtbl = 0;
-resize:
- newsize = iv_index_tbl_newsize(ivup);
- ivtbl = gen_ivtbl_resize(ivtbl, newsize);
- *v = (st_data_t)ivtbl;
+ if (rb_obj_gen_fields_p(obj)) {
+ st_data_t key = (st_data_t)obj, value;
+ switch (BUILTIN_TYPE(obj)) {
+ case T_DATA:
+ if (LIKELY(RTYPEDDATA_P(obj))) {
+ RB_OBJ_WRITE(obj, &RTYPEDDATA(obj)->fields_obj, 0);
+ break;
+ }
+ goto generic_fields;
+ case T_STRUCT:
+ if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) {
+ RSTRUCT_SET_FIELDS_OBJ(obj, 0);
+ break;
+ }
+ goto generic_fields;
+ default:
+ generic_fields:
+ {
+ // Other EC may have stale caches, so fields_obj should be
+ // invalidated and the GC will replace with Qundef
+ rb_execution_context_t *ec = GET_EC();
+ if (ec->gen_fields_cache.obj == obj) {
+ ec->gen_fields_cache.obj = Qundef;
+ ec->gen_fields_cache.fields_obj = Qundef;
+ }
+ RB_VM_LOCKING() {
+ if (!st_delete(generic_fields_tbl_no_ractor_check(), &key, &value)) {
+ rb_bug("Object is missing entry in generic_fields_tbl");
+ }
+ }
+ }
+ }
+ RBASIC_SET_SHAPE_ID(obj, ROOT_SHAPE_ID);
}
- ivup->u.ivtbl = ivtbl;
- return ret;
}
-static VALUE
-generic_ivar_defined(VALUE obj, ID id)
+static void
+rb_obj_set_fields(VALUE obj, VALUE fields_obj, ID field_name, VALUE original_fields_obj)
{
- struct gen_ivtbl *ivtbl;
- st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj));
- st_data_t index;
+ ivar_ractor_check(obj, field_name);
- if (!iv_index_tbl) return Qfalse;
- if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) return Qfalse;
- if (!gen_ivtbl_get(obj, &ivtbl)) return Qfalse;
-
- if ((index < ivtbl->numiv) && (ivtbl->ivptr[index] != Qundef))
- return Qtrue;
+ if (!fields_obj) {
+ RUBY_ASSERT(original_fields_obj);
+ rb_free_generic_ivar(obj);
+ rb_imemo_fields_clear(original_fields_obj);
+ return;
+ }
- return Qfalse;
-}
+ RUBY_ASSERT(IMEMO_TYPE_P(fields_obj, imemo_fields));
+ RUBY_ASSERT(!original_fields_obj || IMEMO_TYPE_P(original_fields_obj, imemo_fields));
-static int
-generic_ivar_remove(VALUE obj, ID id, VALUE *valp)
-{
- struct gen_ivtbl *ivtbl;
- st_data_t key = (st_data_t)id;
- st_data_t index;
- st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj));
+ if (fields_obj != original_fields_obj) {
+ switch (BUILTIN_TYPE(obj)) {
+ case T_DATA:
+ if (LIKELY(RTYPEDDATA_P(obj))) {
+ RB_OBJ_WRITE(obj, &RTYPEDDATA(obj)->fields_obj, fields_obj);
+ break;
+ }
+ goto generic_fields;
+ case T_STRUCT:
+ if (LIKELY(!FL_TEST_RAW(obj, RSTRUCT_GEN_FIELDS))) {
+ RSTRUCT_SET_FIELDS_OBJ(obj, fields_obj);
+ break;
+ }
+ goto generic_fields;
+ default:
+ generic_fields:
+ {
+ RB_VM_LOCKING() {
+ st_insert(generic_fields_tbl_, (st_data_t)obj, (st_data_t)fields_obj);
+ }
+ RB_OBJ_WRITTEN(obj, original_fields_obj, fields_obj);
- if (!iv_index_tbl) return 0;
- if (!st_lookup(iv_index_tbl, key, &index)) return 0;
- if (!gen_ivtbl_get(obj, &ivtbl)) return 0;
+ rb_execution_context_t *ec = GET_EC();
+ if (ec->gen_fields_cache.fields_obj != fields_obj) {
+ ec->gen_fields_cache.obj = obj;
+ ec->gen_fields_cache.fields_obj = fields_obj;
+ }
+ }
+ }
- if (index < ivtbl->numiv) {
- if (ivtbl->ivptr[index] != Qundef) {
- *valp = ivtbl->ivptr[index];
- ivtbl->ivptr[index] = Qundef;
- return 1;
- }
+ if (original_fields_obj) {
+ // Clear root shape to avoid triggering cleanup such as free_object_id.
+ rb_imemo_fields_clear(original_fields_obj);
+ }
}
- return 0;
-}
-
-static void
-gen_ivtbl_mark(const struct gen_ivtbl *ivtbl)
-{
- uint32_t i;
- for (i = 0; i < ivtbl->numiv; i++) {
- rb_gc_mark(ivtbl->ivptr[i]);
- }
+ RBASIC_SET_SHAPE_ID(obj, RBASIC_SHAPE_ID(fields_obj));
}
void
-rb_mark_generic_ivar(VALUE obj)
+rb_obj_replace_fields(VALUE obj, VALUE fields_obj)
{
- struct gen_ivtbl *ivtbl;
-
- if (gen_ivtbl_get(obj, &ivtbl)) {
- gen_ivtbl_mark(ivtbl);
+ RB_VM_LOCKING() {
+ VALUE original_fields_obj = rb_obj_fields_no_ractor_check(obj);
+ rb_obj_set_fields(obj, fields_obj, 0, original_fields_obj);
}
}
-void
-rb_free_generic_ivar(VALUE obj)
+VALUE
+rb_obj_field_get(VALUE obj, shape_id_t target_shape_id)
{
- st_data_t key = (st_data_t)obj;
- struct gen_ivtbl *ivtbl;
+ RUBY_ASSERT(!SPECIAL_CONST_P(obj));
+ RUBY_ASSERT(RSHAPE_TYPE_P(target_shape_id, SHAPE_IVAR) || RSHAPE_TYPE_P(target_shape_id, SHAPE_OBJ_ID));
- if (st_delete(generic_iv_tbl, &key, (st_data_t *)&ivtbl))
- xfree(ivtbl);
+ VALUE fields_obj;
- if (generic_iv_tbl_compat) {
- st_table *tbl;
+ switch (BUILTIN_TYPE(obj)) {
+ case T_CLASS:
+ case T_MODULE:
+ fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ break;
+ case T_OBJECT:
+ fields_obj = obj;
+ break;
+ case T_IMEMO:
+ RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields));
+ fields_obj = obj;
+ break;
+ default:
+ fields_obj = rb_obj_fields(obj, RSHAPE_EDGE_NAME(target_shape_id));
+ break;
+ }
- if (st_delete(generic_iv_tbl_compat, &key, (st_data_t *)&tbl))
- st_free_table(tbl);
+ if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) {
+ st_table *fields_hash = rb_imemo_fields_complex_tbl(fields_obj);
+ VALUE value = Qundef;
+ st_lookup(fields_hash, RSHAPE_EDGE_NAME(target_shape_id), &value);
+ RUBY_ASSERT(!UNDEF_P(value));
+ return value;
}
+
+ attr_index_t index = RSHAPE_INDEX(target_shape_id);
+ return rb_imemo_fields_ptr(fields_obj)[index];
}
-RUBY_FUNC_EXPORTED size_t
-rb_generic_ivar_memsize(VALUE obj)
+VALUE
+rb_ivar_lookup(VALUE obj, ID id, VALUE undef)
{
- struct gen_ivtbl *ivtbl;
+ if (SPECIAL_CONST_P(obj)) return undef;
- if (gen_ivtbl_get(obj, &ivtbl))
- return gen_ivtbl_bytes(ivtbl->numiv);
- return 0;
-}
+ VALUE fields_obj;
-static size_t
-gen_ivtbl_count(const struct gen_ivtbl *ivtbl)
-{
- uint32_t i;
- size_t n = 0;
+ switch (BUILTIN_TYPE(obj)) {
+ case T_CLASS:
+ case T_MODULE:
+ {
+ VALUE val = rb_ivar_lookup(RCLASS_WRITABLE_FIELDS_OBJ(obj), id, undef);
+ if (val != undef &&
+ rb_is_instance_id(id) &&
+ UNLIKELY(!rb_ractor_main_p()) &&
+ !rb_ractor_shareable_p(val)) {
+ rb_raise(rb_eRactorIsolationError,
+ "can not get unshareable values from instance variables of classes/modules from non-main Ractors (%"PRIsVALUE" from %"PRIsVALUE")",
+ rb_id2str(id), obj);
+ }
+ return val;
+ }
+ case T_IMEMO:
+ // Handled like T_OBJECT
+ RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields));
+ fields_obj = obj;
+ break;
+ case T_OBJECT:
+ fields_obj = obj;
+ break;
+ default:
+ fields_obj = rb_obj_fields(obj, id);
+ break;
+ }
+
+ if (!fields_obj) {
+ return undef;
+ }
+
+ shape_id_t shape_id = RBASIC_SHAPE_ID(fields_obj);
- for (i = 0; i < ivtbl->numiv; i++) {
- if (ivtbl->ivptr[i] != Qundef) {
- n++;
- }
+ if (UNLIKELY(rb_shape_too_complex_p(shape_id))) {
+ st_table *iv_table = rb_imemo_fields_complex_tbl(fields_obj);
+ VALUE val;
+ if (rb_st_lookup(iv_table, (st_data_t)id, (st_data_t *)&val)) {
+ return val;
+ }
+ return undef;
}
- return n;
+ attr_index_t index = 0;
+ if (rb_shape_get_iv_index(shape_id, id, &index)) {
+ return rb_imemo_fields_ptr(fields_obj)[index];
+ }
+
+ return undef;
}
VALUE
-rb_ivar_lookup(VALUE obj, ID id, VALUE undef)
+rb_ivar_get(VALUE obj, ID id)
+{
+ VALUE iv = rb_ivar_lookup(obj, id, Qnil);
+ RB_DEBUG_COUNTER_INC(ivar_get_base);
+ return iv;
+}
+
+VALUE
+rb_ivar_get_at(VALUE obj, attr_index_t index, ID id)
{
- VALUE val, *ptr;
- struct st_table *iv_index_tbl;
- uint32_t len;
- st_data_t index;
+ RUBY_ASSERT(rb_is_instance_id(id));
+ // Used by JITs, but never for T_OBJECT.
- if (SPECIAL_CONST_P(obj)) return undef;
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
- len = ROBJECT_NUMIV(obj);
- ptr = ROBJECT_IVPTR(obj);
- iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj);
- if (!iv_index_tbl) break;
- if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break;
- if (len <= index) break;
- val = ptr[index];
- if (val != Qundef)
- return val;
- break;
+ UNREACHABLE_RETURN(Qundef);
case T_CLASS:
case T_MODULE:
- if (RCLASS_IV_TBL(obj) &&
- st_lookup(RCLASS_IV_TBL(obj), (st_data_t)id, &index))
- return (VALUE)index;
- break;
+ {
+ VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ VALUE val = rb_imemo_fields_ptr(fields_obj)[index];
+
+ if (UNLIKELY(!rb_ractor_main_p()) && !rb_ractor_shareable_p(val)) {
+ rb_raise(rb_eRactorIsolationError,
+ "can not get unshareable values from instance variables of classes/modules from non-main Ractors");
+ }
+
+ return val;
+ }
default:
- if (FL_TEST(obj, FL_EXIVAR))
- return generic_ivar_get(obj, id, undef);
- break;
+ {
+ VALUE fields_obj = rb_obj_fields(obj, id);
+ return rb_imemo_fields_ptr(fields_obj)[index];
+ }
}
- return undef;
}
VALUE
-rb_ivar_get(VALUE obj, ID id)
+rb_ivar_get_at_no_ractor_check(VALUE obj, attr_index_t index)
{
- VALUE iv = rb_ivar_lookup(obj, id, Qundef);
- RB_DEBUG_COUNTER_INC(ivar_get_base);
+ // Used by JITs, but never for T_OBJECT.
- if (iv == Qundef) {
- if (RTEST(ruby_verbose))
- rb_warning("instance variable %"PRIsVALUE" not initialized", QUOTE_ID(id));
- iv = Qnil;
+ VALUE fields_obj;
+ switch (BUILTIN_TYPE(obj)) {
+ case T_OBJECT:
+ UNREACHABLE_RETURN(Qundef);
+ case T_CLASS:
+ case T_MODULE:
+ fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ break;
+ default:
+ fields_obj = rb_obj_fields_no_ractor_check(obj);
+ break;
}
- return iv;
+ return rb_imemo_fields_ptr(fields_obj)[index];
}
VALUE
@@ -1228,40 +1552,182 @@ rb_attr_get(VALUE obj, ID id)
return rb_ivar_lookup(obj, id, Qnil);
}
-static VALUE
-rb_ivar_delete(VALUE obj, ID id, VALUE undef)
+void rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table);
+static VALUE imemo_fields_complex_from_obj(VALUE owner, VALUE source_fields_obj, shape_id_t shape_id);
+
+static shape_id_t
+obj_transition_too_complex(VALUE obj, st_table *table)
{
- VALUE val, *ptr;
- struct st_table *iv_index_tbl;
- uint32_t len;
- st_data_t index;
+ RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
+ shape_id_t shape_id = rb_shape_transition_complex(obj);
- rb_check_frozen(obj);
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
- len = ROBJECT_NUMIV(obj);
- ptr = ROBJECT_IVPTR(obj);
- iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj);
- if (!iv_index_tbl) break;
- if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break;
- if (len <= index) break;
- val = ptr[index];
- ptr[index] = Qundef;
- if (val != Qundef)
- return val;
- break;
+ {
+ VALUE *old_fields = NULL;
+ if (FL_TEST_RAW(obj, ROBJECT_HEAP)) {
+ old_fields = ROBJECT_FIELDS(obj);
+ }
+ else {
+ FL_SET_RAW(obj, ROBJECT_HEAP);
+ }
+ RBASIC_SET_SHAPE_ID(obj, shape_id);
+ ROBJECT_SET_FIELDS_HASH(obj, table);
+ if (old_fields) {
+ xfree(old_fields);
+ }
+ }
+ break;
case T_CLASS:
case T_MODULE:
- if (RCLASS_IV_TBL(obj) &&
- st_delete(RCLASS_IV_TBL(obj), (st_data_t *)&id, &index))
- return (VALUE)index;
- break;
+ case T_IMEMO:
+ UNREACHABLE;
+ break;
default:
- if (FL_TEST(obj, FL_EXIVAR))
- return generic_ivar_delete(obj, id, undef);
- break;
+ {
+ VALUE fields_obj = rb_imemo_fields_new_complex_tbl(obj, table, RB_OBJ_SHAREABLE_P(obj));
+ RBASIC_SET_SHAPE_ID(fields_obj, shape_id);
+ rb_obj_replace_fields(obj, fields_obj);
+ }
}
- return undef;
+
+ return shape_id;
+}
+
+// Copy all object fields, including ivars and internal object_id, etc
+static shape_id_t
+rb_evict_fields_to_hash(VALUE obj)
+{
+ RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
+
+ st_table *table = st_init_numtable_with_size(RSHAPE_LEN(RBASIC_SHAPE_ID(obj)));
+ rb_obj_copy_fields_to_hash_table(obj, table);
+ shape_id_t new_shape_id = obj_transition_too_complex(obj, table);
+
+ RUBY_ASSERT(rb_shape_obj_too_complex_p(obj));
+ return new_shape_id;
+}
+
+void
+rb_evict_ivars_to_hash(VALUE obj)
+{
+ RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
+
+ st_table *table = st_init_numtable_with_size(rb_ivar_count(obj));
+
+ // Evacuate all previous values from shape into id_table
+ rb_obj_copy_ivs_to_hash_table(obj, table);
+ obj_transition_too_complex(obj, table);
+
+ RUBY_ASSERT(rb_shape_obj_too_complex_p(obj));
+}
+
+static VALUE
+rb_ivar_delete(VALUE obj, ID id, VALUE undef)
+{
+ rb_check_frozen(obj);
+
+ VALUE val = undef;
+ VALUE fields_obj;
+ bool concurrent = false;
+ int type = BUILTIN_TYPE(obj);
+
+ switch(type) {
+ case T_CLASS:
+ case T_MODULE:
+ IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id);
+
+ fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ if (rb_multi_ractor_p()) {
+ concurrent = true;
+ }
+ break;
+ case T_OBJECT:
+ fields_obj = obj;
+ break;
+ default: {
+ fields_obj = rb_obj_fields(obj, id);
+ break;
+ }
+ }
+
+ if (!fields_obj) {
+ return undef;
+ }
+
+ const VALUE original_fields_obj = fields_obj;
+ if (concurrent) {
+ fields_obj = rb_imemo_fields_clone(fields_obj);
+ }
+
+ shape_id_t old_shape_id = RBASIC_SHAPE_ID(fields_obj);
+ shape_id_t removed_shape_id;
+ shape_id_t next_shape_id = rb_shape_transition_remove_ivar(fields_obj, id, &removed_shape_id);
+
+ if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) {
+ if (UNLIKELY(!rb_shape_too_complex_p(old_shape_id))) {
+ if (type == T_OBJECT) {
+ rb_evict_fields_to_hash(obj);
+ }
+ else {
+ fields_obj = imemo_fields_complex_from_obj(obj, fields_obj, next_shape_id);
+ }
+ }
+ st_data_t key = id;
+ if (!st_delete(rb_imemo_fields_complex_tbl(fields_obj), &key, (st_data_t *)&val)) {
+ val = undef;
+ }
+ }
+ else {
+ if (next_shape_id == old_shape_id) {
+ return undef;
+ }
+
+ RUBY_ASSERT(removed_shape_id != INVALID_SHAPE_ID);
+ RUBY_ASSERT(RSHAPE_LEN(next_shape_id) == RSHAPE_LEN(old_shape_id) - 1);
+
+ VALUE *fields = rb_imemo_fields_ptr(fields_obj);
+ attr_index_t removed_index = RSHAPE_INDEX(removed_shape_id);
+ val = fields[removed_index];
+
+ attr_index_t new_fields_count = RSHAPE_LEN(next_shape_id);
+ if (new_fields_count) {
+ size_t trailing_fields = new_fields_count - removed_index;
+
+ MEMMOVE(&fields[removed_index], &fields[removed_index + 1], VALUE, trailing_fields);
+ RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id);
+
+ if (FL_TEST_RAW(fields_obj, OBJ_FIELD_HEAP) && rb_obj_embedded_size(new_fields_count) <= rb_gc_obj_slot_size(fields_obj)) {
+ // Re-embed objects when instances become small enough
+ // This is necessary because YJIT assumes that objects with the same shape
+ // have the same embeddedness for efficiency (avoid extra checks)
+ FL_UNSET_RAW(fields_obj, ROBJECT_HEAP);
+ MEMCPY(rb_imemo_fields_ptr(fields_obj), fields, VALUE, new_fields_count);
+ xfree(fields);
+ }
+ }
+ else {
+ fields_obj = 0;
+ rb_free_generic_ivar(obj);
+ }
+ }
+
+ RBASIC_SET_SHAPE_ID(obj, next_shape_id);
+ if (fields_obj != original_fields_obj) {
+ switch (type) {
+ case T_OBJECT:
+ break;
+ case T_CLASS:
+ case T_MODULE:
+ RCLASS_WRITABLE_SET_FIELDS_OBJ(obj, fields_obj);
+ break;
+ default:
+ rb_obj_set_fields(obj, fields_obj, id, original_fields_obj);
+ break;
+ }
+ }
+
+ return val;
}
VALUE
@@ -1270,347 +1736,666 @@ rb_attr_delete(VALUE obj, ID id)
return rb_ivar_delete(obj, id, Qnil);
}
-static st_table *
-iv_index_tbl_make(VALUE obj)
+void
+rb_obj_init_too_complex(VALUE obj, st_table *table)
+{
+ // This method is meant to be called on newly allocated object.
+ RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
+ RUBY_ASSERT(rb_shape_canonical_p(RBASIC_SHAPE_ID(obj)));
+ RUBY_ASSERT(RSHAPE_LEN(RBASIC_SHAPE_ID(obj)) == 0);
+
+ obj_transition_too_complex(obj, table);
+}
+
+static int
+imemo_fields_complex_from_obj_i(ID key, VALUE val, st_data_t arg)
{
- VALUE klass = rb_obj_class(obj);
- st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(klass);
+ VALUE fields = (VALUE)arg;
+ st_table *table = rb_imemo_fields_complex_tbl(fields);
+
+ RUBY_ASSERT(!st_lookup(table, (st_data_t)key, NULL));
+ st_add_direct(table, (st_data_t)key, (st_data_t)val);
+ RB_OBJ_WRITTEN(fields, Qundef, val);
- if (!iv_index_tbl) {
- iv_index_tbl = RCLASS_IV_INDEX_TBL(klass) = st_init_numtable();
+ return ST_CONTINUE;
+}
+
+static VALUE
+imemo_fields_complex_from_obj(VALUE owner, VALUE source_fields_obj, shape_id_t shape_id)
+{
+ attr_index_t len = source_fields_obj ? RSHAPE_LEN(RBASIC_SHAPE_ID(source_fields_obj)) : 0;
+ VALUE fields_obj = rb_imemo_fields_new_complex(owner, len + 1, RB_OBJ_SHAREABLE_P(owner));
+
+ rb_field_foreach(source_fields_obj, imemo_fields_complex_from_obj_i, (st_data_t)fields_obj, false);
+ RBASIC_SET_SHAPE_ID(fields_obj, shape_id);
+
+ return fields_obj;
+}
+
+static VALUE
+imemo_fields_copy_capa(VALUE owner, VALUE source_fields_obj, attr_index_t new_size)
+{
+ VALUE fields_obj = rb_imemo_fields_new(owner, new_size, RB_OBJ_SHAREABLE_P(owner));
+ if (source_fields_obj) {
+ attr_index_t fields_count = RSHAPE_LEN(RBASIC_SHAPE_ID(source_fields_obj));
+ VALUE *fields = rb_imemo_fields_ptr(fields_obj);
+ MEMCPY(fields, rb_imemo_fields_ptr(source_fields_obj), VALUE, fields_count);
+ RBASIC_SET_SHAPE_ID(fields_obj, RBASIC_SHAPE_ID(source_fields_obj));
+ for (attr_index_t i = 0; i < fields_count; i++) {
+ RB_OBJ_WRITTEN(fields_obj, Qundef, fields[i]);
+ }
+ }
+ return fields_obj;
+}
+
+static VALUE
+imemo_fields_set(VALUE owner, VALUE fields_obj, shape_id_t target_shape_id, ID field_name, VALUE val, bool concurrent)
+{
+ const VALUE original_fields_obj = fields_obj;
+ shape_id_t current_shape_id = fields_obj ? RBASIC_SHAPE_ID(fields_obj) : ROOT_SHAPE_ID;
+
+ if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) {
+ if (rb_shape_too_complex_p(current_shape_id)) {
+ if (concurrent) {
+ // In multi-ractor case, we must always work on a copy because
+ // even if the field already exist, inserting in a st_table may
+ // cause a rebuild.
+ fields_obj = rb_imemo_fields_clone(fields_obj);
+ }
+ }
+ else {
+ fields_obj = imemo_fields_complex_from_obj(owner, original_fields_obj, target_shape_id);
+ current_shape_id = target_shape_id;
+ }
+
+ st_table *table = rb_imemo_fields_complex_tbl(fields_obj);
+
+ RUBY_ASSERT(field_name);
+ st_insert(table, (st_data_t)field_name, (st_data_t)val);
+ RB_OBJ_WRITTEN(fields_obj, Qundef, val);
+ RBASIC_SET_SHAPE_ID(fields_obj, target_shape_id);
}
+ else {
+ attr_index_t index = RSHAPE_INDEX(target_shape_id);
+ if (concurrent || index >= RSHAPE_CAPACITY(current_shape_id)) {
+ fields_obj = imemo_fields_copy_capa(owner, original_fields_obj, RSHAPE_CAPACITY(target_shape_id));
+ }
- return iv_index_tbl;
+ VALUE *table = rb_imemo_fields_ptr(fields_obj);
+ RB_OBJ_WRITE(fields_obj, &table[index], val);
+
+ if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) {
+ RBASIC_SET_SHAPE_ID(fields_obj, target_shape_id);
+ }
+ }
+
+ return fields_obj;
}
-static void
-iv_index_tbl_extend(struct ivar_update *ivup, ID id)
+static attr_index_t
+generic_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val)
{
- if (st_lookup(ivup->u.iv_index_tbl, (st_data_t)id, &ivup->index)) {
- return;
+ if (!field_name) {
+ field_name = RSHAPE_EDGE_NAME(target_shape_id);
+ RUBY_ASSERT(field_name);
}
- if (ivup->u.iv_index_tbl->num_entries >= INT_MAX) {
- rb_raise(rb_eArgError, "too many instance variables");
+
+ const VALUE original_fields_obj = rb_obj_fields(obj, field_name);
+ VALUE fields_obj = imemo_fields_set(obj, original_fields_obj, target_shape_id, field_name, val, false);
+
+ rb_obj_set_fields(obj, fields_obj, field_name, original_fields_obj);
+ return rb_shape_too_complex_p(target_shape_id) ? ATTR_INDEX_NOT_SET : RSHAPE_INDEX(target_shape_id);
+}
+
+static shape_id_t
+generic_shape_ivar(VALUE obj, ID id, bool *new_ivar_out)
+{
+ bool new_ivar = false;
+ shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj);
+ shape_id_t target_shape_id = current_shape_id;
+
+ if (!rb_shape_too_complex_p(current_shape_id)) {
+ if (!rb_shape_find_ivar(current_shape_id, id, &target_shape_id)) {
+ if (RSHAPE_LEN(current_shape_id) >= SHAPE_MAX_FIELDS) {
+ rb_raise(rb_eArgError, "too many instance variables");
+ }
+
+ new_ivar = true;
+ target_shape_id = rb_shape_transition_add_ivar(obj, id);
+ }
}
- ivup->index = (st_data_t)ivup->u.iv_index_tbl->num_entries;
- st_add_direct(ivup->u.iv_index_tbl, (st_data_t)id, ivup->index);
- ivup->iv_extended = 1;
+
+ *new_ivar_out = new_ivar;
+ return target_shape_id;
}
-static void
+static attr_index_t
generic_ivar_set(VALUE obj, ID id, VALUE val)
{
- struct ivar_update ivup;
+ bool dontcare;
+ shape_id_t target_shape_id = generic_shape_ivar(obj, id, &dontcare);
+ return generic_field_set(obj, target_shape_id, id, val);
+}
- ivup.iv_extended = 0;
- ivup.u.iv_index_tbl = iv_index_tbl_make(obj);
- iv_index_tbl_extend(&ivup, id);
- st_update(generic_iv_tbl, (st_data_t)obj, generic_ivar_update,
- (st_data_t)&ivup);
+void
+rb_ensure_iv_list_size(VALUE obj, uint32_t current_len, uint32_t new_capacity)
+{
+ RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
- ivup.u.ivtbl->ivptr[ivup.index] = val;
+ if (FL_TEST_RAW(obj, ROBJECT_HEAP)) {
+ REALLOC_N(ROBJECT(obj)->as.heap.fields, VALUE, new_capacity);
+ }
+ else {
+ VALUE *ptr = ROBJECT_FIELDS(obj);
+ VALUE *newptr = ALLOC_N(VALUE, new_capacity);
+ MEMCPY(newptr, ptr, VALUE, current_len);
+ FL_SET_RAW(obj, ROBJECT_HEAP);
+ ROBJECT(obj)->as.heap.fields = newptr;
+ }
+}
+
+static int
+rb_obj_copy_ivs_to_hash_table_i(ID key, VALUE val, st_data_t arg)
+{
+ RUBY_ASSERT(!st_lookup((st_table *)arg, (st_data_t)key, NULL));
+
+ st_add_direct((st_table *)arg, (st_data_t)key, (st_data_t)val);
+ return ST_CONTINUE;
+}
+
+void
+rb_obj_copy_ivs_to_hash_table(VALUE obj, st_table *table)
+{
+ rb_ivar_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table);
+}
- RB_OBJ_WRITTEN(obj, Qundef, val);
+void
+rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table)
+{
+ rb_field_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table, false);
}
+static attr_index_t
+obj_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val)
+{
+ shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj);
+
+ if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) {
+ if (UNLIKELY(!rb_shape_too_complex_p(current_shape_id))) {
+ current_shape_id = rb_evict_fields_to_hash(obj);
+ }
+
+ if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) {
+ RBASIC_SET_SHAPE_ID(obj, target_shape_id);
+ }
+
+ if (!field_name) {
+ field_name = RSHAPE_EDGE_NAME(target_shape_id);
+ RUBY_ASSERT(field_name);
+ }
+
+ st_insert(ROBJECT_FIELDS_HASH(obj), (st_data_t)field_name, (st_data_t)val);
+ RB_OBJ_WRITTEN(obj, Qundef, val);
+
+ return ATTR_INDEX_NOT_SET;
+ }
+ else {
+ attr_index_t index = RSHAPE_INDEX(target_shape_id);
+
+ if (index >= RSHAPE_LEN(current_shape_id)) {
+ if (UNLIKELY(index >= RSHAPE_CAPACITY(current_shape_id))) {
+ rb_ensure_iv_list_size(obj, RSHAPE_CAPACITY(current_shape_id), RSHAPE_CAPACITY(target_shape_id));
+ }
+ RBASIC_SET_SHAPE_ID(obj, target_shape_id);
+ }
+
+ RB_OBJ_WRITE(obj, &ROBJECT_FIELDS(obj)[index], val);
+
+ return index;
+ }
+}
+
+static attr_index_t
+obj_ivar_set(VALUE obj, ID id, VALUE val)
+{
+ bool dontcare;
+ shape_id_t target_shape_id = generic_shape_ivar(obj, id, &dontcare);
+ return obj_field_set(obj, target_shape_id, id, val);
+}
+
+/* Set the instance variable +val+ on object +obj+ at ivar name +id+.
+ * This function only works with T_OBJECT objects, so make sure
+ * +obj+ is of type T_OBJECT before using this function.
+ */
VALUE
-rb_ivar_set(VALUE obj, ID id, VALUE val)
+rb_vm_set_ivar_id(VALUE obj, ID id, VALUE val)
+{
+ rb_check_frozen(obj);
+ obj_ivar_set(obj, id, val);
+ return val;
+}
+
+void rb_obj_freeze_inline(VALUE x)
{
- struct ivar_update ivup;
- uint32_t i, len;
+ if (RB_FL_ABLE(x)) {
+ RB_FL_SET_RAW(x, RUBY_FL_FREEZE);
+ if (TYPE(x) == T_STRING) {
+ RB_FL_UNSET_RAW(x, FL_USER2 | FL_USER3); // STR_CHILLED
+ }
+
+ RB_SET_SHAPE_ID(x, rb_shape_transition_frozen(x));
+
+ if (RBASIC_CLASS(x)) {
+ rb_freeze_singleton_class(x);
+ }
+ }
+}
+
+static attr_index_t class_ivar_set(VALUE obj, ID id, VALUE val, bool *new_ivar);
+static attr_index_t
+ivar_set(VALUE obj, ID id, VALUE val)
+{
RB_DEBUG_COUNTER_INC(ivar_set_base);
+ switch (BUILTIN_TYPE(obj)) {
+ case T_OBJECT:
+ return obj_ivar_set(obj, id, val);
+ case T_CLASS:
+ case T_MODULE:
+ {
+ IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id);
+ bool dontcare;
+ return class_ivar_set(obj, id, val, &dontcare);
+ }
+ default:
+ return generic_ivar_set(obj, id, val);
+ }
+}
+
+VALUE
+rb_ivar_set(VALUE obj, ID id, VALUE val)
+{
rb_check_frozen(obj);
+ ivar_set(obj, id, val);
+ return val;
+}
+
+attr_index_t
+rb_ivar_set_index(VALUE obj, ID id, VALUE val)
+{
+ return ivar_set(obj, id, val);
+}
+
+void
+rb_ivar_set_internal(VALUE obj, ID id, VALUE val)
+{
+ // should be internal instance variable name (no @ prefix)
+ VM_ASSERT(!rb_is_instance_id(id));
+ ivar_set(obj, id, val);
+}
+
+attr_index_t
+rb_obj_field_set(VALUE obj, shape_id_t target_shape_id, ID field_name, VALUE val)
+{
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
- ivup.iv_extended = 0;
- ivup.u.iv_index_tbl = iv_index_tbl_make(obj);
- iv_index_tbl_extend(&ivup, id);
- len = ROBJECT_NUMIV(obj);
- if (len <= ivup.index) {
- VALUE *ptr = ROBJECT_IVPTR(obj);
- if (ivup.index < ROBJECT_EMBED_LEN_MAX) {
- RBASIC(obj)->flags |= ROBJECT_EMBED;
- ptr = ROBJECT(obj)->as.ary;
- for (i = 0; i < ROBJECT_EMBED_LEN_MAX; i++) {
- ptr[i] = Qundef;
- }
- }
- else {
- VALUE *newptr;
- uint32_t newsize = iv_index_tbl_newsize(&ivup);
-
- if (RBASIC(obj)->flags & ROBJECT_EMBED) {
- newptr = ALLOC_N(VALUE, newsize);
- MEMCPY(newptr, ptr, VALUE, len);
- RBASIC(obj)->flags &= ~ROBJECT_EMBED;
- ROBJECT(obj)->as.heap.ivptr = newptr;
- }
- else {
- REALLOC_N(ROBJECT(obj)->as.heap.ivptr, VALUE, newsize);
- newptr = ROBJECT(obj)->as.heap.ivptr;
- }
- for (; len < newsize; len++)
- newptr[len] = Qundef;
- ROBJECT(obj)->as.heap.numiv = newsize;
- ROBJECT(obj)->as.heap.iv_index_tbl = ivup.u.iv_index_tbl;
- }
- }
- RB_OBJ_WRITE(obj, &ROBJECT_IVPTR(obj)[ivup.index], val);
- break;
+ return obj_field_set(obj, target_shape_id, field_name, val);
case T_CLASS:
case T_MODULE:
- if (!RCLASS_IV_TBL(obj)) RCLASS_IV_TBL(obj) = st_init_numtable();
- rb_class_ivar_set(obj, id, val);
+ // The only field is object_id and T_CLASS handle it differently.
+ rb_bug("Unreachable");
break;
default:
- generic_ivar_set(obj, id, val);
- break;
+ return generic_field_set(obj, target_shape_id, field_name, val);
+ }
+}
+
+static VALUE
+ivar_defined0(VALUE obj, ID id)
+{
+ attr_index_t index;
+
+ if (rb_shape_obj_too_complex_p(obj)) {
+ VALUE idx;
+ st_table *table = NULL;
+ switch (BUILTIN_TYPE(obj)) {
+ case T_CLASS:
+ case T_MODULE:
+ rb_bug("Unreachable");
+ break;
+
+ case T_IMEMO:
+ RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields));
+ table = rb_imemo_fields_complex_tbl(obj);
+ break;
+
+ case T_OBJECT:
+ table = ROBJECT_FIELDS_HASH(obj);
+ break;
+
+ default: {
+ VALUE fields_obj = rb_obj_fields_no_ractor_check(obj); // defined? doesn't require ractor checks
+ table = rb_imemo_fields_complex_tbl(fields_obj);
+ }
+ }
+
+ if (!table || !rb_st_lookup(table, id, &idx)) {
+ return Qfalse;
+ }
+
+ return Qtrue;
+ }
+ else {
+ return RBOOL(rb_shape_get_iv_index(RBASIC_SHAPE_ID(obj), id, &index));
}
- return val;
}
VALUE
rb_ivar_defined(VALUE obj, ID id)
{
- VALUE val;
- struct st_table *iv_index_tbl;
- st_data_t index;
-
if (SPECIAL_CONST_P(obj)) return Qfalse;
+
+ VALUE defined = Qfalse;
switch (BUILTIN_TYPE(obj)) {
- case T_OBJECT:
- iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj);
- if (!iv_index_tbl) break;
- if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break;
- if (ROBJECT_NUMIV(obj) <= index) break;
- val = ROBJECT_IVPTR(obj)[index];
- if (val != Qundef)
- return Qtrue;
- break;
case T_CLASS:
case T_MODULE:
- if (RCLASS_IV_TBL(obj) && st_lookup(RCLASS_IV_TBL(obj), (st_data_t)id, 0))
- return Qtrue;
- break;
+ {
+ VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ if (fields_obj) {
+ defined = ivar_defined0(fields_obj, id);
+ }
+ }
+ break;
default:
- if (FL_TEST(obj, FL_EXIVAR))
- return generic_ivar_defined(obj, id);
- break;
+ defined = ivar_defined0(obj, id);
+ break;
}
- return Qfalse;
+ return defined;
}
-struct obj_ivar_tag {
+struct iv_itr_data {
VALUE obj;
- int (*func)(ID key, VALUE val, st_data_t arg);
+ struct gen_fields_tbl *fields_tbl;
st_data_t arg;
+ rb_ivar_foreach_callback_func *func;
+ VALUE *fields;
+ bool ivar_only;
};
static int
-obj_ivar_i(st_data_t key, st_data_t index, st_data_t arg)
+iterate_over_shapes_callback(shape_id_t shape_id, void *data)
{
- struct obj_ivar_tag *data = (struct obj_ivar_tag *)arg;
- if (index < ROBJECT_NUMIV(data->obj)) {
- VALUE val = ROBJECT_IVPTR(data->obj)[index];
- if (val != Qundef) {
- return (data->func)((ID)key, val, data->arg);
- }
- }
- return ST_CONTINUE;
-}
+ struct iv_itr_data *itr_data = data;
-static void
-obj_ivar_each(VALUE obj, int (*func)(ANYARGS), st_data_t arg)
-{
- st_table *tbl;
- struct obj_ivar_tag data;
+ if (itr_data->ivar_only && !RSHAPE_TYPE_P(shape_id, SHAPE_IVAR)) {
+ return ST_CONTINUE;
+ }
- tbl = ROBJECT_IV_INDEX_TBL(obj);
- if (!tbl)
- return;
+ VALUE *fields;
+ switch (BUILTIN_TYPE(itr_data->obj)) {
+ case T_OBJECT:
+ RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj));
+ fields = ROBJECT_FIELDS(itr_data->obj);
+ break;
+ case T_IMEMO:
+ RUBY_ASSERT(IMEMO_TYPE_P(itr_data->obj, imemo_fields));
+ RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj));
- data.obj = obj;
- data.func = (int (*)(ID key, VALUE val, st_data_t arg))func;
- data.arg = arg;
+ fields = rb_imemo_fields_ptr(itr_data->obj);
+ break;
+ default:
+ rb_bug("Unreachable");
+ }
- st_foreach_safe(tbl, obj_ivar_i, (st_data_t)&data);
+ VALUE val = fields[RSHAPE_INDEX(shape_id)];
+ return itr_data->func(RSHAPE_EDGE_NAME(shape_id), val, itr_data->arg);
}
-struct gen_ivar_tag {
- struct gen_ivtbl *ivtbl;
- int (*func)(ID key, VALUE val, st_data_t arg);
- st_data_t arg;
-};
+/*
+ * Returns a flag to stop iterating depending on the result of +callback+.
+ */
+static void
+iterate_over_shapes(shape_id_t shape_id, rb_ivar_foreach_callback_func *callback, struct iv_itr_data *itr_data)
+{
+ rb_shape_foreach_field(shape_id, iterate_over_shapes_callback, itr_data);
+}
static int
-gen_ivar_each_i(st_data_t key, st_data_t index, st_data_t data)
+each_hash_iv(st_data_t id, st_data_t val, st_data_t data)
{
- struct gen_ivar_tag *arg = (struct gen_ivar_tag *)data;
-
- if (index < arg->ivtbl->numiv) {
- VALUE val = arg->ivtbl->ivptr[index];
- if (val != Qundef) {
- return (arg->func)((ID)key, val, arg->arg);
- }
+ struct iv_itr_data * itr_data = (struct iv_itr_data *)data;
+ rb_ivar_foreach_callback_func *callback = itr_data->func;
+ if (is_internal_id((ID)id)) {
+ return ST_CONTINUE;
}
- return ST_CONTINUE;
+ return callback((ID)id, (VALUE)val, itr_data->arg);
}
static void
-gen_ivar_each(VALUE obj, int (*func)(ANYARGS), st_data_t arg)
+obj_fields_each(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only)
{
- struct gen_ivar_tag data;
- st_table *iv_index_tbl = RCLASS_IV_INDEX_TBL(rb_obj_class(obj));
-
- if (!iv_index_tbl) return;
- if (!gen_ivtbl_get(obj, &data.ivtbl)) return;
+ struct iv_itr_data itr_data = {
+ .obj = obj,
+ .arg = arg,
+ .func = func,
+ .ivar_only = ivar_only,
+ };
- data.func = (int (*)(ID key, VALUE val, st_data_t arg))func;
- data.arg = arg;
-
- st_foreach_safe(iv_index_tbl, gen_ivar_each_i, (st_data_t)&data);
+ shape_id_t shape_id = RBASIC_SHAPE_ID(obj);
+ if (rb_shape_too_complex_p(shape_id)) {
+ rb_st_foreach(ROBJECT_FIELDS_HASH(obj), each_hash_iv, (st_data_t)&itr_data);
+ }
+ else {
+ itr_data.fields = ROBJECT_FIELDS(obj);
+ iterate_over_shapes(shape_id, func, &itr_data);
+ }
}
-struct givar_copy {
- VALUE obj;
- st_table *iv_index_tbl;
- struct gen_ivtbl *ivtbl;
-};
-
-static int
-gen_ivar_copy(ID id, VALUE val, st_data_t arg)
+static void
+imemo_fields_each(VALUE fields_obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only)
{
- struct givar_copy *c = (struct givar_copy *)arg;
- struct ivar_update ivup;
-
- ivup.iv_extended = 0;
- ivup.u.iv_index_tbl = c->iv_index_tbl;
- iv_index_tbl_extend(&ivup, id);
- if (ivup.index >= c->ivtbl->numiv) {
- uint32_t newsize = iv_index_tbl_newsize(&ivup);
- c->ivtbl = gen_ivtbl_resize(c->ivtbl, newsize);
- }
- c->ivtbl->ivptr[ivup.index] = val;
+ IMEMO_TYPE_P(fields_obj, imemo_fields);
- RB_OBJ_WRITTEN(c->obj, Qundef, val);
+ struct iv_itr_data itr_data = {
+ .obj = fields_obj,
+ .arg = arg,
+ .func = func,
+ .ivar_only = ivar_only,
+ };
- return ST_CONTINUE;
+ shape_id_t shape_id = RBASIC_SHAPE_ID(fields_obj);
+ if (rb_shape_too_complex_p(shape_id)) {
+ rb_st_foreach(rb_imemo_fields_complex_tbl(fields_obj), each_hash_iv, (st_data_t)&itr_data);
+ }
+ else {
+ itr_data.fields = rb_imemo_fields_ptr(fields_obj);
+ iterate_over_shapes(shape_id, func, &itr_data);
+ }
}
void
-rb_copy_generic_ivar(VALUE clone, VALUE obj)
+rb_copy_generic_ivar(VALUE dest, VALUE obj)
{
- struct gen_ivtbl *ivtbl;
+ VALUE new_fields_obj;
- rb_check_frozen(clone);
+ rb_check_frozen(dest);
- if (!FL_TEST(obj, FL_EXIVAR)) {
- clear:
- if (FL_TEST(clone, FL_EXIVAR)) {
- rb_free_generic_ivar(clone);
- FL_UNSET(clone, FL_EXIVAR);
- }
+ if (!rb_obj_gen_fields_p(obj)) {
return;
}
- if (gen_ivtbl_get(obj, &ivtbl)) {
- struct givar_copy c;
- uint32_t i;
- if (gen_ivtbl_count(ivtbl) == 0)
- goto clear;
+ shape_id_t src_shape_id = rb_obj_shape_id(obj);
+
+ VALUE fields_obj = rb_obj_fields_no_ractor_check(obj);
+ if (fields_obj) {
+ unsigned long src_num_ivs = rb_ivar_count(fields_obj);
+ if (!src_num_ivs) {
+ goto clear;
+ }
+
+ if (rb_shape_too_complex_p(src_shape_id)) {
+ rb_shape_copy_complex_ivars(dest, obj, src_shape_id, rb_imemo_fields_complex_tbl(fields_obj));
+ return;
+ }
+
+ shape_id_t dest_shape_id = src_shape_id;
+ shape_id_t initial_shape_id = rb_obj_shape_id(dest);
- if (gen_ivtbl_get(clone, &c.ivtbl)) {
- for (i = 0; i < c.ivtbl->numiv; i++)
- c.ivtbl->ivptr[i] = Qundef;
- }
- else {
- c.ivtbl = gen_ivtbl_resize(0, ivtbl->numiv);
- FL_SET(clone, FL_EXIVAR);
- }
+ if (!rb_shape_canonical_p(src_shape_id)) {
+ RUBY_ASSERT(RSHAPE_TYPE_P(initial_shape_id, SHAPE_ROOT));
+
+ dest_shape_id = rb_shape_rebuild(initial_shape_id, src_shape_id);
+ if (UNLIKELY(rb_shape_too_complex_p(dest_shape_id))) {
+ st_table *table = rb_st_init_numtable_with_size(src_num_ivs);
+ rb_obj_copy_ivs_to_hash_table(obj, table);
+ rb_obj_init_too_complex(dest, table);
+ return;
+ }
+ }
+
+ if (!RSHAPE_LEN(dest_shape_id)) {
+ RBASIC_SET_SHAPE_ID(dest, dest_shape_id);
+ return;
+ }
- c.iv_index_tbl = iv_index_tbl_make(clone);
- c.obj = clone;
- gen_ivar_each(obj, gen_ivar_copy, (st_data_t)&c);
- /*
- * c.ivtbl may change in gen_ivar_copy due to realloc,
- * no need to free
- */
- st_insert(generic_iv_tbl, (st_data_t)clone, (st_data_t)c.ivtbl);
+ new_fields_obj = rb_imemo_fields_new(dest, RSHAPE_CAPACITY(dest_shape_id), RB_OBJ_SHAREABLE_P(dest));
+ VALUE *src_buf = rb_imemo_fields_ptr(fields_obj);
+ VALUE *dest_buf = rb_imemo_fields_ptr(new_fields_obj);
+ rb_shape_copy_fields(new_fields_obj, dest_buf, dest_shape_id, src_buf, src_shape_id);
+ RBASIC_SET_SHAPE_ID(new_fields_obj, dest_shape_id);
+
+ rb_obj_replace_fields(dest, new_fields_obj);
+ }
+ return;
+
+ clear:
+ rb_free_generic_ivar(dest);
+}
+
+void
+rb_replace_generic_ivar(VALUE clone, VALUE obj)
+{
+ RB_VM_LOCKING() {
+ st_data_t fields_tbl, obj_data = (st_data_t)obj;
+ if (st_delete(generic_fields_tbl_, &obj_data, &fields_tbl)) {
+ st_insert(generic_fields_tbl_, (st_data_t)clone, fields_tbl);
+ RB_OBJ_WRITTEN(clone, Qundef, fields_tbl);
+ }
+ else {
+ rb_bug("unreachable");
+ }
}
}
void
-rb_ivar_foreach(VALUE obj, int (*func)(ANYARGS), st_data_t arg)
+rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only)
{
if (SPECIAL_CONST_P(obj)) return;
switch (BUILTIN_TYPE(obj)) {
+ case T_IMEMO:
+ if (IMEMO_TYPE_P(obj, imemo_fields)) {
+ imemo_fields_each(obj, func, arg, ivar_only);
+ }
+ break;
case T_OBJECT:
- obj_ivar_each(obj, func, arg);
- break;
+ obj_fields_each(obj, func, arg, ivar_only);
+ break;
case T_CLASS:
case T_MODULE:
- if (RCLASS_IV_TBL(obj)) {
- st_foreach_safe(RCLASS_IV_TBL(obj), func, arg);
- }
- break;
+ {
+ IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(0);
+ VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ if (fields_obj) {
+ imemo_fields_each(fields_obj, func, arg, ivar_only);
+ }
+ }
+ break;
default:
- if (FL_TEST(obj, FL_EXIVAR)) {
- gen_ivar_each(obj, func, arg);
- }
- break;
+ {
+ VALUE fields_obj = rb_obj_fields_no_ractor_check(obj);
+ if (fields_obj) {
+ imemo_fields_each(fields_obj, func, arg, ivar_only);
+ }
+ }
+ break;
}
}
+void
+rb_ivar_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg)
+{
+ rb_field_foreach(obj, func, arg, true);
+}
+
st_index_t
rb_ivar_count(VALUE obj)
{
- st_table *tbl;
-
if (SPECIAL_CONST_P(obj)) return 0;
+ st_index_t iv_count = 0;
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
- if ((tbl = ROBJECT_IV_INDEX_TBL(obj)) != 0) {
- st_index_t i, count, num = ROBJECT_NUMIV(obj);
- const VALUE *const ivptr = ROBJECT_IVPTR(obj);
- for (i = count = 0; i < num; ++i) {
- if (ivptr[i] != Qundef) {
- count++;
- }
- }
- return count;
- }
- break;
+ iv_count = ROBJECT_FIELDS_COUNT(obj);
+ break;
+
case T_CLASS:
case T_MODULE:
- if ((tbl = RCLASS_IV_TBL(obj)) != 0) {
- return tbl->num_entries;
- }
- break;
+ {
+ VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ if (!fields_obj) {
+ return 0;
+ }
+ if (rb_shape_obj_too_complex_p(fields_obj)) {
+ iv_count = rb_st_table_size(rb_imemo_fields_complex_tbl(fields_obj));
+ }
+ else {
+ iv_count = RBASIC_FIELDS_COUNT(fields_obj);
+ }
+ }
+ break;
+
+ case T_IMEMO:
+ RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_fields));
+
+ if (rb_shape_obj_too_complex_p(obj)) {
+ iv_count = rb_st_table_size(rb_imemo_fields_complex_tbl(obj));
+ }
+ else {
+ iv_count = RBASIC_FIELDS_COUNT(obj);
+ }
+ break;
+
default:
- if (FL_TEST(obj, FL_EXIVAR)) {
- struct gen_ivtbl *ivtbl;
+ {
+ VALUE fields_obj = rb_obj_fields_no_ractor_check(obj);
+ if (fields_obj) {
+ if (rb_shape_obj_too_complex_p(fields_obj)) {
+ rb_st_table_size(rb_imemo_fields_complex_tbl(fields_obj));
+ }
+ else {
+ iv_count = RBASIC_FIELDS_COUNT(obj);
+ }
+ }
+ }
+ break;
+ }
- if (gen_ivtbl_get(obj, &ivtbl)) {
- return gen_ivtbl_count(ivtbl);
- }
- }
- break;
+ if (rb_shape_obj_has_id(obj)) {
+ iv_count--;
}
- return 0;
+
+ return iv_count;
}
static int
-ivar_i(st_data_t k, st_data_t v, st_data_t a)
+ivar_i(ID key, VALUE v, st_data_t a)
{
- ID key = (ID)k;
VALUE ary = (VALUE)a;
if (rb_is_instance_id(key)) {
- rb_ary_push(ary, ID2SYM(key));
+ rb_ary_push(ary, ID2SYM(key));
}
return ST_CONTINUE;
}
@@ -1645,20 +2430,20 @@ rb_obj_instance_variables(VALUE obj)
#define rb_is_constant_id rb_is_const_id
#define rb_is_constant_name rb_is_const_name
#define id_for_var(obj, name, part, type) \
- id_for_var_message(obj, name, type, "`%1$s' is not allowed as "#part" "#type" variable name")
+ id_for_var_message(obj, name, type, "'%1$s' is not allowed as "#part" "#type" variable name")
#define id_for_var_message(obj, name, type, message) \
check_id_type(obj, &(name), rb_is_##type##_id, rb_is_##type##_name, message, strlen(message))
static ID
check_id_type(VALUE obj, VALUE *pname,
- int (*valid_id_p)(ID), int (*valid_name_p)(VALUE),
- const char *message, size_t message_len)
+ int (*valid_id_p)(ID), int (*valid_name_p)(VALUE),
+ const char *message, size_t message_len)
{
ID id = rb_check_id(pname);
VALUE name = *pname;
if (id ? !valid_id_p(id) : !valid_name_p(name)) {
- rb_name_err_raise_str(rb_fstring_new(message, message_len),
- obj, name);
+ rb_name_err_raise_str(rb_fstring_new(message, message_len),
+ obj, name);
}
return id;
}
@@ -1666,9 +2451,10 @@ check_id_type(VALUE obj, VALUE *pname,
/*
* call-seq:
* obj.remove_instance_variable(symbol) -> obj
+ * obj.remove_instance_variable(string) -> obj
*
* Removes the named instance variable from <i>obj</i>, returning that
- * variable's value.
+ * variable's value. The name can be passed as a symbol or as a string.
*
* class Dummy
* attr_reader :var
@@ -1688,49 +2474,21 @@ check_id_type(VALUE obj, VALUE *pname,
VALUE
rb_obj_remove_instance_variable(VALUE obj, VALUE name)
{
- VALUE val = Qnil;
const ID id = id_for_var(obj, name, an, instance);
- st_data_t n, v;
- struct st_table *iv_index_tbl;
- st_data_t index;
+ // Frozen check comes here because it's expected that we raise a
+ // NameError (from the id_for_var check) before we raise a FrozenError
rb_check_frozen(obj);
- if (!id) {
- goto not_defined;
- }
- switch (BUILTIN_TYPE(obj)) {
- case T_OBJECT:
- iv_index_tbl = ROBJECT_IV_INDEX_TBL(obj);
- if (!iv_index_tbl) break;
- if (!st_lookup(iv_index_tbl, (st_data_t)id, &index)) break;
- if (ROBJECT_NUMIV(obj) <= index) break;
- val = ROBJECT_IVPTR(obj)[index];
- if (val != Qundef) {
- ROBJECT_IVPTR(obj)[index] = Qundef;
- return val;
- }
- break;
- case T_CLASS:
- case T_MODULE:
- n = id;
- if (RCLASS_IV_TBL(obj) && st_delete(RCLASS_IV_TBL(obj), &n, &v)) {
- return (VALUE)v;
- }
- break;
- default:
- if (FL_TEST(obj, FL_EXIVAR)) {
- if (generic_ivar_remove(obj, id, &val)) {
- return val;
- }
- }
- break;
+ if (id) {
+ VALUE val = rb_ivar_delete(obj, id, Qundef);
+
+ if (!UNDEF_P(val)) return val;
}
- not_defined:
rb_name_err_raise("instance variable %1$s not defined",
- obj, name);
- UNREACHABLE;
+ obj, name);
+ UNREACHABLE_RETURN(Qnil);
}
NORETURN(static void uninitialized_constant(VALUE, VALUE));
@@ -1738,17 +2496,17 @@ static void
uninitialized_constant(VALUE klass, VALUE name)
{
if (klass && rb_class_real(klass) != rb_cObject)
- rb_name_err_raise("uninitialized constant %2$s::%1$s",
- klass, name);
+ rb_name_err_raise("uninitialized constant %2$s::%1$s",
+ klass, name);
else
- rb_name_err_raise("uninitialized constant %1$s",
- klass, name);
+ rb_name_err_raise("uninitialized constant %1$s",
+ klass, name);
}
VALUE
rb_const_missing(VALUE klass, VALUE name)
{
- VALUE value = rb_funcallv(klass, rb_intern("const_missing"), 1, &name);
+ VALUE value = rb_funcallv(klass, idConst_missing, 1, &name);
rb_vm_inc_const_missing_count();
return value;
}
@@ -1760,8 +2518,7 @@ rb_const_missing(VALUE klass, VALUE name)
*
* Invoked when a reference is made to an undefined constant in
* <i>mod</i>. It is passed a symbol for the undefined constant, and
- * returns a value to be used for that constant. The
- * following code is an example of the same:
+ * returns a value to be used for that constant. For example, consider:
*
* def Foo.const_missing(name)
* name # return the constant name as Symbol
@@ -1769,23 +2526,28 @@ rb_const_missing(VALUE klass, VALUE name)
*
* Foo::UNDEFINED_CONST #=> :UNDEFINED_CONST: symbol returned
*
- * In the next example when a reference is made to an undefined constant,
- * it attempts to load a file whose name is the lowercase version of the
- * constant (thus class <code>Fred</code> is assumed to be in file
- * <code>fred.rb</code>). If found, it returns the loaded class. It
- * therefore implements an autoload feature similar to Kernel#autoload and
- * Module#autoload.
+ * As the example above shows, +const_missing+ is not required to create the
+ * missing constant in <i>mod</i>, though that is often a side-effect. The
+ * caller gets its return value when triggered. If the constant is also defined,
+ * further lookups won't hit +const_missing+ and will return the value stored in
+ * the constant as usual. Otherwise, +const_missing+ will be invoked again.
+ *
+ * In the next example, when a reference is made to an undefined constant,
+ * +const_missing+ attempts to load a file whose path is the lowercase version
+ * of the constant name (thus class <code>Fred</code> is assumed to be in file
+ * <code>fred.rb</code>). If defined as a side-effect of loading the file, the
+ * method returns the value stored in the constant. This implements an autoload
+ * feature similar to Kernel#autoload and Module#autoload, though it differs in
+ * important ways.
*
* def Object.const_missing(name)
* @looked_for ||= {}
* str_name = name.to_s
- * raise "Class not found: #{name}" if @looked_for[str_name]
+ * raise "Constant not found: #{name}" if @looked_for[str_name]
* @looked_for[str_name] = 1
* file = str_name.downcase
* require file
- * klass = const_get(name)
- * return klass if klass
- * raise "Class not found: #{name}"
+ * const_get(name, false)
* end
*
*/
@@ -1793,39 +2555,51 @@ rb_const_missing(VALUE klass, VALUE name)
VALUE
rb_mod_const_missing(VALUE klass, VALUE name)
{
+ rb_execution_context_t *ec = GET_EC();
+ VALUE ref = ec->private_const_reference;
rb_vm_pop_cfunc_frame();
+ if (ref) {
+ ec->private_const_reference = 0;
+ rb_name_err_raise("private constant %2$s::%1$s referenced", ref, name);
+ }
uninitialized_constant(klass, name);
- UNREACHABLE;
+ UNREACHABLE_RETURN(Qnil);
}
static void
-autoload_mark(void *ptr)
+autoload_table_mark(void *ptr)
{
- rb_mark_tbl((st_table *)ptr);
+ rb_mark_tbl_no_pin((st_table *)ptr);
}
static void
-autoload_free(void *ptr)
+autoload_table_free(void *ptr)
{
st_free_table((st_table *)ptr);
}
static size_t
-autoload_memsize(const void *ptr)
+autoload_table_memsize(const void *ptr)
{
const st_table *tbl = ptr;
return st_memsize(tbl);
}
-static const rb_data_type_t autoload_data_type = {
- "autoload",
- {autoload_mark, autoload_free, autoload_memsize,},
- 0, 0, RUBY_TYPED_FREE_IMMEDIATELY
+static void
+autoload_table_compact(void *ptr)
+{
+ rb_gc_ref_update_table_values_only((st_table *)ptr);
+}
+
+static const rb_data_type_t autoload_table_type = {
+ "autoload_table",
+ {autoload_table_mark, autoload_table_free, autoload_table_memsize, autoload_table_compact,},
+ 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED
};
#define check_autoload_table(av) \
- (struct st_table *)rb_check_typeddata((av), &autoload_data_type)
+ (struct st_table *)rb_check_typeddata((av), &autoload_table_type)
static VALUE
autoload_data(VALUE mod, ID id)
@@ -1833,219 +2607,489 @@ autoload_data(VALUE mod, ID id)
struct st_table *tbl;
st_data_t val;
- if (!st_lookup(RCLASS_IV_TBL(mod), autoload, &val) ||
- !(tbl = check_autoload_table((VALUE)val)) ||
- !st_lookup(tbl, (st_data_t)id, &val)) {
- return 0;
+ // If we are called with a non-origin ICLASS, fetch the autoload data from
+ // the original module.
+ if (RB_TYPE_P(mod, T_ICLASS)) {
+ if (RICLASS_IS_ORIGIN_P(mod)) {
+ return 0;
+ }
+ else {
+ mod = RBASIC(mod)->klass;
+ }
}
+
+ RUBY_ASSERT(RB_TYPE_P(mod, T_CLASS) || RB_TYPE_P(mod, T_MODULE));
+
+ // Look up the instance variable table for `autoload`, then index into that table with the given constant name `id`.
+
+ VALUE tbl_value = rb_ivar_lookup(mod, autoload, Qfalse);
+ if (!RTEST(tbl_value) || !(tbl = check_autoload_table(tbl_value)) || !st_lookup(tbl, (st_data_t)id, &val)) {
+ return 0;
+ }
+
return (VALUE)val;
}
-/* always on stack, no need to mark */
-struct autoload_state {
- struct autoload_data_i *ele;
- VALUE mod;
- VALUE result;
- ID id;
- VALUE thread;
- union {
- struct list_node node;
- struct list_head head;
- } waitq;
+// Every autoload constant has exactly one instance of autoload_const, stored in `autoload_features`. Since multiple autoload constants can refer to the same file, every `autoload_const` refers to a de-duplicated `autoload_data`.
+struct autoload_const {
+ // The linked list node of all constants which are loaded by the related autoload feature.
+ struct ccan_list_node cnode; /* <=> autoload_data.constants */
+
+ // The shared "autoload_data" if multiple constants are defined from the same feature.
+ VALUE autoload_data_value;
+
+ // The box object when the autoload is called in a user box
+ // Otherwise, Qnil means the root box
+ VALUE box_value;
+
+ // The module we are loading a constant into.
+ VALUE module;
+
+ // The name of the constant we are loading.
+ ID name;
+
+ // The value of the constant (after it's loaded).
+ VALUE value;
+
+ // The constant entry flags which need to be re-applied after autoloading the feature.
+ rb_const_flag_t flag;
+
+ // The source file and line number that defined this constant (different from feature path).
+ VALUE file;
+ int line;
};
-struct autoload_data_i {
+// Each `autoload_data` uniquely represents a specific feature which can be loaded, and a list of constants which it is able to define. We use a mutex to coordinate multiple threads trying to load the same feature.
+struct autoload_data {
+ // The feature path to require to load this constant.
VALUE feature;
- int safe_level;
- rb_const_flag_t flag;
- VALUE value;
- struct autoload_state *state; /* points to on-stack struct */
+
+ // The mutex which is protecting autoloading this feature.
+ VALUE mutex;
+
+ // The process fork serial number since the autoload mutex will become invalid on fork.
rb_serial_t fork_gen;
+
+ // The linked list of all constants that are going to be loaded by this autoload.
+ struct ccan_list_head constants; /* <=> autoload_const.cnode */
};
static void
-autoload_i_mark(void *ptr)
+autoload_data_mark_and_move(void *ptr)
{
- struct autoload_data_i *p = ptr;
- rb_gc_mark(p->feature);
- rb_gc_mark(p->value);
+ struct autoload_data *p = ptr;
+
+ rb_gc_mark_and_move(&p->feature);
+ rb_gc_mark_and_move(&p->mutex);
+}
+
+static void
+autoload_data_free(void *ptr)
+{
+ struct autoload_data *p = ptr;
+
+ struct autoload_const *autoload_const, *next;
+ ccan_list_for_each_safe(&p->constants, autoload_const, next, cnode) {
+ ccan_list_del_init(&autoload_const->cnode);
+ }
+
+ ruby_xfree(p);
+}
+
+static size_t
+autoload_data_memsize(const void *ptr)
+{
+ return sizeof(struct autoload_data);
+}
+
+static const rb_data_type_t autoload_data_type = {
+ "autoload_data",
+ {autoload_data_mark_and_move, autoload_data_free, autoload_data_memsize, autoload_data_mark_and_move},
+ 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED
+};
+
+static void
+autoload_const_mark_and_move(void *ptr)
+{
+ struct autoload_const *ac = ptr;
+
+ rb_gc_mark_and_move(&ac->module);
+ rb_gc_mark_and_move(&ac->autoload_data_value);
+ rb_gc_mark_and_move(&ac->value);
+ rb_gc_mark_and_move(&ac->file);
+ rb_gc_mark_and_move(&ac->box_value);
}
static size_t
-autoload_i_memsize(const void *ptr)
+autoload_const_memsize(const void *ptr)
+{
+ return sizeof(struct autoload_const);
+}
+
+static void
+autoload_const_free(void *ptr)
{
- return sizeof(struct autoload_data_i);
+ struct autoload_const *autoload_const = ptr;
+
+ ccan_list_del(&autoload_const->cnode);
+ ruby_xfree(ptr);
}
-static const rb_data_type_t autoload_data_i_type = {
- "autoload_i",
- {autoload_i_mark, RUBY_TYPED_DEFAULT_FREE, autoload_i_memsize,},
- 0, 0, RUBY_TYPED_FREE_IMMEDIATELY
+static const rb_data_type_t autoload_const_type = {
+ "autoload_const",
+ {autoload_const_mark_and_move, autoload_const_free, autoload_const_memsize, autoload_const_mark_and_move,},
+ 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED
};
-static struct autoload_data_i *
-get_autoload_data(VALUE av)
+static struct autoload_data *
+get_autoload_data(VALUE autoload_const_value, struct autoload_const **autoload_const_pointer)
{
- struct autoload_data_i *ele = rb_check_typeddata(av, &autoload_data_i_type);
+ struct autoload_const *autoload_const = rb_check_typeddata(autoload_const_value, &autoload_const_type);
+
+ VALUE autoload_data_value = autoload_const->autoload_data_value;
+ struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type);
/* do not reach across stack for ->state after forking: */
- if (ele && ele->state && ele->fork_gen != GET_VM()->fork_gen) {
- ele->state = 0;
- ele->fork_gen = 0;
+ if (autoload_data && autoload_data->fork_gen != GET_VM()->fork_gen) {
+ RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil);
+ autoload_data->fork_gen = 0;
}
- return ele;
+
+ if (autoload_const_pointer) *autoload_const_pointer = autoload_const;
+
+ return autoload_data;
+}
+
+struct autoload_copy_table_data {
+ VALUE dst_tbl_value;
+ struct st_table *dst_tbl;
+ const rb_box_t *box;
+};
+
+static int
+autoload_copy_table_for_box_i(st_data_t key, st_data_t value, st_data_t arg)
+{
+ struct autoload_const *autoload_const;
+ struct autoload_copy_table_data *data = (struct autoload_copy_table_data *)arg;
+ struct st_table *tbl = data->dst_tbl;
+ VALUE tbl_value = data->dst_tbl_value;
+ const rb_box_t *box = data->box;
+
+ VALUE src_value = (VALUE)value;
+ struct autoload_const *src_const = rb_check_typeddata(src_value, &autoload_const_type);
+ // autoload_data can be shared between copies because the feature is equal between copies.
+ VALUE autoload_data_value = src_const->autoload_data_value;
+ struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type);
+
+ VALUE new_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const);
+ RB_OBJ_WRITE(new_value, &autoload_const->box_value, rb_get_box_object((rb_box_t *)box));
+ RB_OBJ_WRITE(new_value, &autoload_const->module, src_const->module);
+ autoload_const->name = src_const->name;
+ RB_OBJ_WRITE(new_value, &autoload_const->value, src_const->value);
+ autoload_const->flag = src_const->flag;
+ RB_OBJ_WRITE(new_value, &autoload_const->autoload_data_value, autoload_data_value);
+ ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode);
+
+ st_insert(tbl, (st_data_t)autoload_const->name, (st_data_t)new_value);
+ RB_OBJ_WRITTEN(tbl_value, Qundef, new_value);
+
+ return ST_CONTINUE;
}
-RUBY_FUNC_EXPORTED void
-rb_autoload(VALUE mod, ID id, const char *file)
+void
+rb_autoload_copy_table_for_box(st_table *iv_ptr, const rb_box_t *box)
{
- if (!file || !*file) {
- rb_raise(rb_eArgError, "empty file name");
+ struct st_table *src_tbl, *dst_tbl;
+ VALUE src_tbl_value, dst_tbl_value;
+ if (!rb_st_lookup(iv_ptr, (st_data_t)autoload, (st_data_t *)&src_tbl_value)) {
+ // the class has no autoload table yet.
+ return;
}
- rb_autoload_str(mod, id, rb_fstring_cstr(file));
+ if (!RTEST(src_tbl_value) || !(src_tbl = check_autoload_table(src_tbl_value))) {
+ // the __autoload__ ivar value isn't autoload table value.
+ return;
+ }
+ src_tbl = check_autoload_table(src_tbl_value);
+
+ dst_tbl_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL);
+ RTYPEDDATA_DATA(dst_tbl_value) = dst_tbl = st_init_numtable();
+
+ struct autoload_copy_table_data data = {
+ .dst_tbl_value = dst_tbl_value,
+ .dst_tbl = dst_tbl,
+ .box = box,
+ };
+
+ st_foreach(src_tbl, autoload_copy_table_for_box_i, (st_data_t)&data);
+ st_insert(iv_ptr, (st_data_t)autoload, (st_data_t)dst_tbl_value);
}
void
-rb_autoload_str(VALUE mod, ID id, VALUE file)
+rb_autoload(VALUE module, ID name, const char *feature)
{
- st_data_t av;
- VALUE ad;
- struct st_table *tbl;
- struct autoload_data_i *ele;
- rb_const_entry_t *ce;
-
- if (!rb_is_const_id(id)) {
- rb_raise(rb_eNameError, "autoload must be constant name: %"PRIsVALUE"",
- QUOTE_ID(id));
+ if (!feature || !*feature) {
+ rb_raise(rb_eArgError, "empty feature name");
}
- Check_Type(file, T_STRING);
- if (!RSTRING_LEN(file)) {
- rb_raise(rb_eArgError, "empty file name");
- }
+ rb_autoload_str(module, name, rb_fstring_cstr(feature));
+}
- ce = rb_const_lookup(mod, id);
- if (ce && ce->value != Qundef) {
- return;
+static void const_set(VALUE klass, ID id, VALUE val);
+static void const_added(VALUE klass, ID const_name);
+
+struct autoload_arguments {
+ VALUE module;
+ ID name;
+ VALUE feature;
+ VALUE box_value;
+};
+
+static VALUE
+autoload_feature_lookup_or_create(VALUE feature, struct autoload_data **autoload_data_pointer)
+{
+ RUBY_ASSERT_MUTEX_OWNED(autoload_mutex);
+ RUBY_ASSERT_CRITICAL_SECTION_ENTER();
+
+ VALUE autoload_data_value = rb_hash_aref(autoload_features, feature);
+ struct autoload_data *autoload_data;
+
+ if (NIL_P(autoload_data_value)) {
+ autoload_data_value = TypedData_Make_Struct(0, struct autoload_data, &autoload_data_type, autoload_data);
+ RB_OBJ_WRITE(autoload_data_value, &autoload_data->feature, feature);
+ RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil);
+ ccan_list_head_init(&autoload_data->constants);
+
+ if (autoload_data_pointer) *autoload_data_pointer = autoload_data;
+
+ rb_hash_aset(autoload_features, feature, autoload_data_value);
+ }
+ else if (autoload_data_pointer) {
+ *autoload_data_pointer = rb_check_typeddata(autoload_data_value, &autoload_data_type);
}
- rb_const_set(mod, id, Qundef);
- tbl = RCLASS_IV_TBL(mod);
- if (tbl && st_lookup(tbl, (st_data_t)autoload, &av)) {
- tbl = check_autoload_table((VALUE)av);
+ RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
+ return autoload_data_value;
+}
+
+static VALUE
+autoload_table_lookup_or_create(VALUE module)
+{
+ VALUE autoload_table_value = rb_ivar_lookup(module, autoload, Qfalse);
+ if (RTEST(autoload_table_value)) {
+ return autoload_table_value;
}
else {
- if (!tbl) tbl = RCLASS_IV_TBL(mod) = st_init_numtable();
- av = (st_data_t)TypedData_Wrap_Struct(0, &autoload_data_type, 0);
- st_add_direct(tbl, (st_data_t)autoload, av);
- RB_OBJ_WRITTEN(mod, Qnil, av);
- DATA_PTR(av) = tbl = st_init_numtable();
+ autoload_table_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL);
+ rb_class_ivar_set(module, autoload, autoload_table_value);
+ RTYPEDDATA_DATA(autoload_table_value) = st_init_numtable();
+ return autoload_table_value;
}
+}
- ad = TypedData_Make_Struct(0, struct autoload_data_i, &autoload_data_i_type, ele);
- if (OBJ_TAINTED(file)) {
- file = rb_str_dup(file);
- FL_UNSET(file, FL_TAINT);
+static VALUE
+autoload_synchronized(VALUE _arguments)
+{
+ struct autoload_arguments *arguments = (struct autoload_arguments *)_arguments;
+
+ rb_const_entry_t *constant_entry = rb_const_lookup(arguments->module, arguments->name);
+ if (constant_entry && !UNDEF_P(constant_entry->value)) {
+ return Qfalse;
}
- ele->feature = rb_fstring(file);
- ele->safe_level = rb_safe_level();
- ele->value = Qundef;
- ele->state = 0;
- ele->flag = CONST_PUBLIC;
- st_insert(tbl, (st_data_t)id, (st_data_t)ad);
+
+ // Reset any state associated with any previous constant:
+ const_set(arguments->module, arguments->name, Qundef);
+
+ VALUE autoload_table_value = autoload_table_lookup_or_create(arguments->module);
+ struct st_table *autoload_table = check_autoload_table(autoload_table_value);
+
+ // Ensure the string is uniqued since we use an identity lookup:
+ VALUE feature = rb_fstring(arguments->feature);
+
+ struct autoload_data *autoload_data;
+ VALUE autoload_data_value = autoload_feature_lookup_or_create(feature, &autoload_data);
+
+ {
+ struct autoload_const *autoload_const;
+ VALUE autoload_const_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const);
+ RB_OBJ_WRITE(autoload_const_value, &autoload_const->box_value, arguments->box_value);
+ RB_OBJ_WRITE(autoload_const_value, &autoload_const->module, arguments->module);
+ autoload_const->name = arguments->name;
+ autoload_const->value = Qundef;
+ autoload_const->flag = CONST_PUBLIC;
+ RB_OBJ_WRITE(autoload_const_value, &autoload_const->autoload_data_value, autoload_data_value);
+ ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode);
+ st_insert(autoload_table, (st_data_t)arguments->name, (st_data_t)autoload_const_value);
+ RB_OBJ_WRITTEN(autoload_table_value, Qundef, autoload_const_value);
+ }
+
+ return Qtrue;
}
-static void
-autoload_delete(VALUE mod, ID id)
+void
+rb_autoload_str(VALUE module, ID name, VALUE feature)
{
- st_data_t val, load = 0, n = id;
+ const rb_box_t *box = rb_current_box();
+ VALUE current_box_value = rb_get_box_object((rb_box_t *)box);
+
+ if (!rb_is_const_id(name)) {
+ rb_raise(rb_eNameError, "autoload must be constant name: %"PRIsVALUE"", QUOTE_ID(name));
+ }
+
+ Check_Type(feature, T_STRING);
+ if (!RSTRING_LEN(feature)) {
+ rb_raise(rb_eArgError, "empty feature name");
+ }
- if (st_lookup(RCLASS_IV_TBL(mod), (st_data_t)autoload, &val)) {
- struct st_table *tbl = check_autoload_table((VALUE)val);
+ struct autoload_arguments arguments = {
+ .module = module,
+ .name = name,
+ .feature = feature,
+ .box_value = current_box_value,
+ };
- st_delete(tbl, &n, &load);
+ VALUE result = rb_mutex_synchronize(autoload_mutex, autoload_synchronized, (VALUE)&arguments);
- if (tbl->num_entries == 0) {
- n = autoload;
- st_delete(RCLASS_IV_TBL(mod), &n, &val);
- }
+ if (result == Qtrue) {
+ const_added(module, name);
}
}
-static VALUE
-autoload_provided(VALUE arg)
+static void
+autoload_delete(VALUE module, ID name)
{
- const char **p = (const char **)arg;
- return rb_feature_provided(*p, p);
+ RUBY_ASSERT_CRITICAL_SECTION_ENTER();
+
+ st_data_t load = 0, key = name;
+
+ RUBY_ASSERT(RB_TYPE_P(module, T_CLASS) || RB_TYPE_P(module, T_MODULE));
+
+ VALUE table_value = rb_ivar_lookup(module, autoload, Qfalse);
+ if (RTEST(table_value)) {
+ struct st_table *table = check_autoload_table(table_value);
+
+ st_delete(table, &key, &load);
+ RB_OBJ_WRITTEN(table_value, load, Qundef);
+
+ /* Qfalse can indicate already deleted */
+ if (load != Qfalse) {
+ struct autoload_const *autoload_const;
+ struct autoload_data *autoload_data = get_autoload_data((VALUE)load, &autoload_const);
+
+ VM_ASSERT(autoload_data);
+ VM_ASSERT(!ccan_list_empty(&autoload_data->constants));
+
+ /*
+ * we must delete here to avoid "already initialized" warnings
+ * with parallel autoload. Using list_del_init here so list_del
+ * works in autoload_const_free
+ */
+ ccan_list_del_init(&autoload_const->cnode);
+
+ if (ccan_list_empty(&autoload_data->constants)) {
+ rb_hash_delete(autoload_features, autoload_data->feature);
+ }
+
+ // If the autoload table is empty, we can delete it.
+ if (table->num_entries == 0) {
+ rb_attr_delete(module, autoload);
+ }
+ }
+ }
+
+ RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
}
-static VALUE
-reset_safe(VALUE safe)
+static int
+autoload_by_someone_else(struct autoload_data *ele)
{
- rb_set_safe_level_force((int)safe);
- return safe;
+ return ele->mutex != Qnil && !rb_mutex_owned_p(ele->mutex);
}
static VALUE
check_autoload_required(VALUE mod, ID id, const char **loadingpath)
{
- VALUE file, load;
- struct autoload_data_i *ele;
+ VALUE autoload_const_value = autoload_data(mod, id);
+ struct autoload_data *autoload_data;
const char *loading;
- int safe;
- if (!(load = autoload_data(mod, id)) || !(ele = get_autoload_data(load))) {
- return 0;
- }
- file = ele->feature;
- Check_Type(file, T_STRING);
- if (!RSTRING_LEN(file) || !*RSTRING_PTR(file)) {
- rb_raise(rb_eArgError, "empty file name");
+ if (!autoload_const_value || !(autoload_data = get_autoload_data(autoload_const_value, 0))) {
+ return 0;
}
+ VALUE feature = autoload_data->feature;
+
/*
* if somebody else is autoloading, we MUST wait for them, since
* rb_provide_feature can provide a feature before autoload_const_set
* completes. We must wait until autoload_const_set finishes in
* the other thread.
*/
- if (ele->state && ele->state->thread != rb_thread_current()) {
- return load;
+ if (autoload_by_someone_else(autoload_data)) {
+ return autoload_const_value;
}
- loading = RSTRING_PTR(file);
- safe = rb_safe_level();
- rb_set_safe_level_force(0);
- if (!rb_ensure(autoload_provided, (VALUE)&loading, reset_safe, (VALUE)safe)) {
- return load;
+ loading = RSTRING_PTR(feature);
+
+ if (!rb_feature_provided(loading, &loading)) {
+ return autoload_const_value;
}
+
if (loadingpath && loading) {
- *loadingpath = loading;
- return load;
+ *loadingpath = loading;
+ return autoload_const_value;
}
+
return 0;
}
+static struct autoload_const *autoloading_const_entry(VALUE mod, ID id);
+
int
rb_autoloading_value(VALUE mod, ID id, VALUE* value, rb_const_flag_t *flag)
{
- VALUE load;
- struct autoload_data_i *ele;
-
- if (!(load = autoload_data(mod, id)) || !(ele = get_autoload_data(load))) {
- return 0;
- }
- if (ele->state && ele->state->thread == rb_thread_current()) {
- if (ele->value != Qundef) {
- if (value) {
- *value = ele->value;
- }
- if (flag) {
- *flag = ele->flag;
- }
- return 1;
- }
+ struct autoload_const *ac = autoloading_const_entry(mod, id);
+ if (!ac) return FALSE;
+
+ if (value) {
+ *value = ac->value;
+ }
+
+ if (flag) {
+ *flag = ac->flag;
+ }
+
+ return TRUE;
+}
+
+static int
+autoload_by_current(struct autoload_data *ele)
+{
+ return ele->mutex != Qnil && rb_mutex_owned_p(ele->mutex);
+}
+
+// If there is an autoloading constant and it has been set by the current
+// execution context, return it. This allows threads which are loading code to
+// refer to their own autoloaded constants.
+struct autoload_const *
+autoloading_const_entry(VALUE mod, ID id)
+{
+ VALUE load = autoload_data(mod, id);
+ struct autoload_data *ele;
+ struct autoload_const *ac;
+
+ // Find the autoloading state:
+ if (!load || !(ele = get_autoload_data(load, &ac))) {
+ // Couldn't be found:
+ return 0;
}
+
+ // Check if it's being loaded by the current thread/fiber:
+ if (autoload_by_current(ele)) {
+ if (!UNDEF_P(ac->value)) {
+ return ac;
+ }
+ }
+
return 0;
}
@@ -2054,258 +3098,346 @@ autoload_defined_p(VALUE mod, ID id)
{
rb_const_entry_t *ce = rb_const_lookup(mod, id);
- if (!ce || ce->value != Qundef) {
- return 0;
+ // If there is no constant or the constant is not undefined (special marker for autoloading):
+ if (!ce || !UNDEF_P(ce->value)) {
+ // We are not autoloading:
+ return 0;
}
+
+ // Otherwise check if there is an autoload in flight right now:
return !rb_autoloading_value(mod, id, NULL, NULL);
}
-struct autoload_const_set_args {
- VALUE mod;
- ID id;
- VALUE value;
- rb_const_flag_t flag;
-};
+static void const_tbl_update(struct autoload_const *, int);
-static void const_tbl_update(struct autoload_const_set_args *);
+struct autoload_load_arguments {
+ VALUE module;
+ ID name;
+ int flag;
-static VALUE
-autoload_const_set(VALUE arg)
-{
- struct autoload_const_set_args* args = (struct autoload_const_set_args *)arg;
- VALUE klass = args->mod;
- ID id = args->id;
- check_before_mod_set(klass, id, args->value, "constant");
- const_tbl_update(args);
- return 0; /* ignored */
-}
+ VALUE mutex;
+
+ // The specific constant which triggered the autoload code to fire:
+ struct autoload_const *autoload_const;
+
+ // The parent autoload data which is shared between multiple constants:
+ struct autoload_data *autoload_data;
+};
static VALUE
-autoload_require(VALUE arg)
+autoload_const_set(struct autoload_const *ac)
{
- struct autoload_state *state = (struct autoload_state *)arg;
+ check_before_mod_set(ac->module, ac->name, ac->value, "constant");
- /* this may release GVL and switch threads: */
- state->result = rb_funcall(rb_vm_top_self(), rb_intern("require"), 1,
- state->ele->feature);
+ RB_VM_LOCKING() {
+ const_tbl_update(ac, true);
+ }
- return state->result;
+ return 0; /* ignored */
}
static VALUE
-autoload_reset(VALUE arg)
+autoload_load_needed(VALUE _arguments)
{
- struct autoload_state *state = (struct autoload_state *)arg;
- int need_wakeups = 0;
+ struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
+
+ const char *loading = 0, *src;
+
+ if (!autoload_defined_p(arguments->module, arguments->name)) {
+ return Qfalse;
+ }
- if (state->ele->state == state) {
- need_wakeups = 1;
- state->ele->state = 0;
- state->ele->fork_gen = 0;
+ VALUE autoload_const_value = check_autoload_required(arguments->module, arguments->name, &loading);
+ if (!autoload_const_value) {
+ return Qfalse;
}
- /* At the last, move a value defined in autoload to constant table */
- if (RTEST(state->result) && state->ele->value != Qundef) {
- int safe_backup;
- struct autoload_const_set_args args;
+ src = rb_sourcefile();
+ if (src && loading && strcmp(src, loading) == 0) {
+ return Qfalse;
+ }
- args.mod = state->mod;
- args.id = state->id;
- args.value = state->ele->value;
- args.flag = state->ele->flag;
- safe_backup = rb_safe_level();
- rb_set_safe_level_force(state->ele->safe_level);
- rb_ensure(autoload_const_set, (VALUE)&args,
- reset_safe, (VALUE)safe_backup);
+ struct autoload_const *autoload_const;
+ struct autoload_data *autoload_data;
+ if (!(autoload_data = get_autoload_data(autoload_const_value, &autoload_const))) {
+ return Qfalse;
}
- /* wakeup any waiters we had */
- if (need_wakeups) {
- struct autoload_state *cur = 0, *nxt;
+ if (NIL_P(autoload_data->mutex)) {
+ RB_OBJ_WRITE(autoload_const->autoload_data_value, &autoload_data->mutex, rb_mutex_new());
+ autoload_data->fork_gen = GET_VM()->fork_gen;
+ }
+ else if (rb_mutex_owned_p(autoload_data->mutex)) {
+ return Qfalse;
+ }
- list_for_each_safe(&state->waitq.head, cur, nxt, waitq.node) {
- VALUE th = cur->thread;
+ arguments->mutex = autoload_data->mutex;
+ arguments->autoload_const = autoload_const;
- cur->thread = Qfalse;
- list_del_init(&cur->waitq.node); /* idempotent */
+ return autoload_const_value;
+}
- /*
- * cur is stored on the stack of cur->waiting_th,
- * do not touch cur after waking up waiting_th
- */
- rb_thread_wakeup_alive(th);
- }
+static VALUE
+autoload_apply_constants(VALUE _arguments)
+{
+ RUBY_ASSERT_CRITICAL_SECTION_ENTER();
+
+ struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
+
+ struct autoload_const *autoload_const = 0; // for ccan_container_off_var()
+ struct autoload_const *next;
+
+ // We use safe iteration here because `autoload_const_set` will eventually invoke
+ // `autoload_delete` which will remove the constant from the linked list. In theory, once
+ // the `autoload_data->constants` linked list is empty, we can remove it.
+
+ // Iterate over all constants and assign them:
+ ccan_list_for_each_safe(&arguments->autoload_data->constants, autoload_const, next, cnode) {
+ if (!UNDEF_P(autoload_const->value)) {
+ autoload_const_set(autoload_const);
+ }
}
- return 0; /* ignored */
+ RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
+
+ return Qtrue;
}
static VALUE
-autoload_sleep(VALUE arg)
+autoload_feature_require(VALUE _arguments)
{
- struct autoload_state *state = (struct autoload_state *)arg;
+ VALUE receiver = rb_vm_top_self();
+
+ struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
+
+ struct autoload_const *autoload_const = arguments->autoload_const;
+ VALUE autoload_box_value = autoload_const->box_value;
+
+ // We save this for later use in autoload_apply_constants:
+ arguments->autoload_data = rb_check_typeddata(autoload_const->autoload_data_value, &autoload_data_type);
+
+ if (rb_box_available() && BOX_OBJ_P(autoload_box_value))
+ receiver = autoload_box_value;
/*
- * autoload_reset in other thread will resume us and remove us
- * from the waitq list
+ * Clear the global cc cache table because the require method can be different from the current
+ * box's one and it may cause inconsistent cc-cme states.
+ * For example, the assertion below may fail in gccct_method_search();
+ * VM_ASSERT(vm_cc_check_cme(cc, rb_callable_method_entry(klass, mid)))
*/
- do {
- rb_thread_sleep_deadly();
- } while (state->thread != Qfalse);
+ rb_gccct_clear_table(Qnil);
- return Qfalse;
+ VALUE result = rb_funcall(receiver, rb_intern("require"), 1, arguments->autoload_data->feature);
+
+ if (RTEST(result)) {
+ return rb_mutex_synchronize(autoload_mutex, autoload_apply_constants, _arguments);
+ }
+ return result;
}
static VALUE
-autoload_sleep_done(VALUE arg)
+autoload_try_load(VALUE _arguments)
{
- struct autoload_state *state = (struct autoload_state *)arg;
+ struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
+
+ VALUE result = autoload_feature_require(_arguments);
+
+ // After we loaded the feature, if the constant is not defined, we remove it completely:
+ rb_const_entry_t *ce = rb_const_lookup(arguments->module, arguments->name);
+
+ if (!ce || UNDEF_P(ce->value)) {
+ result = Qfalse;
+
+ rb_const_remove(arguments->module, arguments->name);
- if (state->thread != Qfalse && rb_thread_to_be_killed(state->thread)) {
- list_del(&state->waitq.node); /* idempotent after list_del_init */
+ if (arguments->module == rb_cObject) {
+ rb_warning(
+ "Expected %"PRIsVALUE" to define %"PRIsVALUE" but it didn't",
+ arguments->autoload_data->feature,
+ ID2SYM(arguments->name)
+ );
+ }
+ else {
+ rb_warning(
+ "Expected %"PRIsVALUE" to define %"PRIsVALUE"::%"PRIsVALUE" but it didn't",
+ arguments->autoload_data->feature,
+ arguments->module,
+ ID2SYM(arguments->name)
+ );
+ }
+ }
+ else {
+ // Otherwise, it was loaded, copy the flags from the autoload constant:
+ ce->flag |= arguments->flag;
}
- return Qfalse;
+ return result;
}
VALUE
-rb_autoload_load(VALUE mod, ID id)
+rb_autoload_load(VALUE module, ID name)
{
- VALUE load, result;
- const char *loading = 0, *src;
- struct autoload_data_i *ele;
- struct autoload_state state;
+ rb_const_entry_t *ce = rb_const_lookup(module, name);
- if (!autoload_defined_p(mod, id)) return Qfalse;
- load = check_autoload_required(mod, id, &loading);
- if (!load) return Qfalse;
- src = rb_sourcefile();
- if (src && loading && strcmp(src, loading) == 0) return Qfalse;
+ // We bail out as early as possible without any synchronisation:
+ if (!ce || !UNDEF_P(ce->value)) {
+ return Qfalse;
+ }
- /* set ele->state for a marker of autoloading thread */
- if (!(ele = get_autoload_data(load))) {
- return Qfalse;
+ // At this point, we assume there might be autoloading, so fail if it's ractor:
+ if (UNLIKELY(!rb_ractor_main_p())) {
+ return rb_ractor_autoload_load(module, name);
}
- state.ele = ele;
- state.mod = mod;
- state.id = id;
- state.thread = rb_thread_current();
- if (!ele->state) {
- ele->state = &state;
- ele->fork_gen = GET_VM()->fork_gen;
+ // This state is stored on the stack and is used during the autoload process.
+ struct autoload_load_arguments arguments = {.module = module, .name = name, .mutex = Qnil};
- /*
- * autoload_reset will wake up any threads added to this
- * iff the GVL is released during autoload_require
- */
- list_head_init(&state.waitq.head);
- }
- else if (state.thread == ele->state->thread) {
- return Qfalse;
- }
- else {
- list_add_tail(&ele->state->waitq.head, &state.waitq.node);
+ // Figure out whether we can autoload the named constant:
+ VALUE autoload_const_value = rb_mutex_synchronize(autoload_mutex, autoload_load_needed, (VALUE)&arguments);
- rb_ensure(autoload_sleep, (VALUE)&state,
- autoload_sleep_done, (VALUE)&state);
- }
+ // This confirms whether autoloading is required or not:
+ if (autoload_const_value == Qfalse) return autoload_const_value;
+
+ arguments.flag = ce->flag & (CONST_DEPRECATED | CONST_VISIBILITY_MASK);
+
+ // Only one thread will enter here at a time:
+ VALUE result = rb_mutex_synchronize(arguments.mutex, autoload_try_load, (VALUE)&arguments);
- /* autoload_data_i can be deleted by another thread while require */
- result = rb_ensure(autoload_require, (VALUE)&state,
- autoload_reset, (VALUE)&state);
+ // If you don't guard this value, it's possible for the autoload constant to
+ // be freed by another thread which loads multiple constants, one of which
+ // resolves to the constant this thread is trying to load, so proteect this
+ // so that it is not freed until we are done with it in `autoload_try_load`:
+ RB_GC_GUARD(autoload_const_value);
- RB_GC_GUARD(load);
return result;
}
VALUE
rb_autoload_p(VALUE mod, ID id)
{
+ return rb_autoload_at_p(mod, id, TRUE);
+}
+
+VALUE
+rb_autoload_at_p(VALUE mod, ID id, int recur)
+{
VALUE load;
- struct autoload_data_i *ele;
+ struct autoload_data *ele;
while (!autoload_defined_p(mod, id)) {
- mod = RCLASS_SUPER(mod);
- if (!mod) return Qnil;
+ if (!recur) return Qnil;
+ mod = RCLASS_SUPER(mod);
+ if (!mod) return Qnil;
}
load = check_autoload_required(mod, id, 0);
if (!load) return Qnil;
- return (ele = get_autoload_data(load)) ? ele->feature : Qnil;
+ return (ele = get_autoload_data(load, 0)) ? ele->feature : Qnil;
}
void
rb_const_warn_if_deprecated(const rb_const_entry_t *ce, VALUE klass, ID id)
{
- if (RB_CONST_DEPRECATED_P(ce)) {
- if (klass == rb_cObject) {
- rb_warn("constant ::%"PRIsVALUE" is deprecated", QUOTE_ID(id));
- }
- else {
- rb_warn("constant %"PRIsVALUE"::%"PRIsVALUE" is deprecated",
- rb_class_name(klass), QUOTE_ID(id));
- }
+ if (RB_CONST_DEPRECATED_P(ce) &&
+ rb_warning_category_enabled_p(RB_WARN_CATEGORY_DEPRECATED)) {
+ if (klass == rb_cObject) {
+ rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant ::%"PRIsVALUE" is deprecated", QUOTE_ID(id));
+ }
+ else {
+ rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant %"PRIsVALUE"::%"PRIsVALUE" is deprecated",
+ rb_class_name(klass), QUOTE_ID(id));
+ }
}
}
static VALUE
rb_const_get_0(VALUE klass, ID id, int exclude, int recurse, int visibility)
{
- VALUE c = rb_const_search(klass, id, exclude, recurse, visibility);
- if (c != Qundef) return c;
+ VALUE found_in;
+ VALUE c = rb_const_search(klass, id, exclude, recurse, visibility, &found_in);
+ if (!UNDEF_P(c)) {
+ if (UNLIKELY(!rb_ractor_main_p())) {
+ if (!rb_ractor_shareable_p(c)) {
+ rb_raise(rb_eRactorIsolationError, "can not access non-shareable objects in constant %"PRIsVALUE"::%"PRIsVALUE" by non-main Ractor.", rb_class_path(found_in), rb_id2str(id));
+ }
+ }
+ return c;
+ }
return rb_const_missing(klass, ID2SYM(id));
}
static VALUE
-rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility)
-{
- VALUE value, tmp, av;
- rb_const_flag_t flag;
- int mod_retry = 0;
+rb_const_search_from(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in)
+{
+ VALUE value, current;
+ bool first_iteration = true;
+
+ for (current = klass;
+ RTEST(current);
+ current = RCLASS_SUPER(current), first_iteration = false) {
+ VALUE tmp;
+ VALUE am = 0;
+ rb_const_entry_t *ce;
+
+ if (!first_iteration && RCLASS_ORIGIN(current) != current) {
+ // This item in the super chain has an origin iclass
+ // that comes later in the chain. Skip this item so
+ // prepended modules take precedence.
+ continue;
+ }
- tmp = klass;
- retry:
- while (RTEST(tmp)) {
- VALUE am = 0;
- rb_const_entry_t *ce;
-
- while ((ce = rb_const_lookup(tmp, id))) {
- if (visibility && RB_CONST_PRIVATE_P(ce)) {
- if (BUILTIN_TYPE(tmp) == T_ICLASS) tmp = RBASIC(tmp)->klass;
- rb_name_err_raise("private constant %2$s::%1$s referenced",
- tmp, ID2SYM(id));
- }
- rb_const_warn_if_deprecated(ce, tmp, id);
- value = ce->value;
- if (value == Qundef) {
- if (am == tmp) break;
- am = tmp;
- if (rb_autoloading_value(tmp, id, &av, &flag)) return av;
- rb_autoload_load(tmp, id);
- continue;
- }
- if (exclude && tmp == rb_cObject && klass != rb_cObject) {
-#if 0
- rb_warn("toplevel constant %"PRIsVALUE" referenced by %"PRIsVALUE"::%"PRIsVALUE"",
- QUOTE_ID(id), rb_class_name(klass), QUOTE_ID(id));
-#else
- return Qundef;
-#endif
- }
- return value;
- }
- if (!recurse) break;
- tmp = RCLASS_SUPER(tmp);
- }
- if (!exclude && !mod_retry && BUILTIN_TYPE(klass) == T_MODULE) {
- mod_retry = 1;
- tmp = rb_cObject;
- goto retry;
+ // Do lookup in original class or module in case we are at an origin
+ // iclass in the chain.
+ tmp = current;
+ if (BUILTIN_TYPE(tmp) == T_ICLASS) tmp = RBASIC(tmp)->klass;
+
+ // Do the lookup. Loop in case of autoload.
+ while ((ce = rb_const_lookup(tmp, id))) {
+ if (visibility && RB_CONST_PRIVATE_P(ce)) {
+ GET_EC()->private_const_reference = tmp;
+ return Qundef;
+ }
+ rb_const_warn_if_deprecated(ce, tmp, id);
+ value = ce->value;
+ if (UNDEF_P(value)) {
+ struct autoload_const *ac;
+ if (am == tmp) break;
+ am = tmp;
+ ac = autoloading_const_entry(tmp, id);
+ if (ac) {
+ if (found_in) { *found_in = tmp; }
+ return ac->value;
+ }
+ rb_autoload_load(tmp, id);
+ continue;
+ }
+ if (exclude && tmp == rb_cObject) {
+ goto not_found;
+ }
+ if (found_in) { *found_in = tmp; }
+ return value;
+ }
+ if (!recurse) break;
}
+ not_found:
+ GET_EC()->private_const_reference = 0;
return Qundef;
}
+static VALUE
+rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility, VALUE *found_in)
+{
+ VALUE value;
+
+ if (klass == rb_cObject) exclude = FALSE;
+ value = rb_const_search_from(klass, id, exclude, recurse, visibility, found_in);
+ if (!UNDEF_P(value)) return value;
+ if (exclude) return value;
+ if (BUILTIN_TYPE(klass) != T_MODULE) return value;
+ /* search global const too, if klass is a module */
+ return rb_const_search_from(rb_cObject, id, FALSE, recurse, visibility, found_in);
+}
+
VALUE
rb_const_get_from(VALUE klass, ID id)
{
@@ -2331,15 +3463,80 @@ rb_public_const_get_from(VALUE klass, ID id)
}
VALUE
-rb_public_const_get(VALUE klass, ID id)
+rb_public_const_get_at(VALUE klass, ID id)
+{
+ return rb_const_get_0(klass, id, TRUE, FALSE, TRUE);
+}
+
+NORETURN(static void undefined_constant(VALUE mod, VALUE name));
+static void
+undefined_constant(VALUE mod, VALUE name)
{
- return rb_const_get_0(klass, id, FALSE, TRUE, TRUE);
+ rb_name_err_raise("constant %2$s::%1$s not defined",
+ mod, name);
+}
+
+static VALUE
+rb_const_location_from(VALUE klass, ID id, int exclude, int recurse, int visibility)
+{
+ while (RTEST(klass)) {
+ rb_const_entry_t *ce;
+
+ while ((ce = rb_const_lookup(klass, id))) {
+ if (visibility && RB_CONST_PRIVATE_P(ce)) {
+ return Qnil;
+ }
+ if (exclude && klass == rb_cObject) {
+ goto not_found;
+ }
+
+ if (UNDEF_P(ce->value)) { // autoload
+ VALUE autoload_const_value = autoload_data(klass, id);
+ if (RTEST(autoload_const_value)) {
+ struct autoload_const *autoload_const;
+ struct autoload_data *autoload_data = get_autoload_data(autoload_const_value, &autoload_const);
+
+ if (!UNDEF_P(autoload_const->value) && RTEST(rb_mutex_owned_p(autoload_data->mutex))) {
+ return rb_assoc_new(autoload_const->file, INT2NUM(autoload_const->line));
+ }
+ }
+ }
+
+ if (NIL_P(ce->file)) return rb_ary_new();
+ return rb_assoc_new(ce->file, INT2NUM(ce->line));
+ }
+ if (!recurse) break;
+ klass = RCLASS_SUPER(klass);
+ }
+
+ not_found:
+ return Qnil;
+}
+
+static VALUE
+rb_const_location(VALUE klass, ID id, int exclude, int recurse, int visibility)
+{
+ VALUE loc;
+
+ if (klass == rb_cObject) exclude = FALSE;
+ loc = rb_const_location_from(klass, id, exclude, recurse, visibility);
+ if (!NIL_P(loc)) return loc;
+ if (exclude) return loc;
+ if (BUILTIN_TYPE(klass) != T_MODULE) return loc;
+ /* search global const too, if klass is a module */
+ return rb_const_location_from(rb_cObject, id, FALSE, recurse, visibility);
}
VALUE
-rb_public_const_get_at(VALUE klass, ID id)
+rb_const_source_location(VALUE klass, ID id)
{
- return rb_const_get_0(klass, id, TRUE, FALSE, TRUE);
+ return rb_const_location(klass, id, FALSE, TRUE, FALSE);
+}
+
+VALUE
+rb_const_source_location_at(VALUE klass, ID id)
+{
+ return rb_const_location(klass, id, TRUE, FALSE, FALSE);
}
/*
@@ -2358,12 +3555,13 @@ rb_mod_remove_const(VALUE mod, VALUE name)
const ID id = id_for_var(mod, name, a, constant);
if (!id) {
- rb_name_err_raise("constant %2$s::%1$s not defined",
- mod, name);
+ undefined_constant(mod, name);
}
return rb_const_remove(mod, id);
}
+static rb_const_entry_t * const_lookup(struct rb_id_table *tbl, ID id);
+
VALUE
rb_const_remove(VALUE mod, ID id)
{
@@ -2371,24 +3569,40 @@ rb_const_remove(VALUE mod, ID id)
rb_const_entry_t *ce;
rb_check_frozen(mod);
+
ce = rb_const_lookup(mod, id);
- if (!ce || !rb_id_table_delete(RCLASS_CONST_TBL(mod), id)) {
- if (rb_const_defined_at(mod, id)) {
- rb_name_err_raise("cannot remove %2$s::%1$s",
- mod, ID2SYM(id));
- }
- rb_name_err_raise("constant %2$s::%1$s not defined",
- mod, ID2SYM(id));
+
+ if (!ce) {
+ if (rb_const_defined_at(mod, id)) {
+ rb_name_err_raise("cannot remove %2$s::%1$s", mod, ID2SYM(id));
+ }
+
+ undefined_constant(mod, ID2SYM(id));
}
- rb_clear_constant_cache();
+ VALUE writable_ce = 0;
+ if (rb_id_table_lookup(RCLASS_WRITABLE_CONST_TBL(mod), id, &writable_ce)) {
+ rb_id_table_delete(RCLASS_WRITABLE_CONST_TBL(mod), id);
+ if ((rb_const_entry_t *)writable_ce != ce) {
+ xfree((rb_const_entry_t *)writable_ce);
+ }
+ }
+
+ rb_const_warn_if_deprecated(ce, mod, id);
+ rb_clear_constant_cache_for_id(id);
val = ce->value;
- if (val == Qundef) {
- autoload_delete(mod, id);
- val = Qnil;
+
+ if (UNDEF_P(val)) {
+ autoload_delete(mod, id);
+ val = Qnil;
+ }
+
+ if (ce != const_lookup(RCLASS_PRIME_CONST_TBL(mod), id)) {
+ ruby_xfree(ce);
}
- xfree(ce);
+ // else - skip free'ing the ce because it still exists in the prime classext
+
return val;
}
@@ -2407,7 +3621,7 @@ sv_i(ID key, VALUE v, void *a)
st_table *tbl = a;
if (rb_is_const_id(key)) {
- st_update(tbl, (st_data_t)key, cv_i_update, (st_data_t)ce);
+ st_update(tbl, (st_data_t)key, cv_i_update, (st_data_t)ce);
}
return ID_TABLE_CONTINUE;
}
@@ -2416,7 +3630,7 @@ static enum rb_id_table_iterator_result
rb_local_constants_i(ID const_name, VALUE const_value, void *ary)
{
if (rb_is_const_id(const_name) && !RB_CONST_PRIVATE_P((rb_const_entry_t *)const_value)) {
- rb_ary_push((VALUE)ary, ID2SYM(const_name));
+ rb_ary_push((VALUE)ary, ID2SYM(const_name));
}
return ID_TABLE_CONTINUE;
}
@@ -2429,8 +3643,11 @@ rb_local_constants(VALUE mod)
if (!tbl) return rb_ary_new2(0);
- ary = rb_ary_new2(rb_id_table_size(tbl));
- rb_id_table_foreach(tbl, rb_local_constants_i, (void *)ary);
+ RB_VM_LOCKING() {
+ ary = rb_ary_new2(rb_id_table_size(tbl));
+ rb_id_table_foreach(tbl, rb_local_constants_i, (void *)ary);
+ }
+
return ary;
}
@@ -2439,10 +3656,12 @@ rb_mod_const_at(VALUE mod, void *data)
{
st_table *tbl = data;
if (!tbl) {
- tbl = st_init_numtable();
+ tbl = st_init_numtable();
}
if (RCLASS_CONST_TBL(mod)) {
- rb_id_table_foreach(RCLASS_CONST_TBL(mod), sv_i, tbl);
+ RB_VM_LOCKING() {
+ rb_id_table_foreach(RCLASS_CONST_TBL(mod), sv_i, tbl);
+ }
}
return tbl;
}
@@ -2452,10 +3671,10 @@ rb_mod_const_of(VALUE mod, void *data)
{
VALUE tmp = mod;
for (;;) {
- data = rb_mod_const_at(tmp, data);
- tmp = RCLASS_SUPER(tmp);
- if (!tmp) break;
- if (tmp == rb_cObject && mod != rb_cObject) break;
+ data = rb_mod_const_at(tmp, data);
+ tmp = RCLASS_SUPER(tmp);
+ if (!tmp) break;
+ if (tmp == rb_cObject && mod != rb_cObject) break;
}
return data;
}
@@ -2498,26 +3717,21 @@ rb_const_list(void *data)
* IO.constants.include?(:SYNC) #=> true
* IO.constants(false).include?(:SYNC) #=> false
*
- * Also see <code>Module::const_defined?</code>.
+ * Also see Module#const_defined?.
*/
VALUE
rb_mod_constants(int argc, const VALUE *argv, VALUE mod)
{
- VALUE inherit;
+ bool inherit = true;
- if (argc == 0) {
- inherit = Qtrue;
- }
- else {
- rb_scan_args(argc, argv, "01", &inherit);
- }
+ if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]);
- if (RTEST(inherit)) {
- return rb_const_list(rb_mod_const_of(mod, 0));
+ if (inherit) {
+ return rb_const_list(rb_mod_const_of(mod, 0));
}
else {
- return rb_local_constants(mod);
+ return rb_local_constants(mod);
}
}
@@ -2531,27 +3745,27 @@ rb_const_defined_0(VALUE klass, ID id, int exclude, int recurse, int visibility)
tmp = klass;
retry:
while (tmp) {
- if ((ce = rb_const_lookup(tmp, id))) {
- if (visibility && RB_CONST_PRIVATE_P(ce)) {
- return (int)Qfalse;
- }
- if (ce->value == Qundef && !check_autoload_required(tmp, id, 0) &&
- !rb_autoloading_value(tmp, id, NULL, NULL))
- return (int)Qfalse;
-
- if (exclude && tmp == rb_cObject && klass != rb_cObject) {
- return (int)Qfalse;
- }
-
- return (int)Qtrue;
- }
- if (!recurse) break;
- tmp = RCLASS_SUPER(tmp);
+ if ((ce = rb_const_lookup(tmp, id))) {
+ if (visibility && RB_CONST_PRIVATE_P(ce)) {
+ return (int)Qfalse;
+ }
+ if (UNDEF_P(ce->value) && !check_autoload_required(tmp, id, 0) &&
+ !rb_autoloading_value(tmp, id, NULL, NULL))
+ return (int)Qfalse;
+
+ if (exclude && tmp == rb_cObject && klass != rb_cObject) {
+ return (int)Qfalse;
+ }
+
+ return (int)Qtrue;
+ }
+ if (!recurse) break;
+ tmp = RCLASS_SUPER(tmp);
}
if (!exclude && !mod_retry && BUILTIN_TYPE(klass) == T_MODULE) {
- mod_retry = 1;
- tmp = rb_cObject;
- goto retry;
+ mod_retry = 1;
+ tmp = rb_cObject;
+ goto retry;
}
return (int)Qfalse;
}
@@ -2580,148 +3794,223 @@ rb_public_const_defined_from(VALUE klass, ID id)
return rb_const_defined_0(klass, id, TRUE, TRUE, TRUE);
}
-int
-rb_public_const_defined(VALUE klass, ID id)
+static void
+check_before_mod_set(VALUE klass, ID id, VALUE val, const char *dest)
{
- return rb_const_defined_0(klass, id, FALSE, TRUE, TRUE);
+ rb_check_frozen(klass);
}
-int
-rb_public_const_defined_at(VALUE klass, ID id)
+static void set_namespace_path(VALUE named_namespace, VALUE name);
+
+static enum rb_id_table_iterator_result
+set_namespace_path_i(ID id, VALUE v, void *payload)
{
- return rb_const_defined_0(klass, id, TRUE, FALSE, TRUE);
+ rb_const_entry_t *ce = (rb_const_entry_t *)v;
+ VALUE value = ce->value;
+ VALUE parental_path = *((VALUE *) payload);
+ if (!rb_is_const_id(id) || !rb_namespace_p(value)) {
+ return ID_TABLE_CONTINUE;
+ }
+
+ bool has_permanent_classpath;
+ classname(value, &has_permanent_classpath);
+ if (has_permanent_classpath) {
+ return ID_TABLE_CONTINUE;
+ }
+ set_namespace_path(value, build_const_path(parental_path, id));
+
+ if (!RCLASS_PERMANENT_CLASSPATH_P(value)) {
+ RCLASS_WRITE_CLASSPATH(value, 0, false);
+ }
+
+ return ID_TABLE_CONTINUE;
}
+/*
+ * Assign permanent classpaths to all namespaces that are directly or indirectly
+ * nested under +named_namespace+. +named_namespace+ must have a permanent
+ * classpath.
+ */
static void
-check_before_mod_set(VALUE klass, ID id, VALUE val, const char *dest)
+set_namespace_path(VALUE named_namespace, VALUE namespace_path)
{
- rb_check_frozen(klass);
+ struct rb_id_table *const_table = RCLASS_CONST_TBL(named_namespace);
+ RB_OBJ_SET_SHAREABLE(namespace_path);
+
+ RB_VM_LOCKING() {
+ RCLASS_WRITE_CLASSPATH(named_namespace, namespace_path, true);
+
+ if (const_table) {
+ rb_id_table_foreach(const_table, set_namespace_path_i, &namespace_path);
+ }
+ }
}
-void
-rb_const_set(VALUE klass, ID id, VALUE val)
+static void
+const_added(VALUE klass, ID const_name)
+{
+ if (GET_VM()->running) {
+ VALUE name = ID2SYM(const_name);
+ rb_funcallv(klass, idConst_added, 1, &name);
+ }
+}
+
+static void
+const_set(VALUE klass, ID id, VALUE val)
{
rb_const_entry_t *ce;
- struct rb_id_table *tbl = RCLASS_CONST_TBL(klass);
if (NIL_P(klass)) {
- rb_raise(rb_eTypeError, "no class/module to define constant %"PRIsVALUE"",
- QUOTE_ID(id));
+ rb_raise(rb_eTypeError, "no class/module to define constant %"PRIsVALUE"",
+ QUOTE_ID(id));
}
- check_before_mod_set(klass, id, val, "constant");
- if (!tbl) {
- RCLASS_CONST_TBL(klass) = tbl = rb_id_table_create(0);
- rb_clear_constant_cache();
- ce = ZALLOC(rb_const_entry_t);
- rb_id_table_insert(tbl, id, (VALUE)ce);
- setup_const_entry(ce, klass, val, CONST_PUBLIC);
+ if (!rb_ractor_main_p() && !rb_ractor_shareable_p(val)) {
+ rb_raise(rb_eRactorIsolationError, "can not set constants with non-shareable objects by non-main Ractors");
}
- else {
- struct autoload_const_set_args args;
- args.mod = klass;
- args.id = id;
- args.value = val;
- args.flag = CONST_PUBLIC;
- const_tbl_update(&args);
+
+ check_before_mod_set(klass, id, val, "constant");
+
+ RB_VM_LOCKING() {
+ struct rb_id_table *tbl = RCLASS_WRITABLE_CONST_TBL(klass);
+ if (!tbl) {
+ tbl = rb_id_table_create(0);
+ RCLASS_WRITE_CONST_TBL(klass, tbl, false);
+ rb_clear_constant_cache_for_id(id);
+ ce = ZALLOC(rb_const_entry_t);
+ rb_id_table_insert(tbl, id, (VALUE)ce);
+ setup_const_entry(ce, klass, val, CONST_PUBLIC);
+ }
+ else {
+ struct autoload_const ac = {
+ .module = klass, .name = id,
+ .value = val, .flag = CONST_PUBLIC,
+ /* fill the rest with 0 */
+ };
+ ac.file = rb_source_location(&ac.line);
+ const_tbl_update(&ac, false);
+ }
}
+
/*
* Resolve and cache class name immediately to resolve ambiguity
* and avoid order-dependency on const_tbl
*/
- if (rb_cObject && (RB_TYPE_P(val, T_MODULE) || RB_TYPE_P(val, T_CLASS))) {
- if (NIL_P(rb_class_path_cached(val))) {
- if (klass == rb_cObject) {
- rb_ivar_set(val, classpath, rb_id2str(id));
- rb_name_class(val, id);
- }
- else {
- VALUE path;
- ID pathid;
- st_data_t n;
- st_table *ivtbl = RCLASS_IV_TBL(klass);
- if (ivtbl &&
- (st_lookup(ivtbl, (st_data_t)(pathid = classpath), &n) ||
- st_lookup(ivtbl, (st_data_t)(pathid = tmp_classpath), &n))) {
- path = rb_str_dup((VALUE)n);
- rb_str_append(rb_str_cat2(path, "::"), rb_id2str(id));
- OBJ_FREEZE(path);
- rb_ivar_set(val, pathid, path);
- rb_name_class(val, id);
- }
- }
- }
- }
-}
-
-static struct autoload_data_i *
-current_autoload_data(VALUE mod, ID id)
-{
- struct autoload_data_i *ele;
- VALUE load = autoload_data(mod, id);
- if (!load) return 0;
- ele = get_autoload_data(load);
- if (!ele) return 0;
+ if (rb_cObject && rb_namespace_p(val)) {
+ bool val_path_permanent;
+ VALUE val_path = classname(val, &val_path_permanent);
+ if (NIL_P(val_path) || !val_path_permanent) {
+ if (klass == rb_cObject) {
+ set_namespace_path(val, rb_id2str(id));
+ }
+ else {
+ bool parental_path_permanent;
+ VALUE parental_path = classname(klass, &parental_path_permanent);
+ if (NIL_P(parental_path)) {
+ bool throwaway;
+ parental_path = rb_tmp_class_path(klass, &throwaway, make_temporary_path);
+ }
+ if (parental_path_permanent && !val_path_permanent) {
+ set_namespace_path(val, build_const_path(parental_path, id));
+ }
+ else if (!parental_path_permanent && NIL_P(val_path)) {
+ VALUE path = build_const_path(parental_path, id);
+ RCLASS_SET_CLASSPATH(val, path, false);
+ }
+ }
+ }
+ }
+}
+
+void
+rb_const_set(VALUE klass, ID id, VALUE val)
+{
+ const_set(klass, id, val);
+ const_added(klass, id);
+}
+
+static VALUE
+autoload_const_value_for_named_constant(VALUE module, ID name, struct autoload_const **autoload_const_pointer)
+{
+ VALUE autoload_const_value = autoload_data(module, name);
+ if (!autoload_const_value) return Qfalse;
+
+ struct autoload_data *autoload_data = get_autoload_data(autoload_const_value, autoload_const_pointer);
+ if (!autoload_data) return Qfalse;
+
/* for autoloading thread, keep the defined value to autoloading storage */
- if (ele->state && (ele->state->thread == rb_thread_current())) {
- return ele;
+ if (autoload_by_current(autoload_data)) {
+ return autoload_const_value;
}
- return 0;
+
+ return Qfalse;
}
static void
-const_tbl_update(struct autoload_const_set_args *args)
+const_tbl_update(struct autoload_const *ac, int autoload_force)
{
VALUE value;
- VALUE klass = args->mod;
- VALUE val = args->value;
- ID id = args->id;
+ VALUE klass = ac->module;
+ VALUE val = ac->value;
+ ID id = ac->name;
struct rb_id_table *tbl = RCLASS_CONST_TBL(klass);
- rb_const_flag_t visibility = args->flag;
+ rb_const_flag_t visibility = ac->flag;
rb_const_entry_t *ce;
if (rb_id_table_lookup(tbl, id, &value)) {
- ce = (rb_const_entry_t *)value;
- if (ce->value == Qundef) {
- struct autoload_data_i *ele = current_autoload_data(klass, id);
-
- if (ele) {
- rb_clear_constant_cache();
-
- ele->value = val; /* autoload_i is non-WB-protected */
- return;
- }
- /* otherwise, allow to override */
- autoload_delete(klass, id);
- }
- else {
- VALUE name = QUOTE_ID(id);
- visibility = ce->flag;
- if (klass == rb_cObject)
- rb_warn("already initialized constant %"PRIsVALUE"", name);
- else
- rb_warn("already initialized constant %"PRIsVALUE"::%"PRIsVALUE"",
- rb_class_name(klass), name);
- if (!NIL_P(ce->file) && ce->line) {
- rb_compile_warn(RSTRING_PTR(ce->file), ce->line,
- "previous definition of %"PRIsVALUE" was here", name);
- }
- }
- rb_clear_constant_cache();
- setup_const_entry(ce, klass, val, visibility);
+ ce = (rb_const_entry_t *)value;
+ if (UNDEF_P(ce->value)) {
+ RUBY_ASSERT_CRITICAL_SECTION_ENTER();
+ VALUE file = ac->file;
+ int line = ac->line;
+ VALUE autoload_const_value = autoload_const_value_for_named_constant(klass, id, &ac);
+
+ if (!autoload_force && autoload_const_value) {
+ rb_clear_constant_cache_for_id(id);
+
+ RB_OBJ_WRITE(autoload_const_value, &ac->value, val);
+ RB_OBJ_WRITE(autoload_const_value, &ac->file, rb_source_location(&ac->line));
+ }
+ else {
+ /* otherwise autoloaded constant, allow to override */
+ autoload_delete(klass, id);
+ ce->flag = visibility;
+ RB_OBJ_WRITE(klass, &ce->value, val);
+ RB_OBJ_WRITE(klass, &ce->file, file);
+ ce->line = line;
+ }
+ RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
+ return;
+ }
+ else {
+ VALUE name = QUOTE_ID(id);
+ visibility = ce->flag;
+ if (klass == rb_cObject)
+ rb_warn("already initialized constant %"PRIsVALUE"", name);
+ else
+ rb_warn("already initialized constant %"PRIsVALUE"::%"PRIsVALUE"",
+ rb_class_name(klass), name);
+ if (!NIL_P(ce->file) && ce->line) {
+ rb_compile_warn(RSTRING_PTR(ce->file), ce->line,
+ "previous definition of %"PRIsVALUE" was here", name);
+ }
+ }
+ rb_clear_constant_cache_for_id(id);
+ setup_const_entry(ce, klass, val, visibility);
}
else {
- rb_clear_constant_cache();
+ tbl = RCLASS_WRITABLE_CONST_TBL(klass);
+ rb_clear_constant_cache_for_id(id);
- ce = ZALLOC(rb_const_entry_t);
- rb_id_table_insert(tbl, id, (VALUE)ce);
- setup_const_entry(ce, klass, val, visibility);
+ ce = ZALLOC(rb_const_entry_t);
+ rb_id_table_insert(tbl, id, (VALUE)ce);
+ setup_const_entry(ce, klass, val, visibility);
}
}
static void
setup_const_entry(rb_const_entry_t *ce, VALUE klass, VALUE val,
- rb_const_flag_t visibility)
+ rb_const_flag_t visibility)
{
ce->flag = visibility;
RB_OBJ_WRITE(klass, &ce->value, val);
@@ -2734,7 +4023,10 @@ rb_define_const(VALUE klass, const char *name, VALUE val)
ID id = rb_intern(name);
if (!rb_is_const_id(id)) {
- rb_warn("rb_define_const: invalid name `%s' for constant", name);
+ rb_warn("rb_define_const: invalid name '%s' for constant", name);
+ }
+ if (!RB_SPECIAL_CONST_P(val)) {
+ rb_vm_register_global_object(val);
}
rb_const_set(klass, id, val);
}
@@ -2747,50 +4039,41 @@ rb_define_global_const(const char *name, VALUE val)
static void
set_const_visibility(VALUE mod, int argc, const VALUE *argv,
- rb_const_flag_t flag, rb_const_flag_t mask)
+ rb_const_flag_t flag, rb_const_flag_t mask)
{
int i;
rb_const_entry_t *ce;
ID id;
- rb_frozen_class_p(mod);
+ rb_class_modify_check(mod);
if (argc == 0) {
- rb_warning("%"PRIsVALUE" with no argument is just ignored",
- QUOTE_ID(rb_frame_callee()));
- return;
+ rb_warning("%"PRIsVALUE" with no argument is just ignored",
+ QUOTE_ID(rb_frame_callee()));
+ return;
}
for (i = 0; i < argc; i++) {
- VALUE val = argv[i];
- id = rb_check_id(&val);
- if (!id) {
- if (i > 0) {
- rb_clear_constant_cache();
- }
-
- rb_name_err_raise("constant %2$s::%1$s not defined",
- mod, val);
- }
- if ((ce = rb_const_lookup(mod, id))) {
- ce->flag &= ~mask;
- ce->flag |= flag;
- if (ce->value == Qundef) {
- struct autoload_data_i *ele = current_autoload_data(mod, id);
- if (ele) {
- ele->flag &= ~mask;
- ele->flag |= flag;
- }
- }
- }
- else {
- if (i > 0) {
- rb_clear_constant_cache();
- }
- rb_name_err_raise("constant %2$s::%1$s not defined",
- mod, ID2SYM(id));
- }
- }
- rb_clear_constant_cache();
+ struct autoload_const *ac;
+ VALUE val = argv[i];
+ id = rb_check_id(&val);
+ if (!id) {
+ undefined_constant(mod, val);
+ }
+ if ((ce = rb_const_lookup(mod, id))) {
+ ce->flag &= ~mask;
+ ce->flag |= flag;
+ if (UNDEF_P(ce->value)) {
+ if (autoload_const_value_for_named_constant(mod, id, &ac)) {
+ ac->flag &= ~mask;
+ ac->flag |= flag;
+ }
+ }
+ rb_clear_constant_cache_for_id(id);
+ }
+ else {
+ undefined_constant(mod, ID2SYM(id));
+ }
+ }
}
void
@@ -2800,11 +4083,12 @@ rb_deprecate_constant(VALUE mod, const char *name)
ID id;
long len = strlen(name);
- rb_frozen_class_p(mod);
- if (!(id = rb_check_id_cstr(name, len, NULL)) ||
- !(ce = rb_const_lookup(mod, id))) {
- rb_name_err_raise("constant %2$s::%1$s not defined",
- mod, rb_fstring_new(name, len));
+ rb_class_modify_check(mod);
+ if (!(id = rb_check_id_cstr(name, len, NULL))) {
+ undefined_constant(mod, rb_fstring_new(name, len));
+ }
+ if (!(ce = rb_const_lookup(mod, id))) {
+ undefined_constant(mod, ID2SYM(id));
}
ce->flag |= CONST_DEPRECATED;
}
@@ -2841,7 +4125,19 @@ rb_mod_public_constant(int argc, const VALUE *argv, VALUE obj)
* call-seq:
* mod.deprecate_constant(symbol, ...) => mod
*
- * Makes a list of existing constants deprecated.
+ * Makes a list of existing constants deprecated. Attempt
+ * to refer to them will produce a warning.
+ *
+ * module HTTP
+ * NotFound = Exception.new
+ * NOT_FOUND = NotFound # previous version of the library used this name
+ *
+ * deprecate_constant :NOT_FOUND
+ * end
+ *
+ * HTTP::NOT_FOUND
+ * # warning: constant HTTP::NOT_FOUND is deprecated
+ *
*/
VALUE
@@ -2855,41 +4151,98 @@ static VALUE
original_module(VALUE c)
{
if (RB_TYPE_P(c, T_ICLASS))
- return RBASIC(c)->klass;
+ return RBASIC(c)->klass;
return c;
}
static int
cvar_lookup_at(VALUE klass, ID id, st_data_t *v)
{
- if (!RCLASS_IV_TBL(klass)) return 0;
- return st_lookup(RCLASS_IV_TBL(klass), (st_data_t)id, v);
+ if (RB_TYPE_P(klass, T_ICLASS)) {
+ if (RICLASS_IS_ORIGIN_P(klass)) {
+ return 0;
+ }
+ else {
+ // check the original module
+ klass = RBASIC(klass)->klass;
+ }
+ }
+
+ VALUE n = rb_ivar_lookup(klass, id, Qundef);
+ if (UNDEF_P(n)) return 0;
+
+ if (v) *v = n;
+ return 1;
}
static VALUE
cvar_front_klass(VALUE klass)
{
- if (FL_TEST(klass, FL_SINGLETON)) {
- VALUE obj = rb_ivar_get(klass, id__attached__);
- if (RB_TYPE_P(obj, T_MODULE) || RB_TYPE_P(obj, T_CLASS)) {
- return obj;
- }
+ if (RCLASS_SINGLETON_P(klass)) {
+ VALUE obj = RCLASS_ATTACHED_OBJECT(klass);
+ if (rb_namespace_p(obj)) {
+ return obj;
+ }
}
return RCLASS_SUPER(klass);
}
+static void
+cvar_overtaken(VALUE front, VALUE target, ID id)
+{
+ if (front && target != front) {
+ if (original_module(front) != original_module(target)) {
+ rb_raise(rb_eRuntimeError,
+ "class variable % "PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"",
+ ID2SYM(id), rb_class_name(original_module(front)),
+ rb_class_name(original_module(target)));
+ }
+ if (BUILTIN_TYPE(front) == T_CLASS) {
+ rb_ivar_delete(front, id, Qundef);
+ }
+ }
+}
+
#define CVAR_FOREACH_ANCESTORS(klass, v, r) \
for (klass = cvar_front_klass(klass); klass; klass = RCLASS_SUPER(klass)) { \
- if (cvar_lookup_at(klass, id, (v))) { \
- r; \
- } \
+ if (cvar_lookup_at(klass, id, (v))) { \
+ r; \
+ } \
}
#define CVAR_LOOKUP(v,r) do {\
+ CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(klass, id); \
if (cvar_lookup_at(klass, id, (v))) {r;}\
CVAR_FOREACH_ANCESTORS(klass, v, r);\
} while(0)
+static VALUE
+find_cvar(VALUE klass, VALUE * front, VALUE * target, ID id)
+{
+ VALUE v = Qundef;
+ CVAR_LOOKUP(&v, {
+ if (!*front) {
+ *front = klass;
+ }
+ *target = klass;
+ });
+
+ return v;
+}
+
+static void
+check_for_cvar_table(VALUE subclass, VALUE key)
+{
+ // Must not check ivar on ICLASS
+ if (!RB_TYPE_P(subclass, T_ICLASS) && RTEST(rb_ivar_defined(subclass, key))) {
+ RB_DEBUG_COUNTER_INC(cvar_class_invalidate);
+ ruby_vm_global_cvar_state++;
+ return;
+ }
+
+ rb_class_foreach_subclass(subclass, check_for_cvar_table, key);
+}
+
void
rb_cvar_set(VALUE klass, ID id, VALUE val)
{
@@ -2898,59 +4251,77 @@ rb_cvar_set(VALUE klass, ID id, VALUE val)
tmp = klass;
CVAR_LOOKUP(0, {if (!front) front = klass; target = klass;});
if (target) {
- if (front && target != front) {
- st_data_t did = id;
-
- if (RTEST(ruby_verbose)) {
- rb_warning("class variable %"PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"",
- QUOTE_ID(id), rb_class_name(original_module(front)),
- rb_class_name(original_module(target)));
- }
- if (BUILTIN_TYPE(front) == T_CLASS) {
- st_delete(RCLASS_IV_TBL(front),&did,0);
- }
- }
+ cvar_overtaken(front, target, id);
}
else {
- target = tmp;
+ target = tmp;
}
+ if (RB_TYPE_P(target, T_ICLASS)) {
+ target = RBASIC(target)->klass;
+ }
check_before_mod_set(target, id, val, "class variable");
- if (!RCLASS_IV_TBL(target)) {
- RCLASS_IV_TBL(target) = st_init_numtable();
+
+ bool new_cvar = rb_class_ivar_set(target, id, val);
+
+ struct rb_id_table *rb_cvc_tbl = RCLASS_WRITABLE_CVC_TBL(target);
+
+ if (!rb_cvc_tbl) {
+ rb_cvc_tbl = rb_id_table_create(2);
+ RCLASS_WRITE_CVC_TBL(target, rb_cvc_tbl);
+ }
+
+ struct rb_cvar_class_tbl_entry *ent;
+ VALUE ent_data;
+
+ if (!rb_id_table_lookup(rb_cvc_tbl, id, &ent_data)) {
+ ent = ALLOC(struct rb_cvar_class_tbl_entry);
+ ent->class_value = target;
+ ent->global_cvar_state = GET_GLOBAL_CVAR_STATE();
+ ent->cref = 0;
+ rb_id_table_insert(rb_cvc_tbl, id, (VALUE)ent);
+ RB_DEBUG_COUNTER_INC(cvar_inline_miss);
+ }
+ else {
+ ent = (void *)ent_data;
+ ent->global_cvar_state = GET_GLOBAL_CVAR_STATE();
}
- rb_class_ivar_set(target, id, val);
+ // Break the cvar cache if this is a new class variable
+ // and target is a module or a subclass with the same
+ // cvar in this lookup.
+ if (new_cvar) {
+ if (RB_TYPE_P(target, T_CLASS)) {
+ if (RCLASS_SUBCLASSES_FIRST(target)) {
+ rb_class_foreach_subclass(target, check_for_cvar_table, id);
+ }
+ }
+ }
}
VALUE
-rb_cvar_get(VALUE klass, ID id)
+rb_cvar_find(VALUE klass, ID id, VALUE *front)
{
- VALUE tmp, front = 0, target = 0;
- st_data_t value;
+ VALUE target = 0;
+ VALUE value;
- tmp = klass;
- CVAR_LOOKUP(&value, {if (!front) front = klass; target = klass;});
+ value = find_cvar(klass, front, &target, id);
if (!target) {
- rb_name_err_raise("uninitialized class variable %1$s in %2$s",
- tmp, ID2SYM(id));
- }
- if (front && target != front) {
- st_data_t did = id;
-
- if (RTEST(ruby_verbose)) {
- rb_warning("class variable %"PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"",
- QUOTE_ID(id), rb_class_name(original_module(front)),
- rb_class_name(original_module(target)));
- }
- if (BUILTIN_TYPE(front) == T_CLASS) {
- st_delete(RCLASS_IV_TBL(front),&did,0);
- }
+ rb_name_err_raise("uninitialized class variable %1$s in %2$s",
+ klass, ID2SYM(id));
}
+ cvar_overtaken(*front, target, id);
return (VALUE)value;
}
VALUE
+rb_cvar_get(VALUE klass, ID id)
+{
+ VALUE front = 0;
+ return rb_cvar_find(klass, id, &front);
+}
+
+VALUE
rb_cvar_defined(VALUE klass, ID id)
{
if (!klass) return Qfalse;
@@ -2963,8 +4334,8 @@ cv_intern(VALUE klass, const char *name)
{
ID id = rb_intern(name);
if (!rb_is_class_id(id)) {
- rb_name_err_raise("wrong class variable name %1$s",
- klass, rb_str_new_cstr(name));
+ rb_name_err_raise("wrong class variable name %1$s",
+ klass, rb_str_new_cstr(name));
}
return id;
}
@@ -2986,18 +4357,16 @@ rb_cv_get(VALUE klass, const char *name)
void
rb_define_class_variable(VALUE klass, const char *name, VALUE val)
{
- ID id = cv_intern(klass, name);
- rb_cvar_set(klass, id, val);
+ rb_cv_set(klass, name, val);
}
static int
-cv_i(st_data_t k, st_data_t v, st_data_t a)
+cv_i(ID key, VALUE v, st_data_t a)
{
- ID key = (ID)k;
st_table *tbl = (st_table *)a;
if (rb_is_class_id(key)) {
- st_update(tbl, (st_data_t)key, cv_i_update, 0);
+ st_update(tbl, (st_data_t)key, cv_i_update, 0);
}
return ST_CONTINUE;
}
@@ -3007,11 +4376,11 @@ mod_cvar_at(VALUE mod, void *data)
{
st_table *tbl = data;
if (!tbl) {
- tbl = st_init_numtable();
- }
- if (RCLASS_IV_TBL(mod)) {
- st_foreach_safe(RCLASS_IV_TBL(mod), cv_i, (st_data_t)tbl);
+ tbl = st_init_numtable();
}
+ mod = original_module(mod);
+
+ rb_ivar_foreach(mod, cv_i, (st_data_t)tbl);
return tbl;
}
@@ -3019,10 +4388,16 @@ static void*
mod_cvar_of(VALUE mod, void *data)
{
VALUE tmp = mod;
+ if (RCLASS_SINGLETON_P(mod)) {
+ if (rb_namespace_p(RCLASS_ATTACHED_OBJECT(mod))) {
+ data = mod_cvar_at(tmp, data);
+ tmp = cvar_front_klass(tmp);
+ }
+ }
for (;;) {
- data = mod_cvar_at(tmp, data);
- tmp = RCLASS_SUPER(tmp);
- if (!tmp) break;
+ data = mod_cvar_at(tmp, data);
+ tmp = RCLASS_SUPER(tmp);
+ if (!tmp) break;
}
return data;
}
@@ -3072,20 +4447,15 @@ cvar_list(void *data)
VALUE
rb_mod_class_variables(int argc, const VALUE *argv, VALUE mod)
{
- VALUE inherit;
+ bool inherit = true;
st_table *tbl;
- if (argc == 0) {
- inherit = Qtrue;
+ if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]);
+ if (inherit) {
+ tbl = mod_cvar_of(mod, 0);
}
else {
- rb_scan_args(argc, argv, "01", &inherit);
- }
- if (RTEST(inherit)) {
- tbl = mod_cvar_of(mod, 0);
- }
- else {
- tbl = mod_cvar_at(mod, 0);
+ tbl = mod_cvar_at(mod, 0);
}
return cvar_list(tbl);
}
@@ -3094,13 +4464,12 @@ rb_mod_class_variables(int argc, const VALUE *argv, VALUE mod)
* call-seq:
* remove_class_variable(sym) -> obj
*
- * Removes the definition of the <i>sym</i>, returning that
- * constant's value.
+ * Removes the named class variable from the receiver, returning that
+ * variable's value.
*
- * class Dummy
+ * class Example
* @@var = 99
- * puts @@var
- * remove_class_variable(:@@var)
+ * puts remove_class_variable(:@@var)
* p(defined? @@var)
* end
*
@@ -3114,28 +4483,33 @@ VALUE
rb_mod_remove_cvar(VALUE mod, VALUE name)
{
const ID id = id_for_var_message(mod, name, class, "wrong class variable name %1$s");
- st_data_t val, n = id;
+ st_data_t val;
if (!id) {
- not_defined:
- rb_name_err_raise("class variable %1$s not defined for %2$s",
- mod, name);
+ goto not_defined;
}
rb_check_frozen(mod);
- if (RCLASS_IV_TBL(mod) && st_delete(RCLASS_IV_TBL(mod), &n, &val)) {
- return (VALUE)val;
+ val = rb_ivar_delete(mod, id, Qundef);
+ if (!UNDEF_P(val)) {
+ return (VALUE)val;
}
if (rb_cvar_defined(mod, id)) {
- rb_name_err_raise("cannot remove %1$s for %2$s", mod, ID2SYM(id));
+ rb_name_err_raise("cannot remove %1$s for %2$s", mod, ID2SYM(id));
}
- goto not_defined;
+ not_defined:
+ rb_name_err_raise("class variable %1$s not defined for %2$s",
+ mod, name);
+ UNREACHABLE_RETURN(Qundef);
}
VALUE
rb_iv_get(VALUE obj, const char *name)
{
- ID id = rb_intern(name);
+ ID id = rb_check_id_cstr(name, strlen(name), rb_usascii_encoding());
+ if (!id) {
+ return Qnil;
+ }
return rb_ivar_get(obj, id);
}
@@ -3147,39 +4521,143 @@ rb_iv_set(VALUE obj, const char *name, VALUE val)
return rb_ivar_set(obj, id, val);
}
-/* tbl = xx(obj); tbl[key] = value; */
-int
-rb_class_ivar_set(VALUE obj, ID key, VALUE value)
+static attr_index_t
+class_fields_ivar_set(VALUE klass, VALUE fields_obj, ID id, VALUE val, bool concurrent, VALUE *new_fields_obj, bool *new_ivar_out)
{
- st_table *tbl = RCLASS_IV_TBL(obj);
- int result = st_insert(tbl, (st_data_t)key, (st_data_t)value);
- RB_OBJ_WRITTEN(obj, Qundef, value);
- return result;
+ const VALUE original_fields_obj = fields_obj;
+ fields_obj = original_fields_obj ? original_fields_obj : rb_imemo_fields_new(klass, 1, true);
+
+ shape_id_t current_shape_id = RBASIC_SHAPE_ID(fields_obj);
+ shape_id_t next_shape_id = current_shape_id; // for too_complex
+ if (UNLIKELY(rb_shape_too_complex_p(current_shape_id))) {
+ goto too_complex;
+ }
+
+ bool new_ivar;
+ next_shape_id = generic_shape_ivar(fields_obj, id, &new_ivar);
+
+ if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) {
+ fields_obj = imemo_fields_complex_from_obj(klass, fields_obj, next_shape_id);
+ goto too_complex;
+ }
+
+ attr_index_t index = RSHAPE_INDEX(next_shape_id);
+ if (new_ivar) {
+ if (index >= RSHAPE_CAPACITY(current_shape_id)) {
+ // We allocate a new fields_obj even when concurrency isn't a concern
+ // so that we're embedded as long as possible.
+ fields_obj = imemo_fields_copy_capa(klass, fields_obj, RSHAPE_CAPACITY(next_shape_id));
+ }
+ }
+
+ VALUE *fields = rb_imemo_fields_ptr(fields_obj);
+
+ if (concurrent && original_fields_obj == fields_obj) {
+ // In the concurrent case, if we're mutating the existing
+ // fields_obj, we must use an atomic write, because if we're
+ // adding a new field, the shape_id must be written after the field
+ // and if we're updating an existing field, we at least need a relaxed
+ // write to avoid reaping.
+ RB_OBJ_ATOMIC_WRITE(fields_obj, &fields[index], val);
+ }
+ else {
+ RB_OBJ_WRITE(fields_obj, &fields[index], val);
+ }
+
+ if (new_ivar) {
+ RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id);
+ }
+
+ *new_fields_obj = fields_obj;
+ *new_ivar_out = new_ivar;
+ return index;
+
+too_complex:
+ {
+ if (concurrent && fields_obj == original_fields_obj) {
+ // In multi-ractor case, we must always work on a copy because
+ // even if the field already exist, inserting in a st_table may
+ // cause a rebuild.
+ fields_obj = rb_imemo_fields_clone(fields_obj);
+ }
+
+ st_table *table = rb_imemo_fields_complex_tbl(fields_obj);
+ new_ivar = !st_insert(table, (st_data_t)id, (st_data_t)val);
+ RB_OBJ_WRITTEN(fields_obj, Qundef, val);
+
+ if (fields_obj != original_fields_obj) {
+ RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id);
+ }
+ }
+
+ *new_fields_obj = fields_obj;
+ *new_ivar_out = new_ivar;
+ return ATTR_INDEX_NOT_SET;
}
-static int
-tbl_copy_i(st_data_t key, st_data_t value, st_data_t data)
+static attr_index_t
+class_ivar_set(VALUE obj, ID id, VALUE val, bool *new_ivar)
{
- RB_OBJ_WRITTEN((VALUE)data, Qundef, (VALUE)value);
- return ST_CONTINUE;
+ rb_class_ensure_writable(obj);
+
+ const VALUE original_fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(obj);
+ VALUE new_fields_obj = 0;
+
+ attr_index_t index = class_fields_ivar_set(obj, original_fields_obj, id, val, rb_multi_ractor_p(), &new_fields_obj, new_ivar);
+
+ if (new_fields_obj != original_fields_obj) {
+ RCLASS_WRITABLE_SET_FIELDS_OBJ(obj, new_fields_obj);
+ }
+
+ // TODO: What should we set as the T_CLASS shape_id?
+ // In most case we can replicate the single `fields_obj` shape
+ // but in namespaced case? Perhaps INVALID_SHAPE_ID?
+ RBASIC_SET_SHAPE_ID(obj, RBASIC_SHAPE_ID(new_fields_obj));
+ return index;
}
-st_table *
-rb_st_copy(VALUE obj, struct st_table *orig_tbl)
+bool
+rb_class_ivar_set(VALUE obj, ID id, VALUE val)
{
- st_table *new_tbl = st_copy(orig_tbl);
- st_foreach(new_tbl, tbl_copy_i, (st_data_t)obj);
- return new_tbl;
+ RUBY_ASSERT(RB_TYPE_P(obj, T_CLASS) || RB_TYPE_P(obj, T_MODULE));
+ rb_check_frozen(obj);
+
+ bool new_ivar;
+ class_ivar_set(obj, id, val, &new_ivar);
+ return new_ivar;
}
-rb_const_entry_t *
-rb_const_lookup(VALUE klass, ID id)
+void
+rb_fields_tbl_copy(VALUE dst, VALUE src)
{
- struct rb_id_table *tbl = RCLASS_CONST_TBL(klass);
- VALUE val;
+ RUBY_ASSERT(rb_type(dst) == rb_type(src));
+ RUBY_ASSERT(RB_TYPE_P(dst, T_CLASS) || RB_TYPE_P(dst, T_MODULE));
+ RUBY_ASSERT(RSHAPE_TYPE_P(RBASIC_SHAPE_ID(dst), SHAPE_ROOT));
- if (tbl && rb_id_table_lookup(tbl, id, &val)) {
- return (rb_const_entry_t *)val;
+ VALUE fields_obj = RCLASS_WRITABLE_FIELDS_OBJ(src);
+ if (fields_obj) {
+ RCLASS_WRITABLE_SET_FIELDS_OBJ(dst, rb_imemo_fields_clone(fields_obj));
+ RBASIC_SET_SHAPE_ID(dst, RBASIC_SHAPE_ID(src));
}
- return 0;
+}
+
+static rb_const_entry_t *
+const_lookup(struct rb_id_table *tbl, ID id)
+{
+ if (tbl) {
+ VALUE val;
+ bool r;
+ RB_VM_LOCKING() {
+ r = rb_id_table_lookup(tbl, id, &val);
+ }
+
+ if (r) return (rb_const_entry_t *)val;
+ }
+ return NULL;
+}
+
+rb_const_entry_t *
+rb_const_lookup(VALUE klass, ID id)
+{
+ return const_lookup(RCLASS_CONST_TBL(klass), id);
}