ruby/variable.c
Jean Boussier 3abdd4241f Turn rb_classext_t.fields into a T_IMEMO/class_fields
This behave almost exactly as a T_OBJECT, the layout is entirely
compatible.

This aims to solve two problems.

First, it solves the problem of namspaced classes having
a single `shape_id`. Now each namespaced classext
has an object that can hold the namespace specific
shape.

Second, it open the door to later make class instance variable
writes atomics, hence be able to read class variables
without locking the VM.
In the future, in multi-ractor mode, we can do the write
on a copy of the `fields_obj` and then atomically swap it.

Considerations:

  - Right now the `RClass` shape_id is always synchronized,
    but with namespace we should likely mark classes that have
    multiple namespace with a specific shape flag.
2025-06-12 07:58:16 +02:00

4854 lines
130 KiB
C

/**********************************************************************
variable.c -
$Author$
created at: Tue Apr 19 23:55:15 JST 1994
Copyright (C) 1993-2007 Yukihiro Matsumoto
Copyright (C) 2000 Network Applied Communication Laboratory, Inc.
Copyright (C) 2000 Information-technology Promotion Agency, Japan
**********************************************************************/
#include "ruby/internal/config.h"
#include <stddef.h>
#include "ruby/internal/stdbool.h"
#include "ccan/list/list.h"
#include "constant.h"
#include "debug_counter.h"
#include "id.h"
#include "id_table.h"
#include "internal.h"
#include "internal/class.h"
#include "internal/compilers.h"
#include "internal/error.h"
#include "internal/eval.h"
#include "internal/hash.h"
#include "internal/namespace.h"
#include "internal/object.h"
#include "internal/gc.h"
#include "internal/re.h"
#include "internal/symbol.h"
#include "internal/thread.h"
#include "internal/variable.h"
#include "ruby/encoding.h"
#include "ruby/st.h"
#include "ruby/util.h"
#include "shape.h"
#include "symbol.h"
#include "variable.h"
#include "vm_core.h"
#include "ractor_core.h"
#include "vm_sync.h"
RUBY_EXTERN rb_serial_t ruby_vm_global_cvar_state;
#define GET_GLOBAL_CVAR_STATE() (ruby_vm_global_cvar_state)
typedef void rb_gvar_compact_t(void *var);
static struct rb_id_table *rb_global_tbl;
static ID autoload;
// This hash table maps file paths to loadable features. We use this to track
// autoload state until it's no longer needed.
// feature (file path) => struct autoload_data
static VALUE autoload_features;
// This mutex is used to protect autoloading state. We use a global mutex which
// is held until a per-feature mutex can be created. This ensures there are no
// race conditions relating to autoload state.
static VALUE autoload_mutex;
static void check_before_mod_set(VALUE, ID, VALUE, const char *);
static void setup_const_entry(rb_const_entry_t *, VALUE, VALUE, rb_const_flag_t);
static VALUE rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility);
static st_table *generic_fields_tbl_;
typedef int rb_ivar_foreach_callback_func(ID key, VALUE val, st_data_t arg);
static void rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only);
void
Init_var_tables(void)
{
rb_global_tbl = rb_id_table_create(0);
generic_fields_tbl_ = st_init_numtable();
autoload = rb_intern_const("__autoload__");
autoload_mutex = rb_mutex_new();
rb_obj_hide(autoload_mutex);
rb_vm_register_global_object(autoload_mutex);
autoload_features = rb_ident_hash_new();
rb_obj_hide(autoload_features);
rb_vm_register_global_object(autoload_features);
}
static inline bool
rb_namespace_p(VALUE obj)
{
if (RB_SPECIAL_CONST_P(obj)) return false;
switch (RB_BUILTIN_TYPE(obj)) {
case T_MODULE: case T_CLASS: return true;
default: break;
}
return false;
}
/**
* Returns +classpath+ of _klass_, if it is named, or +nil+ for
* anonymous +class+/+module+. A named +classpath+ may contain
* an anonymous component, but the last component is guaranteed
* to not be anonymous. <code>*permanent</code> is set to 1
* if +classpath+ has no anonymous components. There is no builtin
* Ruby level APIs that can change a permanent +classpath+.
*
* YJIT needs this function to not allocate.
*/
static VALUE
classname(VALUE klass, bool *permanent)
{
*permanent = false;
VALUE classpath = RCLASS_CLASSPATH(klass);
if (classpath == 0) return Qnil;
*permanent = RCLASS_PERMANENT_CLASSPATH_P(klass);
return classpath;
}
VALUE
rb_mod_name0(VALUE klass, bool *permanent)
{
return classname(klass, permanent);
}
/*
* call-seq:
* mod.name -> string or nil
*
* Returns the name of the module <i>mod</i>. Returns +nil+ for anonymous modules.
*/
VALUE
rb_mod_name(VALUE mod)
{
// YJIT needs this function to not allocate.
bool permanent;
return classname(mod, &permanent);
}
// Similar to logic in rb_mod_const_get().
static bool
is_constant_path(VALUE name)
{
const char *path = RSTRING_PTR(name);
const char *pend = RSTRING_END(name);
rb_encoding *enc = rb_enc_get(name);
const char *p = path;
if (p >= pend || !*p) {
return false;
}
while (p < pend) {
if (p + 2 <= pend && p[0] == ':' && p[1] == ':') {
p += 2;
}
const char *pbeg = p;
while (p < pend && *p != ':') p++;
if (pbeg == p) return false;
if (rb_enc_symname_type(pbeg, p - pbeg, enc, 0) != ID_CONST) {
return false;
}
}
return true;
}
struct sub_temporary_name_args {
VALUE names;
ID last;
};
static VALUE build_const_path(VALUE head, ID tail);
static void set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name);
static VALUE
set_sub_temporary_name_recursive(VALUE mod, VALUE data, int recursive)
{
if (recursive) return Qfalse;
struct sub_temporary_name_args *args = (void *)data;
VALUE name = 0;
if (args->names) {
name = build_const_path(rb_ary_last(0, 0, args->names), args->last);
}
set_sub_temporary_name_foreach(mod, args, name);
return Qtrue;
}
static VALUE
set_sub_temporary_name_topmost(VALUE mod, VALUE data, int recursive)
{
if (recursive) return Qfalse;
struct sub_temporary_name_args *args = (void *)data;
VALUE name = args->names;
if (name) {
args->names = rb_ary_hidden_new(0);
}
set_sub_temporary_name_foreach(mod, args, name);
return Qtrue;
}
static enum rb_id_table_iterator_result
set_sub_temporary_name_i(ID id, VALUE val, void *data)
{
val = ((rb_const_entry_t *)val)->value;
if (rb_namespace_p(val) && !RCLASS_PERMANENT_CLASSPATH_P(val)) {
VALUE arg = (VALUE)data;
struct sub_temporary_name_args *args = data;
args->last = id;
rb_exec_recursive_paired(set_sub_temporary_name_recursive, val, arg, arg);
}
return ID_TABLE_CONTINUE;
}
static void
set_sub_temporary_name_foreach(VALUE mod, struct sub_temporary_name_args *args, VALUE name)
{
RCLASS_WRITE_CLASSPATH(mod, name, FALSE);
struct rb_id_table *tbl = RCLASS_CONST_TBL(mod);
if (!tbl) return;
if (!name) {
rb_id_table_foreach(tbl, set_sub_temporary_name_i, args);
}
else {
long names_len = RARRAY_LEN(args->names); // paranoiac check?
rb_ary_push(args->names, name);
rb_id_table_foreach(tbl, set_sub_temporary_name_i, args);
rb_ary_set_len(args->names, names_len);
}
}
static void
set_sub_temporary_name(VALUE mod, VALUE name)
{
struct sub_temporary_name_args args = {name};
VALUE arg = (VALUE)&args;
rb_exec_recursive_paired(set_sub_temporary_name_topmost, mod, arg, arg);
}
/*
* call-seq:
* mod.set_temporary_name(string) -> self
* mod.set_temporary_name(nil) -> self
*
* Sets the temporary name of the module. This name is reflected in
* introspection of the module and the values that are related to it, such
* as instances, constants, and methods.
*
* The name should be +nil+ or a non-empty string that is not a valid constant
* path (to avoid confusing between permanent and temporary names).
*
* The method can be useful to distinguish dynamically generated classes and
* modules without assigning them to constants.
*
* If the module is given a permanent name by assigning it to a constant,
* the temporary name is discarded. A temporary name can't be assigned to
* modules that have a permanent name.
*
* If the given name is +nil+, the module becomes anonymous again.
*
* Example:
*
* m = Module.new # => #<Module:0x0000000102c68f38>
* m.name #=> nil
*
* m.set_temporary_name("fake_name") # => fake_name
* m.name #=> "fake_name"
*
* m.set_temporary_name(nil) # => #<Module:0x0000000102c68f38>
* m.name #=> nil
*
* c = Class.new
* c.set_temporary_name("MyClass(with description)")
*
* c.new # => #<MyClass(with description):0x0....>
*
* c::M = m
* c::M.name #=> "MyClass(with description)::M"
*
* # Assigning to a constant replaces the name with a permanent one
* C = c
*
* C.name #=> "C"
* C::M.name #=> "C::M"
* c.new # => #<C:0x0....>
*/
VALUE
rb_mod_set_temporary_name(VALUE mod, VALUE name)
{
// We don't allow setting the name if the classpath is already permanent:
if (RCLASS_PERMANENT_CLASSPATH_P(mod)) {
rb_raise(rb_eRuntimeError, "can't change permanent name");
}
if (NIL_P(name)) {
// Set the temporary classpath to NULL (anonymous):
RB_VM_LOCKING() { set_sub_temporary_name(mod, 0);}
}
else {
// Ensure the name is a string:
StringValue(name);
if (RSTRING_LEN(name) == 0) {
rb_raise(rb_eArgError, "empty class/module name");
}
if (is_constant_path(name)) {
rb_raise(rb_eArgError, "the temporary name must not be a constant path to avoid confusion");
}
name = rb_str_new_frozen(name);
// Set the temporary classpath to the given name:
RB_VM_LOCKING() { set_sub_temporary_name(mod, name);}
}
return mod;
}
static VALUE
make_temporary_path(VALUE obj, VALUE klass)
{
VALUE path;
switch (klass) {
case Qnil:
path = rb_sprintf("#<Class:%p>", (void*)obj);
break;
case Qfalse:
path = rb_sprintf("#<Module:%p>", (void*)obj);
break;
default:
path = rb_sprintf("#<%"PRIsVALUE":%p>", klass, (void*)obj);
break;
}
OBJ_FREEZE(path);
return path;
}
typedef VALUE (*fallback_func)(VALUE obj, VALUE name);
static VALUE
rb_tmp_class_path(VALUE klass, bool *permanent, fallback_func fallback)
{
VALUE path = classname(klass, permanent);
if (!NIL_P(path)) {
return path;
}
if (RB_TYPE_P(klass, T_MODULE)) {
if (rb_obj_class(klass) == rb_cModule) {
path = Qfalse;
}
else {
bool perm;
path = rb_tmp_class_path(RBASIC(klass)->klass, &perm, fallback);
}
}
*permanent = false;
return fallback(klass, path);
}
VALUE
rb_class_path(VALUE klass)
{
bool permanent;
VALUE path = rb_tmp_class_path(klass, &permanent, make_temporary_path);
if (!NIL_P(path)) path = rb_str_dup(path);
return path;
}
VALUE
rb_class_path_cached(VALUE klass)
{
return rb_mod_name(klass);
}
static VALUE
no_fallback(VALUE obj, VALUE name)
{
return name;
}
VALUE
rb_search_class_path(VALUE klass)
{
bool permanent;
return rb_tmp_class_path(klass, &permanent, no_fallback);
}
static VALUE
build_const_pathname(VALUE head, VALUE tail)
{
VALUE path = rb_str_dup(head);
rb_str_cat2(path, "::");
rb_str_append(path, tail);
return rb_fstring(path);
}
static VALUE
build_const_path(VALUE head, ID tail)
{
return build_const_pathname(head, rb_id2str(tail));
}
void
rb_set_class_path_string(VALUE klass, VALUE under, VALUE name)
{
bool permanent = true;
VALUE str;
if (under == rb_cObject) {
str = rb_str_new_frozen(name);
}
else {
str = rb_tmp_class_path(under, &permanent, make_temporary_path);
str = build_const_pathname(str, name);
}
RCLASS_SET_CLASSPATH(klass, str, permanent);
}
void
rb_set_class_path(VALUE klass, VALUE under, const char *name)
{
VALUE str = rb_str_new2(name);
OBJ_FREEZE(str);
rb_set_class_path_string(klass, under, str);
}
VALUE
rb_path_to_class(VALUE pathname)
{
rb_encoding *enc = rb_enc_get(pathname);
const char *pbeg, *pend, *p, *path = RSTRING_PTR(pathname);
ID id;
VALUE c = rb_cObject;
if (!rb_enc_asciicompat(enc)) {
rb_raise(rb_eArgError, "invalid class path encoding (non ASCII)");
}
pbeg = p = path;
pend = path + RSTRING_LEN(pathname);
if (path == pend || path[0] == '#') {
rb_raise(rb_eArgError, "can't retrieve anonymous class %"PRIsVALUE,
QUOTE(pathname));
}
while (p < pend) {
while (p < pend && *p != ':') p++;
id = rb_check_id_cstr(pbeg, p-pbeg, enc);
if (p < pend && p[0] == ':') {
if ((size_t)(pend - p) < 2 || p[1] != ':') goto undefined_class;
p += 2;
pbeg = p;
}
if (!id) {
goto undefined_class;
}
c = rb_const_search(c, id, TRUE, FALSE, FALSE);
if (UNDEF_P(c)) goto undefined_class;
if (!rb_namespace_p(c)) {
rb_raise(rb_eTypeError, "%"PRIsVALUE" does not refer to class/module",
pathname);
}
}
RB_GC_GUARD(pathname);
return c;
undefined_class:
rb_raise(rb_eArgError, "undefined class/module % "PRIsVALUE,
rb_str_subseq(pathname, 0, p-path));
UNREACHABLE_RETURN(Qundef);
}
VALUE
rb_path2class(const char *path)
{
return rb_path_to_class(rb_str_new_cstr(path));
}
VALUE
rb_class_name(VALUE klass)
{
return rb_class_path(rb_class_real(klass));
}
const char *
rb_class2name(VALUE klass)
{
bool permanent;
VALUE path = rb_tmp_class_path(rb_class_real(klass), &permanent, make_temporary_path);
if (NIL_P(path)) return NULL;
return RSTRING_PTR(path);
}
const char *
rb_obj_classname(VALUE obj)
{
return rb_class2name(CLASS_OF(obj));
}
struct trace_var {
int removed;
void (*func)(VALUE arg, VALUE val);
VALUE data;
struct trace_var *next;
};
struct rb_global_variable {
int counter;
int block_trace;
VALUE *data;
rb_gvar_getter_t *getter;
rb_gvar_setter_t *setter;
rb_gvar_marker_t *marker;
rb_gvar_compact_t *compactor;
struct trace_var *trace;
bool namespace_ready;
};
struct rb_global_entry {
struct rb_global_variable *var;
ID id;
bool ractor_local;
};
static void
free_global_variable(struct rb_global_variable *var)
{
RUBY_ASSERT(var->counter == 0);
struct trace_var *trace = var->trace;
while (trace) {
struct trace_var *next = trace->next;
xfree(trace);
trace = next;
}
xfree(var);
}
static enum rb_id_table_iterator_result
free_global_entry_i(VALUE val, void *arg)
{
struct rb_global_entry *entry = (struct rb_global_entry *)val;
entry->var->counter--;
if (entry->var->counter == 0) {
free_global_variable(entry->var);
}
ruby_xfree(entry);
return ID_TABLE_DELETE;
}
void
rb_free_rb_global_tbl(void)
{
rb_id_table_foreach_values(rb_global_tbl, free_global_entry_i, 0);
rb_id_table_free(rb_global_tbl);
}
void
rb_free_generic_fields_tbl_(void)
{
st_free_table(generic_fields_tbl_);
}
static struct rb_global_entry*
rb_find_global_entry(ID id)
{
struct rb_global_entry *entry;
VALUE data;
if (!rb_id_table_lookup(rb_global_tbl, id, &data)) {
entry = NULL;
}
else {
entry = (struct rb_global_entry *)data;
RUBY_ASSERT(entry != NULL);
}
if (UNLIKELY(!rb_ractor_main_p()) && (!entry || !entry->ractor_local)) {
rb_raise(rb_eRactorIsolationError, "can not access global variables %s from non-main Ractors", rb_id2name(id));
}
return entry;
}
void
rb_gvar_ractor_local(const char *name)
{
struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name));
entry->ractor_local = true;
}
void
rb_gvar_namespace_ready(const char *name)
{
struct rb_global_entry *entry = rb_find_global_entry(rb_intern(name));
entry->var->namespace_ready = true;
}
static void
rb_gvar_undef_compactor(void *var)
{
}
static struct rb_global_entry*
rb_global_entry(ID id)
{
struct rb_global_entry *entry = rb_find_global_entry(id);
if (!entry) {
struct rb_global_variable *var;
entry = ALLOC(struct rb_global_entry);
var = ALLOC(struct rb_global_variable);
entry->id = id;
entry->var = var;
entry->ractor_local = false;
var->counter = 1;
var->data = 0;
var->getter = rb_gvar_undef_getter;
var->setter = rb_gvar_undef_setter;
var->marker = rb_gvar_undef_marker;
var->compactor = rb_gvar_undef_compactor;
var->block_trace = 0;
var->trace = 0;
var->namespace_ready = false;
rb_id_table_insert(rb_global_tbl, id, (VALUE)entry);
}
return entry;
}
VALUE
rb_gvar_undef_getter(ID id, VALUE *_)
{
rb_warning("global variable '%"PRIsVALUE"' not initialized", QUOTE_ID(id));
return Qnil;
}
static void
rb_gvar_val_compactor(void *_var)
{
struct rb_global_variable *var = (struct rb_global_variable *)_var;
VALUE obj = (VALUE)var->data;
if (obj) {
VALUE new = rb_gc_location(obj);
if (new != obj) {
var->data = (void*)new;
}
}
}
void
rb_gvar_undef_setter(VALUE val, ID id, VALUE *_)
{
struct rb_global_variable *var = rb_global_entry(id)->var;
var->getter = rb_gvar_val_getter;
var->setter = rb_gvar_val_setter;
var->marker = rb_gvar_val_marker;
var->compactor = rb_gvar_val_compactor;
var->data = (void*)val;
}
void
rb_gvar_undef_marker(VALUE *var)
{
}
VALUE
rb_gvar_val_getter(ID id, VALUE *data)
{
return (VALUE)data;
}
void
rb_gvar_val_setter(VALUE val, ID id, VALUE *_)
{
struct rb_global_variable *var = rb_global_entry(id)->var;
var->data = (void*)val;
}
void
rb_gvar_val_marker(VALUE *var)
{
VALUE data = (VALUE)var;
if (data) rb_gc_mark_movable(data);
}
VALUE
rb_gvar_var_getter(ID id, VALUE *var)
{
if (!var) return Qnil;
return *var;
}
void
rb_gvar_var_setter(VALUE val, ID id, VALUE *data)
{
*data = val;
}
void
rb_gvar_var_marker(VALUE *var)
{
if (var) rb_gc_mark_maybe(*var);
}
void
rb_gvar_readonly_setter(VALUE v, ID id, VALUE *_)
{
rb_name_error(id, "%"PRIsVALUE" is a read-only variable", QUOTE_ID(id));
}
static enum rb_id_table_iterator_result
mark_global_entry(VALUE v, void *ignored)
{
struct rb_global_entry *entry = (struct rb_global_entry *)v;
struct trace_var *trace;
struct rb_global_variable *var = entry->var;
(*var->marker)(var->data);
trace = var->trace;
while (trace) {
if (trace->data) rb_gc_mark_maybe(trace->data);
trace = trace->next;
}
return ID_TABLE_CONTINUE;
}
#define gc_mark_table(task) \
if (rb_global_tbl) { rb_id_table_foreach_values(rb_global_tbl, task##_global_entry, 0); }
void
rb_gc_mark_global_tbl(void)
{
gc_mark_table(mark);
}
static enum rb_id_table_iterator_result
update_global_entry(VALUE v, void *ignored)
{
struct rb_global_entry *entry = (struct rb_global_entry *)v;
struct rb_global_variable *var = entry->var;
(*var->compactor)(var);
return ID_TABLE_CONTINUE;
}
void
rb_gc_update_global_tbl(void)
{
gc_mark_table(update);
}
static ID
global_id(const char *name)
{
ID id;
if (name[0] == '$') id = rb_intern(name);
else {
size_t len = strlen(name);
VALUE vbuf = 0;
char *buf = ALLOCV_N(char, vbuf, len+1);
buf[0] = '$';
memcpy(buf+1, name, len);
id = rb_intern2(buf, len+1);
ALLOCV_END(vbuf);
}
return id;
}
static ID
find_global_id(const char *name)
{
ID id;
size_t len = strlen(name);
if (name[0] == '$') {
id = rb_check_id_cstr(name, len, NULL);
}
else {
VALUE vbuf = 0;
char *buf = ALLOCV_N(char, vbuf, len+1);
buf[0] = '$';
memcpy(buf+1, name, len);
id = rb_check_id_cstr(buf, len+1, NULL);
ALLOCV_END(vbuf);
}
return id;
}
void
rb_define_hooked_variable(
const char *name,
VALUE *var,
rb_gvar_getter_t *getter,
rb_gvar_setter_t *setter)
{
volatile VALUE tmp = var ? *var : Qnil;
ID id = global_id(name);
struct rb_global_variable *gvar = rb_global_entry(id)->var;
gvar->data = (void*)var;
gvar->getter = getter ? (rb_gvar_getter_t *)getter : rb_gvar_var_getter;
gvar->setter = setter ? (rb_gvar_setter_t *)setter : rb_gvar_var_setter;
gvar->marker = rb_gvar_var_marker;
RB_GC_GUARD(tmp);
}
void
rb_define_variable(const char *name, VALUE *var)
{
rb_define_hooked_variable(name, var, 0, 0);
}
void
rb_define_readonly_variable(const char *name, const VALUE *var)
{
rb_define_hooked_variable(name, (VALUE *)var, 0, rb_gvar_readonly_setter);
}
void
rb_define_virtual_variable(
const char *name,
rb_gvar_getter_t *getter,
rb_gvar_setter_t *setter)
{
if (!getter) getter = rb_gvar_val_getter;
if (!setter) setter = rb_gvar_readonly_setter;
rb_define_hooked_variable(name, 0, getter, setter);
}
static void
rb_trace_eval(VALUE cmd, VALUE val)
{
rb_eval_cmd_kw(cmd, rb_ary_new3(1, val), RB_NO_KEYWORDS);
}
VALUE
rb_f_trace_var(int argc, const VALUE *argv)
{
VALUE var, cmd;
struct rb_global_entry *entry;
struct trace_var *trace;
if (rb_scan_args(argc, argv, "11", &var, &cmd) == 1) {
cmd = rb_block_proc();
}
if (NIL_P(cmd)) {
return rb_f_untrace_var(argc, argv);
}
entry = rb_global_entry(rb_to_id(var));
trace = ALLOC(struct trace_var);
trace->next = entry->var->trace;
trace->func = rb_trace_eval;
trace->data = cmd;
trace->removed = 0;
entry->var->trace = trace;
return Qnil;
}
static void
remove_trace(struct rb_global_variable *var)
{
struct trace_var *trace = var->trace;
struct trace_var t;
struct trace_var *next;
t.next = trace;
trace = &t;
while (trace->next) {
next = trace->next;
if (next->removed) {
trace->next = next->next;
xfree(next);
}
else {
trace = next;
}
}
var->trace = t.next;
}
VALUE
rb_f_untrace_var(int argc, const VALUE *argv)
{
VALUE var, cmd;
ID id;
struct rb_global_entry *entry;
struct trace_var *trace;
rb_scan_args(argc, argv, "11", &var, &cmd);
id = rb_check_id(&var);
if (!id) {
rb_name_error_str(var, "undefined global variable %"PRIsVALUE"", QUOTE(var));
}
if ((entry = rb_find_global_entry(id)) == NULL) {
rb_name_error(id, "undefined global variable %"PRIsVALUE"", QUOTE_ID(id));
}
trace = entry->var->trace;
if (NIL_P(cmd)) {
VALUE ary = rb_ary_new();
while (trace) {
struct trace_var *next = trace->next;
rb_ary_push(ary, (VALUE)trace->data);
trace->removed = 1;
trace = next;
}
if (!entry->var->block_trace) remove_trace(entry->var);
return ary;
}
else {
while (trace) {
if (trace->data == cmd) {
trace->removed = 1;
if (!entry->var->block_trace) remove_trace(entry->var);
return rb_ary_new3(1, cmd);
}
trace = trace->next;
}
}
return Qnil;
}
struct trace_data {
struct trace_var *trace;
VALUE val;
};
static VALUE
trace_ev(VALUE v)
{
struct trace_data *data = (void *)v;
struct trace_var *trace = data->trace;
while (trace) {
(*trace->func)(trace->data, data->val);
trace = trace->next;
}
return Qnil;
}
static VALUE
trace_en(VALUE v)
{
struct rb_global_variable *var = (void *)v;
var->block_trace = 0;
remove_trace(var);
return Qnil; /* not reached */
}
static VALUE
rb_gvar_set_entry(struct rb_global_entry *entry, VALUE val)
{
struct trace_data trace;
struct rb_global_variable *var = entry->var;
(*var->setter)(val, entry->id, var->data);
if (var->trace && !var->block_trace) {
var->block_trace = 1;
trace.trace = var->trace;
trace.val = val;
rb_ensure(trace_ev, (VALUE)&trace, trace_en, (VALUE)var);
}
return val;
}
#define USE_NAMESPACE_GVAR_TBL(ns,entry) \
(NAMESPACE_OPTIONAL_P(ns) && \
(!entry || !entry->var->namespace_ready || entry->var->setter != rb_gvar_readonly_setter))
VALUE
rb_gvar_set(ID id, VALUE val)
{
VALUE retval;
struct rb_global_entry *entry;
const rb_namespace_t *ns = rb_current_namespace();
entry = rb_global_entry(id);
if (USE_NAMESPACE_GVAR_TBL(ns, entry)) {
rb_hash_aset(ns->gvar_tbl, rb_id2sym(entry->id), val);
retval = val;
// TODO: think about trace
}
else {
retval = rb_gvar_set_entry(entry, val);
}
return retval;
}
VALUE
rb_gv_set(const char *name, VALUE val)
{
return rb_gvar_set(global_id(name), val);
}
VALUE
rb_gvar_get(ID id)
{
VALUE retval, gvars, key;
struct rb_global_entry *entry = rb_global_entry(id);
struct rb_global_variable *var = entry->var;
const rb_namespace_t *ns = rb_current_namespace();
if (USE_NAMESPACE_GVAR_TBL(ns, entry)) {
gvars = ns->gvar_tbl;
key = rb_id2sym(entry->id);
if (RTEST(rb_hash_has_key(gvars, key))) { // this gvar is already cached
retval = rb_hash_aref(gvars, key);
}
else {
retval = (*var->getter)(entry->id, var->data);
if (rb_obj_respond_to(retval, rb_intern("clone"), 1)) {
retval = rb_funcall(retval, rb_intern("clone"), 0);
}
rb_hash_aset(gvars, key, retval);
}
}
else {
retval = (*var->getter)(entry->id, var->data);
}
return retval;
}
VALUE
rb_gv_get(const char *name)
{
ID id = find_global_id(name);
if (!id) {
rb_warning("global variable '%s' not initialized", name);
return Qnil;
}
return rb_gvar_get(id);
}
VALUE
rb_gvar_defined(ID id)
{
struct rb_global_entry *entry = rb_global_entry(id);
return RBOOL(entry->var->getter != rb_gvar_undef_getter);
}
rb_gvar_getter_t *
rb_gvar_getter_function_of(ID id)
{
const struct rb_global_entry *entry = rb_global_entry(id);
return entry->var->getter;
}
rb_gvar_setter_t *
rb_gvar_setter_function_of(ID id)
{
const struct rb_global_entry *entry = rb_global_entry(id);
return entry->var->setter;
}
static enum rb_id_table_iterator_result
gvar_i(ID key, VALUE val, void *a)
{
VALUE ary = (VALUE)a;
rb_ary_push(ary, ID2SYM(key));
return ID_TABLE_CONTINUE;
}
VALUE
rb_f_global_variables(void)
{
VALUE ary = rb_ary_new();
VALUE sym, backref = rb_backref_get();
if (!rb_ractor_main_p()) {
rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors");
}
/* gvar access (get/set) in namespaces creates gvar entries globally */
rb_id_table_foreach(rb_global_tbl, gvar_i, (void *)ary);
if (!NIL_P(backref)) {
char buf[2];
int i, nmatch = rb_match_count(backref);
buf[0] = '$';
for (i = 1; i <= nmatch; ++i) {
if (!RTEST(rb_reg_nth_defined(i, backref))) continue;
if (i < 10) {
/* probably reused, make static ID */
buf[1] = (char)(i + '0');
sym = ID2SYM(rb_intern2(buf, 2));
}
else {
/* dynamic symbol */
sym = rb_str_intern(rb_sprintf("$%d", i));
}
rb_ary_push(ary, sym);
}
}
return ary;
}
void
rb_alias_variable(ID name1, ID name2)
{
struct rb_global_entry *entry1, *entry2;
VALUE data1;
struct rb_id_table *gtbl = rb_global_tbl;
if (!rb_ractor_main_p()) {
rb_raise(rb_eRactorIsolationError, "can not access global variables from non-main Ractors");
}
entry2 = rb_global_entry(name2);
if (!rb_id_table_lookup(gtbl, name1, &data1)) {
entry1 = ALLOC(struct rb_global_entry);
entry1->id = name1;
rb_id_table_insert(gtbl, name1, (VALUE)entry1);
}
else if ((entry1 = (struct rb_global_entry *)data1)->var != entry2->var) {
struct rb_global_variable *var = entry1->var;
if (var->block_trace) {
rb_raise(rb_eRuntimeError, "can't alias in tracer");
}
var->counter--;
if (var->counter == 0) {
free_global_variable(var);
}
}
else {
return;
}
entry2->var->counter++;
entry1->var = entry2->var;
}
static void
IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(ID id)
{
if (UNLIKELY(!rb_ractor_main_p())) {
if (rb_is_instance_id(id)) { // check only normal ivars
rb_raise(rb_eRactorIsolationError, "can not set instance variables of classes/modules by non-main Ractors");
}
}
}
#define CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR() \
if (UNLIKELY(!rb_ractor_main_p())) { \
rb_raise(rb_eRactorIsolationError, "can not access class variables from non-main Ractors"); \
}
static inline struct st_table *
generic_fields_tbl(VALUE obj, ID id, bool force_check_ractor)
{
ASSERT_vm_locking();
if ((force_check_ractor || LIKELY(rb_is_instance_id(id)) /* not internal ID */ ) &&
!RB_OBJ_FROZEN_RAW(obj) &&
UNLIKELY(!rb_ractor_main_p()) &&
UNLIKELY(rb_ractor_shareable_p(obj))) {
rb_raise(rb_eRactorIsolationError, "can not access instance variables of shareable objects from non-main Ractors");
}
return generic_fields_tbl_;
}
static inline struct st_table *
generic_fields_tbl_no_ractor_check(VALUE obj)
{
return generic_fields_tbl(obj, 0, false);
}
struct st_table *
rb_generic_fields_tbl_get(void)
{
return generic_fields_tbl_;
}
int
rb_gen_fields_tbl_get(VALUE obj, ID id, struct gen_fields_tbl **fields_tbl)
{
RUBY_ASSERT(!RB_TYPE_P(obj, T_ICLASS));
st_data_t data;
int r = 0;
RB_VM_LOCKING() {
if (st_lookup(generic_fields_tbl(obj, id, false), (st_data_t)obj, &data)) {
*fields_tbl = (struct gen_fields_tbl *)data;
r = 1;
}
}
return r;
}
int
rb_ivar_generic_fields_tbl_lookup(VALUE obj, struct gen_fields_tbl **fields_tbl)
{
return rb_gen_fields_tbl_get(obj, 0, fields_tbl);
}
static size_t
gen_fields_tbl_bytes(size_t n)
{
return offsetof(struct gen_fields_tbl, as.shape.fields) + n * sizeof(VALUE);
}
static struct gen_fields_tbl *
gen_fields_tbl_resize(struct gen_fields_tbl *old, uint32_t new_capa)
{
RUBY_ASSERT(new_capa > 0);
return xrealloc(old, gen_fields_tbl_bytes(new_capa));
}
void
rb_mark_generic_ivar(VALUE obj)
{
st_data_t data;
if (st_lookup(generic_fields_tbl_no_ractor_check(obj), (st_data_t)obj, &data)) {
struct gen_fields_tbl *fields_tbl = (struct gen_fields_tbl *)data;
if (rb_shape_obj_too_complex_p(obj)) {
rb_mark_tbl_no_pin(fields_tbl->as.complex.table);
}
else {
uint32_t fields_count = RSHAPE_LEN(RBASIC_SHAPE_ID(obj));
for (uint32_t i = 0; i < fields_count; i++) {
rb_gc_mark_movable(fields_tbl->as.shape.fields[i]);
}
}
}
}
void
rb_free_generic_ivar(VALUE obj)
{
st_data_t key = (st_data_t)obj, value;
bool too_complex = rb_shape_obj_too_complex_p(obj);
RB_VM_LOCKING() {
if (st_delete(generic_fields_tbl_no_ractor_check(obj), &key, &value)) {
struct gen_fields_tbl *fields_tbl = (struct gen_fields_tbl *)value;
if (UNLIKELY(too_complex)) {
st_free_table(fields_tbl->as.complex.table);
}
xfree(fields_tbl);
}
}
}
size_t
rb_generic_ivar_memsize(VALUE obj)
{
struct gen_fields_tbl *fields_tbl;
if (rb_gen_fields_tbl_get(obj, 0, &fields_tbl)) {
if (rb_shape_obj_too_complex_p(obj)) {
return sizeof(struct gen_fields_tbl) + st_memsize(fields_tbl->as.complex.table);
}
else {
return gen_fields_tbl_bytes(RSHAPE_CAPACITY(RBASIC_SHAPE_ID(obj)));
}
}
return 0;
}
static size_t
gen_fields_tbl_count(VALUE obj, const struct gen_fields_tbl *fields_tbl)
{
if (rb_shape_obj_too_complex_p(obj)) {
return st_table_size(fields_tbl->as.complex.table);
}
else {
return RSHAPE_LEN(RBASIC_SHAPE_ID(obj));
}
}
VALUE
rb_obj_field_get(VALUE obj, shape_id_t target_shape_id)
{
RUBY_ASSERT(!SPECIAL_CONST_P(obj));
RUBY_ASSERT(RSHAPE_TYPE_P(target_shape_id, SHAPE_IVAR) || RSHAPE_TYPE_P(target_shape_id, SHAPE_OBJ_ID));
if (BUILTIN_TYPE(obj) == T_CLASS || BUILTIN_TYPE(obj) == T_MODULE) {
ASSERT_vm_locking();
VALUE field_obj = RCLASS_FIELDS_OBJ(obj);
if (field_obj) {
return rb_obj_field_get(field_obj, target_shape_id);
}
return Qundef;
}
if (rb_shape_too_complex_p(target_shape_id)) {
st_table *fields_hash;
switch (BUILTIN_TYPE(obj)) {
case T_CLASS:
case T_MODULE:
rb_bug("Unreachable");
break;
case T_OBJECT:
fields_hash = ROBJECT_FIELDS_HASH(obj);
break;
default:
RUBY_ASSERT(FL_TEST_RAW(obj, FL_EXIVAR));
struct gen_fields_tbl *fields_tbl = NULL;
rb_ivar_generic_fields_tbl_lookup(obj, &fields_tbl);
RUBY_ASSERT(fields_tbl);
fields_hash = fields_tbl->as.complex.table;
break;
}
VALUE value = Qundef;
st_lookup(fields_hash, RSHAPE_EDGE_NAME(target_shape_id), &value);
#if RUBY_DEBUG
if (UNDEF_P(value)) {
rb_bug("Object's shape includes object_id, but it's missing %s", rb_obj_info(obj));
}
#endif
RUBY_ASSERT(!UNDEF_P(value));
return value;
}
attr_index_t attr_index = RSHAPE_INDEX(target_shape_id);
VALUE *fields;
switch (BUILTIN_TYPE(obj)) {
case T_CLASS:
case T_MODULE:
rb_bug("Unreachable");
break;
case T_OBJECT:
fields = ROBJECT_FIELDS(obj);
break;
default:
RUBY_ASSERT(FL_TEST_RAW(obj, FL_EXIVAR));
struct gen_fields_tbl *fields_tbl = NULL;
rb_ivar_generic_fields_tbl_lookup(obj, &fields_tbl);
RUBY_ASSERT(fields_tbl);
fields = fields_tbl->as.shape.fields;
break;
}
return fields[attr_index];
}
VALUE
rb_ivar_lookup(VALUE obj, ID id, VALUE undef)
{
if (SPECIAL_CONST_P(obj)) return undef;
if (BUILTIN_TYPE(obj) == T_CLASS || BUILTIN_TYPE(obj) == T_MODULE) {
VALUE val = undef;
RB_VM_LOCK_ENTER();
{
VALUE fields_obj = RCLASS_FIELDS_OBJ(obj);
if (fields_obj) {
val = rb_ivar_lookup(fields_obj, id, undef);
}
}
RB_VM_LOCK_LEAVE();
if (val != undef &&
rb_is_instance_id(id) &&
UNLIKELY(!rb_ractor_main_p()) &&
!rb_ractor_shareable_p(val)) {
rb_raise(rb_eRactorIsolationError,
"can not get unshareable values from instance variables of classes/modules from non-main Ractors");
}
return val;
}
shape_id_t shape_id;
VALUE * ivar_list;
shape_id = RBASIC_SHAPE_ID(obj);
switch (BUILTIN_TYPE(obj)) {
case T_CLASS:
case T_MODULE:
{
rb_bug("Unreachable");
}
case T_IMEMO:
// Handled like T_OBJECT
{
RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_class_fields));
if (rb_shape_too_complex_p(shape_id)) {
st_table * iv_table = rb_imemo_class_fields_complex_tbl(obj);
VALUE val;
if (rb_st_lookup(iv_table, (st_data_t)id, (st_data_t *)&val)) {
return val;
}
else {
return undef;
}
}
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
ivar_list = rb_imemo_class_fields_ptr(obj);
break;
}
case T_OBJECT:
{
if (rb_shape_too_complex_p(shape_id)) {
st_table * iv_table = ROBJECT_FIELDS_HASH(obj);
VALUE val;
if (rb_st_lookup(iv_table, (st_data_t)id, (st_data_t *)&val)) {
return val;
}
else {
return undef;
}
}
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
ivar_list = ROBJECT_FIELDS(obj);
break;
}
default:
if (FL_TEST_RAW(obj, FL_EXIVAR)) {
struct gen_fields_tbl *fields_tbl;
rb_gen_fields_tbl_get(obj, id, &fields_tbl);
if (rb_shape_obj_too_complex_p(obj)) {
VALUE val;
if (rb_st_lookup(fields_tbl->as.complex.table, (st_data_t)id, (st_data_t *)&val)) {
return val;
}
else {
return undef;
}
}
ivar_list = fields_tbl->as.shape.fields;
}
else {
return undef;
}
break;
}
attr_index_t index = 0;
if (rb_shape_get_iv_index(shape_id, id, &index)) {
return ivar_list[index];
}
return undef;
}
VALUE
rb_ivar_get(VALUE obj, ID id)
{
VALUE iv = rb_ivar_lookup(obj, id, Qnil);
RB_DEBUG_COUNTER_INC(ivar_get_base);
return iv;
}
VALUE
rb_attr_get(VALUE obj, ID id)
{
return rb_ivar_lookup(obj, id, Qnil);
}
static VALUE
rb_ivar_delete(VALUE obj, ID id, VALUE undef)
{
rb_check_frozen(obj);
VALUE val = undef;
if (BUILTIN_TYPE(obj) == T_CLASS || BUILTIN_TYPE(obj) == T_MODULE) {
IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id);
VALUE fields_obj = RCLASS_FIELDS_OBJ(obj);
if (fields_obj) {
RB_VM_LOCK_ENTER();
{
val = rb_ivar_delete(fields_obj, id, undef);
}
RB_VM_LOCK_LEAVE();
return val;
}
}
shape_id_t old_shape_id = rb_obj_shape_id(obj);
if (rb_shape_too_complex_p(old_shape_id)) {
goto too_complex;
}
shape_id_t removed_shape_id = 0;
shape_id_t next_shape_id = rb_shape_transition_remove_ivar(obj, id, &removed_shape_id);
if (next_shape_id == old_shape_id) {
return undef;
}
if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) {
rb_evict_fields_to_hash(obj);
goto too_complex;
}
RUBY_ASSERT(RSHAPE_LEN(next_shape_id) == RSHAPE_LEN(old_shape_id) - 1);
VALUE *fields;
switch(BUILTIN_TYPE(obj)) {
case T_CLASS:
case T_MODULE:
rb_bug("Unreachable");
break;
case T_IMEMO:
RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_class_fields));
fields = rb_imemo_class_fields_ptr(obj);
break;
case T_OBJECT:
fields = ROBJECT_FIELDS(obj);
break;
default: {
struct gen_fields_tbl *fields_tbl;
rb_gen_fields_tbl_get(obj, id, &fields_tbl);
fields = fields_tbl->as.shape.fields;
break;
}
}
RUBY_ASSERT(removed_shape_id != INVALID_SHAPE_ID);
attr_index_t new_fields_count = RSHAPE_LEN(next_shape_id);
attr_index_t removed_index = RSHAPE_INDEX(removed_shape_id);
val = fields[removed_index];
size_t trailing_fields = new_fields_count - removed_index;
MEMMOVE(&fields[removed_index], &fields[removed_index + 1], VALUE, trailing_fields);
if (RB_TYPE_P(obj, T_OBJECT) &&
!RB_FL_TEST_RAW(obj, ROBJECT_EMBED) &&
rb_obj_embedded_size(new_fields_count) <= rb_gc_obj_slot_size(obj)) {
// Re-embed objects when instances become small enough
// This is necessary because YJIT assumes that objects with the same shape
// have the same embeddedness for efficiency (avoid extra checks)
RB_FL_SET_RAW(obj, ROBJECT_EMBED);
MEMCPY(ROBJECT_FIELDS(obj), fields, VALUE, new_fields_count);
xfree(fields);
}
rb_obj_set_shape_id(obj, next_shape_id);
return val;
too_complex:
{
st_table *table = NULL;
switch (BUILTIN_TYPE(obj)) {
case T_CLASS:
case T_MODULE:
rb_bug("Unreachable");
break;
case T_IMEMO:
RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_class_fields));
table = rb_imemo_class_fields_complex_tbl(obj);
break;
case T_OBJECT:
table = ROBJECT_FIELDS_HASH(obj);
break;
default: {
struct gen_fields_tbl *fields_tbl;
if (rb_gen_fields_tbl_get(obj, 0, &fields_tbl)) {
table = fields_tbl->as.complex.table;
}
break;
}
}
if (table) {
if (!st_delete(table, (st_data_t *)&id, (st_data_t *)&val)) {
val = undef;
}
}
}
return val;
}
VALUE
rb_attr_delete(VALUE obj, ID id)
{
return rb_ivar_delete(obj, id, Qnil);
}
static shape_id_t
obj_transition_too_complex(VALUE obj, st_table *table)
{
if (BUILTIN_TYPE(obj) == T_CLASS || BUILTIN_TYPE(obj) == T_MODULE) {
RUBY_ASSERT(RCLASS_FIELDS_OBJ(obj));
return obj_transition_too_complex(RCLASS_FIELDS_OBJ(obj), table);
}
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
shape_id_t shape_id = rb_shape_transition_complex(obj);
VALUE *old_fields = NULL;
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
if (!(RBASIC(obj)->flags & ROBJECT_EMBED)) {
old_fields = ROBJECT_FIELDS(obj);
}
RBASIC_SET_SHAPE_ID(obj, shape_id);
ROBJECT_SET_FIELDS_HASH(obj, table);
break;
case T_CLASS:
case T_MODULE:
rb_bug("Unreachable");
break;
default:
RB_VM_LOCKING() {
struct st_table *gen_ivs = generic_fields_tbl_no_ractor_check(obj);
struct gen_fields_tbl *old_fields_tbl = NULL;
st_lookup(gen_ivs, (st_data_t)obj, (st_data_t *)&old_fields_tbl);
if (old_fields_tbl) {
/* We need to modify old_fields_tbl to have the too complex shape
* and hold the table because the xmalloc could trigger a GC
* compaction. We want the table to be updated rather than
* the original fields. */
rb_obj_set_shape_id(obj, shape_id);
old_fields_tbl->as.complex.table = table;
old_fields = (VALUE *)old_fields_tbl;
}
struct gen_fields_tbl *fields_tbl = xmalloc(sizeof(struct gen_fields_tbl));
fields_tbl->as.complex.table = table;
st_insert(gen_ivs, (st_data_t)obj, (st_data_t)fields_tbl);
RBASIC_SET_SHAPE_ID(obj, shape_id);
}
}
xfree(old_fields);
return shape_id;
}
void
rb_obj_init_too_complex(VALUE obj, st_table *table)
{
// This method is meant to be called on newly allocated object.
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
RUBY_ASSERT(rb_shape_canonical_p(RBASIC_SHAPE_ID(obj)));
RUBY_ASSERT(RSHAPE_LEN(RBASIC_SHAPE_ID(obj)) == 0);
obj_transition_too_complex(obj, table);
}
// Copy all object fields, including ivars and internal object_id, etc
shape_id_t
rb_evict_fields_to_hash(VALUE obj)
{
void rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table);
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
st_table *table = st_init_numtable_with_size(RSHAPE_LEN(RBASIC_SHAPE_ID(obj)));
rb_obj_copy_fields_to_hash_table(obj, table);
shape_id_t new_shape_id = obj_transition_too_complex(obj, table);
RUBY_ASSERT(rb_shape_obj_too_complex_p(obj));
return new_shape_id;
}
void
rb_evict_ivars_to_hash(VALUE obj)
{
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
st_table *table = st_init_numtable_with_size(rb_ivar_count(obj));
// Evacuate all previous values from shape into id_table
rb_obj_copy_ivs_to_hash_table(obj, table);
obj_transition_too_complex(obj, table);
RUBY_ASSERT(rb_shape_obj_too_complex_p(obj));
}
struct general_ivar_set_result {
attr_index_t index;
bool existing;
};
static struct general_ivar_set_result
general_ivar_set(VALUE obj, ID id, VALUE val, void *data,
VALUE *(*shape_fields_func)(VALUE, void *),
void (*shape_resize_fields_func)(VALUE, attr_index_t, attr_index_t, void *),
void (*set_shape_id_func)(VALUE, shape_id_t, void *),
shape_id_t (*transition_too_complex_func)(VALUE, void *),
st_table *(*too_complex_table_func)(VALUE, void *))
{
struct general_ivar_set_result result = {
.index = 0,
.existing = true
};
shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj);
if (UNLIKELY(rb_shape_too_complex_p(current_shape_id))) {
goto too_complex;
}
attr_index_t index;
if (!rb_shape_get_iv_index(current_shape_id, id, &index)) {
result.existing = false;
index = RSHAPE_LEN(current_shape_id);
if (index >= SHAPE_MAX_FIELDS) {
rb_raise(rb_eArgError, "too many instance variables");
}
shape_id_t next_shape_id = rb_shape_transition_add_ivar(obj, id);
if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) {
current_shape_id = transition_too_complex_func(obj, data);
goto too_complex;
}
else if (UNLIKELY(RSHAPE_CAPACITY(next_shape_id) != RSHAPE_CAPACITY(current_shape_id))) {
RUBY_ASSERT(RSHAPE_CAPACITY(next_shape_id) > RSHAPE_CAPACITY(current_shape_id));
shape_resize_fields_func(obj, RSHAPE_CAPACITY(current_shape_id), RSHAPE_CAPACITY(next_shape_id), data);
}
RUBY_ASSERT(RSHAPE_TYPE_P(next_shape_id, SHAPE_IVAR));
RUBY_ASSERT(index == (RSHAPE_INDEX(next_shape_id)));
set_shape_id_func(obj, next_shape_id, data);
}
VALUE *table = shape_fields_func(obj, data);
RB_OBJ_WRITE(obj, &table[index], val);
result.index = index;
return result;
too_complex:
{
RUBY_ASSERT(rb_shape_obj_too_complex_p(obj));
st_table *table = too_complex_table_func(obj, data);
result.existing = st_insert(table, (st_data_t)id, (st_data_t)val);
result.index = 0;
RB_OBJ_WRITTEN(obj, Qundef, val);
}
return result;
}
static void
general_field_set(VALUE obj, shape_id_t target_shape_id, VALUE val, void *data,
VALUE *(*shape_fields_func)(VALUE, void *),
void (*shape_resize_fields_func)(VALUE, attr_index_t, attr_index_t, void *),
void (*set_shape_id_func)(VALUE, shape_id_t, void *),
shape_id_t (*transition_too_complex_func)(VALUE, void *),
st_table *(*too_complex_table_func)(VALUE, void *))
{
shape_id_t current_shape_id = RBASIC_SHAPE_ID(obj);
if (UNLIKELY(rb_shape_too_complex_p(target_shape_id))) {
if (UNLIKELY(!rb_shape_too_complex_p(current_shape_id))) {
current_shape_id = transition_too_complex_func(obj, data);
}
st_table *table = too_complex_table_func(obj, data);
if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) {
RBASIC_SET_SHAPE_ID(obj, target_shape_id);
}
RUBY_ASSERT(RSHAPE_EDGE_NAME(target_shape_id));
st_insert(table, (st_data_t)RSHAPE_EDGE_NAME(target_shape_id), (st_data_t)val);
RB_OBJ_WRITTEN(obj, Qundef, val);
}
else {
attr_index_t index = RSHAPE_INDEX(target_shape_id);
if (index >= RSHAPE_CAPACITY(current_shape_id)) {
shape_resize_fields_func(obj, RSHAPE_CAPACITY(current_shape_id), RSHAPE_CAPACITY(target_shape_id), data);
}
if (RSHAPE_LEN(target_shape_id) > RSHAPE_LEN(current_shape_id)) {
set_shape_id_func(obj, target_shape_id, data);
}
VALUE *table = shape_fields_func(obj, data);
RB_OBJ_WRITE(obj, &table[index], val);
}
}
struct gen_fields_lookup_ensure_size {
VALUE obj;
ID id;
struct gen_fields_tbl *fields_tbl;
shape_id_t shape_id;
bool resize;
};
static int
generic_fields_lookup_ensure_size(st_data_t *k, st_data_t *v, st_data_t u, int existing)
{
ASSERT_vm_locking();
struct gen_fields_lookup_ensure_size *fields_lookup = (struct gen_fields_lookup_ensure_size *)u;
struct gen_fields_tbl *fields_tbl = existing ? (struct gen_fields_tbl *)*v : NULL;
if (!existing || fields_lookup->resize) {
if (existing) {
RUBY_ASSERT(RSHAPE_TYPE_P(fields_lookup->shape_id, SHAPE_IVAR) || RSHAPE_TYPE_P(fields_lookup->shape_id, SHAPE_OBJ_ID));
RUBY_ASSERT(RSHAPE_CAPACITY(RSHAPE_PARENT(fields_lookup->shape_id)) < RSHAPE_CAPACITY(fields_lookup->shape_id));
}
else {
FL_SET_RAW((VALUE)*k, FL_EXIVAR);
}
fields_tbl = gen_fields_tbl_resize(fields_tbl, RSHAPE_CAPACITY(fields_lookup->shape_id));
*v = (st_data_t)fields_tbl;
}
RUBY_ASSERT(FL_TEST((VALUE)*k, FL_EXIVAR));
fields_lookup->fields_tbl = fields_tbl;
if (fields_lookup->shape_id) {
rb_obj_set_shape_id(fields_lookup->obj, fields_lookup->shape_id);
}
return ST_CONTINUE;
}
static VALUE *
generic_ivar_set_shape_fields(VALUE obj, void *data)
{
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
struct gen_fields_lookup_ensure_size *fields_lookup = data;
RB_VM_LOCKING() {
st_update(generic_fields_tbl(obj, fields_lookup->id, false), (st_data_t)obj, generic_fields_lookup_ensure_size, (st_data_t)fields_lookup);
}
FL_SET_RAW(obj, FL_EXIVAR);
return fields_lookup->fields_tbl->as.shape.fields;
}
static void
generic_ivar_set_shape_resize_fields(VALUE obj, attr_index_t _old_capa, attr_index_t new_capa, void *data)
{
struct gen_fields_lookup_ensure_size *fields_lookup = data;
fields_lookup->resize = true;
}
static void
generic_ivar_set_set_shape_id(VALUE obj, shape_id_t shape_id, void *data)
{
struct gen_fields_lookup_ensure_size *fields_lookup = data;
fields_lookup->shape_id = shape_id;
}
static shape_id_t
generic_ivar_set_transition_too_complex(VALUE obj, void *_data)
{
shape_id_t new_shape_id = rb_evict_fields_to_hash(obj);
FL_SET_RAW(obj, FL_EXIVAR);
return new_shape_id;
}
static st_table *
generic_ivar_set_too_complex_table(VALUE obj, void *data)
{
struct gen_fields_lookup_ensure_size *fields_lookup = data;
struct gen_fields_tbl *fields_tbl;
if (!rb_gen_fields_tbl_get(obj, 0, &fields_tbl)) {
fields_tbl = xmalloc(sizeof(struct gen_fields_tbl));
fields_tbl->as.complex.table = st_init_numtable_with_size(1);
RB_VM_LOCKING() {
st_insert(generic_fields_tbl(obj, fields_lookup->id, false), (st_data_t)obj, (st_data_t)fields_tbl);
}
FL_SET_RAW(obj, FL_EXIVAR);
}
RUBY_ASSERT(rb_shape_obj_too_complex_p(obj));
return fields_tbl->as.complex.table;
}
static void
generic_ivar_set(VALUE obj, ID id, VALUE val)
{
struct gen_fields_lookup_ensure_size fields_lookup = {
.obj = obj,
.id = id,
.resize = false,
};
general_ivar_set(obj, id, val, &fields_lookup,
generic_ivar_set_shape_fields,
generic_ivar_set_shape_resize_fields,
generic_ivar_set_set_shape_id,
generic_ivar_set_transition_too_complex,
generic_ivar_set_too_complex_table);
}
static void
generic_field_set(VALUE obj, shape_id_t target_shape_id, VALUE val)
{
struct gen_fields_lookup_ensure_size fields_lookup = {
.obj = obj,
.resize = false,
};
general_field_set(obj, target_shape_id, val, &fields_lookup,
generic_ivar_set_shape_fields,
generic_ivar_set_shape_resize_fields,
generic_ivar_set_set_shape_id,
generic_ivar_set_transition_too_complex,
generic_ivar_set_too_complex_table);
}
void
rb_ensure_iv_list_size(VALUE obj, uint32_t current_len, uint32_t new_capacity)
{
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
if (RBASIC(obj)->flags & ROBJECT_EMBED) {
VALUE *ptr = ROBJECT_FIELDS(obj);
VALUE *newptr = ALLOC_N(VALUE, new_capacity);
MEMCPY(newptr, ptr, VALUE, current_len);
RB_FL_UNSET_RAW(obj, ROBJECT_EMBED);
ROBJECT(obj)->as.heap.fields = newptr;
}
else {
REALLOC_N(ROBJECT(obj)->as.heap.fields, VALUE, new_capacity);
}
}
static int
rb_obj_copy_ivs_to_hash_table_i(ID key, VALUE val, st_data_t arg)
{
RUBY_ASSERT(!st_lookup((st_table *)arg, (st_data_t)key, NULL));
st_add_direct((st_table *)arg, (st_data_t)key, (st_data_t)val);
return ST_CONTINUE;
}
void
rb_obj_copy_ivs_to_hash_table(VALUE obj, st_table *table)
{
rb_ivar_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table);
}
void
rb_obj_copy_fields_to_hash_table(VALUE obj, st_table *table)
{
rb_field_foreach(obj, rb_obj_copy_ivs_to_hash_table_i, (st_data_t)table, false);
}
static VALUE *
obj_ivar_set_shape_fields(VALUE obj, void *_data)
{
RUBY_ASSERT(!rb_shape_obj_too_complex_p(obj));
return ROBJECT_FIELDS(obj);
}
static void
obj_ivar_set_shape_resize_fields(VALUE obj, attr_index_t old_capa, attr_index_t new_capa, void *_data)
{
rb_ensure_iv_list_size(obj, old_capa, new_capa);
}
static void
obj_ivar_set_set_shape_id(VALUE obj, shape_id_t shape_id, void *_data)
{
rb_obj_set_shape_id(obj, shape_id);
}
static shape_id_t
obj_ivar_set_transition_too_complex(VALUE obj, void *_data)
{
return rb_evict_fields_to_hash(obj);
}
static st_table *
obj_ivar_set_too_complex_table(VALUE obj, void *_data)
{
RUBY_ASSERT(rb_shape_obj_too_complex_p(obj));
return ROBJECT_FIELDS_HASH(obj);
}
attr_index_t
rb_obj_ivar_set(VALUE obj, ID id, VALUE val)
{
return general_ivar_set(obj, id, val, NULL,
obj_ivar_set_shape_fields,
obj_ivar_set_shape_resize_fields,
obj_ivar_set_set_shape_id,
obj_ivar_set_transition_too_complex,
obj_ivar_set_too_complex_table).index;
}
static void
obj_field_set(VALUE obj, shape_id_t target_shape_id, VALUE val)
{
general_field_set(obj, target_shape_id, val, NULL,
obj_ivar_set_shape_fields,
obj_ivar_set_shape_resize_fields,
obj_ivar_set_set_shape_id,
obj_ivar_set_transition_too_complex,
obj_ivar_set_too_complex_table);
}
/* Set the instance variable +val+ on object +obj+ at ivar name +id+.
* This function only works with T_OBJECT objects, so make sure
* +obj+ is of type T_OBJECT before using this function.
*/
VALUE
rb_vm_set_ivar_id(VALUE obj, ID id, VALUE val)
{
rb_check_frozen(obj);
rb_obj_ivar_set(obj, id, val);
return val;
}
bool
rb_obj_set_shape_id(VALUE obj, shape_id_t shape_id)
{
shape_id_t old_shape_id = rb_obj_shape_id(obj);
if (old_shape_id == shape_id) {
return false;
}
if (BUILTIN_TYPE(obj) == T_CLASS || BUILTIN_TYPE(obj) == T_MODULE) {
// Avoid creating the fields_obj just to freeze the class
if (!(shape_id == SPECIAL_CONST_SHAPE_ID && old_shape_id == ROOT_SHAPE_ID)) {
RBASIC_SET_SHAPE_ID(RCLASS_ENSURE_FIELDS_OBJ(obj), shape_id);
}
}
// FIXME: How to do multi-shape?
RBASIC_SET_SHAPE_ID(obj, shape_id);
return true;
}
void rb_obj_freeze_inline(VALUE x)
{
if (RB_FL_ABLE(x)) {
RB_FL_SET_RAW(x, RUBY_FL_FREEZE);
if (TYPE(x) == T_STRING) {
RB_FL_UNSET_RAW(x, FL_USER2 | FL_USER3); // STR_CHILLED
}
shape_id_t next_shape_id = rb_shape_transition_frozen(x);
rb_obj_set_shape_id(x, next_shape_id);
if (RBASIC_CLASS(x)) {
rb_freeze_singleton_class(x);
}
}
}
static void
ivar_set(VALUE obj, ID id, VALUE val)
{
RB_DEBUG_COUNTER_INC(ivar_set_base);
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
{
rb_obj_ivar_set(obj, id, val);
break;
}
case T_CLASS:
case T_MODULE:
IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(id);
rb_class_ivar_set(obj, id, val);
break;
default:
generic_ivar_set(obj, id, val);
break;
}
}
VALUE
rb_ivar_set(VALUE obj, ID id, VALUE val)
{
rb_check_frozen(obj);
ivar_set(obj, id, val);
return val;
}
void
rb_ivar_set_internal(VALUE obj, ID id, VALUE val)
{
// should be internal instance variable name (no @ prefix)
VM_ASSERT(!rb_is_instance_id(id));
ivar_set(obj, id, val);
}
static void class_field_set(VALUE obj, shape_id_t target_shape_id, VALUE val);
void
rb_obj_field_set(VALUE obj, shape_id_t target_shape_id, VALUE val)
{
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
obj_field_set(obj, target_shape_id, val);
break;
case T_CLASS:
case T_MODULE:
ASSERT_vm_locking();
class_field_set(obj, target_shape_id, val);
break;
default:
generic_field_set(obj, target_shape_id, val);
break;
}
}
static VALUE
ivar_defined0(VALUE obj, ID id)
{
attr_index_t index;
if (rb_shape_obj_too_complex_p(obj)) {
VALUE idx;
st_table *table = NULL;
switch (BUILTIN_TYPE(obj)) {
case T_CLASS:
case T_MODULE:
rb_bug("Unreachable");
break;
case T_IMEMO:
RUBY_ASSERT(IMEMO_TYPE_P(obj, imemo_class_fields));
table = rb_imemo_class_fields_complex_tbl(obj);
break;
case T_OBJECT:
table = ROBJECT_FIELDS_HASH(obj);
break;
default: {
struct gen_fields_tbl *fields_tbl;
if (rb_gen_fields_tbl_get(obj, 0, &fields_tbl)) {
table = fields_tbl->as.complex.table;
}
break;
}
}
if (!table || !rb_st_lookup(table, id, &idx)) {
return Qfalse;
}
return Qtrue;
}
else {
return RBOOL(rb_shape_get_iv_index(RBASIC_SHAPE_ID(obj), id, &index));
}
}
VALUE
rb_ivar_defined(VALUE obj, ID id)
{
if (SPECIAL_CONST_P(obj)) return Qfalse;
VALUE defined = Qfalse;
switch (BUILTIN_TYPE(obj)) {
case T_CLASS:
case T_MODULE:
RB_VM_LOCKING() {
VALUE fields_obj = RCLASS_FIELDS_OBJ(obj);
if (fields_obj) {
defined = ivar_defined0(fields_obj, id);
}
}
break;
default:
defined = ivar_defined0(obj, id);
break;
}
return defined;
}
struct iv_itr_data {
VALUE obj;
struct gen_fields_tbl *fields_tbl;
st_data_t arg;
rb_ivar_foreach_callback_func *func;
VALUE *fields;
bool ivar_only;
};
static int
iterate_over_shapes_callback(shape_id_t shape_id, void *data)
{
struct iv_itr_data *itr_data = data;
if (itr_data->ivar_only && !RSHAPE_TYPE_P(shape_id, SHAPE_IVAR)) {
return ST_CONTINUE;
}
VALUE *iv_list;
switch (BUILTIN_TYPE(itr_data->obj)) {
case T_OBJECT:
RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj));
iv_list = ROBJECT_FIELDS(itr_data->obj);
break;
case T_CLASS:
case T_MODULE:
rb_bug("Unreachable");
case T_IMEMO:
RUBY_ASSERT(IMEMO_TYPE_P(itr_data->obj, imemo_class_fields));
RUBY_ASSERT(!rb_shape_obj_too_complex_p(itr_data->obj));
iv_list = rb_imemo_class_fields_ptr(itr_data->obj);
break;
default:
iv_list = itr_data->fields_tbl->as.shape.fields;
break;
}
VALUE val = iv_list[RSHAPE_INDEX(shape_id)];
return itr_data->func(RSHAPE_EDGE_NAME(shape_id), val, itr_data->arg);
}
/*
* Returns a flag to stop iterating depending on the result of +callback+.
*/
static void
iterate_over_shapes(shape_id_t shape_id, rb_ivar_foreach_callback_func *callback, struct iv_itr_data *itr_data)
{
rb_shape_foreach_field(shape_id, iterate_over_shapes_callback, itr_data);
}
static int
each_hash_iv(st_data_t id, st_data_t val, st_data_t data)
{
struct iv_itr_data * itr_data = (struct iv_itr_data *)data;
rb_ivar_foreach_callback_func *callback = itr_data->func;
return callback((ID)id, (VALUE)val, itr_data->arg);
}
static void
obj_fields_each(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only)
{
struct iv_itr_data itr_data = {
.obj = obj,
.arg = arg,
.func = func,
.ivar_only = ivar_only,
};
shape_id_t shape_id = RBASIC_SHAPE_ID(obj);
if (rb_shape_too_complex_p(shape_id)) {
rb_st_foreach(ROBJECT_FIELDS_HASH(obj), each_hash_iv, (st_data_t)&itr_data);
}
else {
itr_data.fields = ROBJECT_FIELDS(obj);
iterate_over_shapes(shape_id, func, &itr_data);
}
}
static void
gen_fields_each(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only)
{
struct gen_fields_tbl *fields_tbl;
if (!rb_gen_fields_tbl_get(obj, 0, &fields_tbl)) return;
struct iv_itr_data itr_data = {
.obj = obj,
.fields_tbl = fields_tbl,
.arg = arg,
.func = func,
.ivar_only = ivar_only,
};
shape_id_t shape_id = RBASIC_SHAPE_ID(obj);
if (rb_shape_too_complex_p(shape_id)) {
rb_st_foreach(fields_tbl->as.complex.table, each_hash_iv, (st_data_t)&itr_data);
}
else {
itr_data.fields = fields_tbl->as.shape.fields;
iterate_over_shapes(shape_id, func, &itr_data);
}
}
static void
class_fields_each(VALUE fields_obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only)
{
IMEMO_TYPE_P(fields_obj, imemo_class_fields);
struct iv_itr_data itr_data = {
.obj = fields_obj,
.arg = arg,
.func = func,
.ivar_only = ivar_only,
};
shape_id_t shape_id = RBASIC_SHAPE_ID(fields_obj);
if (rb_shape_too_complex_p(shape_id)) {
rb_st_foreach(rb_imemo_class_fields_complex_tbl(fields_obj), each_hash_iv, (st_data_t)&itr_data);
}
else {
itr_data.fields = rb_imemo_class_fields_ptr(fields_obj);
iterate_over_shapes(shape_id, func, &itr_data);
}
}
void
rb_copy_generic_ivar(VALUE dest, VALUE obj)
{
struct gen_fields_tbl *obj_fields_tbl;
struct gen_fields_tbl *new_fields_tbl;
rb_check_frozen(dest);
if (!FL_TEST(obj, FL_EXIVAR)) {
goto clear;
}
unsigned long src_num_ivs = rb_ivar_count(obj);
if (!src_num_ivs) {
goto clear;
}
shape_id_t src_shape_id = rb_obj_shape_id(obj);
if (rb_gen_fields_tbl_get(obj, 0, &obj_fields_tbl)) {
if (gen_fields_tbl_count(obj, obj_fields_tbl) == 0)
goto clear;
FL_SET(dest, FL_EXIVAR);
if (rb_shape_too_complex_p(src_shape_id)) {
rb_shape_copy_complex_ivars(dest, obj, src_shape_id, obj_fields_tbl->as.complex.table);
return;
}
shape_id_t dest_shape_id = src_shape_id;
shape_id_t initial_shape_id = rb_obj_shape_id(dest);
if (!rb_shape_canonical_p(src_shape_id)) {
RUBY_ASSERT(RSHAPE_TYPE_P(initial_shape_id, SHAPE_ROOT));
dest_shape_id = rb_shape_rebuild(initial_shape_id, src_shape_id);
if (UNLIKELY(rb_shape_too_complex_p(dest_shape_id))) {
st_table *table = rb_st_init_numtable_with_size(src_num_ivs);
rb_obj_copy_ivs_to_hash_table(obj, table);
rb_obj_init_too_complex(dest, table);
return;
}
}
if (!RSHAPE_LEN(dest_shape_id)) {
rb_obj_set_shape_id(dest, dest_shape_id);
FL_UNSET(dest, FL_EXIVAR);
return;
}
new_fields_tbl = gen_fields_tbl_resize(0, RSHAPE_CAPACITY(dest_shape_id));
VALUE *src_buf = obj_fields_tbl->as.shape.fields;
VALUE *dest_buf = new_fields_tbl->as.shape.fields;
rb_shape_copy_fields(dest, dest_buf, dest_shape_id, obj, src_buf, src_shape_id);
/*
* c.fields_tbl may change in gen_fields_copy due to realloc,
* no need to free
*/
RB_VM_LOCKING() {
generic_fields_tbl_no_ractor_check(dest);
st_insert(generic_fields_tbl_no_ractor_check(obj), (st_data_t)dest, (st_data_t)new_fields_tbl);
}
rb_obj_set_shape_id(dest, dest_shape_id);
}
return;
clear:
if (FL_TEST(dest, FL_EXIVAR)) {
RBASIC_SET_SHAPE_ID(dest, ROOT_SHAPE_ID);
rb_free_generic_ivar(dest);
FL_UNSET(dest, FL_EXIVAR);
}
}
void
rb_replace_generic_ivar(VALUE clone, VALUE obj)
{
RUBY_ASSERT(FL_TEST(obj, FL_EXIVAR));
RB_VM_LOCKING() {
st_data_t fields_tbl, obj_data = (st_data_t)obj;
if (st_delete(generic_fields_tbl_, &obj_data, &fields_tbl)) {
FL_UNSET_RAW(obj, FL_EXIVAR);
st_insert(generic_fields_tbl_, (st_data_t)clone, fields_tbl);
FL_SET_RAW(clone, FL_EXIVAR);
}
else {
rb_bug("unreachable");
}
}
}
void
rb_field_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg, bool ivar_only)
{
if (SPECIAL_CONST_P(obj)) return;
switch (BUILTIN_TYPE(obj)) {
case T_IMEMO:
if (IMEMO_TYPE_P(obj, imemo_class_fields)) {
class_fields_each(obj, func, arg, ivar_only);
}
break;
case T_OBJECT:
obj_fields_each(obj, func, arg, ivar_only);
break;
case T_CLASS:
case T_MODULE:
IVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(0);
RB_VM_LOCKING() {
VALUE fields_obj = RCLASS_FIELDS_OBJ(obj);
if (fields_obj) {
class_fields_each(fields_obj, func, arg, ivar_only);
}
}
break;
default:
if (FL_TEST_RAW(obj, FL_EXIVAR)) {
gen_fields_each(obj, func, arg, ivar_only);
}
break;
}
}
void
rb_ivar_foreach(VALUE obj, rb_ivar_foreach_callback_func *func, st_data_t arg)
{
rb_field_foreach(obj, func, arg, true);
}
st_index_t
rb_ivar_count(VALUE obj)
{
if (SPECIAL_CONST_P(obj)) return 0;
st_index_t iv_count = 0;
switch (BUILTIN_TYPE(obj)) {
case T_OBJECT:
iv_count = ROBJECT_FIELDS_COUNT(obj);
break;
case T_CLASS:
case T_MODULE:
{
VALUE fields_obj = RCLASS_FIELDS_OBJ(obj);
if (!fields_obj) {
return 0;
}
if (rb_shape_obj_too_complex_p(fields_obj)) {
return rb_st_table_size(rb_imemo_class_fields_complex_tbl(fields_obj));
}
return RBASIC_FIELDS_COUNT(fields_obj);
}
default:
if (FL_TEST(obj, FL_EXIVAR)) {
struct gen_fields_tbl *fields_tbl;
if (rb_gen_fields_tbl_get(obj, 0, &fields_tbl)) {
iv_count = gen_fields_tbl_count(obj, fields_tbl);
}
}
break;
}
if (rb_shape_obj_has_id(obj)) {
iv_count--;
}
return iv_count;
}
static int
ivar_i(ID key, VALUE v, st_data_t a)
{
VALUE ary = (VALUE)a;
if (rb_is_instance_id(key)) {
rb_ary_push(ary, ID2SYM(key));
}
return ST_CONTINUE;
}
/*
* call-seq:
* obj.instance_variables -> array
*
* Returns an array of instance variable names for the receiver. Note
* that simply defining an accessor does not create the corresponding
* instance variable.
*
* class Fred
* attr_accessor :a1
* def initialize
* @iv = 3
* end
* end
* Fred.new.instance_variables #=> [:@iv]
*/
VALUE
rb_obj_instance_variables(VALUE obj)
{
VALUE ary;
ary = rb_ary_new();
rb_ivar_foreach(obj, ivar_i, ary);
return ary;
}
#define rb_is_constant_id rb_is_const_id
#define rb_is_constant_name rb_is_const_name
#define id_for_var(obj, name, part, type) \
id_for_var_message(obj, name, type, "'%1$s' is not allowed as "#part" "#type" variable name")
#define id_for_var_message(obj, name, type, message) \
check_id_type(obj, &(name), rb_is_##type##_id, rb_is_##type##_name, message, strlen(message))
static ID
check_id_type(VALUE obj, VALUE *pname,
int (*valid_id_p)(ID), int (*valid_name_p)(VALUE),
const char *message, size_t message_len)
{
ID id = rb_check_id(pname);
VALUE name = *pname;
if (id ? !valid_id_p(id) : !valid_name_p(name)) {
rb_name_err_raise_str(rb_fstring_new(message, message_len),
obj, name);
}
return id;
}
/*
* call-seq:
* obj.remove_instance_variable(symbol) -> obj
* obj.remove_instance_variable(string) -> obj
*
* Removes the named instance variable from <i>obj</i>, returning that
* variable's value. The name can be passed as a symbol or as a string.
*
* class Dummy
* attr_reader :var
* def initialize
* @var = 99
* end
* def remove
* remove_instance_variable(:@var)
* end
* end
* d = Dummy.new
* d.var #=> 99
* d.remove #=> 99
* d.var #=> nil
*/
VALUE
rb_obj_remove_instance_variable(VALUE obj, VALUE name)
{
const ID id = id_for_var(obj, name, an, instance);
// Frozen check comes here because it's expected that we raise a
// NameError (from the id_for_var check) before we raise a FrozenError
rb_check_frozen(obj);
if (id) {
VALUE val = rb_ivar_delete(obj, id, Qundef);
if (!UNDEF_P(val)) return val;
}
rb_name_err_raise("instance variable %1$s not defined",
obj, name);
UNREACHABLE_RETURN(Qnil);
}
NORETURN(static void uninitialized_constant(VALUE, VALUE));
static void
uninitialized_constant(VALUE klass, VALUE name)
{
if (klass && rb_class_real(klass) != rb_cObject)
rb_name_err_raise("uninitialized constant %2$s::%1$s",
klass, name);
else
rb_name_err_raise("uninitialized constant %1$s",
klass, name);
}
VALUE
rb_const_missing(VALUE klass, VALUE name)
{
VALUE value = rb_funcallv(klass, idConst_missing, 1, &name);
rb_vm_inc_const_missing_count();
return value;
}
/*
* call-seq:
* mod.const_missing(sym) -> obj
*
* Invoked when a reference is made to an undefined constant in
* <i>mod</i>. It is passed a symbol for the undefined constant, and
* returns a value to be used for that constant. For example, consider:
*
* def Foo.const_missing(name)
* name # return the constant name as Symbol
* end
*
* Foo::UNDEFINED_CONST #=> :UNDEFINED_CONST: symbol returned
*
* As the example above shows, +const_missing+ is not required to create the
* missing constant in <i>mod</i>, though that is often a side-effect. The
* caller gets its return value when triggered. If the constant is also defined,
* further lookups won't hit +const_missing+ and will return the value stored in
* the constant as usual. Otherwise, +const_missing+ will be invoked again.
*
* In the next example, when a reference is made to an undefined constant,
* +const_missing+ attempts to load a file whose path is the lowercase version
* of the constant name (thus class <code>Fred</code> is assumed to be in file
* <code>fred.rb</code>). If defined as a side-effect of loading the file, the
* method returns the value stored in the constant. This implements an autoload
* feature similar to Kernel#autoload and Module#autoload, though it differs in
* important ways.
*
* def Object.const_missing(name)
* @looked_for ||= {}
* str_name = name.to_s
* raise "Constant not found: #{name}" if @looked_for[str_name]
* @looked_for[str_name] = 1
* file = str_name.downcase
* require file
* const_get(name, false)
* end
*
*/
VALUE
rb_mod_const_missing(VALUE klass, VALUE name)
{
rb_execution_context_t *ec = GET_EC();
VALUE ref = ec->private_const_reference;
rb_vm_pop_cfunc_frame();
if (ref) {
ec->private_const_reference = 0;
rb_name_err_raise("private constant %2$s::%1$s referenced", ref, name);
}
uninitialized_constant(klass, name);
UNREACHABLE_RETURN(Qnil);
}
static void
autoload_table_mark(void *ptr)
{
rb_mark_tbl_no_pin((st_table *)ptr);
}
static void
autoload_table_free(void *ptr)
{
st_free_table((st_table *)ptr);
}
static size_t
autoload_table_memsize(const void *ptr)
{
const st_table *tbl = ptr;
return st_memsize(tbl);
}
static void
autoload_table_compact(void *ptr)
{
rb_gc_ref_update_table_values_only((st_table *)ptr);
}
static const rb_data_type_t autoload_table_type = {
"autoload_table",
{autoload_table_mark, autoload_table_free, autoload_table_memsize, autoload_table_compact,},
0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED
};
#define check_autoload_table(av) \
(struct st_table *)rb_check_typeddata((av), &autoload_table_type)
static VALUE
autoload_data(VALUE mod, ID id)
{
struct st_table *tbl;
st_data_t val;
// If we are called with a non-origin ICLASS, fetch the autoload data from
// the original module.
if (RB_TYPE_P(mod, T_ICLASS)) {
if (RICLASS_IS_ORIGIN_P(mod)) {
return 0;
}
else {
mod = RBASIC(mod)->klass;
}
}
RUBY_ASSERT(RB_TYPE_P(mod, T_CLASS) || RB_TYPE_P(mod, T_MODULE));
// Look up the instance variable table for `autoload`, then index into that table with the given constant name `id`.
VALUE tbl_value = rb_ivar_lookup(mod, autoload, Qfalse);
if (!RTEST(tbl_value) || !(tbl = check_autoload_table(tbl_value)) || !st_lookup(tbl, (st_data_t)id, &val)) {
return 0;
}
return (VALUE)val;
}
// Every autoload constant has exactly one instance of autoload_const, stored in `autoload_features`. Since multiple autoload constants can refer to the same file, every `autoload_const` refers to a de-duplicated `autoload_data`.
struct autoload_const {
// The linked list node of all constants which are loaded by the related autoload feature.
struct ccan_list_node cnode; /* <=> autoload_data.constants */
// The shared "autoload_data" if multiple constants are defined from the same feature.
VALUE autoload_data_value;
// The namespace object when the autoload is called in a user namespace
// Otherwise, Qnil means the builtin namespace, Qfalse means unspecified.
VALUE namespace;
// The module we are loading a constant into.
VALUE module;
// The name of the constant we are loading.
ID name;
// The value of the constant (after it's loaded).
VALUE value;
// The constant entry flags which need to be re-applied after autoloading the feature.
rb_const_flag_t flag;
// The source file and line number that defined this constant (different from feature path).
VALUE file;
int line;
};
// Each `autoload_data` uniquely represents a specific feature which can be loaded, and a list of constants which it is able to define. We use a mutex to coordinate multiple threads trying to load the same feature.
struct autoload_data {
// The feature path to require to load this constant.
VALUE feature;
// The mutex which is protecting autoloading this feature.
VALUE mutex;
// The process fork serial number since the autoload mutex will become invalid on fork.
rb_serial_t fork_gen;
// The linked list of all constants that are going to be loaded by this autoload.
struct ccan_list_head constants; /* <=> autoload_const.cnode */
};
static void
autoload_data_compact(void *ptr)
{
struct autoload_data *p = ptr;
p->feature = rb_gc_location(p->feature);
p->mutex = rb_gc_location(p->mutex);
}
static void
autoload_data_mark(void *ptr)
{
struct autoload_data *p = ptr;
rb_gc_mark_movable(p->feature);
rb_gc_mark_movable(p->mutex);
}
static void
autoload_data_free(void *ptr)
{
struct autoload_data *p = ptr;
struct autoload_const *autoload_const, *next;
ccan_list_for_each_safe(&p->constants, autoload_const, next, cnode) {
ccan_list_del_init(&autoload_const->cnode);
}
ruby_xfree(p);
}
static size_t
autoload_data_memsize(const void *ptr)
{
return sizeof(struct autoload_data);
}
static const rb_data_type_t autoload_data_type = {
"autoload_data",
{autoload_data_mark, autoload_data_free, autoload_data_memsize, autoload_data_compact},
0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED
};
static void
autoload_const_compact(void *ptr)
{
struct autoload_const *ac = ptr;
ac->module = rb_gc_location(ac->module);
ac->autoload_data_value = rb_gc_location(ac->autoload_data_value);
ac->value = rb_gc_location(ac->value);
ac->file = rb_gc_location(ac->file);
ac->namespace = rb_gc_location(ac->namespace);
}
static void
autoload_const_mark(void *ptr)
{
struct autoload_const *ac = ptr;
rb_gc_mark_movable(ac->module);
rb_gc_mark_movable(ac->autoload_data_value);
rb_gc_mark_movable(ac->value);
rb_gc_mark_movable(ac->file);
rb_gc_mark_movable(ac->namespace);
}
static size_t
autoload_const_memsize(const void *ptr)
{
return sizeof(struct autoload_const);
}
static void
autoload_const_free(void *ptr)
{
struct autoload_const *autoload_const = ptr;
ccan_list_del(&autoload_const->cnode);
ruby_xfree(ptr);
}
static const rb_data_type_t autoload_const_type = {
"autoload_const",
{autoload_const_mark, autoload_const_free, autoload_const_memsize, autoload_const_compact,},
0, 0, RUBY_TYPED_FREE_IMMEDIATELY
};
static struct autoload_data *
get_autoload_data(VALUE autoload_const_value, struct autoload_const **autoload_const_pointer)
{
struct autoload_const *autoload_const = rb_check_typeddata(autoload_const_value, &autoload_const_type);
VALUE autoload_data_value = autoload_const->autoload_data_value;
struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type);
/* do not reach across stack for ->state after forking: */
if (autoload_data && autoload_data->fork_gen != GET_VM()->fork_gen) {
RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil);
autoload_data->fork_gen = 0;
}
if (autoload_const_pointer) *autoload_const_pointer = autoload_const;
return autoload_data;
}
struct autoload_copy_table_data {
VALUE dst_tbl_value;
struct st_table *dst_tbl;
const rb_namespace_t *ns;
};
static int
autoload_copy_table_for_namespace_i(st_data_t key, st_data_t value, st_data_t arg)
{
struct autoload_const *autoload_const;
struct autoload_copy_table_data *data = (struct autoload_copy_table_data *)arg;
struct st_table *tbl = data->dst_tbl;
VALUE tbl_value = data->dst_tbl_value;
const rb_namespace_t *ns = data->ns;
VALUE src_value = (VALUE)value;
struct autoload_const *src_const = rb_check_typeddata(src_value, &autoload_const_type);
// autoload_data can be shared between copies because the feature is equal between copies.
VALUE autoload_data_value = src_const->autoload_data_value;
struct autoload_data *autoload_data = rb_check_typeddata(autoload_data_value, &autoload_data_type);
VALUE new_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const);
autoload_const->namespace = rb_get_namespace_object((rb_namespace_t *)ns);
autoload_const->module = src_const->module;
autoload_const->name = src_const->name;
autoload_const->value = src_const->value;
autoload_const->flag = src_const->flag;
autoload_const->autoload_data_value = autoload_data_value;
ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode);
st_insert(tbl, (st_data_t)autoload_const->name, (st_data_t)new_value);
RB_OBJ_WRITTEN(tbl_value, Qundef, new_value);
return ST_CONTINUE;
}
void
rb_autoload_copy_table_for_namespace(st_table *iv_ptr, const rb_namespace_t *ns)
{
struct st_table *src_tbl, *dst_tbl;
VALUE src_tbl_value, dst_tbl_value;
if (!rb_st_lookup(iv_ptr, (st_data_t)autoload, (st_data_t *)&src_tbl_value)) {
// the class has no autoload table yet.
return;
}
if (!RTEST(src_tbl_value) || !(src_tbl = check_autoload_table(src_tbl_value))) {
// the __autoload__ ivar value isn't autoload table value.
return;
}
src_tbl = check_autoload_table(src_tbl_value);
dst_tbl_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL);
RTYPEDDATA_DATA(dst_tbl_value) = dst_tbl = st_init_numtable();
struct autoload_copy_table_data data = {
.dst_tbl_value = dst_tbl_value,
.dst_tbl = dst_tbl,
.ns = ns,
};
st_foreach(src_tbl, autoload_copy_table_for_namespace_i, (st_data_t)&data);
st_insert(iv_ptr, (st_data_t)autoload, (st_data_t)dst_tbl_value);
}
void
rb_autoload(VALUE module, ID name, const char *feature)
{
if (!feature || !*feature) {
rb_raise(rb_eArgError, "empty feature name");
}
rb_autoload_str(module, name, rb_fstring_cstr(feature));
}
static void const_set(VALUE klass, ID id, VALUE val);
static void const_added(VALUE klass, ID const_name);
struct autoload_arguments {
VALUE module;
ID name;
VALUE feature;
VALUE namespace;
};
static VALUE
autoload_feature_lookup_or_create(VALUE feature, struct autoload_data **autoload_data_pointer)
{
RUBY_ASSERT_MUTEX_OWNED(autoload_mutex);
RUBY_ASSERT_CRITICAL_SECTION_ENTER();
VALUE autoload_data_value = rb_hash_aref(autoload_features, feature);
struct autoload_data *autoload_data;
if (NIL_P(autoload_data_value)) {
autoload_data_value = TypedData_Make_Struct(0, struct autoload_data, &autoload_data_type, autoload_data);
RB_OBJ_WRITE(autoload_data_value, &autoload_data->feature, feature);
RB_OBJ_WRITE(autoload_data_value, &autoload_data->mutex, Qnil);
ccan_list_head_init(&autoload_data->constants);
if (autoload_data_pointer) *autoload_data_pointer = autoload_data;
rb_hash_aset(autoload_features, feature, autoload_data_value);
}
else if (autoload_data_pointer) {
*autoload_data_pointer = rb_check_typeddata(autoload_data_value, &autoload_data_type);
}
RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
return autoload_data_value;
}
static VALUE
autoload_table_lookup_or_create(VALUE module)
{
VALUE autoload_table_value = rb_ivar_lookup(module, autoload, Qfalse);
if (RTEST(autoload_table_value)) {
return autoload_table_value;
}
else {
autoload_table_value = TypedData_Wrap_Struct(0, &autoload_table_type, NULL);
rb_class_ivar_set(module, autoload, autoload_table_value);
RTYPEDDATA_DATA(autoload_table_value) = st_init_numtable();
return autoload_table_value;
}
}
static VALUE
autoload_synchronized(VALUE _arguments)
{
struct autoload_arguments *arguments = (struct autoload_arguments *)_arguments;
rb_const_entry_t *constant_entry = rb_const_lookup(arguments->module, arguments->name);
if (constant_entry && !UNDEF_P(constant_entry->value)) {
return Qfalse;
}
// Reset any state associated with any previous constant:
const_set(arguments->module, arguments->name, Qundef);
VALUE autoload_table_value = autoload_table_lookup_or_create(arguments->module);
struct st_table *autoload_table = check_autoload_table(autoload_table_value);
// Ensure the string is uniqued since we use an identity lookup:
VALUE feature = rb_fstring(arguments->feature);
struct autoload_data *autoload_data;
VALUE autoload_data_value = autoload_feature_lookup_or_create(feature, &autoload_data);
{
struct autoload_const *autoload_const;
VALUE autoload_const_value = TypedData_Make_Struct(0, struct autoload_const, &autoload_const_type, autoload_const);
autoload_const->namespace = arguments->namespace;
autoload_const->module = arguments->module;
autoload_const->name = arguments->name;
autoload_const->value = Qundef;
autoload_const->flag = CONST_PUBLIC;
autoload_const->autoload_data_value = autoload_data_value;
ccan_list_add_tail(&autoload_data->constants, &autoload_const->cnode);
st_insert(autoload_table, (st_data_t)arguments->name, (st_data_t)autoload_const_value);
RB_OBJ_WRITTEN(autoload_table_value, Qundef, autoload_const_value);
}
return Qtrue;
}
void
rb_autoload_str(VALUE module, ID name, VALUE feature)
{
const rb_namespace_t *ns = rb_current_namespace();
VALUE current_namespace = rb_get_namespace_object((rb_namespace_t *)ns);
if (!rb_is_const_id(name)) {
rb_raise(rb_eNameError, "autoload must be constant name: %"PRIsVALUE"", QUOTE_ID(name));
}
Check_Type(feature, T_STRING);
if (!RSTRING_LEN(feature)) {
rb_raise(rb_eArgError, "empty feature name");
}
struct autoload_arguments arguments = {
.module = module,
.name = name,
.feature = feature,
.namespace = current_namespace,
};
VALUE result = rb_mutex_synchronize(autoload_mutex, autoload_synchronized, (VALUE)&arguments);
if (result == Qtrue) {
const_added(module, name);
}
}
static void
autoload_delete(VALUE module, ID name)
{
RUBY_ASSERT_CRITICAL_SECTION_ENTER();
st_data_t load = 0, key = name;
RUBY_ASSERT(RB_TYPE_P(module, T_CLASS) || RB_TYPE_P(module, T_MODULE));
VALUE table_value = rb_ivar_lookup(module, autoload, Qfalse);
if (RTEST(table_value)) {
struct st_table *table = check_autoload_table(table_value);
st_delete(table, &key, &load);
RB_OBJ_WRITTEN(table_value, load, Qundef);
/* Qfalse can indicate already deleted */
if (load != Qfalse) {
struct autoload_const *autoload_const;
struct autoload_data *autoload_data = get_autoload_data((VALUE)load, &autoload_const);
VM_ASSERT(autoload_data);
VM_ASSERT(!ccan_list_empty(&autoload_data->constants));
/*
* we must delete here to avoid "already initialized" warnings
* with parallel autoload. Using list_del_init here so list_del
* works in autoload_const_free
*/
ccan_list_del_init(&autoload_const->cnode);
if (ccan_list_empty(&autoload_data->constants)) {
rb_hash_delete(autoload_features, autoload_data->feature);
}
// If the autoload table is empty, we can delete it.
if (table->num_entries == 0) {
rb_attr_delete(module, autoload);
}
}
}
RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
}
static int
autoload_by_someone_else(struct autoload_data *ele)
{
return ele->mutex != Qnil && !rb_mutex_owned_p(ele->mutex);
}
static VALUE
check_autoload_required(VALUE mod, ID id, const char **loadingpath)
{
VALUE autoload_const_value = autoload_data(mod, id);
struct autoload_data *autoload_data;
const char *loading;
if (!autoload_const_value || !(autoload_data = get_autoload_data(autoload_const_value, 0))) {
return 0;
}
VALUE feature = autoload_data->feature;
/*
* if somebody else is autoloading, we MUST wait for them, since
* rb_provide_feature can provide a feature before autoload_const_set
* completes. We must wait until autoload_const_set finishes in
* the other thread.
*/
if (autoload_by_someone_else(autoload_data)) {
return autoload_const_value;
}
loading = RSTRING_PTR(feature);
if (!rb_feature_provided(loading, &loading)) {
return autoload_const_value;
}
if (loadingpath && loading) {
*loadingpath = loading;
return autoload_const_value;
}
return 0;
}
static struct autoload_const *autoloading_const_entry(VALUE mod, ID id);
int
rb_autoloading_value(VALUE mod, ID id, VALUE* value, rb_const_flag_t *flag)
{
struct autoload_const *ac = autoloading_const_entry(mod, id);
if (!ac) return FALSE;
if (value) {
*value = ac->value;
}
if (flag) {
*flag = ac->flag;
}
return TRUE;
}
static int
autoload_by_current(struct autoload_data *ele)
{
return ele->mutex != Qnil && rb_mutex_owned_p(ele->mutex);
}
// If there is an autoloading constant and it has been set by the current
// execution context, return it. This allows threads which are loading code to
// refer to their own autoloaded constants.
struct autoload_const *
autoloading_const_entry(VALUE mod, ID id)
{
VALUE load = autoload_data(mod, id);
struct autoload_data *ele;
struct autoload_const *ac;
// Find the autoloading state:
if (!load || !(ele = get_autoload_data(load, &ac))) {
// Couldn't be found:
return 0;
}
// Check if it's being loaded by the current thread/fiber:
if (autoload_by_current(ele)) {
if (!UNDEF_P(ac->value)) {
return ac;
}
}
return 0;
}
static int
autoload_defined_p(VALUE mod, ID id)
{
rb_const_entry_t *ce = rb_const_lookup(mod, id);
// If there is no constant or the constant is not undefined (special marker for autoloading):
if (!ce || !UNDEF_P(ce->value)) {
// We are not autoloading:
return 0;
}
// Otherwise check if there is an autoload in flight right now:
return !rb_autoloading_value(mod, id, NULL, NULL);
}
static void const_tbl_update(struct autoload_const *, int);
struct autoload_load_arguments {
VALUE module;
ID name;
int flag;
VALUE mutex;
// The specific constant which triggered the autoload code to fire:
struct autoload_const *autoload_const;
// The parent autoload data which is shared between multiple constants:
struct autoload_data *autoload_data;
};
static VALUE
autoload_const_set(struct autoload_const *ac)
{
check_before_mod_set(ac->module, ac->name, ac->value, "constant");
RB_VM_LOCKING() {
const_tbl_update(ac, true);
}
return 0; /* ignored */
}
static VALUE
autoload_load_needed(VALUE _arguments)
{
struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
const char *loading = 0, *src;
if (!autoload_defined_p(arguments->module, arguments->name)) {
return Qfalse;
}
VALUE autoload_const_value = check_autoload_required(arguments->module, arguments->name, &loading);
if (!autoload_const_value) {
return Qfalse;
}
src = rb_sourcefile();
if (src && loading && strcmp(src, loading) == 0) {
return Qfalse;
}
struct autoload_const *autoload_const;
struct autoload_data *autoload_data;
if (!(autoload_data = get_autoload_data(autoload_const_value, &autoload_const))) {
return Qfalse;
}
if (NIL_P(autoload_data->mutex)) {
RB_OBJ_WRITE(autoload_const->autoload_data_value, &autoload_data->mutex, rb_mutex_new());
autoload_data->fork_gen = GET_VM()->fork_gen;
}
else if (rb_mutex_owned_p(autoload_data->mutex)) {
return Qfalse;
}
arguments->mutex = autoload_data->mutex;
arguments->autoload_const = autoload_const;
return autoload_const_value;
}
static VALUE
autoload_apply_constants(VALUE _arguments)
{
RUBY_ASSERT_CRITICAL_SECTION_ENTER();
struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
struct autoload_const *autoload_const = 0; // for ccan_container_off_var()
struct autoload_const *next;
// We use safe iteration here because `autoload_const_set` will eventually invoke
// `autoload_delete` which will remove the constant from the linked list. In theory, once
// the `autoload_data->constants` linked list is empty, we can remove it.
// Iterate over all constants and assign them:
ccan_list_for_each_safe(&arguments->autoload_data->constants, autoload_const, next, cnode) {
if (!UNDEF_P(autoload_const->value)) {
autoload_const_set(autoload_const);
}
}
RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
return Qtrue;
}
struct autoload_feature_require_data {
struct autoload_load_arguments *arguments;
VALUE receiver;
VALUE feature;
};
static VALUE
autoload_feature_require_in_builtin(VALUE arg)
{
struct autoload_feature_require_data *data = (struct autoload_feature_require_data *)arg;
VALUE result = rb_funcall(data->receiver, rb_intern("require"), 1, data->feature);
if (RTEST(result)) {
return rb_mutex_synchronize(autoload_mutex, autoload_apply_constants, (VALUE)data->arguments);
}
return Qnil;
}
static VALUE
autoload_feature_require_ensure_in_builtin(VALUE _arg)
{
/*
* The gccct should be cleared again after the rb_funcall() to remove
* the inconsistent cache entry against the current namespace.
*/
rb_gccct_clear_table(Qnil);
rb_namespace_disable_builtin();
return Qnil;
}
static VALUE
autoload_feature_require_in_builtin_wrap(VALUE arg)
{
return rb_ensure(autoload_feature_require_in_builtin, arg,
autoload_feature_require_ensure_in_builtin, Qnil);
}
static VALUE
autoload_feature_require(VALUE _arguments)
{
VALUE receiver = rb_vm_top_self();
struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
struct autoload_const *autoload_const = arguments->autoload_const;
VALUE autoload_namespace = autoload_const->namespace;
// We save this for later use in autoload_apply_constants:
arguments->autoload_data = rb_check_typeddata(autoload_const->autoload_data_value, &autoload_data_type);
if (NIL_P(autoload_namespace)) {
rb_namespace_enable_builtin();
/*
* Clear the global cc cache table because the require method can be different from the current
* namespace's one and it may cause inconsistent cc-cme states.
* For example, the assertion below may fail in gccct_method_search();
* VM_ASSERT(vm_cc_check_cme(cc, rb_callable_method_entry(klass, mid)))
*/
rb_gccct_clear_table(Qnil);
struct autoload_feature_require_data data = {
.arguments = arguments,
.receiver = receiver,
.feature = arguments->autoload_data->feature,
};
return rb_namespace_exec(rb_builtin_namespace(), autoload_feature_require_in_builtin_wrap, (VALUE)&data);
}
if (RTEST(autoload_namespace) && NAMESPACE_OPTIONAL_P(rb_get_namespace_t(autoload_namespace))) {
receiver = autoload_namespace;
}
VALUE result = rb_funcall(receiver, rb_intern("require"), 1, arguments->autoload_data->feature);
if (RTEST(result)) {
return rb_mutex_synchronize(autoload_mutex, autoload_apply_constants, _arguments);
}
return result;
}
static VALUE
autoload_try_load(VALUE _arguments)
{
struct autoload_load_arguments *arguments = (struct autoload_load_arguments*)_arguments;
VALUE result = autoload_feature_require(_arguments);
// After we loaded the feature, if the constant is not defined, we remove it completely:
rb_const_entry_t *ce = rb_const_lookup(arguments->module, arguments->name);
if (!ce || UNDEF_P(ce->value)) {
result = Qfalse;
rb_const_remove(arguments->module, arguments->name);
if (arguments->module == rb_cObject) {
rb_warning(
"Expected %"PRIsVALUE" to define %"PRIsVALUE" but it didn't",
arguments->autoload_data->feature,
ID2SYM(arguments->name)
);
}
else {
rb_warning(
"Expected %"PRIsVALUE" to define %"PRIsVALUE"::%"PRIsVALUE" but it didn't",
arguments->autoload_data->feature,
arguments->module,
ID2SYM(arguments->name)
);
}
}
else {
// Otherwise, it was loaded, copy the flags from the autoload constant:
ce->flag |= arguments->flag;
}
return result;
}
VALUE
rb_autoload_load(VALUE module, ID name)
{
rb_const_entry_t *ce = rb_const_lookup(module, name);
// We bail out as early as possible without any synchronisation:
if (!ce || !UNDEF_P(ce->value)) {
return Qfalse;
}
// At this point, we assume there might be autoloading, so fail if it's ractor:
if (UNLIKELY(!rb_ractor_main_p())) {
return rb_ractor_autoload_load(module, name);
}
// This state is stored on the stack and is used during the autoload process.
struct autoload_load_arguments arguments = {.module = module, .name = name, .mutex = Qnil};
// Figure out whether we can autoload the named constant:
VALUE autoload_const_value = rb_mutex_synchronize(autoload_mutex, autoload_load_needed, (VALUE)&arguments);
// This confirms whether autoloading is required or not:
if (autoload_const_value == Qfalse) return autoload_const_value;
arguments.flag = ce->flag & (CONST_DEPRECATED | CONST_VISIBILITY_MASK);
// Only one thread will enter here at a time:
VALUE result = rb_mutex_synchronize(arguments.mutex, autoload_try_load, (VALUE)&arguments);
// If you don't guard this value, it's possible for the autoload constant to
// be freed by another thread which loads multiple constants, one of which
// resolves to the constant this thread is trying to load, so proteect this
// so that it is not freed until we are done with it in `autoload_try_load`:
RB_GC_GUARD(autoload_const_value);
return result;
}
VALUE
rb_autoload_p(VALUE mod, ID id)
{
return rb_autoload_at_p(mod, id, TRUE);
}
VALUE
rb_autoload_at_p(VALUE mod, ID id, int recur)
{
VALUE load;
struct autoload_data *ele;
while (!autoload_defined_p(mod, id)) {
if (!recur) return Qnil;
mod = RCLASS_SUPER(mod);
if (!mod) return Qnil;
}
load = check_autoload_required(mod, id, 0);
if (!load) return Qnil;
return (ele = get_autoload_data(load, 0)) ? ele->feature : Qnil;
}
void
rb_const_warn_if_deprecated(const rb_const_entry_t *ce, VALUE klass, ID id)
{
if (RB_CONST_DEPRECATED_P(ce) &&
rb_warning_category_enabled_p(RB_WARN_CATEGORY_DEPRECATED)) {
if (klass == rb_cObject) {
rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant ::%"PRIsVALUE" is deprecated", QUOTE_ID(id));
}
else {
rb_category_warn(RB_WARN_CATEGORY_DEPRECATED, "constant %"PRIsVALUE"::%"PRIsVALUE" is deprecated",
rb_class_name(klass), QUOTE_ID(id));
}
}
}
static VALUE
rb_const_get_0(VALUE klass, ID id, int exclude, int recurse, int visibility)
{
VALUE c = rb_const_search(klass, id, exclude, recurse, visibility);
if (!UNDEF_P(c)) {
if (UNLIKELY(!rb_ractor_main_p())) {
if (!rb_ractor_shareable_p(c)) {
rb_raise(rb_eRactorIsolationError, "can not access non-shareable objects in constant %"PRIsVALUE"::%s by non-main Ractor.", rb_class_path(klass), rb_id2name(id));
}
}
return c;
}
return rb_const_missing(klass, ID2SYM(id));
}
static VALUE
rb_const_search_from(VALUE klass, ID id, int exclude, int recurse, int visibility)
{
VALUE value, current;
bool first_iteration = true;
for (current = klass;
RTEST(current);
current = RCLASS_SUPER(current), first_iteration = false) {
VALUE tmp;
VALUE am = 0;
rb_const_entry_t *ce;
if (!first_iteration && RCLASS_ORIGIN(current) != current) {
// This item in the super chain has an origin iclass
// that comes later in the chain. Skip this item so
// prepended modules take precedence.
continue;
}
// Do lookup in original class or module in case we are at an origin
// iclass in the chain.
tmp = current;
if (BUILTIN_TYPE(tmp) == T_ICLASS) tmp = RBASIC(tmp)->klass;
// Do the lookup. Loop in case of autoload.
while ((ce = rb_const_lookup(tmp, id))) {
if (visibility && RB_CONST_PRIVATE_P(ce)) {
GET_EC()->private_const_reference = tmp;
return Qundef;
}
rb_const_warn_if_deprecated(ce, tmp, id);
value = ce->value;
if (UNDEF_P(value)) {
struct autoload_const *ac;
if (am == tmp) break;
am = tmp;
ac = autoloading_const_entry(tmp, id);
if (ac) return ac->value;
rb_autoload_load(tmp, id);
continue;
}
if (exclude && tmp == rb_cObject) {
goto not_found;
}
return value;
}
if (!recurse) break;
}
not_found:
GET_EC()->private_const_reference = 0;
return Qundef;
}
static VALUE
rb_const_search(VALUE klass, ID id, int exclude, int recurse, int visibility)
{
VALUE value;
if (klass == rb_cObject) exclude = FALSE;
value = rb_const_search_from(klass, id, exclude, recurse, visibility);
if (!UNDEF_P(value)) return value;
if (exclude) return value;
if (BUILTIN_TYPE(klass) != T_MODULE) return value;
/* search global const too, if klass is a module */
return rb_const_search_from(rb_cObject, id, FALSE, recurse, visibility);
}
VALUE
rb_const_get_from(VALUE klass, ID id)
{
return rb_const_get_0(klass, id, TRUE, TRUE, FALSE);
}
VALUE
rb_const_get(VALUE klass, ID id)
{
return rb_const_get_0(klass, id, FALSE, TRUE, FALSE);
}
VALUE
rb_const_get_at(VALUE klass, ID id)
{
return rb_const_get_0(klass, id, TRUE, FALSE, FALSE);
}
VALUE
rb_public_const_get_from(VALUE klass, ID id)
{
return rb_const_get_0(klass, id, TRUE, TRUE, TRUE);
}
VALUE
rb_public_const_get_at(VALUE klass, ID id)
{
return rb_const_get_0(klass, id, TRUE, FALSE, TRUE);
}
NORETURN(static void undefined_constant(VALUE mod, VALUE name));
static void
undefined_constant(VALUE mod, VALUE name)
{
rb_name_err_raise("constant %2$s::%1$s not defined",
mod, name);
}
static VALUE
rb_const_location_from(VALUE klass, ID id, int exclude, int recurse, int visibility)
{
while (RTEST(klass)) {
rb_const_entry_t *ce;
while ((ce = rb_const_lookup(klass, id))) {
if (visibility && RB_CONST_PRIVATE_P(ce)) {
return Qnil;
}
if (exclude && klass == rb_cObject) {
goto not_found;
}
if (UNDEF_P(ce->value)) { // autoload
VALUE autoload_const_value = autoload_data(klass, id);
if (RTEST(autoload_const_value)) {
struct autoload_const *autoload_const;
struct autoload_data *autoload_data = get_autoload_data(autoload_const_value, &autoload_const);
if (!UNDEF_P(autoload_const->value) && RTEST(rb_mutex_owned_p(autoload_data->mutex))) {
return rb_assoc_new(autoload_const->file, INT2NUM(autoload_const->line));
}
}
}
if (NIL_P(ce->file)) return rb_ary_new();
return rb_assoc_new(ce->file, INT2NUM(ce->line));
}
if (!recurse) break;
klass = RCLASS_SUPER(klass);
}
not_found:
return Qnil;
}
static VALUE
rb_const_location(VALUE klass, ID id, int exclude, int recurse, int visibility)
{
VALUE loc;
if (klass == rb_cObject) exclude = FALSE;
loc = rb_const_location_from(klass, id, exclude, recurse, visibility);
if (!NIL_P(loc)) return loc;
if (exclude) return loc;
if (BUILTIN_TYPE(klass) != T_MODULE) return loc;
/* search global const too, if klass is a module */
return rb_const_location_from(rb_cObject, id, FALSE, recurse, visibility);
}
VALUE
rb_const_source_location(VALUE klass, ID id)
{
return rb_const_location(klass, id, FALSE, TRUE, FALSE);
}
VALUE
rb_const_source_location_at(VALUE klass, ID id)
{
return rb_const_location(klass, id, TRUE, FALSE, FALSE);
}
/*
* call-seq:
* remove_const(sym) -> obj
*
* Removes the definition of the given constant, returning that
* constant's previous value. If that constant referred to
* a module, this will not change that module's name and can lead
* to confusion.
*/
VALUE
rb_mod_remove_const(VALUE mod, VALUE name)
{
const ID id = id_for_var(mod, name, a, constant);
if (!id) {
undefined_constant(mod, name);
}
return rb_const_remove(mod, id);
}
static rb_const_entry_t * const_lookup(struct rb_id_table *tbl, ID id);
VALUE
rb_const_remove(VALUE mod, ID id)
{
VALUE val;
rb_const_entry_t *ce;
rb_check_frozen(mod);
ce = rb_const_lookup(mod, id);
if (!ce || !rb_id_table_delete(RCLASS_WRITABLE_CONST_TBL(mod), id)) {
if (rb_const_defined_at(mod, id)) {
rb_name_err_raise("cannot remove %2$s::%1$s", mod, ID2SYM(id));
}
undefined_constant(mod, ID2SYM(id));
}
rb_const_warn_if_deprecated(ce, mod, id);
rb_clear_constant_cache_for_id(id);
val = ce->value;
if (UNDEF_P(val)) {
autoload_delete(mod, id);
val = Qnil;
}
if (ce != const_lookup(RCLASS_PRIME_CONST_TBL(mod), id)) {
ruby_xfree(ce);
}
// else - skip free'ing the ce because it still exists in the prime classext
return val;
}
static int
cv_i_update(st_data_t *k, st_data_t *v, st_data_t a, int existing)
{
if (existing) return ST_STOP;
*v = a;
return ST_CONTINUE;
}
static enum rb_id_table_iterator_result
sv_i(ID key, VALUE v, void *a)
{
rb_const_entry_t *ce = (rb_const_entry_t *)v;
st_table *tbl = a;
if (rb_is_const_id(key)) {
st_update(tbl, (st_data_t)key, cv_i_update, (st_data_t)ce);
}
return ID_TABLE_CONTINUE;
}
static enum rb_id_table_iterator_result
rb_local_constants_i(ID const_name, VALUE const_value, void *ary)
{
if (rb_is_const_id(const_name) && !RB_CONST_PRIVATE_P((rb_const_entry_t *)const_value)) {
rb_ary_push((VALUE)ary, ID2SYM(const_name));
}
return ID_TABLE_CONTINUE;
}
static VALUE
rb_local_constants(VALUE mod)
{
struct rb_id_table *tbl = RCLASS_CONST_TBL(mod);
VALUE ary;
if (!tbl) return rb_ary_new2(0);
RB_VM_LOCKING() {
ary = rb_ary_new2(rb_id_table_size(tbl));
rb_id_table_foreach(tbl, rb_local_constants_i, (void *)ary);
}
return ary;
}
void*
rb_mod_const_at(VALUE mod, void *data)
{
st_table *tbl = data;
if (!tbl) {
tbl = st_init_numtable();
}
if (RCLASS_CONST_TBL(mod)) {
RB_VM_LOCKING() {
rb_id_table_foreach(RCLASS_CONST_TBL(mod), sv_i, tbl);
}
}
return tbl;
}
void*
rb_mod_const_of(VALUE mod, void *data)
{
VALUE tmp = mod;
for (;;) {
data = rb_mod_const_at(tmp, data);
tmp = RCLASS_SUPER(tmp);
if (!tmp) break;
if (tmp == rb_cObject && mod != rb_cObject) break;
}
return data;
}
static int
list_i(st_data_t key, st_data_t value, VALUE ary)
{
ID sym = (ID)key;
rb_const_entry_t *ce = (rb_const_entry_t *)value;
if (RB_CONST_PUBLIC_P(ce)) rb_ary_push(ary, ID2SYM(sym));
return ST_CONTINUE;
}
VALUE
rb_const_list(void *data)
{
st_table *tbl = data;
VALUE ary;
if (!tbl) return rb_ary_new2(0);
ary = rb_ary_new2(tbl->num_entries);
st_foreach_safe(tbl, list_i, ary);
st_free_table(tbl);
return ary;
}
/*
* call-seq:
* mod.constants(inherit=true) -> array
*
* Returns an array of the names of the constants accessible in
* <i>mod</i>. This includes the names of constants in any included
* modules (example at start of section), unless the <i>inherit</i>
* parameter is set to <code>false</code>.
*
* The implementation makes no guarantees about the order in which the
* constants are yielded.
*
* IO.constants.include?(:SYNC) #=> true
* IO.constants(false).include?(:SYNC) #=> false
*
* Also see Module#const_defined?.
*/
VALUE
rb_mod_constants(int argc, const VALUE *argv, VALUE mod)
{
bool inherit = true;
if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]);
if (inherit) {
return rb_const_list(rb_mod_const_of(mod, 0));
}
else {
return rb_local_constants(mod);
}
}
static int
rb_const_defined_0(VALUE klass, ID id, int exclude, int recurse, int visibility)
{
VALUE tmp;
int mod_retry = 0;
rb_const_entry_t *ce;
tmp = klass;
retry:
while (tmp) {
if ((ce = rb_const_lookup(tmp, id))) {
if (visibility && RB_CONST_PRIVATE_P(ce)) {
return (int)Qfalse;
}
if (UNDEF_P(ce->value) && !check_autoload_required(tmp, id, 0) &&
!rb_autoloading_value(tmp, id, NULL, NULL))
return (int)Qfalse;
if (exclude && tmp == rb_cObject && klass != rb_cObject) {
return (int)Qfalse;
}
return (int)Qtrue;
}
if (!recurse) break;
tmp = RCLASS_SUPER(tmp);
}
if (!exclude && !mod_retry && BUILTIN_TYPE(klass) == T_MODULE) {
mod_retry = 1;
tmp = rb_cObject;
goto retry;
}
return (int)Qfalse;
}
int
rb_const_defined_from(VALUE klass, ID id)
{
return rb_const_defined_0(klass, id, TRUE, TRUE, FALSE);
}
int
rb_const_defined(VALUE klass, ID id)
{
return rb_const_defined_0(klass, id, FALSE, TRUE, FALSE);
}
int
rb_const_defined_at(VALUE klass, ID id)
{
return rb_const_defined_0(klass, id, TRUE, FALSE, FALSE);
}
int
rb_public_const_defined_from(VALUE klass, ID id)
{
return rb_const_defined_0(klass, id, TRUE, TRUE, TRUE);
}
static void
check_before_mod_set(VALUE klass, ID id, VALUE val, const char *dest)
{
rb_check_frozen(klass);
}
static void set_namespace_path(VALUE named_namespace, VALUE name);
static enum rb_id_table_iterator_result
set_namespace_path_i(ID id, VALUE v, void *payload)
{
rb_const_entry_t *ce = (rb_const_entry_t *)v;
VALUE value = ce->value;
VALUE parental_path = *((VALUE *) payload);
if (!rb_is_const_id(id) || !rb_namespace_p(value)) {
return ID_TABLE_CONTINUE;
}
bool has_permanent_classpath;
classname(value, &has_permanent_classpath);
if (has_permanent_classpath) {
return ID_TABLE_CONTINUE;
}
set_namespace_path(value, build_const_path(parental_path, id));
if (!RCLASS_PERMANENT_CLASSPATH_P(value)) {
RCLASS_WRITE_CLASSPATH(value, 0, false);
}
return ID_TABLE_CONTINUE;
}
/*
* Assign permanent classpaths to all namespaces that are directly or indirectly
* nested under +named_namespace+. +named_namespace+ must have a permanent
* classpath.
*/
static void
set_namespace_path(VALUE named_namespace, VALUE namespace_path)
{
struct rb_id_table *const_table = RCLASS_CONST_TBL(named_namespace);
RB_VM_LOCKING() {
RCLASS_WRITE_CLASSPATH(named_namespace, namespace_path, true);
if (const_table) {
rb_id_table_foreach(const_table, set_namespace_path_i, &namespace_path);
}
}
}
static void
const_added(VALUE klass, ID const_name)
{
if (GET_VM()->running) {
VALUE name = ID2SYM(const_name);
rb_funcallv(klass, idConst_added, 1, &name);
}
}
static void
const_set(VALUE klass, ID id, VALUE val)
{
rb_const_entry_t *ce;
if (NIL_P(klass)) {
rb_raise(rb_eTypeError, "no class/module to define constant %"PRIsVALUE"",
QUOTE_ID(id));
}
if (!rb_ractor_main_p() && !rb_ractor_shareable_p(val)) {
rb_raise(rb_eRactorIsolationError, "can not set constants with non-shareable objects by non-main Ractors");
}
check_before_mod_set(klass, id, val, "constant");
RB_VM_LOCKING() {
struct rb_id_table *tbl = RCLASS_WRITABLE_CONST_TBL(klass);
if (!tbl) {
tbl = rb_id_table_create(0);
RCLASS_WRITE_CONST_TBL(klass, tbl, false);
rb_clear_constant_cache_for_id(id);
ce = ZALLOC(rb_const_entry_t);
rb_id_table_insert(tbl, id, (VALUE)ce);
setup_const_entry(ce, klass, val, CONST_PUBLIC);
}
else {
struct autoload_const ac = {
.module = klass, .name = id,
.value = val, .flag = CONST_PUBLIC,
/* fill the rest with 0 */
};
ac.file = rb_source_location(&ac.line);
const_tbl_update(&ac, false);
}
}
/*
* Resolve and cache class name immediately to resolve ambiguity
* and avoid order-dependency on const_tbl
*/
if (rb_cObject && rb_namespace_p(val)) {
bool val_path_permanent;
VALUE val_path = classname(val, &val_path_permanent);
if (NIL_P(val_path) || !val_path_permanent) {
if (klass == rb_cObject) {
set_namespace_path(val, rb_id2str(id));
}
else {
bool parental_path_permanent;
VALUE parental_path = classname(klass, &parental_path_permanent);
if (NIL_P(parental_path)) {
bool throwaway;
parental_path = rb_tmp_class_path(klass, &throwaway, make_temporary_path);
}
if (parental_path_permanent && !val_path_permanent) {
set_namespace_path(val, build_const_path(parental_path, id));
}
else if (!parental_path_permanent && NIL_P(val_path)) {
RCLASS_SET_CLASSPATH(val, build_const_path(parental_path, id), false);
}
}
}
}
}
void
rb_const_set(VALUE klass, ID id, VALUE val)
{
const_set(klass, id, val);
const_added(klass, id);
}
static struct autoload_data *
autoload_data_for_named_constant(VALUE module, ID name, struct autoload_const **autoload_const_pointer)
{
VALUE autoload_data_value = autoload_data(module, name);
if (!autoload_data_value) return 0;
struct autoload_data *autoload_data = get_autoload_data(autoload_data_value, autoload_const_pointer);
if (!autoload_data) return 0;
/* for autoloading thread, keep the defined value to autoloading storage */
if (autoload_by_current(autoload_data)) {
return autoload_data;
}
return 0;
}
static void
const_tbl_update(struct autoload_const *ac, int autoload_force)
{
VALUE value;
VALUE klass = ac->module;
VALUE val = ac->value;
ID id = ac->name;
struct rb_id_table *tbl = RCLASS_CONST_TBL(klass);
rb_const_flag_t visibility = ac->flag;
rb_const_entry_t *ce;
if (rb_id_table_lookup(tbl, id, &value)) {
ce = (rb_const_entry_t *)value;
if (UNDEF_P(ce->value)) {
RUBY_ASSERT_CRITICAL_SECTION_ENTER();
VALUE file = ac->file;
int line = ac->line;
struct autoload_data *ele = autoload_data_for_named_constant(klass, id, &ac);
if (!autoload_force && ele) {
rb_clear_constant_cache_for_id(id);
ac->value = val; /* autoload_data is non-WB-protected */
ac->file = rb_source_location(&ac->line);
}
else {
/* otherwise autoloaded constant, allow to override */
autoload_delete(klass, id);
ce->flag = visibility;
RB_OBJ_WRITE(klass, &ce->value, val);
RB_OBJ_WRITE(klass, &ce->file, file);
ce->line = line;
}
RUBY_ASSERT_CRITICAL_SECTION_LEAVE();
return;
}
else {
VALUE name = QUOTE_ID(id);
visibility = ce->flag;
if (klass == rb_cObject)
rb_warn("already initialized constant %"PRIsVALUE"", name);
else
rb_warn("already initialized constant %"PRIsVALUE"::%"PRIsVALUE"",
rb_class_name(klass), name);
if (!NIL_P(ce->file) && ce->line) {
rb_compile_warn(RSTRING_PTR(ce->file), ce->line,
"previous definition of %"PRIsVALUE" was here", name);
}
}
rb_clear_constant_cache_for_id(id);
setup_const_entry(ce, klass, val, visibility);
}
else {
tbl = RCLASS_WRITABLE_CONST_TBL(klass);
rb_clear_constant_cache_for_id(id);
ce = ZALLOC(rb_const_entry_t);
rb_id_table_insert(tbl, id, (VALUE)ce);
setup_const_entry(ce, klass, val, visibility);
}
}
static void
setup_const_entry(rb_const_entry_t *ce, VALUE klass, VALUE val,
rb_const_flag_t visibility)
{
ce->flag = visibility;
RB_OBJ_WRITE(klass, &ce->value, val);
RB_OBJ_WRITE(klass, &ce->file, rb_source_location(&ce->line));
}
void
rb_define_const(VALUE klass, const char *name, VALUE val)
{
ID id = rb_intern(name);
if (!rb_is_const_id(id)) {
rb_warn("rb_define_const: invalid name '%s' for constant", name);
}
if (!RB_SPECIAL_CONST_P(val)) {
rb_vm_register_global_object(val);
}
rb_const_set(klass, id, val);
}
void
rb_define_global_const(const char *name, VALUE val)
{
rb_define_const(rb_cObject, name, val);
}
static void
set_const_visibility(VALUE mod, int argc, const VALUE *argv,
rb_const_flag_t flag, rb_const_flag_t mask)
{
int i;
rb_const_entry_t *ce;
ID id;
rb_class_modify_check(mod);
if (argc == 0) {
rb_warning("%"PRIsVALUE" with no argument is just ignored",
QUOTE_ID(rb_frame_callee()));
return;
}
for (i = 0; i < argc; i++) {
struct autoload_const *ac;
VALUE val = argv[i];
id = rb_check_id(&val);
if (!id) {
undefined_constant(mod, val);
}
if ((ce = rb_const_lookup(mod, id))) {
ce->flag &= ~mask;
ce->flag |= flag;
if (UNDEF_P(ce->value)) {
struct autoload_data *ele;
ele = autoload_data_for_named_constant(mod, id, &ac);
if (ele) {
ac->flag &= ~mask;
ac->flag |= flag;
}
}
rb_clear_constant_cache_for_id(id);
}
else {
undefined_constant(mod, ID2SYM(id));
}
}
}
void
rb_deprecate_constant(VALUE mod, const char *name)
{
rb_const_entry_t *ce;
ID id;
long len = strlen(name);
rb_class_modify_check(mod);
if (!(id = rb_check_id_cstr(name, len, NULL))) {
undefined_constant(mod, rb_fstring_new(name, len));
}
if (!(ce = rb_const_lookup(mod, id))) {
undefined_constant(mod, ID2SYM(id));
}
ce->flag |= CONST_DEPRECATED;
}
/*
* call-seq:
* mod.private_constant(symbol, ...) => mod
*
* Makes a list of existing constants private.
*/
VALUE
rb_mod_private_constant(int argc, const VALUE *argv, VALUE obj)
{
set_const_visibility(obj, argc, argv, CONST_PRIVATE, CONST_VISIBILITY_MASK);
return obj;
}
/*
* call-seq:
* mod.public_constant(symbol, ...) => mod
*
* Makes a list of existing constants public.
*/
VALUE
rb_mod_public_constant(int argc, const VALUE *argv, VALUE obj)
{
set_const_visibility(obj, argc, argv, CONST_PUBLIC, CONST_VISIBILITY_MASK);
return obj;
}
/*
* call-seq:
* mod.deprecate_constant(symbol, ...) => mod
*
* Makes a list of existing constants deprecated. Attempt
* to refer to them will produce a warning.
*
* module HTTP
* NotFound = Exception.new
* NOT_FOUND = NotFound # previous version of the library used this name
*
* deprecate_constant :NOT_FOUND
* end
*
* HTTP::NOT_FOUND
* # warning: constant HTTP::NOT_FOUND is deprecated
*
*/
VALUE
rb_mod_deprecate_constant(int argc, const VALUE *argv, VALUE obj)
{
set_const_visibility(obj, argc, argv, CONST_DEPRECATED, CONST_DEPRECATED);
return obj;
}
static VALUE
original_module(VALUE c)
{
if (RB_TYPE_P(c, T_ICLASS))
return RBASIC(c)->klass;
return c;
}
static int
cvar_lookup_at(VALUE klass, ID id, st_data_t *v)
{
if (RB_TYPE_P(klass, T_ICLASS)) {
if (RICLASS_IS_ORIGIN_P(klass)) {
return 0;
}
else {
// check the original module
klass = RBASIC(klass)->klass;
}
}
VALUE n = rb_ivar_lookup(klass, id, Qundef);
if (UNDEF_P(n)) return 0;
if (v) *v = n;
return 1;
}
static VALUE
cvar_front_klass(VALUE klass)
{
if (RCLASS_SINGLETON_P(klass)) {
VALUE obj = RCLASS_ATTACHED_OBJECT(klass);
if (rb_namespace_p(obj)) {
return obj;
}
}
return RCLASS_SUPER(klass);
}
static void
cvar_overtaken(VALUE front, VALUE target, ID id)
{
if (front && target != front) {
if (original_module(front) != original_module(target)) {
rb_raise(rb_eRuntimeError,
"class variable % "PRIsVALUE" of %"PRIsVALUE" is overtaken by %"PRIsVALUE"",
ID2SYM(id), rb_class_name(original_module(front)),
rb_class_name(original_module(target)));
}
if (BUILTIN_TYPE(front) == T_CLASS) {
rb_ivar_delete(front, id, Qundef);
}
}
}
#define CVAR_FOREACH_ANCESTORS(klass, v, r) \
for (klass = cvar_front_klass(klass); klass; klass = RCLASS_SUPER(klass)) { \
if (cvar_lookup_at(klass, id, (v))) { \
r; \
} \
}
#define CVAR_LOOKUP(v,r) do {\
CVAR_ACCESSOR_SHOULD_BE_MAIN_RACTOR(); \
if (cvar_lookup_at(klass, id, (v))) {r;}\
CVAR_FOREACH_ANCESTORS(klass, v, r);\
} while(0)
static VALUE
find_cvar(VALUE klass, VALUE * front, VALUE * target, ID id)
{
VALUE v = Qundef;
CVAR_LOOKUP(&v, {
if (!*front) {
*front = klass;
}
*target = klass;
});
return v;
}
static void
check_for_cvar_table(VALUE subclass, VALUE key)
{
// Must not check ivar on ICLASS
if (!RB_TYPE_P(subclass, T_ICLASS) && RTEST(rb_ivar_defined(subclass, key))) {
RB_DEBUG_COUNTER_INC(cvar_class_invalidate);
ruby_vm_global_cvar_state++;
return;
}
rb_class_foreach_subclass(subclass, check_for_cvar_table, key);
}
void
rb_cvar_set(VALUE klass, ID id, VALUE val)
{
VALUE tmp, front = 0, target = 0;
tmp = klass;
CVAR_LOOKUP(0, {if (!front) front = klass; target = klass;});
if (target) {
cvar_overtaken(front, target, id);
}
else {
target = tmp;
}
if (RB_TYPE_P(target, T_ICLASS)) {
target = RBASIC(target)->klass;
}
check_before_mod_set(target, id, val, "class variable");
int result = rb_class_ivar_set(target, id, val);
struct rb_id_table *rb_cvc_tbl = RCLASS_WRITABLE_CVC_TBL(target);
if (!rb_cvc_tbl) {
rb_cvc_tbl = rb_id_table_create(2);
RCLASS_WRITE_CVC_TBL(target, rb_cvc_tbl);
}
struct rb_cvar_class_tbl_entry *ent;
VALUE ent_data;
if (!rb_id_table_lookup(rb_cvc_tbl, id, &ent_data)) {
ent = ALLOC(struct rb_cvar_class_tbl_entry);
ent->class_value = target;
ent->global_cvar_state = GET_GLOBAL_CVAR_STATE();
ent->cref = 0;
rb_id_table_insert(rb_cvc_tbl, id, (VALUE)ent);
RB_DEBUG_COUNTER_INC(cvar_inline_miss);
}
else {
ent = (void *)ent_data;
ent->global_cvar_state = GET_GLOBAL_CVAR_STATE();
}
// Break the cvar cache if this is a new class variable
// and target is a module or a subclass with the same
// cvar in this lookup.
if (result == 0) {
if (RB_TYPE_P(target, T_CLASS)) {
if (RCLASS_SUBCLASSES_FIRST(target)) {
rb_class_foreach_subclass(target, check_for_cvar_table, id);
}
}
}
}
VALUE
rb_cvar_find(VALUE klass, ID id, VALUE *front)
{
VALUE target = 0;
VALUE value;
value = find_cvar(klass, front, &target, id);
if (!target) {
rb_name_err_raise("uninitialized class variable %1$s in %2$s",
klass, ID2SYM(id));
}
cvar_overtaken(*front, target, id);
return (VALUE)value;
}
VALUE
rb_cvar_get(VALUE klass, ID id)
{
VALUE front = 0;
return rb_cvar_find(klass, id, &front);
}
VALUE
rb_cvar_defined(VALUE klass, ID id)
{
if (!klass) return Qfalse;
CVAR_LOOKUP(0,return Qtrue);
return Qfalse;
}
static ID
cv_intern(VALUE klass, const char *name)
{
ID id = rb_intern(name);
if (!rb_is_class_id(id)) {
rb_name_err_raise("wrong class variable name %1$s",
klass, rb_str_new_cstr(name));
}
return id;
}
void
rb_cv_set(VALUE klass, const char *name, VALUE val)
{
ID id = cv_intern(klass, name);
rb_cvar_set(klass, id, val);
}
VALUE
rb_cv_get(VALUE klass, const char *name)
{
ID id = cv_intern(klass, name);
return rb_cvar_get(klass, id);
}
void
rb_define_class_variable(VALUE klass, const char *name, VALUE val)
{
rb_cv_set(klass, name, val);
}
static int
cv_i(ID key, VALUE v, st_data_t a)
{
st_table *tbl = (st_table *)a;
if (rb_is_class_id(key)) {
st_update(tbl, (st_data_t)key, cv_i_update, 0);
}
return ST_CONTINUE;
}
static void*
mod_cvar_at(VALUE mod, void *data)
{
st_table *tbl = data;
if (!tbl) {
tbl = st_init_numtable();
}
mod = original_module(mod);
rb_ivar_foreach(mod, cv_i, (st_data_t)tbl);
return tbl;
}
static void*
mod_cvar_of(VALUE mod, void *data)
{
VALUE tmp = mod;
if (RCLASS_SINGLETON_P(mod)) {
if (rb_namespace_p(RCLASS_ATTACHED_OBJECT(mod))) {
data = mod_cvar_at(tmp, data);
tmp = cvar_front_klass(tmp);
}
}
for (;;) {
data = mod_cvar_at(tmp, data);
tmp = RCLASS_SUPER(tmp);
if (!tmp) break;
}
return data;
}
static int
cv_list_i(st_data_t key, st_data_t value, VALUE ary)
{
ID sym = (ID)key;
rb_ary_push(ary, ID2SYM(sym));
return ST_CONTINUE;
}
static VALUE
cvar_list(void *data)
{
st_table *tbl = data;
VALUE ary;
if (!tbl) return rb_ary_new2(0);
ary = rb_ary_new2(tbl->num_entries);
st_foreach_safe(tbl, cv_list_i, ary);
st_free_table(tbl);
return ary;
}
/*
* call-seq:
* mod.class_variables(inherit=true) -> array
*
* Returns an array of the names of class variables in <i>mod</i>.
* This includes the names of class variables in any included
* modules, unless the <i>inherit</i> parameter is set to
* <code>false</code>.
*
* class One
* @@var1 = 1
* end
* class Two < One
* @@var2 = 2
* end
* One.class_variables #=> [:@@var1]
* Two.class_variables #=> [:@@var2, :@@var1]
* Two.class_variables(false) #=> [:@@var2]
*/
VALUE
rb_mod_class_variables(int argc, const VALUE *argv, VALUE mod)
{
bool inherit = true;
st_table *tbl;
if (rb_check_arity(argc, 0, 1)) inherit = RTEST(argv[0]);
if (inherit) {
tbl = mod_cvar_of(mod, 0);
}
else {
tbl = mod_cvar_at(mod, 0);
}
return cvar_list(tbl);
}
/*
* call-seq:
* remove_class_variable(sym) -> obj
*
* Removes the named class variable from the receiver, returning that
* variable's value.
*
* class Example
* @@var = 99
* puts remove_class_variable(:@@var)
* p(defined? @@var)
* end
*
* <em>produces:</em>
*
* 99
* nil
*/
VALUE
rb_mod_remove_cvar(VALUE mod, VALUE name)
{
const ID id = id_for_var_message(mod, name, class, "wrong class variable name %1$s");
st_data_t val;
if (!id) {
goto not_defined;
}
rb_check_frozen(mod);
val = rb_ivar_delete(mod, id, Qundef);
if (!UNDEF_P(val)) {
return (VALUE)val;
}
if (rb_cvar_defined(mod, id)) {
rb_name_err_raise("cannot remove %1$s for %2$s", mod, ID2SYM(id));
}
not_defined:
rb_name_err_raise("class variable %1$s not defined for %2$s",
mod, name);
UNREACHABLE_RETURN(Qundef);
}
VALUE
rb_iv_get(VALUE obj, const char *name)
{
ID id = rb_check_id_cstr(name, strlen(name), rb_usascii_encoding());
if (!id) {
return Qnil;
}
return rb_ivar_get(obj, id);
}
VALUE
rb_iv_set(VALUE obj, const char *name, VALUE val)
{
ID id = rb_intern(name);
return rb_ivar_set(obj, id, val);
}
static int
class_ivar_set(VALUE obj, ID id, VALUE val)
{
bool existing = true;
const VALUE original_fields_obj = RCLASS_FIELDS_OBJ(obj);
VALUE fields_obj = original_fields_obj ? original_fields_obj : rb_imemo_class_fields_new(obj, 1);
shape_id_t next_shape_id = 0;
shape_id_t current_shape_id = RBASIC_SHAPE_ID(fields_obj);
if (UNLIKELY(rb_shape_too_complex_p(current_shape_id))) {
goto too_complex;
}
attr_index_t index;
if (!rb_shape_get_iv_index(current_shape_id, id, &index)) {
existing = false;
index = RSHAPE_LEN(current_shape_id);
if (index >= SHAPE_MAX_FIELDS) {
rb_raise(rb_eArgError, "too many instance variables");
}
next_shape_id = rb_shape_transition_add_ivar(fields_obj, id);
if (UNLIKELY(rb_shape_too_complex_p(next_shape_id))) {
attr_index_t current_len = RSHAPE_LEN(current_shape_id);
fields_obj = rb_imemo_class_fields_new_complex(obj, current_len + 1);
if (current_len) {
rb_obj_copy_fields_to_hash_table(original_fields_obj, rb_imemo_class_fields_complex_tbl(fields_obj));
RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id);
}
goto too_complex;
}
attr_index_t next_capacity = RSHAPE_CAPACITY(next_shape_id);
attr_index_t current_capacity = RSHAPE_CAPACITY(current_shape_id);
if (UNLIKELY(next_capacity != current_capacity)) {
RUBY_ASSERT(next_capacity > current_capacity);
// We allocate a new fields_obj so that we're embedded as long as possible
fields_obj = rb_imemo_class_fields_new(obj, next_capacity);
if (original_fields_obj) {
MEMCPY(rb_imemo_class_fields_ptr(fields_obj), rb_imemo_class_fields_ptr(original_fields_obj), VALUE, RSHAPE_LEN(current_shape_id));
}
}
RUBY_ASSERT(RSHAPE(next_shape_id)->type == SHAPE_IVAR);
RUBY_ASSERT(index == (RSHAPE_LEN(next_shape_id) - 1));
}
VALUE *fields = rb_imemo_class_fields_ptr(fields_obj);
RB_OBJ_WRITE(fields_obj, &fields[index], val);
if (!existing) {
RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id);
}
if (fields_obj != original_fields_obj) {
RCLASS_SET_FIELDS_OBJ(obj, fields_obj);
// TODO: What should we set as the T_CLASS shape_id?
// In most case we can replicate the single `fields_obj` shape
// but in namespaced case?
// Perhaps INVALID_SHAPE_ID?
RBASIC_SET_SHAPE_ID(obj, next_shape_id);
}
RB_GC_GUARD(fields_obj);
return existing;
too_complex:
{
st_table *table = rb_imemo_class_fields_complex_tbl(fields_obj);
existing = st_insert(table, (st_data_t)id, (st_data_t)val);
RB_OBJ_WRITTEN(fields_obj, Qundef, val);
if (fields_obj != original_fields_obj) {
RBASIC_SET_SHAPE_ID(fields_obj, next_shape_id);
RCLASS_SET_FIELDS_OBJ(obj, fields_obj);
// TODO: What should we set as the T_CLASS shape_id?
// In most case we can replicate the single `fields_obj` shape
// but in namespaced case?
// Perhaps INVALID_SHAPE_ID?
RBASIC_SET_SHAPE_ID(obj, next_shape_id);
}
}
RB_GC_GUARD(fields_obj);
return existing;
}
int
rb_class_ivar_set(VALUE obj, ID id, VALUE val)
{
RUBY_ASSERT(RB_TYPE_P(obj, T_CLASS) || RB_TYPE_P(obj, T_MODULE));
bool existing = false;
rb_check_frozen(obj);
rb_class_ensure_writable(obj);
RB_VM_LOCKING() {
existing = class_ivar_set(obj, id, val);
}
return existing;
}
static void
class_field_set(VALUE obj, shape_id_t target_shape_id, VALUE val)
{
RUBY_ASSERT(RB_TYPE_P(obj, T_CLASS) || RB_TYPE_P(obj, T_MODULE));
obj_field_set(RCLASS_ENSURE_FIELDS_OBJ(obj), target_shape_id, val);
}
static int
tbl_copy_i(ID key, VALUE val, st_data_t dest)
{
rb_class_ivar_set((VALUE)dest, key, val);
return ST_CONTINUE;
}
void
rb_fields_tbl_copy(VALUE dst, VALUE src)
{
RUBY_ASSERT(rb_type(dst) == rb_type(src));
RUBY_ASSERT(RB_TYPE_P(dst, T_CLASS) || RB_TYPE_P(dst, T_MODULE));
RUBY_ASSERT(RSHAPE_TYPE_P(RBASIC_SHAPE_ID(dst), SHAPE_ROOT));
rb_ivar_foreach(src, tbl_copy_i, dst);
}
static rb_const_entry_t *
const_lookup(struct rb_id_table *tbl, ID id)
{
if (tbl) {
VALUE val;
bool r;
RB_VM_LOCKING() {
r = rb_id_table_lookup(tbl, id, &val);
}
if (r) return (rb_const_entry_t *)val;
}
return NULL;
}
rb_const_entry_t *
rb_const_lookup(VALUE klass, ID id)
{
return const_lookup(RCLASS_CONST_TBL(klass), id);
}