summaryrefslogtreecommitdiff
path: root/rjit.c
diff options
context:
space:
mode:
Diffstat (limited to 'rjit.c')
-rw-r--r--rjit.c501
1 files changed, 501 insertions, 0 deletions
diff --git a/rjit.c b/rjit.c
new file mode 100644
index 0000000000..72660394b3
--- /dev/null
+++ b/rjit.c
@@ -0,0 +1,501 @@
+/**********************************************************************
+
+ rjit.c - Ruby JIT compiler functions
+
+ Copyright (C) 2023 Takashi Kokubun <k0kubun@ruby-lang.org>.
+
+**********************************************************************/
+
+#include "rjit.h" // defines USE_RJIT
+
+#if USE_RJIT
+
+#include "constant.h"
+#include "id_table.h"
+#include "internal.h"
+#include "internal/class.h"
+#include "internal/cmdlineopt.h"
+#include "internal/cont.h"
+#include "internal/file.h"
+#include "internal/hash.h"
+#include "internal/process.h"
+#include "internal/warnings.h"
+#include "vm_sync.h"
+#include "ractor_core.h"
+
+#ifdef __sun
+#define __EXTENSIONS__ 1
+#endif
+
+#include "vm_core.h"
+#include "vm_callinfo.h"
+#include "rjit_c.h"
+#include "ruby_assert.h"
+#include "ruby/debug.h"
+#include "ruby/thread.h"
+#include "ruby/version.h"
+#include "builtin.h"
+#include "insns.inc"
+#include "insns_info.inc"
+#include "internal/compile.h"
+#include "internal/gc.h"
+
+#include <sys/wait.h>
+#include <sys/time.h>
+#include <dlfcn.h>
+#include <errno.h>
+#ifdef HAVE_FCNTL_H
+#include <fcntl.h>
+#endif
+#ifdef HAVE_SYS_PARAM_H
+# include <sys/param.h>
+#endif
+#include "dln.h"
+
+// For mmapp(), sysconf()
+#ifndef _WIN32
+#include <unistd.h>
+#include <sys/mman.h>
+#endif
+
+#include "ruby/util.h"
+
+// A copy of RJIT portion of MRI options since RJIT initialization. We
+// need them as RJIT threads still can work when the most MRI data were
+// freed.
+struct rb_rjit_options rb_rjit_opts;
+
+// true if RJIT is enabled.
+bool rb_rjit_enabled = false;
+// true if --rjit-stats (used before rb_rjit_opts is set)
+bool rb_rjit_stats_enabled = false;
+// true if --rjit-trace-exits (used before rb_rjit_opts is set)
+bool rb_rjit_trace_exits_enabled = false;
+// true if JIT-ed code should be called. When `ruby_vm_event_enabled_global_flags & ISEQ_TRACE_EVENTS`
+// and `rb_rjit_call_p == false`, any JIT-ed code execution is cancelled as soon as possible.
+bool rb_rjit_call_p = false;
+// A flag to communicate that rb_rjit_call_p should be disabled while it's temporarily false.
+static bool rjit_cancel_p = false;
+
+// `rb_ec_ractor_hooks(ec)->events` is moved to this variable during compilation.
+rb_event_flag_t rb_rjit_global_events = 0;
+
+// Basically rb_rjit_opts.stats, but this becomes false during RJIT compilation.
+static bool rjit_stats_p = false;
+
+// RubyVM::RJIT
+static VALUE rb_mRJIT = 0;
+// RubyVM::RJIT::C
+static VALUE rb_mRJITC = 0;
+// RubyVM::RJIT::Compiler
+static VALUE rb_RJITCompiler = 0;
+// RubyVM::RJIT::CPointer::Struct_rb_iseq_t
+static VALUE rb_cRJITIseqPtr = 0;
+// RubyVM::RJIT::CPointer::Struct_rb_control_frame_t
+static VALUE rb_cRJITCfpPtr = 0;
+// RubyVM::RJIT::Hooks
+static VALUE rb_mRJITHooks = 0;
+
+// Frames for --rjit-trace-exits
+VALUE rb_rjit_raw_samples = 0;
+// Line numbers for --rjit-trace-exits
+VALUE rb_rjit_line_samples = 0;
+
+// Postponed job handle for triggering rjit_iseq_update_references
+static rb_postponed_job_handle_t rjit_iseq_update_references_pjob;
+
+// A default threshold used to add iseq to JIT.
+#define DEFAULT_CALL_THRESHOLD 10
+// Size of executable memory block in MiB.
+#define DEFAULT_EXEC_MEM_SIZE 64
+
+#define opt_match_noarg(s, l, name) \
+ opt_match(s, l, name) && (*(s) ? (rb_warn("argument to --rjit-" name " is ignored"), 1) : 1)
+#define opt_match_arg(s, l, name) \
+ opt_match(s, l, name) && (*(s) ? 1 : (rb_raise(rb_eRuntimeError, "--rjit-" name " needs an argument"), 0))
+
+void
+rb_rjit_setup_options(const char *s, struct rb_rjit_options *rjit_opt)
+{
+ const size_t l = strlen(s);
+ if (l == 0) {
+ return;
+ }
+ else if (opt_match_arg(s, l, "exec-mem-size")) {
+ rjit_opt->exec_mem_size = atoi(s + 1);
+ }
+ else if (opt_match_arg(s, l, "call-threshold")) {
+ rjit_opt->call_threshold = atoi(s + 1);
+ }
+ else if (opt_match_noarg(s, l, "stats")) {
+ rjit_opt->stats = true;
+ }
+ else if (opt_match_noarg(s, l, "disable")) {
+ rjit_opt->disable = true;
+ }
+ else if (opt_match_noarg(s, l, "trace")) {
+ rjit_opt->trace = true;
+ }
+ else if (opt_match_noarg(s, l, "trace-exits")) {
+ rjit_opt->trace_exits = true;
+ }
+ else if (opt_match_noarg(s, l, "dump-disasm")) {
+ rjit_opt->dump_disasm = true;
+ }
+ else if (opt_match_noarg(s, l, "verify-ctx")) {
+ rjit_opt->verify_ctx = true;
+ }
+ else {
+ rb_raise(rb_eRuntimeError,
+ "invalid RJIT option '%s' (--help will show valid RJIT options)", s);
+ }
+}
+
+#define M(shortopt, longopt, desc) RUBY_OPT_MESSAGE(shortopt, longopt, desc)
+const struct ruby_opt_message rb_rjit_option_messages[] = {
+ M("--rjit-exec-mem-size=num", "", "Size of executable memory block in MiB (default: " STRINGIZE(DEFAULT_EXEC_MEM_SIZE) ")."),
+ M("--rjit-call-threshold=num", "", "Number of calls to trigger JIT (default: " STRINGIZE(DEFAULT_CALL_THRESHOLD) ")."),
+ M("--rjit-stats", "", "Enable collecting RJIT statistics."),
+ M("--rjit-disable", "", "Disable RJIT for lazily enabling it with RubyVM::RJIT.enable."),
+ M("--rjit-trace", "", "Allow TracePoint during JIT compilation."),
+ M("--rjit-trace-exits", "", "Trace side exit locations."),
+#ifdef HAVE_LIBCAPSTONE
+ M("--rjit-dump-disasm", "", "Dump all JIT code"),
+#endif
+ {0}
+};
+#undef M
+
+struct rb_rjit_runtime_counters rb_rjit_counters = { 0 };
+
+extern VALUE rb_gc_enable(void);
+extern VALUE rb_gc_disable(void);
+extern uint64_t rb_vm_insns_count;
+
+// Disable GC, TracePoint, JIT, stats, and $!
+#define WITH_RJIT_ISOLATED_USING_PC(using_pc, stmt) do { \
+ VALUE was_disabled = rb_gc_disable(); \
+ \
+ rb_hook_list_t *global_hooks = rb_ec_ractor_hooks(GET_EC()); \
+ rb_rjit_global_events = global_hooks->events; \
+ \
+ const VALUE *pc = NULL; \
+ if (rb_rjit_opts.trace) { \
+ pc = GET_EC()->cfp->pc; \
+ if (!using_pc) GET_EC()->cfp->pc = 0; /* avoid crashing on calc_lineno */ \
+ } \
+ else global_hooks->events = 0; \
+ \
+ bool original_call_p = rb_rjit_call_p; \
+ rb_rjit_call_p = false; \
+ \
+ rjit_stats_p = false; \
+ uint64_t insns_count = rb_vm_insns_count; \
+ \
+ VALUE err = rb_errinfo(); \
+ \
+ stmt; \
+ \
+ rb_set_errinfo(err); \
+ \
+ rb_vm_insns_count = insns_count; \
+ rjit_stats_p = rb_rjit_opts.stats; \
+ \
+ rb_rjit_call_p = (rjit_cancel_p ? false : original_call_p); \
+ \
+ if (rb_rjit_opts.trace) GET_EC()->cfp->pc = pc; \
+ else global_hooks->events = rb_rjit_global_events; \
+ \
+ if (!was_disabled) rb_gc_enable(); \
+} while (0);
+#define WITH_RJIT_ISOLATED(stmt) WITH_RJIT_ISOLATED_USING_PC(false, stmt)
+
+void
+rb_rjit_cancel_all(const char *reason)
+{
+ if (!rb_rjit_enabled)
+ return;
+
+ rb_rjit_call_p = false;
+ rjit_cancel_p = true;
+}
+
+void
+rb_rjit_bop_redefined(int redefined_flag, enum ruby_basic_operators bop)
+{
+ if (!rb_rjit_call_p) return;
+ rb_rjit_call_p = false;
+}
+
+static void
+rjit_cme_invalidate(void *data)
+{
+ if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
+ WITH_RJIT_ISOLATED({
+ rb_funcall(rb_mRJITHooks, rb_intern("on_cme_invalidate"), 1, SIZET2NUM((size_t)data));
+ });
+}
+
+extern int rb_workqueue_register(unsigned flags, rb_postponed_job_func_t func, void *data);
+
+void
+rb_rjit_cme_invalidate(rb_callable_method_entry_t *cme)
+{
+ if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
+ // Asynchronously hook the Ruby code since running Ruby in the middle of cme invalidation is dangerous.
+ rb_workqueue_register(0, rjit_cme_invalidate, (void *)cme);
+}
+
+void
+rb_rjit_before_ractor_spawn(void)
+{
+ if (!rb_rjit_call_p) return;
+ rb_rjit_call_p = false;
+}
+
+static void
+rjit_constant_state_changed(void *data)
+{
+ if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
+ RB_VM_LOCK_ENTER();
+ rb_vm_barrier();
+
+ WITH_RJIT_ISOLATED({
+ rb_funcall(rb_mRJITHooks, rb_intern("on_constant_state_changed"), 1, SIZET2NUM((size_t)data));
+ });
+
+ RB_VM_LOCK_LEAVE();
+}
+
+void
+rb_rjit_constant_state_changed(ID id)
+{
+ if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
+ // Asynchronously hook the Ruby code since this is hooked during a "Ruby critical section".
+ rb_workqueue_register(0, rjit_constant_state_changed, (void *)id);
+}
+
+void
+rb_rjit_constant_ic_update(const rb_iseq_t *const iseq, IC ic, unsigned insn_idx)
+{
+ if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
+
+ RB_VM_LOCK_ENTER();
+ rb_vm_barrier();
+
+ WITH_RJIT_ISOLATED({
+ rb_funcall(rb_mRJITHooks, rb_intern("on_constant_ic_update"), 3,
+ SIZET2NUM((size_t)iseq), SIZET2NUM((size_t)ic), UINT2NUM(insn_idx));
+ });
+
+ RB_VM_LOCK_LEAVE();
+}
+
+void
+rb_rjit_tracing_invalidate_all(rb_event_flag_t new_iseq_events)
+{
+ if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
+ WITH_RJIT_ISOLATED({
+ rb_funcall(rb_mRJITHooks, rb_intern("on_tracing_invalidate_all"), 1, UINT2NUM(new_iseq_events));
+ });
+}
+
+static void
+rjit_iseq_update_references(void *data)
+{
+ if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
+ WITH_RJIT_ISOLATED({
+ rb_funcall(rb_mRJITHooks, rb_intern("on_update_references"), 0);
+ });
+}
+
+void
+rb_rjit_iseq_update_references(struct rb_iseq_constant_body *const body)
+{
+ if (!rb_rjit_enabled) return;
+
+ if (body->rjit_blocks) {
+ body->rjit_blocks = rb_gc_location(body->rjit_blocks);
+ }
+
+ // Asynchronously hook the Ruby code to avoid allocation during GC.compact.
+ // Using _one because it's too slow to invalidate all for each ISEQ. Thus
+ // not giving an ISEQ pointer.
+ rb_postponed_job_trigger(rjit_iseq_update_references_pjob);
+}
+
+void
+rb_rjit_iseq_mark(VALUE rjit_blocks)
+{
+ if (!rb_rjit_enabled) return;
+
+ // Note: This wasn't enough for some reason.
+ // We actually rely on RubyVM::RJIT::GC_REFS to mark this.
+ if (rjit_blocks) {
+ rb_gc_mark_movable(rjit_blocks);
+ }
+}
+
+// Called by rb_vm_mark()
+void
+rb_rjit_mark(void)
+{
+ if (!rb_rjit_enabled)
+ return;
+ RUBY_MARK_ENTER("rjit");
+
+ // Pin object pointers used in this file
+ rb_gc_mark(rb_RJITCompiler);
+ rb_gc_mark(rb_cRJITIseqPtr);
+ rb_gc_mark(rb_cRJITCfpPtr);
+ rb_gc_mark(rb_mRJITHooks);
+ rb_gc_mark(rb_rjit_raw_samples);
+ rb_gc_mark(rb_rjit_line_samples);
+
+ RUBY_MARK_LEAVE("rjit");
+}
+
+void
+rb_rjit_free_iseq(const rb_iseq_t *iseq)
+{
+ // TODO: implement this. GC_REFS should remove this iseq's mjit_blocks
+}
+
+// TODO: Use this in more places
+VALUE
+rb_rjit_iseq_new(rb_iseq_t *iseq)
+{
+ return rb_funcall(rb_cRJITIseqPtr, rb_intern("new"), 1, SIZET2NUM((size_t)iseq));
+}
+
+void
+rb_rjit_compile(const rb_iseq_t *iseq)
+{
+ RB_VM_LOCK_ENTER();
+ rb_vm_barrier();
+
+ WITH_RJIT_ISOLATED_USING_PC(true, {
+ VALUE iseq_ptr = rb_funcall(rb_cRJITIseqPtr, rb_intern("new"), 1, SIZET2NUM((size_t)iseq));
+ VALUE cfp_ptr = rb_funcall(rb_cRJITCfpPtr, rb_intern("new"), 1, SIZET2NUM((size_t)GET_EC()->cfp));
+ rb_funcall(rb_RJITCompiler, rb_intern("compile"), 2, iseq_ptr, cfp_ptr);
+ });
+
+ RB_VM_LOCK_LEAVE();
+}
+
+void *
+rb_rjit_entry_stub_hit(VALUE branch_stub)
+{
+ VALUE result;
+
+ RB_VM_LOCK_ENTER();
+ rb_vm_barrier();
+
+ rb_control_frame_t *cfp = GET_EC()->cfp;
+
+ WITH_RJIT_ISOLATED_USING_PC(true, {
+ VALUE cfp_ptr = rb_funcall(rb_cRJITCfpPtr, rb_intern("new"), 1, SIZET2NUM((size_t)cfp));
+ result = rb_funcall(rb_RJITCompiler, rb_intern("entry_stub_hit"), 2, branch_stub, cfp_ptr);
+ });
+
+ RB_VM_LOCK_LEAVE();
+
+ return (void *)NUM2SIZET(result);
+}
+
+void *
+rb_rjit_branch_stub_hit(VALUE branch_stub, int sp_offset, int target0_p)
+{
+ VALUE result;
+
+ RB_VM_LOCK_ENTER();
+ rb_vm_barrier();
+
+ rb_control_frame_t *cfp = GET_EC()->cfp;
+ cfp->sp += sp_offset; // preserve stack values, also using the actual sp_offset to make jit.peek_at_stack work
+
+ WITH_RJIT_ISOLATED({
+ VALUE cfp_ptr = rb_funcall(rb_cRJITCfpPtr, rb_intern("new"), 1, SIZET2NUM((size_t)cfp));
+ result = rb_funcall(rb_RJITCompiler, rb_intern("branch_stub_hit"), 3, branch_stub, cfp_ptr, RBOOL(target0_p));
+ });
+
+ cfp->sp -= sp_offset; // reset for consistency with the code without the stub
+
+ RB_VM_LOCK_LEAVE();
+
+ return (void *)NUM2SIZET(result);
+}
+
+void
+rb_rjit_init(const struct rb_rjit_options *opts)
+{
+ VM_ASSERT(rb_rjit_enabled);
+
+ // Normalize options
+ rb_rjit_opts = *opts;
+ if (rb_rjit_opts.exec_mem_size == 0)
+ rb_rjit_opts.exec_mem_size = DEFAULT_EXEC_MEM_SIZE;
+ if (rb_rjit_opts.call_threshold == 0)
+ rb_rjit_opts.call_threshold = DEFAULT_CALL_THRESHOLD;
+#ifndef HAVE_LIBCAPSTONE
+ if (rb_rjit_opts.dump_disasm)
+ rb_warn("libcapstone has not been linked. Ignoring --rjit-dump-disasm.");
+#endif
+
+ // RJIT doesn't support miniruby, but it might reach here by RJIT_FORCE_ENABLE.
+ rb_mRJIT = rb_const_get(rb_cRubyVM, rb_intern("RJIT"));
+ if (!rb_const_defined(rb_mRJIT, rb_intern("Compiler"))) {
+ rb_warn("Disabling RJIT because RubyVM::RJIT::Compiler is not defined");
+ rb_rjit_enabled = false;
+ return;
+ }
+ rjit_iseq_update_references_pjob = rb_postponed_job_preregister(0, rjit_iseq_update_references, NULL);
+ if (rjit_iseq_update_references_pjob == POSTPONED_JOB_HANDLE_INVALID) {
+ rb_bug("Could not preregister postponed job for RJIT");
+ }
+ rb_mRJITC = rb_const_get(rb_mRJIT, rb_intern("C"));
+ VALUE rb_cRJITCompiler = rb_const_get(rb_mRJIT, rb_intern("Compiler"));
+ rb_RJITCompiler = rb_funcall(rb_cRJITCompiler, rb_intern("new"), 0);
+ rb_cRJITIseqPtr = rb_funcall(rb_mRJITC, rb_intern("rb_iseq_t"), 0);
+ rb_cRJITCfpPtr = rb_funcall(rb_mRJITC, rb_intern("rb_control_frame_t"), 0);
+ rb_mRJITHooks = rb_const_get(rb_mRJIT, rb_intern("Hooks"));
+ if (rb_rjit_opts.trace_exits) {
+ rb_rjit_raw_samples = rb_ary_new();
+ rb_rjit_line_samples = rb_ary_new();
+ }
+
+ // Enable RJIT and stats from here
+ rb_rjit_call_p = !rb_rjit_opts.disable;
+ rjit_stats_p = rb_rjit_opts.stats;
+}
+
+//
+// Primitive for rjit.rb
+//
+
+// Same as `rb_rjit_opts.stats`, but this is used before rb_rjit_opts is set.
+static VALUE
+rjit_stats_enabled_p(rb_execution_context_t *ec, VALUE self)
+{
+ return RBOOL(rb_rjit_stats_enabled);
+}
+
+// Same as `rb_rjit_opts.trace_exits`, but this is used before rb_rjit_opts is set.
+static VALUE
+rjit_trace_exits_enabled_p(rb_execution_context_t *ec, VALUE self)
+{
+ return RBOOL(rb_rjit_trace_exits_enabled);
+}
+
+// Disable anything that could impact stats. It ends up disabling JIT calls as well.
+static VALUE
+rjit_stop_stats(rb_execution_context_t *ec, VALUE self)
+{
+ rb_rjit_call_p = false;
+ rjit_stats_p = false;
+ return Qnil;
+}
+
+#include "rjit.rbinc"
+
+#endif // USE_RJIT