summaryrefslogtreecommitdiff
path: root/yjit
diff options
context:
space:
mode:
authorAlan Wu <XrXr@users.noreply.github.com>2022-12-05 17:43:17 -0500
committerGitHub <noreply@github.com>2022-12-05 17:43:17 -0500
commit235fc50447c13b054d8849ef2fcdac55ef1f5f9c (patch)
treed04b088ab0c8cc13c63eb691b86a25c2ec3d494f /yjit
parente4aba8f519ca0b63fb3da11d6573443545aaef9b (diff)
YJIT: Remove --yjit-code-page-size (#6865)
Certain code page sizes don't work and can cause crashes, so having this value available as a command-line option is a bit dangerous. Remove it and turn it into a constant instead.
Notes
Notes: Merged-By: maximecb <maximecb@ruby-lang.org>
Diffstat (limited to 'yjit')
-rw-r--r--yjit/src/asm/mod.rs59
-rw-r--r--yjit/src/codegen.rs9
-rw-r--r--yjit/src/options.rs20
3 files changed, 33 insertions, 55 deletions
diff --git a/yjit/src/asm/mod.rs b/yjit/src/asm/mod.rs
index 47e2dca137..44ae74d470 100644
--- a/yjit/src/asm/mod.rs
+++ b/yjit/src/asm/mod.rs
@@ -24,6 +24,9 @@ pub mod x86_64;
pub mod arm64;
+/// Size of a code page in bytes. Each code page is split into an inlined and an outlined portion.
+const CODE_PAGE_SIZE: usize = 16 * 1024;
+
//
// TODO: need a field_size_of macro, to compute the size of a struct field in bytes
//
@@ -57,9 +60,6 @@ pub struct CodeBlock {
// Current writing position
write_pos: usize,
- // Size of a code page (inlined + outlined)
- page_size: usize,
-
// Size reserved for writing a jump to the next page
page_end_reserve: usize,
@@ -94,13 +94,12 @@ pub struct LabelState {
impl CodeBlock {
/// Make a new CodeBlock
- pub fn new(mem_block: Rc<RefCell<VirtualMem>>, page_size: usize, outlined: bool) -> Self {
+ pub fn new(mem_block: Rc<RefCell<VirtualMem>>, outlined: bool) -> Self {
let mem_size = mem_block.borrow().virtual_region_size();
let mut cb = Self {
mem_block,
mem_size,
write_pos: 0,
- page_size,
page_end_reserve: JMP_PTR_BYTES,
label_addrs: Vec::new(),
label_names: Vec::new(),
@@ -122,10 +121,10 @@ impl CodeBlock {
// Use the freed_pages list if code GC has been used. Otherwise use the next page.
let next_page_idx = if let Some(freed_pages) = CodegenGlobals::get_freed_pages() {
- let current_page = self.write_pos / self.page_size;
+ let current_page = self.write_pos / CODE_PAGE_SIZE;
freed_pages.iter().find(|&&page| current_page < page).map(|&page| page)
} else {
- Some(self.write_pos / self.page_size + 1)
+ Some(self.write_pos / CODE_PAGE_SIZE + 1)
};
// Move self to the next page
@@ -162,7 +161,7 @@ impl CodeBlock {
// but you need to waste some space for keeping write_pos for every single page.
// It doesn't seem necessary for performance either. So we're currently not doing it.
let dst_pos = self.get_page_pos(page_idx);
- if self.page_size * page_idx < self.mem_size && self.write_pos < dst_pos {
+ if CODE_PAGE_SIZE * page_idx < self.mem_size && self.write_pos < dst_pos {
// Reset dropped_bytes
self.dropped_bytes = false;
@@ -200,14 +199,14 @@ impl CodeBlock {
}
// Free the grouped pages at once
- let start_ptr = self.mem_block.borrow().start_ptr().add_bytes(page_idx * self.page_size);
- let batch_size = self.page_size * batch_idxs.len();
+ let start_ptr = self.mem_block.borrow().start_ptr().add_bytes(page_idx * CODE_PAGE_SIZE);
+ let batch_size = CODE_PAGE_SIZE * batch_idxs.len();
self.mem_block.borrow_mut().free_bytes(start_ptr, batch_size as u32);
}
}
pub fn page_size(&self) -> usize {
- self.page_size
+ CODE_PAGE_SIZE
}
pub fn mapped_region_size(&self) -> usize {
@@ -217,16 +216,16 @@ impl CodeBlock {
/// Return the number of code pages that have been mapped by the VirtualMemory.
pub fn num_mapped_pages(&self) -> usize {
// CodeBlock's page size != VirtualMem's page size on Linux,
- // so mapped_region_size % self.page_size may not be 0
- ((self.mapped_region_size() - 1) / self.page_size) + 1
+ // so mapped_region_size % CODE_PAGE_SIZE may not be 0
+ ((self.mapped_region_size() - 1) / CODE_PAGE_SIZE) + 1
}
/// Return the number of code pages that have been reserved by the VirtualMemory.
pub fn num_virtual_pages(&self) -> usize {
let virtual_region_size = self.mem_block.borrow().virtual_region_size();
// CodeBlock's page size != VirtualMem's page size on Linux,
- // so mapped_region_size % self.page_size may not be 0
- ((virtual_region_size - 1) / self.page_size) + 1
+ // so mapped_region_size % CODE_PAGE_SIZE may not be 0
+ ((virtual_region_size - 1) / CODE_PAGE_SIZE) + 1
}
/// Return the number of code pages that have been freed and not used yet.
@@ -236,17 +235,17 @@ impl CodeBlock {
pub fn has_freed_page(&self, page_idx: usize) -> bool {
CodegenGlobals::get_freed_pages().as_ref().map_or(false, |pages| pages.contains(&page_idx)) && // code GCed
- self.write_pos < page_idx * self.page_size // and not written yet
+ self.write_pos < page_idx * CODE_PAGE_SIZE // and not written yet
}
/// Convert a page index to the write_pos for the page start.
fn get_page_pos(&self, page_idx: usize) -> usize {
- self.page_size * page_idx + self.page_start()
+ CODE_PAGE_SIZE * page_idx + self.page_start()
}
/// write_pos of the current page start
pub fn page_start_pos(&self) -> usize {
- self.get_write_pos() / self.page_size * self.page_size + self.page_start()
+ self.get_write_pos() / CODE_PAGE_SIZE * CODE_PAGE_SIZE + self.page_start()
}
/// Offset of each page where CodeBlock should start writing
@@ -254,7 +253,7 @@ impl CodeBlock {
let mut start = if self.inline() {
0
} else {
- self.page_size / 2
+ CODE_PAGE_SIZE / 2
};
if cfg!(debug_assertions) && !cfg!(test) {
// Leave illegal instructions at the beginning of each page to assert
@@ -267,9 +266,9 @@ impl CodeBlock {
/// Offset of each page where CodeBlock should stop writing (exclusive)
pub fn page_end(&self) -> usize {
let page_end = if self.inline() {
- self.page_size / 2
+ CODE_PAGE_SIZE / 2
} else {
- self.page_size
+ CODE_PAGE_SIZE
};
page_end - self.page_end_reserve // reserve space to jump to the next page
}
@@ -299,14 +298,14 @@ impl CodeBlock {
let mut addrs = vec![];
while start < end {
- let page_idx = start.saturating_sub(region_start) / self.page_size;
- let current_page = region_start + (page_idx * self.page_size);
+ let page_idx = start.saturating_sub(region_start) / CODE_PAGE_SIZE;
+ let current_page = region_start + (page_idx * CODE_PAGE_SIZE);
let page_end = std::cmp::min(end, current_page + self.page_end());
// If code GC has been used, skip pages that are used by past on-stack code
if freed_pages.map_or(true, |pages| pages.contains(&page_idx)) {
addrs.push((start, page_end));
}
- start = current_page + self.page_size + self.page_start();
+ start = current_page + CODE_PAGE_SIZE + self.page_start();
}
addrs
}
@@ -314,11 +313,11 @@ impl CodeBlock {
/// Return the code size that has been used by this CodeBlock.
pub fn code_size(&self) -> usize {
let mut size = 0;
- let current_page_idx = self.write_pos / self.page_size;
+ let current_page_idx = self.write_pos / CODE_PAGE_SIZE;
for page_idx in 0..self.num_mapped_pages() {
if page_idx == current_page_idx {
// Count only actually used bytes for the current page.
- size += (self.write_pos % self.page_size).saturating_sub(self.page_start());
+ size += (self.write_pos % CODE_PAGE_SIZE).saturating_sub(self.page_start());
} else if !self.has_freed_page(page_idx) {
// Count an entire range for any non-freed pages that have been used.
size += self.page_end() - self.page_start() + self.page_end_reserve;
@@ -329,7 +328,7 @@ impl CodeBlock {
/// Check if this code block has sufficient remaining capacity
pub fn has_capacity(&self, num_bytes: usize) -> bool {
- let page_offset = self.write_pos % self.page_size;
+ let page_offset = self.write_pos % CODE_PAGE_SIZE;
let capacity = self.page_end().saturating_sub(page_offset);
num_bytes <= capacity
}
@@ -407,8 +406,8 @@ impl CodeBlock {
return vec![];
}
- let start_page = (start_addr.into_usize() - mem_start) / self.page_size;
- let end_page = (end_addr.into_usize() - mem_start - 1) / self.page_size;
+ let start_page = (start_addr.into_usize() - mem_start) / CODE_PAGE_SIZE;
+ let end_page = (end_addr.into_usize() - mem_start - 1) / CODE_PAGE_SIZE;
(start_page..=end_page).collect() // TODO: consider returning an iterator
}
@@ -644,7 +643,7 @@ impl CodeBlock {
let mem_start: *const u8 = alloc.mem_start();
let virt_mem = VirtualMem::new(alloc, 1, NonNull::new(mem_start as *mut u8).unwrap(), mem_size);
- Self::new(Rc::new(RefCell::new(virt_mem)), 16 * 1024, false)
+ Self::new(Rc::new(RefCell::new(virt_mem)), false)
}
}
diff --git a/yjit/src/codegen.rs b/yjit/src/codegen.rs
index 97768eb5fe..e01266ff21 100644
--- a/yjit/src/codegen.rs
+++ b/yjit/src/codegen.rs
@@ -7067,8 +7067,6 @@ impl CodegenGlobals {
use std::cell::RefCell;
use std::rc::Rc;
- let code_page_size = get_option!(code_page_size);
-
let virt_block: *mut u8 = unsafe { rb_yjit_reserve_addr_space(mem_size as u32) };
// Memory protection syscalls need page-aligned addresses, so check it here. Assuming
@@ -7083,7 +7081,6 @@ impl CodegenGlobals {
virt_block as usize % page_size.as_usize(), 0,
"Start of virtual address block should be page-aligned",
);
- assert_eq!(code_page_size % page_size.as_usize(), 0, "code_page_size was not page-aligned");
use crate::virtualmem::*;
use std::ptr::NonNull;
@@ -7096,8 +7093,10 @@ impl CodegenGlobals {
);
let mem_block = Rc::new(RefCell::new(mem_block));
- let cb = CodeBlock::new(mem_block.clone(), code_page_size, false);
- let ocb = OutlinedCb::wrap(CodeBlock::new(mem_block, code_page_size, true));
+ let cb = CodeBlock::new(mem_block.clone(), false);
+ let ocb = OutlinedCb::wrap(CodeBlock::new(mem_block, true));
+
+ assert_eq!(cb.page_size() % page_size.as_usize(), 0, "code page size is not page-aligned");
(cb, ocb)
};
diff --git a/yjit/src/options.rs b/yjit/src/options.rs
index 268e141ed6..d734bcc40b 100644
--- a/yjit/src/options.rs
+++ b/yjit/src/options.rs
@@ -8,10 +8,6 @@ pub struct Options {
// Note that the command line argument is expressed in MiB and not bytes
pub exec_mem_size: usize,
- // Size of each executable memory code page in bytes
- // Note that the command line argument is expressed in KiB and not bytes
- pub code_page_size: usize,
-
// Number of method calls after which to start generating code
// Threshold==1 means compile on first execution
pub call_threshold: usize,
@@ -54,7 +50,6 @@ pub struct Options {
// Initialize the options to default values
pub static mut OPTIONS: Options = Options {
exec_mem_size: 128 * 1024 * 1024,
- code_page_size: 16 * 1024,
call_threshold: 30,
greedy_versioning: false,
no_type_prop: false,
@@ -130,21 +125,6 @@ pub fn parse_option(str_ptr: *const std::os::raw::c_char) -> Option<()> {
}
},
- ("code-page-size", _) => match opt_val.parse::<usize>() {
- Ok(n) => {
- // Enforce bounds checks and that n is divisible by 4KiB
- if n < 4 || n > 256 || n % 4 != 0 {
- return None
- }
-
- // Convert from KiB to bytes internally for convenience
- unsafe { OPTIONS.code_page_size = n * 1024 }
- }
- Err(_) => {
- return None;
- }
- },
-
("call-threshold", _) => match opt_val.parse() {
Ok(n) => unsafe { OPTIONS.call_threshold = n },
Err(_) => {