YJIT: Remove --yjit-code-page-size (#6865)

Certain code page sizes don't work and can cause crashes, so having this
value available as a command-line option is a bit dangerous. Remove it
and turn it into a constant instead.
This commit is contained in:
Alan Wu 2022-12-05 17:43:17 -05:00 committed by GitHub
parent e4aba8f519
commit 235fc50447
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
Notes: git 2022-12-05 22:43:36 +00:00
Merged-By: maximecb <maximecb@ruby-lang.org>
3 changed files with 33 additions and 55 deletions

View File

@ -24,6 +24,9 @@ pub mod x86_64;
pub mod arm64;
/// Size of a code page in bytes. Each code page is split into an inlined and an outlined portion.
const CODE_PAGE_SIZE: usize = 16 * 1024;
//
// TODO: need a field_size_of macro, to compute the size of a struct field in bytes
//
@ -57,9 +60,6 @@ pub struct CodeBlock {
// Current writing position
write_pos: usize,
// Size of a code page (inlined + outlined)
page_size: usize,
// Size reserved for writing a jump to the next page
page_end_reserve: usize,
@ -94,13 +94,12 @@ pub struct LabelState {
impl CodeBlock {
/// Make a new CodeBlock
pub fn new(mem_block: Rc<RefCell<VirtualMem>>, page_size: usize, outlined: bool) -> Self {
pub fn new(mem_block: Rc<RefCell<VirtualMem>>, outlined: bool) -> Self {
let mem_size = mem_block.borrow().virtual_region_size();
let mut cb = Self {
mem_block,
mem_size,
write_pos: 0,
page_size,
page_end_reserve: JMP_PTR_BYTES,
label_addrs: Vec::new(),
label_names: Vec::new(),
@ -122,10 +121,10 @@ impl CodeBlock {
// Use the freed_pages list if code GC has been used. Otherwise use the next page.
let next_page_idx = if let Some(freed_pages) = CodegenGlobals::get_freed_pages() {
let current_page = self.write_pos / self.page_size;
let current_page = self.write_pos / CODE_PAGE_SIZE;
freed_pages.iter().find(|&&page| current_page < page).map(|&page| page)
} else {
Some(self.write_pos / self.page_size + 1)
Some(self.write_pos / CODE_PAGE_SIZE + 1)
};
// Move self to the next page
@ -162,7 +161,7 @@ impl CodeBlock {
// but you need to waste some space for keeping write_pos for every single page.
// It doesn't seem necessary for performance either. So we're currently not doing it.
let dst_pos = self.get_page_pos(page_idx);
if self.page_size * page_idx < self.mem_size && self.write_pos < dst_pos {
if CODE_PAGE_SIZE * page_idx < self.mem_size && self.write_pos < dst_pos {
// Reset dropped_bytes
self.dropped_bytes = false;
@ -200,14 +199,14 @@ impl CodeBlock {
}
// Free the grouped pages at once
let start_ptr = self.mem_block.borrow().start_ptr().add_bytes(page_idx * self.page_size);
let batch_size = self.page_size * batch_idxs.len();
let start_ptr = self.mem_block.borrow().start_ptr().add_bytes(page_idx * CODE_PAGE_SIZE);
let batch_size = CODE_PAGE_SIZE * batch_idxs.len();
self.mem_block.borrow_mut().free_bytes(start_ptr, batch_size as u32);
}
}
pub fn page_size(&self) -> usize {
self.page_size
CODE_PAGE_SIZE
}
pub fn mapped_region_size(&self) -> usize {
@ -217,16 +216,16 @@ impl CodeBlock {
/// Return the number of code pages that have been mapped by the VirtualMemory.
pub fn num_mapped_pages(&self) -> usize {
// CodeBlock's page size != VirtualMem's page size on Linux,
// so mapped_region_size % self.page_size may not be 0
((self.mapped_region_size() - 1) / self.page_size) + 1
// so mapped_region_size % CODE_PAGE_SIZE may not be 0
((self.mapped_region_size() - 1) / CODE_PAGE_SIZE) + 1
}
/// Return the number of code pages that have been reserved by the VirtualMemory.
pub fn num_virtual_pages(&self) -> usize {
let virtual_region_size = self.mem_block.borrow().virtual_region_size();
// CodeBlock's page size != VirtualMem's page size on Linux,
// so mapped_region_size % self.page_size may not be 0
((virtual_region_size - 1) / self.page_size) + 1
// so mapped_region_size % CODE_PAGE_SIZE may not be 0
((virtual_region_size - 1) / CODE_PAGE_SIZE) + 1
}
/// Return the number of code pages that have been freed and not used yet.
@ -236,17 +235,17 @@ impl CodeBlock {
pub fn has_freed_page(&self, page_idx: usize) -> bool {
CodegenGlobals::get_freed_pages().as_ref().map_or(false, |pages| pages.contains(&page_idx)) && // code GCed
self.write_pos < page_idx * self.page_size // and not written yet
self.write_pos < page_idx * CODE_PAGE_SIZE // and not written yet
}
/// Convert a page index to the write_pos for the page start.
fn get_page_pos(&self, page_idx: usize) -> usize {
self.page_size * page_idx + self.page_start()
CODE_PAGE_SIZE * page_idx + self.page_start()
}
/// write_pos of the current page start
pub fn page_start_pos(&self) -> usize {
self.get_write_pos() / self.page_size * self.page_size + self.page_start()
self.get_write_pos() / CODE_PAGE_SIZE * CODE_PAGE_SIZE + self.page_start()
}
/// Offset of each page where CodeBlock should start writing
@ -254,7 +253,7 @@ impl CodeBlock {
let mut start = if self.inline() {
0
} else {
self.page_size / 2
CODE_PAGE_SIZE / 2
};
if cfg!(debug_assertions) && !cfg!(test) {
// Leave illegal instructions at the beginning of each page to assert
@ -267,9 +266,9 @@ impl CodeBlock {
/// Offset of each page where CodeBlock should stop writing (exclusive)
pub fn page_end(&self) -> usize {
let page_end = if self.inline() {
self.page_size / 2
CODE_PAGE_SIZE / 2
} else {
self.page_size
CODE_PAGE_SIZE
};
page_end - self.page_end_reserve // reserve space to jump to the next page
}
@ -299,14 +298,14 @@ impl CodeBlock {
let mut addrs = vec![];
while start < end {
let page_idx = start.saturating_sub(region_start) / self.page_size;
let current_page = region_start + (page_idx * self.page_size);
let page_idx = start.saturating_sub(region_start) / CODE_PAGE_SIZE;
let current_page = region_start + (page_idx * CODE_PAGE_SIZE);
let page_end = std::cmp::min(end, current_page + self.page_end());
// If code GC has been used, skip pages that are used by past on-stack code
if freed_pages.map_or(true, |pages| pages.contains(&page_idx)) {
addrs.push((start, page_end));
}
start = current_page + self.page_size + self.page_start();
start = current_page + CODE_PAGE_SIZE + self.page_start();
}
addrs
}
@ -314,11 +313,11 @@ impl CodeBlock {
/// Return the code size that has been used by this CodeBlock.
pub fn code_size(&self) -> usize {
let mut size = 0;
let current_page_idx = self.write_pos / self.page_size;
let current_page_idx = self.write_pos / CODE_PAGE_SIZE;
for page_idx in 0..self.num_mapped_pages() {
if page_idx == current_page_idx {
// Count only actually used bytes for the current page.
size += (self.write_pos % self.page_size).saturating_sub(self.page_start());
size += (self.write_pos % CODE_PAGE_SIZE).saturating_sub(self.page_start());
} else if !self.has_freed_page(page_idx) {
// Count an entire range for any non-freed pages that have been used.
size += self.page_end() - self.page_start() + self.page_end_reserve;
@ -329,7 +328,7 @@ impl CodeBlock {
/// Check if this code block has sufficient remaining capacity
pub fn has_capacity(&self, num_bytes: usize) -> bool {
let page_offset = self.write_pos % self.page_size;
let page_offset = self.write_pos % CODE_PAGE_SIZE;
let capacity = self.page_end().saturating_sub(page_offset);
num_bytes <= capacity
}
@ -407,8 +406,8 @@ impl CodeBlock {
return vec![];
}
let start_page = (start_addr.into_usize() - mem_start) / self.page_size;
let end_page = (end_addr.into_usize() - mem_start - 1) / self.page_size;
let start_page = (start_addr.into_usize() - mem_start) / CODE_PAGE_SIZE;
let end_page = (end_addr.into_usize() - mem_start - 1) / CODE_PAGE_SIZE;
(start_page..=end_page).collect() // TODO: consider returning an iterator
}
@ -644,7 +643,7 @@ impl CodeBlock {
let mem_start: *const u8 = alloc.mem_start();
let virt_mem = VirtualMem::new(alloc, 1, NonNull::new(mem_start as *mut u8).unwrap(), mem_size);
Self::new(Rc::new(RefCell::new(virt_mem)), 16 * 1024, false)
Self::new(Rc::new(RefCell::new(virt_mem)), false)
}
}

View File

@ -7067,8 +7067,6 @@ impl CodegenGlobals {
use std::cell::RefCell;
use std::rc::Rc;
let code_page_size = get_option!(code_page_size);
let virt_block: *mut u8 = unsafe { rb_yjit_reserve_addr_space(mem_size as u32) };
// Memory protection syscalls need page-aligned addresses, so check it here. Assuming
@ -7083,7 +7081,6 @@ impl CodegenGlobals {
virt_block as usize % page_size.as_usize(), 0,
"Start of virtual address block should be page-aligned",
);
assert_eq!(code_page_size % page_size.as_usize(), 0, "code_page_size was not page-aligned");
use crate::virtualmem::*;
use std::ptr::NonNull;
@ -7096,8 +7093,10 @@ impl CodegenGlobals {
);
let mem_block = Rc::new(RefCell::new(mem_block));
let cb = CodeBlock::new(mem_block.clone(), code_page_size, false);
let ocb = OutlinedCb::wrap(CodeBlock::new(mem_block, code_page_size, true));
let cb = CodeBlock::new(mem_block.clone(), false);
let ocb = OutlinedCb::wrap(CodeBlock::new(mem_block, true));
assert_eq!(cb.page_size() % page_size.as_usize(), 0, "code page size is not page-aligned");
(cb, ocb)
};

View File

@ -8,10 +8,6 @@ pub struct Options {
// Note that the command line argument is expressed in MiB and not bytes
pub exec_mem_size: usize,
// Size of each executable memory code page in bytes
// Note that the command line argument is expressed in KiB and not bytes
pub code_page_size: usize,
// Number of method calls after which to start generating code
// Threshold==1 means compile on first execution
pub call_threshold: usize,
@ -54,7 +50,6 @@ pub struct Options {
// Initialize the options to default values
pub static mut OPTIONS: Options = Options {
exec_mem_size: 128 * 1024 * 1024,
code_page_size: 16 * 1024,
call_threshold: 30,
greedy_versioning: false,
no_type_prop: false,
@ -130,21 +125,6 @@ pub fn parse_option(str_ptr: *const std::os::raw::c_char) -> Option<()> {
}
},
("code-page-size", _) => match opt_val.parse::<usize>() {
Ok(n) => {
// Enforce bounds checks and that n is divisible by 4KiB
if n < 4 || n > 256 || n % 4 != 0 {
return None
}
// Convert from KiB to bytes internally for convenience
unsafe { OPTIONS.code_page_size = n * 1024 }
}
Err(_) => {
return None;
}
},
("call-threshold", _) => match opt_val.parse() {
Ok(n) => unsafe { OPTIONS.call_threshold = n },
Err(_) => {