Auto merge of #83214 - cjgillot:dep-map, r=michaelwoerister
Mmap the incremental data instead of reading it. Instead of reading the full incremental state using `fs::read_file`, we memmap it using a private read-only file-backed map. This allows the system to reclaim any memory we are not using, while ensuring we are not polluted by outside modifications to the file. Suggested in https://github.com/rust-lang/rust/pull/83036#issuecomment-800458082 by `@bjorn3`
This commit is contained in:
commit
11bbb52313
5 changed files with 123 additions and 88 deletions
|
@ -12,10 +12,12 @@
|
|||
use std::env;
|
||||
use std::fs;
|
||||
use std::io::{self, Read};
|
||||
use std::path::Path;
|
||||
use std::path::{Path, PathBuf};
|
||||
|
||||
use rustc_data_structures::memmap::Mmap;
|
||||
use rustc_serialize::opaque::{FileEncodeResult, FileEncoder};
|
||||
use rustc_serialize::Encoder;
|
||||
use rustc_session::Session;
|
||||
|
||||
/// The first few bytes of files generated by incremental compilation.
|
||||
const FILE_MAGIC: &[u8] = b"RSIC";
|
||||
|
@ -28,7 +30,7 @@ const HEADER_FORMAT_VERSION: u16 = 0;
|
|||
/// the Git commit hash.
|
||||
const RUSTC_VERSION: Option<&str> = option_env!("CFG_VERSION");
|
||||
|
||||
pub fn write_file_header(stream: &mut FileEncoder, nightly_build: bool) -> FileEncodeResult {
|
||||
pub(crate) fn write_file_header(stream: &mut FileEncoder, nightly_build: bool) -> FileEncodeResult {
|
||||
stream.emit_raw_bytes(FILE_MAGIC)?;
|
||||
stream.emit_raw_bytes(&[
|
||||
(HEADER_FORMAT_VERSION >> 0) as u8,
|
||||
|
@ -41,6 +43,61 @@ pub fn write_file_header(stream: &mut FileEncoder, nightly_build: bool) -> FileE
|
|||
stream.emit_raw_bytes(rustc_version.as_bytes())
|
||||
}
|
||||
|
||||
pub(crate) fn save_in<F>(sess: &Session, path_buf: PathBuf, name: &str, encode: F)
|
||||
where
|
||||
F: FnOnce(&mut FileEncoder) -> FileEncodeResult,
|
||||
{
|
||||
debug!("save: storing data in {}", path_buf.display());
|
||||
|
||||
// Delete the old file, if any.
|
||||
// Note: It's important that we actually delete the old file and not just
|
||||
// truncate and overwrite it, since it might be a shared hard-link, the
|
||||
// underlying data of which we don't want to modify.
|
||||
//
|
||||
// We have to ensure we have dropped the memory maps to this file
|
||||
// before performing this removal.
|
||||
match fs::remove_file(&path_buf) {
|
||||
Ok(()) => {
|
||||
debug!("save: remove old file");
|
||||
}
|
||||
Err(err) if err.kind() == io::ErrorKind::NotFound => (),
|
||||
Err(err) => {
|
||||
sess.err(&format!(
|
||||
"unable to delete old {} at `{}`: {}",
|
||||
name,
|
||||
path_buf.display(),
|
||||
err
|
||||
));
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
let mut encoder = match FileEncoder::new(&path_buf) {
|
||||
Ok(encoder) => encoder,
|
||||
Err(err) => {
|
||||
sess.err(&format!("failed to create {} at `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
};
|
||||
|
||||
if let Err(err) = write_file_header(&mut encoder, sess.is_nightly_build()) {
|
||||
sess.err(&format!("failed to write {} header to `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
|
||||
if let Err(err) = encode(&mut encoder) {
|
||||
sess.err(&format!("failed to write {} to `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
|
||||
if let Err(err) = encoder.flush() {
|
||||
sess.err(&format!("failed to flush {} to `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
|
||||
debug!("save: data written to disk successfully");
|
||||
}
|
||||
|
||||
/// Reads the contents of a file with a file header as defined in this module.
|
||||
///
|
||||
/// - Returns `Ok(Some(data, pos))` if the file existed and was generated by a
|
||||
|
@ -54,14 +111,21 @@ pub fn read_file(
|
|||
report_incremental_info: bool,
|
||||
path: &Path,
|
||||
nightly_build: bool,
|
||||
) -> io::Result<Option<(Vec<u8>, usize)>> {
|
||||
let data = match fs::read(path) {
|
||||
Ok(data) => data,
|
||||
) -> io::Result<Option<(Mmap, usize)>> {
|
||||
let file = match fs::File::open(path) {
|
||||
Ok(file) => file,
|
||||
Err(err) if err.kind() == io::ErrorKind::NotFound => return Ok(None),
|
||||
Err(err) => return Err(err),
|
||||
};
|
||||
// SAFETY: This process must not modify nor remove the backing file while the memory map lives.
|
||||
// For the dep-graph and the work product index, it is as soon as the decoding is done.
|
||||
// For the query result cache, the memory map is dropped in save_dep_graph before calling
|
||||
// save_in and trying to remove the backing file.
|
||||
//
|
||||
// There is no way to prevent another process from modifying this file.
|
||||
let mmap = unsafe { Mmap::map(file) }?;
|
||||
|
||||
let mut file = io::Cursor::new(data);
|
||||
let mut file = io::Cursor::new(&*mmap);
|
||||
|
||||
// Check FILE_MAGIC
|
||||
{
|
||||
|
@ -103,7 +167,7 @@ pub fn read_file(
|
|||
}
|
||||
|
||||
let post_header_start_pos = file.position() as usize;
|
||||
Ok(Some((file.into_inner(), post_header_start_pos)))
|
||||
Ok(Some((mmap, post_header_start_pos)))
|
||||
}
|
||||
|
||||
fn report_format_mismatch(report_incremental_info: bool, file: &Path, message: &str) {
|
||||
|
|
|
@ -1,6 +1,7 @@
|
|||
//! Code to save/load the dep-graph from files.
|
||||
|
||||
use rustc_data_structures::fx::FxHashMap;
|
||||
use rustc_data_structures::memmap::Mmap;
|
||||
use rustc_middle::dep_graph::{SerializedDepGraph, WorkProduct, WorkProductId};
|
||||
use rustc_middle::ty::OnDiskCache;
|
||||
use rustc_serialize::opaque::Decoder;
|
||||
|
@ -48,7 +49,7 @@ fn load_data(
|
|||
report_incremental_info: bool,
|
||||
path: &Path,
|
||||
nightly_build: bool,
|
||||
) -> LoadResult<(Vec<u8>, usize)> {
|
||||
) -> LoadResult<(Mmap, usize)> {
|
||||
match file_format::read_file(report_incremental_info, path, nightly_build) {
|
||||
Ok(Some(data_and_pos)) => LoadResult::Ok { data: data_and_pos },
|
||||
Ok(None) => {
|
||||
|
|
|
@ -6,8 +6,6 @@ use rustc_serialize::opaque::{FileEncodeResult, FileEncoder};
|
|||
use rustc_serialize::Encodable as RustcEncodable;
|
||||
use rustc_session::Session;
|
||||
use std::fs;
|
||||
use std::io;
|
||||
use std::path::PathBuf;
|
||||
|
||||
use super::data::*;
|
||||
use super::dirty_clean;
|
||||
|
@ -44,7 +42,14 @@ pub fn save_dep_graph(tcx: TyCtxt<'_>) {
|
|||
join(
|
||||
move || {
|
||||
sess.time("incr_comp_persist_result_cache", || {
|
||||
save_in(sess, query_cache_path, "query cache", |e| encode_query_cache(tcx, e));
|
||||
// Drop the memory map so that we can remove the file and write to it.
|
||||
if let Some(odc) = &tcx.on_disk_cache {
|
||||
odc.drop_serialized_data(tcx);
|
||||
}
|
||||
|
||||
file_format::save_in(sess, query_cache_path, "query cache", |e| {
|
||||
encode_query_cache(tcx, e)
|
||||
});
|
||||
});
|
||||
},
|
||||
move || {
|
||||
|
@ -86,7 +91,9 @@ pub fn save_work_product_index(
|
|||
debug!("save_work_product_index()");
|
||||
dep_graph.assert_ignored();
|
||||
let path = work_products_path(sess);
|
||||
save_in(sess, path, "work product index", |e| encode_work_product_index(&new_work_products, e));
|
||||
file_format::save_in(sess, path, "work product index", |e| {
|
||||
encode_work_product_index(&new_work_products, e)
|
||||
});
|
||||
|
||||
// We also need to clean out old work-products, as not all of them are
|
||||
// deleted during invalidation. Some object files don't change their
|
||||
|
@ -113,58 +120,6 @@ pub fn save_work_product_index(
|
|||
});
|
||||
}
|
||||
|
||||
pub(crate) fn save_in<F>(sess: &Session, path_buf: PathBuf, name: &str, encode: F)
|
||||
where
|
||||
F: FnOnce(&mut FileEncoder) -> FileEncodeResult,
|
||||
{
|
||||
debug!("save: storing data in {}", path_buf.display());
|
||||
|
||||
// Delete the old file, if any.
|
||||
// Note: It's important that we actually delete the old file and not just
|
||||
// truncate and overwrite it, since it might be a shared hard-link, the
|
||||
// underlying data of which we don't want to modify
|
||||
match fs::remove_file(&path_buf) {
|
||||
Ok(()) => {
|
||||
debug!("save: remove old file");
|
||||
}
|
||||
Err(err) if err.kind() == io::ErrorKind::NotFound => (),
|
||||
Err(err) => {
|
||||
sess.err(&format!(
|
||||
"unable to delete old {} at `{}`: {}",
|
||||
name,
|
||||
path_buf.display(),
|
||||
err
|
||||
));
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
let mut encoder = match FileEncoder::new(&path_buf) {
|
||||
Ok(encoder) => encoder,
|
||||
Err(err) => {
|
||||
sess.err(&format!("failed to create {} at `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
};
|
||||
|
||||
if let Err(err) = file_format::write_file_header(&mut encoder, sess.is_nightly_build()) {
|
||||
sess.err(&format!("failed to write {} header to `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
|
||||
if let Err(err) = encode(&mut encoder) {
|
||||
sess.err(&format!("failed to write {} to `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
|
||||
if let Err(err) = encoder.flush() {
|
||||
sess.err(&format!("failed to flush {} to `{}`: {}", name, path_buf.display(), err));
|
||||
return;
|
||||
}
|
||||
|
||||
debug!("save: data written to disk successfully");
|
||||
}
|
||||
|
||||
fn encode_work_product_index(
|
||||
work_products: &FxHashMap<WorkProductId, WorkProduct>,
|
||||
encoder: &mut FileEncoder,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue