mirror of
				https://github.com/rust-lang/rust.git
				synced 2025-10-31 13:04:42 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			262 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			Rust
		
	
	
	
	
	
			
		
		
	
	
			262 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			Rust
		
	
	
	
	
	
| use std::ffi::OsStr;
 | |
| use std::hash::{Hash, Hasher};
 | |
| use std::path::{Path, PathBuf};
 | |
| use std::process::Command;
 | |
| use std::{fs, io};
 | |
| 
 | |
| use crate::path::{Dirs, RelPath};
 | |
| use crate::utils::{copy_dir_recursively, ensure_empty_dir, spawn_and_wait};
 | |
| 
 | |
| pub(crate) fn prepare(dirs: &Dirs) {
 | |
|     std::fs::create_dir_all(&dirs.download_dir).unwrap();
 | |
|     crate::tests::RAND_REPO.fetch(dirs);
 | |
|     crate::tests::REGEX_REPO.fetch(dirs);
 | |
| }
 | |
| 
 | |
| pub(crate) struct GitRepo {
 | |
|     url: GitRepoUrl,
 | |
|     rev: &'static str,
 | |
|     content_hash: &'static str,
 | |
|     patch_name: &'static str,
 | |
| }
 | |
| 
 | |
| enum GitRepoUrl {
 | |
|     Github { user: &'static str, repo: &'static str },
 | |
| }
 | |
| 
 | |
| // Note: This uses a hasher which is not cryptographically secure. This is fine as the hash is meant
 | |
| // to protect against accidental modification and outdated downloads, not against manipulation.
 | |
| fn hash_file(file: &std::path::Path) -> u64 {
 | |
|     let contents = std::fs::read(file).unwrap();
 | |
|     #[allow(deprecated)]
 | |
|     let mut hasher = std::hash::SipHasher::new();
 | |
|     // The following is equivalent to
 | |
|     //   std::hash::Hash::hash(&contents, &mut hasher);
 | |
|     // but gives the same result independent of host byte order.
 | |
|     hasher.write_usize(contents.len().to_le());
 | |
|     Hash::hash_slice(&contents, &mut hasher);
 | |
|     std::hash::Hasher::finish(&hasher)
 | |
| }
 | |
| 
 | |
| fn hash_dir(dir: &std::path::Path) -> u64 {
 | |
|     let mut sub_hashes = std::collections::BTreeMap::new();
 | |
|     for entry in std::fs::read_dir(dir).unwrap() {
 | |
|         let entry = entry.unwrap();
 | |
|         if entry.file_type().unwrap().is_dir() {
 | |
|             sub_hashes.insert(
 | |
|                 entry.file_name().to_str().unwrap().to_owned(),
 | |
|                 hash_dir(&entry.path()).to_le(),
 | |
|             );
 | |
|         } else {
 | |
|             sub_hashes.insert(
 | |
|                 entry.file_name().to_str().unwrap().to_owned(),
 | |
|                 hash_file(&entry.path()).to_le(),
 | |
|             );
 | |
|         }
 | |
|     }
 | |
|     #[allow(deprecated)]
 | |
|     let mut hasher = std::hash::SipHasher::new();
 | |
|     // The following is equivalent to
 | |
|     //   std::hash::Hash::hash(&sub_hashes, &mut hasher);
 | |
|     // but gives the same result independent of host byte order.
 | |
|     hasher.write_usize(sub_hashes.len().to_le());
 | |
|     for elt in sub_hashes {
 | |
|         elt.hash(&mut hasher);
 | |
|     }
 | |
|     std::hash::Hasher::finish(&hasher)
 | |
| }
 | |
| 
 | |
| impl GitRepo {
 | |
|     pub(crate) const fn github(
 | |
|         user: &'static str,
 | |
|         repo: &'static str,
 | |
|         rev: &'static str,
 | |
|         content_hash: &'static str,
 | |
|         patch_name: &'static str,
 | |
|     ) -> GitRepo {
 | |
|         GitRepo { url: GitRepoUrl::Github { user, repo }, rev, content_hash, patch_name }
 | |
|     }
 | |
| 
 | |
|     fn download_dir(&self, dirs: &Dirs) -> PathBuf {
 | |
|         match self.url {
 | |
|             GitRepoUrl::Github { user: _, repo } => dirs.download_dir.join(repo),
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     pub(crate) const fn source_dir(&self) -> RelPath {
 | |
|         match self.url {
 | |
|             GitRepoUrl::Github { user: _, repo } => RelPath::build(repo),
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     fn verify_checksum(&self, dirs: &Dirs) {
 | |
|         let download_dir = self.download_dir(dirs);
 | |
|         let actual_hash = format!("{:016x}", hash_dir(&download_dir));
 | |
|         if actual_hash != self.content_hash {
 | |
|             eprintln!(
 | |
|                 "Mismatched content hash for {download_dir}: {actual_hash} != {content_hash}. Please run ./y.sh prepare again.",
 | |
|                 download_dir = download_dir.display(),
 | |
|                 content_hash = self.content_hash,
 | |
|             );
 | |
|             std::process::exit(1);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     pub(crate) fn fetch(&self, dirs: &Dirs) {
 | |
|         let download_dir = self.download_dir(dirs);
 | |
| 
 | |
|         if download_dir.exists() {
 | |
|             let actual_hash = format!("{:016x}", hash_dir(&download_dir));
 | |
|             if actual_hash == self.content_hash {
 | |
|                 eprintln!("[FRESH] {}", download_dir.display());
 | |
|                 return;
 | |
|             } else {
 | |
|                 eprintln!(
 | |
|                     "Mismatched content hash for {download_dir}: {actual_hash} != {content_hash}. Downloading again.",
 | |
|                     download_dir = download_dir.display(),
 | |
|                     content_hash = self.content_hash,
 | |
|                 );
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         match self.url {
 | |
|             GitRepoUrl::Github { user, repo } => {
 | |
|                 clone_repo(
 | |
|                     &download_dir,
 | |
|                     &format!("https://github.com/{}/{}.git", user, repo),
 | |
|                     self.rev,
 | |
|                 );
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         let source_lockfile =
 | |
|             dirs.source_dir.join("patches").join(format!("{}-lock.toml", self.patch_name));
 | |
|         let target_lockfile = download_dir.join("Cargo.lock");
 | |
|         if source_lockfile.exists() {
 | |
|             assert!(!target_lockfile.exists());
 | |
|             fs::copy(source_lockfile, target_lockfile).unwrap();
 | |
|         } else {
 | |
|             assert!(target_lockfile.exists());
 | |
|         }
 | |
| 
 | |
|         self.verify_checksum(dirs);
 | |
|     }
 | |
| 
 | |
|     pub(crate) fn patch(&self, dirs: &Dirs) {
 | |
|         self.verify_checksum(dirs);
 | |
|         apply_patches(
 | |
|             dirs,
 | |
|             self.patch_name,
 | |
|             &self.download_dir(dirs),
 | |
|             &self.source_dir().to_path(dirs),
 | |
|         );
 | |
|     }
 | |
| }
 | |
| 
 | |
| fn clone_repo(download_dir: &Path, repo: &str, rev: &str) {
 | |
|     eprintln!("[CLONE] {}", repo);
 | |
| 
 | |
|     match fs::remove_dir_all(download_dir) {
 | |
|         Ok(()) => {}
 | |
|         Err(err) if err.kind() == io::ErrorKind::NotFound => {}
 | |
|         Err(err) => panic!("Failed to remove {path}: {err}", path = download_dir.display()),
 | |
|     }
 | |
| 
 | |
|     // Ignore exit code as the repo may already have been checked out
 | |
|     git_command(None, "clone").arg(repo).arg(download_dir).spawn().unwrap().wait().unwrap();
 | |
| 
 | |
|     let mut clean_cmd = git_command(download_dir, "checkout");
 | |
|     clean_cmd.arg("--").arg(".");
 | |
|     spawn_and_wait(clean_cmd);
 | |
| 
 | |
|     let mut checkout_cmd = git_command(download_dir, "checkout");
 | |
|     checkout_cmd.arg("-q").arg(rev);
 | |
|     spawn_and_wait(checkout_cmd);
 | |
| 
 | |
|     std::fs::remove_dir_all(download_dir.join(".git")).unwrap();
 | |
| }
 | |
| 
 | |
| fn init_git_repo(repo_dir: &Path) {
 | |
|     let mut git_init_cmd = git_command(repo_dir, "init");
 | |
|     git_init_cmd.arg("-q");
 | |
|     spawn_and_wait(git_init_cmd);
 | |
| 
 | |
|     let mut git_add_cmd = git_command(repo_dir, "add");
 | |
|     git_add_cmd.arg(".");
 | |
|     spawn_and_wait(git_add_cmd);
 | |
| 
 | |
|     let mut git_commit_cmd = git_command(repo_dir, "commit");
 | |
|     git_commit_cmd.arg("-m").arg("Initial commit").arg("-q").arg("--no-verify");
 | |
|     spawn_and_wait(git_commit_cmd);
 | |
| }
 | |
| 
 | |
| fn get_patches(dirs: &Dirs, crate_name: &str) -> Vec<PathBuf> {
 | |
|     let mut patches: Vec<_> = fs::read_dir(dirs.source_dir.join("patches"))
 | |
|         .unwrap()
 | |
|         .map(|entry| entry.unwrap().path())
 | |
|         .filter(|path| path.extension() == Some(OsStr::new("patch")))
 | |
|         .filter(|path| {
 | |
|             path.file_name()
 | |
|                 .unwrap()
 | |
|                 .to_str()
 | |
|                 .unwrap()
 | |
|                 .split_once("-")
 | |
|                 .unwrap()
 | |
|                 .1
 | |
|                 .starts_with(crate_name)
 | |
|         })
 | |
|         .collect();
 | |
|     patches.sort();
 | |
|     patches
 | |
| }
 | |
| 
 | |
| pub(crate) fn apply_patches(dirs: &Dirs, crate_name: &str, source_dir: &Path, target_dir: &Path) {
 | |
|     // FIXME avoid copy and patch if src, patches and target are unchanged
 | |
| 
 | |
|     eprintln!("[COPY] {crate_name} source");
 | |
| 
 | |
|     ensure_empty_dir(target_dir);
 | |
|     if crate_name == "stdlib" {
 | |
|         fs::create_dir(target_dir.join("library")).unwrap();
 | |
|         copy_dir_recursively(&source_dir.join("library"), &target_dir.join("library"));
 | |
|     } else {
 | |
|         copy_dir_recursively(source_dir, target_dir);
 | |
|     }
 | |
| 
 | |
|     init_git_repo(target_dir);
 | |
| 
 | |
|     if crate_name == "<none>" {
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     for patch in get_patches(dirs, crate_name) {
 | |
|         eprintln!(
 | |
|             "[PATCH] {:?} <- {:?}",
 | |
|             target_dir.file_name().unwrap(),
 | |
|             patch.file_name().unwrap()
 | |
|         );
 | |
|         let mut apply_patch_cmd = git_command(target_dir, "am");
 | |
|         apply_patch_cmd.arg(patch).arg("-q");
 | |
|         spawn_and_wait(apply_patch_cmd);
 | |
|     }
 | |
| }
 | |
| 
 | |
| #[must_use]
 | |
| fn git_command<'a>(repo_dir: impl Into<Option<&'a Path>>, cmd: &str) -> Command {
 | |
|     let mut git_cmd = Command::new("git");
 | |
|     git_cmd
 | |
|         .arg("-c")
 | |
|         .arg("user.name=Dummy")
 | |
|         .arg("-c")
 | |
|         .arg("user.email=dummy@example.com")
 | |
|         .arg("-c")
 | |
|         .arg("core.autocrlf=false")
 | |
|         .arg("-c")
 | |
|         .arg("commit.gpgSign=false")
 | |
|         .arg(cmd);
 | |
|     if let Some(repo_dir) = repo_dir.into() {
 | |
|         git_cmd.current_dir(repo_dir);
 | |
|     }
 | |
|     git_cmd
 | |
| }
 | 
