| //! Checks the licenses of third-party dependencies. |
| |
| use std::collections::{HashMap, HashSet}; |
| use std::fs::{File, read_dir}; |
| use std::io::Write; |
| use std::path::Path; |
| |
| use build_helper::ci::CiEnv; |
| use cargo_metadata::semver::Version; |
| use cargo_metadata::{Metadata, Package, PackageId}; |
| |
| use crate::diagnostics::{RunningCheck, TidyCtx}; |
| |
| #[path = "../../../bootstrap/src/utils/proc_macro_deps.rs"] |
| mod proc_macro_deps; |
| |
| /// These are licenses that are allowed for all crates, including the runtime, |
| /// rustc, tools, etc. |
| #[rustfmt::skip] |
| const LICENSES: &[&str] = &[ |
| // tidy-alphabetical-start |
| "0BSD OR MIT OR Apache-2.0", // adler2 license |
| "Apache-2.0 / MIT", |
| "Apache-2.0 OR ISC OR MIT", |
| "Apache-2.0 OR MIT", |
| "Apache-2.0 WITH LLVM-exception OR Apache-2.0 OR MIT", // wasi license |
| "Apache-2.0/MIT", |
| "BSD-2-Clause OR Apache-2.0 OR MIT", // zerocopy |
| "BSD-2-Clause OR MIT OR Apache-2.0", |
| "BSD-3-Clause/MIT", |
| "CC0-1.0 OR MIT-0 OR Apache-2.0", |
| "ISC", |
| "MIT / Apache-2.0", |
| "MIT AND (MIT OR Apache-2.0)", |
| "MIT AND Apache-2.0 WITH LLVM-exception AND (MIT OR Apache-2.0)", // compiler-builtins |
| "MIT OR Apache-2.0 OR BSD-1-Clause", |
| "MIT OR Apache-2.0 OR LGPL-2.1-or-later", // r-efi, r-efi-alloc; LGPL is not acceptable, but we use it under MIT OR Apache-2.0 |
| "MIT OR Apache-2.0 OR Zlib", // tinyvec_macros |
| "MIT OR Apache-2.0", |
| "MIT OR Zlib OR Apache-2.0", // miniz_oxide |
| "MIT", |
| "MIT/Apache-2.0", |
| "Unlicense OR MIT", |
| "Unlicense/MIT", |
| // tidy-alphabetical-end |
| ]; |
| |
| /// These are licenses that are allowed for rustc, tools, etc. But not for the runtime! |
| #[rustfmt::skip] |
| const LICENSES_TOOLS: &[&str] = &[ |
| // tidy-alphabetical-start |
| "(Apache-2.0 OR MIT) AND BSD-3-Clause", |
| "(MIT OR Apache-2.0) AND Unicode-3.0", // unicode_ident (1.0.14) |
| "(MIT OR Apache-2.0) AND Unicode-DFS-2016", // unicode_ident (1.0.12) |
| "0BSD", |
| "Apache-2.0 AND ISC", |
| "Apache-2.0 OR BSL-1.0", // BSL is not acceptable, but we use it under Apache-2.0 |
| "Apache-2.0 WITH LLVM-exception", |
| "Apache-2.0", |
| "BSD-2-Clause", |
| "BSD-3-Clause", |
| "CC0-1.0 OR Apache-2.0 OR Apache-2.0 WITH LLVM-exception", |
| "CC0-1.0", |
| "Unicode-3.0", // icu4x |
| "Unicode-DFS-2016", // tinystr |
| "Zlib OR Apache-2.0 OR MIT", // tinyvec |
| "Zlib", |
| // tidy-alphabetical-end |
| ]; |
| |
| type ExceptionList = &'static [(&'static str, &'static str)]; |
| |
| #[derive(Clone, Copy)] |
| pub(crate) struct WorkspaceInfo<'a> { |
| /// Path to the directory containing the workspace root Cargo.toml file. |
| pub(crate) path: &'a str, |
| /// The list of license exceptions. |
| pub(crate) exceptions: ExceptionList, |
| /// Optionally: |
| /// * A list of crates for which dependencies need to be explicitly allowed. |
| /// * The list of allowed dependencies. |
| /// * The source code location of the allowed dependencies list |
| crates_and_deps: Option<(&'a [&'a str], &'a [&'a str], ListLocation)>, |
| /// Submodules required for the workspace |
| pub(crate) submodules: &'a [&'a str], |
| } |
| |
| /// The workspaces to check for licensing and optionally permitted dependencies. |
| // FIXME auto detect all cargo workspaces |
| pub(crate) const WORKSPACES: &[WorkspaceInfo<'static>] = &[ |
| // The root workspace has to be first for check_rustfix to work. |
| WorkspaceInfo { |
| path: ".", |
| exceptions: EXCEPTIONS, |
| crates_and_deps: Some(( |
| &["rustc-main"], |
| PERMITTED_RUSTC_DEPENDENCIES, |
| PERMITTED_RUSTC_DEPS_LOCATION, |
| )), |
| submodules: &[], |
| }, |
| WorkspaceInfo { |
| path: "library", |
| exceptions: EXCEPTIONS_STDLIB, |
| crates_and_deps: Some(( |
| &["sysroot"], |
| PERMITTED_STDLIB_DEPENDENCIES, |
| PERMITTED_STDLIB_DEPS_LOCATION, |
| )), |
| submodules: &[], |
| }, |
| WorkspaceInfo { |
| path: "compiler/rustc_codegen_cranelift", |
| exceptions: EXCEPTIONS_CRANELIFT, |
| crates_and_deps: Some(( |
| &["rustc_codegen_cranelift"], |
| PERMITTED_CRANELIFT_DEPENDENCIES, |
| PERMITTED_CRANELIFT_DEPS_LOCATION, |
| )), |
| submodules: &[], |
| }, |
| WorkspaceInfo { |
| path: "compiler/rustc_codegen_gcc", |
| exceptions: EXCEPTIONS_GCC, |
| crates_and_deps: None, |
| submodules: &[], |
| }, |
| WorkspaceInfo { |
| path: "src/bootstrap", |
| exceptions: EXCEPTIONS_BOOTSTRAP, |
| crates_and_deps: None, |
| submodules: &[], |
| }, |
| WorkspaceInfo { |
| path: "src/tools/cargo", |
| exceptions: EXCEPTIONS_CARGO, |
| crates_and_deps: None, |
| submodules: &["src/tools/cargo"], |
| }, |
| // FIXME uncomment once all deps are vendored |
| // WorkspaceInfo { |
| // path: "src/tools/miri/test-cargo-miri", |
| // crates_and_deps: None |
| // submodules: &[], |
| // }, |
| // WorkspaceInfo { |
| // path: "src/tools/miri/test_dependencies", |
| // crates_and_deps: None, |
| // submodules: &[], |
| // } |
| WorkspaceInfo { |
| path: "src/tools/rust-analyzer", |
| exceptions: EXCEPTIONS_RUST_ANALYZER, |
| crates_and_deps: None, |
| submodules: &[], |
| }, |
| WorkspaceInfo { |
| path: "src/tools/rustbook", |
| exceptions: EXCEPTIONS_RUSTBOOK, |
| crates_and_deps: None, |
| submodules: &["src/doc/book", "src/doc/reference"], |
| }, |
| WorkspaceInfo { |
| path: "src/tools/rustc-perf", |
| exceptions: EXCEPTIONS_RUSTC_PERF, |
| crates_and_deps: None, |
| submodules: &["src/tools/rustc-perf"], |
| }, |
| WorkspaceInfo { |
| path: "src/tools/test-float-parse", |
| exceptions: EXCEPTIONS, |
| crates_and_deps: None, |
| submodules: &[], |
| }, |
| WorkspaceInfo { |
| path: "tests/run-make-cargo/uefi-qemu/uefi_qemu_test", |
| exceptions: EXCEPTIONS_UEFI_QEMU_TEST, |
| crates_and_deps: None, |
| submodules: &[], |
| }, |
| ]; |
| |
| /// These are exceptions to Rust's permissive licensing policy, and |
| /// should be considered bugs. Exceptions are only allowed in Rust |
| /// tooling. It is _crucial_ that no exception crates be dependencies |
| /// of the Rust runtime (std/test). |
| #[rustfmt::skip] |
| const EXCEPTIONS: ExceptionList = &[ |
| // tidy-alphabetical-start |
| ("colored", "MPL-2.0"), // rustfmt |
| ("option-ext", "MPL-2.0"), // cargo-miri (via `directories`) |
| // tidy-alphabetical-end |
| ]; |
| |
| /// These are exceptions to Rust's permissive licensing policy, and |
| /// should be considered bugs. Exceptions are only allowed in Rust |
| /// tooling. It is _crucial_ that no exception crates be dependencies |
| /// of the Rust runtime (std/test). |
| #[rustfmt::skip] |
| const EXCEPTIONS_STDLIB: ExceptionList = &[ |
| // tidy-alphabetical-start |
| ("fortanix-sgx-abi", "MPL-2.0"), // libstd but only for `sgx` target. FIXME: this dependency violates the documentation comment above. |
| // tidy-alphabetical-end |
| ]; |
| |
| const EXCEPTIONS_CARGO: ExceptionList = &[ |
| // tidy-alphabetical-start |
| ("bitmaps", "MPL-2.0+"), |
| ("im-rc", "MPL-2.0+"), |
| ("sized-chunks", "MPL-2.0+"), |
| // tidy-alphabetical-end |
| ]; |
| |
| const EXCEPTIONS_RUST_ANALYZER: ExceptionList = &[ |
| // tidy-alphabetical-start |
| ("option-ext", "MPL-2.0"), |
| // tidy-alphabetical-end |
| ]; |
| |
| const EXCEPTIONS_RUSTC_PERF: ExceptionList = &[ |
| // tidy-alphabetical-start |
| ("inferno", "CDDL-1.0"), |
| ("option-ext", "MPL-2.0"), |
| // tidy-alphabetical-end |
| ]; |
| |
| const EXCEPTIONS_RUSTBOOK: ExceptionList = &[ |
| // tidy-alphabetical-start |
| ("cssparser", "MPL-2.0"), |
| ("cssparser-macros", "MPL-2.0"), |
| ("dtoa-short", "MPL-2.0"), |
| ("mdbook", "MPL-2.0"), |
| // tidy-alphabetical-end |
| ]; |
| |
| const EXCEPTIONS_CRANELIFT: ExceptionList = &[]; |
| |
| const EXCEPTIONS_GCC: ExceptionList = &[ |
| // tidy-alphabetical-start |
| ("gccjit", "GPL-3.0"), |
| ("gccjit_sys", "GPL-3.0"), |
| // tidy-alphabetical-end |
| ]; |
| |
| const EXCEPTIONS_BOOTSTRAP: ExceptionList = &[]; |
| |
| const EXCEPTIONS_UEFI_QEMU_TEST: ExceptionList = &[]; |
| |
| #[derive(Clone, Copy)] |
| struct ListLocation { |
| path: &'static str, |
| line: u32, |
| } |
| |
| /// Creates a [`ListLocation`] for the current location (with an additional offset to the actual list start); |
| macro_rules! location { |
| (+ $offset:literal) => { |
| ListLocation { path: file!(), line: line!() + $offset } |
| }; |
| } |
| |
| const PERMITTED_RUSTC_DEPS_LOCATION: ListLocation = location!(+6); |
| |
| /// Crates rustc is allowed to depend on. Avoid adding to the list if possible. |
| /// |
| /// This list is here to provide a speed-bump to adding a new dependency to |
| /// rustc. Please check with the compiler team before adding an entry. |
| const PERMITTED_RUSTC_DEPENDENCIES: &[&str] = &[ |
| // tidy-alphabetical-start |
| "adler2", |
| "aho-corasick", |
| "allocator-api2", // FIXME: only appears in Cargo.lock due to https://github.com/rust-lang/cargo/issues/10801 |
| "annotate-snippets", |
| "anstream", |
| "anstyle", |
| "anstyle-parse", |
| "anstyle-query", |
| "anstyle-wincon", |
| "ar_archive_writer", |
| "arrayref", |
| "arrayvec", |
| "autocfg", |
| "bitflags", |
| "blake3", |
| "block-buffer", |
| "bstr", |
| "cc", |
| "cfg-if", |
| "cfg_aliases", |
| "colorchoice", |
| "constant_time_eq", |
| "cpufeatures", |
| "crc32fast", |
| "crossbeam-deque", |
| "crossbeam-epoch", |
| "crossbeam-utils", |
| "crypto-common", |
| "ctrlc", |
| "darling", |
| "darling_core", |
| "darling_macro", |
| "datafrog", |
| "derive-where", |
| "derive_setters", |
| "digest", |
| "displaydoc", |
| "dissimilar", |
| "dyn-clone", |
| "either", |
| "elsa", |
| "ena", |
| "equivalent", |
| "errno", |
| "expect-test", |
| "fallible-iterator", // dependency of `thorin` |
| "fastrand", |
| "find-msvc-tools", |
| "flate2", |
| "fluent-bundle", |
| "fluent-langneg", |
| "fluent-syntax", |
| "fnv", |
| "foldhash", |
| "generic-array", |
| "getopts", |
| "getrandom", |
| "gimli", |
| "gsgdt", |
| "hashbrown", |
| "icu_collections", |
| "icu_list", |
| "icu_locale", |
| "icu_locale_core", |
| "icu_locale_data", |
| "icu_provider", |
| "ident_case", |
| "indexmap", |
| "intl-memoizer", |
| "intl_pluralrules", |
| "is_terminal_polyfill", |
| "itertools", |
| "itoa", |
| "jiff", |
| "jiff-static", |
| "jobserver", |
| "lazy_static", |
| "leb128", |
| "libc", |
| "libloading", |
| "linux-raw-sys", |
| "litemap", |
| "lock_api", |
| "log", |
| "matchers", |
| "md-5", |
| "measureme", |
| "memchr", |
| "memmap2", |
| "miniz_oxide", |
| "nix", |
| "nu-ansi-term", |
| "object", |
| "odht", |
| "once_cell", |
| "once_cell_polyfill", |
| "overload", |
| "parking_lot", |
| "parking_lot_core", |
| "pathdiff", |
| "perf-event-open-sys", |
| "pin-project-lite", |
| "polonius-engine", |
| "portable-atomic", // dependency for platforms doesn't support `AtomicU64` in std |
| "portable-atomic-util", |
| "potential_utf", |
| "ppv-lite86", |
| "proc-macro-hack", |
| "proc-macro2", |
| "psm", |
| "pulldown-cmark", |
| "pulldown-cmark-escape", |
| "punycode", |
| "quote", |
| "r-efi", |
| "rand", |
| "rand_chacha", |
| "rand_core", |
| "rand_xorshift", // dependency for doc-tests in rustc_thread_pool |
| "rand_xoshiro", |
| "redox_syscall", |
| "ref-cast", |
| "ref-cast-impl", |
| "regex", |
| "regex-automata", |
| "regex-syntax", |
| "rustc-demangle", |
| "rustc-hash", |
| "rustc-literal-escaper", |
| "rustc-stable-hash", |
| "rustc_apfloat", |
| "rustix", |
| "ruzstd", // via object in thorin-dwp |
| "ryu", |
| "schemars", |
| "schemars_derive", |
| "scoped-tls", |
| "scopeguard", |
| "self_cell", |
| "serde", |
| "serde_core", |
| "serde_derive", |
| "serde_derive_internals", |
| "serde_json", |
| "serde_path_to_error", |
| "sha1", |
| "sha2", |
| "sharded-slab", |
| "shlex", |
| "smallvec", |
| "stable_deref_trait", |
| "stacker", |
| "static_assertions", |
| "strsim", |
| "syn", |
| "synstructure", |
| "tempfile", |
| "termize", |
| "thin-vec", |
| "thiserror", |
| "thiserror-impl", |
| "thorin-dwp", |
| "thread_local", |
| "tikv-jemalloc-sys", |
| "tinystr", |
| "tinyvec", |
| "tinyvec_macros", |
| "tracing", |
| "tracing-attributes", |
| "tracing-core", |
| "tracing-log", |
| "tracing-subscriber", |
| "tracing-tree", |
| "twox-hash", |
| "type-map", |
| "typenum", |
| "unic-langid", |
| "unic-langid-impl", |
| "unic-langid-macros", |
| "unic-langid-macros-impl", |
| "unicase", |
| "unicode-ident", |
| "unicode-normalization", |
| "unicode-properties", |
| "unicode-script", |
| "unicode-security", |
| "unicode-width", |
| "unicode-xid", |
| "utf8parse", |
| "valuable", |
| "version_check", |
| "wasi", |
| "wasm-encoder", |
| "wasmparser", |
| "winapi", |
| "winapi-i686-pc-windows-gnu", |
| "winapi-x86_64-pc-windows-gnu", |
| "windows", |
| "windows-collections", |
| "windows-core", |
| "windows-future", |
| "windows-implement", |
| "windows-interface", |
| "windows-link", |
| "windows-numerics", |
| "windows-result", |
| "windows-strings", |
| "windows-sys", |
| "windows-targets", |
| "windows-threading", |
| "windows_aarch64_gnullvm", |
| "windows_aarch64_msvc", |
| "windows_i686_gnu", |
| "windows_i686_gnullvm", |
| "windows_i686_msvc", |
| "windows_x86_64_gnu", |
| "windows_x86_64_gnullvm", |
| "windows_x86_64_msvc", |
| "wit-bindgen-rt@0.39.0", // pinned to a specific version due to using a binary blob: <https://github.com/rust-lang/rust/pull/136395#issuecomment-2692769062> |
| "writeable", |
| "yoke", |
| "yoke-derive", |
| "zerocopy", |
| "zerocopy-derive", |
| "zerofrom", |
| "zerofrom-derive", |
| "zerotrie", |
| "zerovec", |
| "zerovec-derive", |
| // tidy-alphabetical-end |
| ]; |
| |
| const PERMITTED_STDLIB_DEPS_LOCATION: ListLocation = location!(+2); |
| |
| const PERMITTED_STDLIB_DEPENDENCIES: &[&str] = &[ |
| // tidy-alphabetical-start |
| "addr2line", |
| "adler2", |
| "cc", |
| "cfg-if", |
| "compiler_builtins", |
| "dlmalloc", |
| "fortanix-sgx-abi", |
| "getopts", |
| "gimli", |
| "hashbrown", |
| "hermit-abi", |
| "libc", |
| "memchr", |
| "miniz_oxide", |
| "moto-rt", |
| "object", |
| "r-efi", |
| "r-efi-alloc", |
| "rand", |
| "rand_core", |
| "rand_xorshift", |
| "rustc-demangle", |
| "rustc-literal-escaper", |
| "shlex", |
| "unwinding", |
| "vex-sdk", |
| "wasi", |
| "windows-sys", |
| "windows-targets", |
| "windows_aarch64_gnullvm", |
| "windows_aarch64_msvc", |
| "windows_i686_gnu", |
| "windows_i686_gnullvm", |
| "windows_i686_msvc", |
| "windows_x86_64_gnu", |
| "windows_x86_64_gnullvm", |
| "windows_x86_64_msvc", |
| "wit-bindgen", |
| // tidy-alphabetical-end |
| ]; |
| |
| const PERMITTED_CRANELIFT_DEPS_LOCATION: ListLocation = location!(+2); |
| |
| const PERMITTED_CRANELIFT_DEPENDENCIES: &[&str] = &[ |
| // tidy-alphabetical-start |
| "allocator-api2", |
| "anyhow", |
| "arbitrary", |
| "bitflags", |
| "bumpalo", |
| "cfg-if", |
| "cranelift-assembler-x64", |
| "cranelift-assembler-x64-meta", |
| "cranelift-bforest", |
| "cranelift-bitset", |
| "cranelift-codegen", |
| "cranelift-codegen-meta", |
| "cranelift-codegen-shared", |
| "cranelift-control", |
| "cranelift-entity", |
| "cranelift-frontend", |
| "cranelift-isle", |
| "cranelift-jit", |
| "cranelift-module", |
| "cranelift-native", |
| "cranelift-object", |
| "cranelift-srcgen", |
| "crc32fast", |
| "equivalent", |
| "fallible-iterator", |
| "foldhash", |
| "gimli", |
| "hashbrown", |
| "indexmap", |
| "libc", |
| "libloading", |
| "libm", |
| "log", |
| "mach2", |
| "memchr", |
| "object", |
| "proc-macro2", |
| "quote", |
| "regalloc2", |
| "region", |
| "rustc-hash", |
| "serde", |
| "serde_derive", |
| "smallvec", |
| "stable_deref_trait", |
| "syn", |
| "target-lexicon", |
| "unicode-ident", |
| "wasmtime-jit-icache-coherence", |
| "wasmtime-math", |
| "windows-sys", |
| "windows-targets", |
| "windows_aarch64_gnullvm", |
| "windows_aarch64_msvc", |
| "windows_i686_gnu", |
| "windows_i686_gnullvm", |
| "windows_i686_msvc", |
| "windows_x86_64_gnu", |
| "windows_x86_64_gnullvm", |
| "windows_x86_64_msvc", |
| // tidy-alphabetical-end |
| ]; |
| |
| /// Dependency checks. |
| /// |
| /// `root` is path to the directory with the root `Cargo.toml` (for the workspace). `cargo` is path |
| /// to the cargo executable. |
| pub fn check(root: &Path, cargo: &Path, tidy_ctx: TidyCtx) { |
| let mut check = tidy_ctx.start_check("deps"); |
| let bless = tidy_ctx.is_bless_enabled(); |
| |
| let mut checked_runtime_licenses = false; |
| |
| check_proc_macro_dep_list(root, cargo, bless, &mut check); |
| |
| for &WorkspaceInfo { path, exceptions, crates_and_deps, submodules } in WORKSPACES { |
| if has_missing_submodule(root, submodules) { |
| continue; |
| } |
| |
| if !root.join(path).join("Cargo.lock").exists() { |
| check.error(format!("the `{path}` workspace doesn't have a Cargo.lock")); |
| continue; |
| } |
| |
| let mut cmd = cargo_metadata::MetadataCommand::new(); |
| cmd.cargo_path(cargo) |
| .manifest_path(root.join(path).join("Cargo.toml")) |
| .features(cargo_metadata::CargoOpt::AllFeatures) |
| .other_options(vec!["--locked".to_owned()]); |
| let metadata = t!(cmd.exec()); |
| |
| check_license_exceptions(&metadata, path, exceptions, &mut check); |
| if let Some((crates, permitted_deps, location)) = crates_and_deps { |
| let descr = crates.get(0).unwrap_or(&path); |
| check_permitted_dependencies( |
| &metadata, |
| descr, |
| permitted_deps, |
| crates, |
| location, |
| &mut check, |
| ); |
| } |
| |
| if path == "library" { |
| check_runtime_license_exceptions(&metadata, &mut check); |
| check_runtime_no_duplicate_dependencies(&metadata, &mut check); |
| check_runtime_no_proc_macros(&metadata, &mut check); |
| checked_runtime_licenses = true; |
| } |
| } |
| |
| // Sanity check to ensure we don't accidentally remove the workspace containing the runtime |
| // crates. |
| assert!(checked_runtime_licenses); |
| } |
| |
| /// Ensure the list of proc-macro crate transitive dependencies is up to date |
| fn check_proc_macro_dep_list(root: &Path, cargo: &Path, bless: bool, check: &mut RunningCheck) { |
| let mut cmd = cargo_metadata::MetadataCommand::new(); |
| cmd.cargo_path(cargo) |
| .manifest_path(root.join("Cargo.toml")) |
| .features(cargo_metadata::CargoOpt::AllFeatures) |
| .other_options(vec!["--locked".to_owned()]); |
| let metadata = t!(cmd.exec()); |
| let is_proc_macro_pkg = |pkg: &Package| pkg.targets.iter().any(|target| target.is_proc_macro()); |
| |
| let mut proc_macro_deps = HashSet::new(); |
| for pkg in metadata.packages.iter().filter(|pkg| is_proc_macro_pkg(pkg)) { |
| deps_of(&metadata, &pkg.id, &mut proc_macro_deps); |
| } |
| // Remove the proc-macro crates themselves |
| proc_macro_deps.retain(|pkg| !is_proc_macro_pkg(&metadata[pkg])); |
| |
| let proc_macro_deps: HashSet<_> = |
| proc_macro_deps.into_iter().map(|dep| metadata[dep].name.as_ref()).collect(); |
| let expected = proc_macro_deps::CRATES.iter().copied().collect::<HashSet<_>>(); |
| |
| let needs_blessing = proc_macro_deps.difference(&expected).next().is_some() |
| || expected.difference(&proc_macro_deps).next().is_some(); |
| |
| if needs_blessing && bless { |
| let mut proc_macro_deps: Vec<_> = proc_macro_deps.into_iter().collect(); |
| proc_macro_deps.sort(); |
| let mut file = File::create(root.join("src/bootstrap/src/utils/proc_macro_deps.rs")) |
| .expect("`proc_macro_deps` should exist"); |
| writeln!( |
| &mut file, |
| "/// Do not update manually - use `./x.py test tidy --bless` |
| /// Holds all direct and indirect dependencies of proc-macro crates in tree. |
| /// See <https://github.com/rust-lang/rust/issues/134863> |
| pub static CRATES: &[&str] = &[ |
| // tidy-alphabetical-start" |
| ) |
| .unwrap(); |
| for dep in proc_macro_deps { |
| writeln!(&mut file, " {dep:?},").unwrap(); |
| } |
| writeln!( |
| &mut file, |
| " // tidy-alphabetical-end |
| ];" |
| ) |
| .unwrap(); |
| } else { |
| let mut error_found = false; |
| |
| for missing in proc_macro_deps.difference(&expected) { |
| error_found = true; |
| check.error(format!( |
| "proc-macro crate dependency `{missing}` is not registered in `src/bootstrap/src/utils/proc_macro_deps.rs`", |
| )); |
| } |
| for extra in expected.difference(&proc_macro_deps) { |
| error_found = true; |
| check.error(format!( |
| "`{extra}` is registered in `src/bootstrap/src/utils/proc_macro_deps.rs`, but is not a proc-macro crate dependency", |
| )); |
| } |
| if error_found { |
| check.message("Run `./x.py test tidy --bless` to regenerate the list"); |
| } |
| } |
| } |
| |
| /// Used to skip a check if a submodule is not checked out, and not in a CI environment. |
| /// |
| /// This helps prevent enforcing developers to fetch submodules for tidy. |
| pub fn has_missing_submodule(root: &Path, submodules: &[&str]) -> bool { |
| !CiEnv::is_ci() |
| && submodules.iter().any(|submodule| { |
| let path = root.join(submodule); |
| !path.exists() |
| // If the directory is empty, we can consider it as an uninitialized submodule. |
| || read_dir(path).unwrap().next().is_none() |
| }) |
| } |
| |
| /// Check that all licenses of runtime dependencies are in the valid list in `LICENSES`. |
| /// |
| /// Unlike for tools we don't allow exceptions to the `LICENSES` list for the runtime with the sole |
| /// exception of `fortanix-sgx-abi` which is only used on x86_64-fortanix-unknown-sgx. |
| fn check_runtime_license_exceptions(metadata: &Metadata, check: &mut RunningCheck) { |
| for pkg in &metadata.packages { |
| if pkg.source.is_none() { |
| // No need to check local packages. |
| continue; |
| } |
| let license = match &pkg.license { |
| Some(license) => license, |
| None => { |
| check |
| .error(format!("dependency `{}` does not define a license expression", pkg.id)); |
| continue; |
| } |
| }; |
| if !LICENSES.contains(&license.as_str()) { |
| // This is a specific exception because SGX is considered "third party". |
| // See https://github.com/rust-lang/rust/issues/62620 for more. |
| // In general, these should never be added and this exception |
| // should not be taken as precedent for any new target. |
| if *pkg.name == "fortanix-sgx-abi" && pkg.license.as_deref() == Some("MPL-2.0") { |
| continue; |
| } |
| |
| check.error(format!("invalid license `{}` in `{}`", license, pkg.id)); |
| } |
| } |
| } |
| |
| /// Check that all licenses of tool dependencies are in the valid list in `LICENSES`. |
| /// |
| /// Packages listed in `exceptions` are allowed for tools. |
| fn check_license_exceptions( |
| metadata: &Metadata, |
| workspace: &str, |
| exceptions: &[(&str, &str)], |
| check: &mut RunningCheck, |
| ) { |
| // Validate the EXCEPTIONS list hasn't changed. |
| for (name, license) in exceptions { |
| // Check that the package actually exists. |
| if !metadata.packages.iter().any(|p| *p.name == *name) { |
| check.error(format!( |
| "could not find exception package `{name}` in workspace `{workspace}`\n\ |
| Remove from EXCEPTIONS list if it is no longer used.", |
| )); |
| } |
| // Check that the license hasn't changed. |
| for pkg in metadata.packages.iter().filter(|p| *p.name == *name) { |
| match &pkg.license { |
| None => { |
| check.error(format!( |
| "dependency exception `{}` in workspace `{workspace}` does not declare a license expression", |
| pkg.id |
| )); |
| } |
| Some(pkg_license) => { |
| if pkg_license.as_str() != *license { |
| check.error(format!(r#"dependency exception `{name}` license in workspace `{workspace}` has changed |
| previously `{license}` now `{pkg_license}` |
| update EXCEPTIONS for the new license |
| "#)); |
| } |
| } |
| } |
| } |
| if LICENSES.contains(license) || LICENSES_TOOLS.contains(license) { |
| check.error(format!( |
| "dependency exception `{name}` is not necessary. `{license}` is an allowed license" |
| )); |
| } |
| } |
| |
| let exception_names: Vec<_> = exceptions.iter().map(|(name, _license)| *name).collect(); |
| |
| // Check if any package does not have a valid license. |
| for pkg in &metadata.packages { |
| if pkg.source.is_none() { |
| // No need to check local packages. |
| continue; |
| } |
| if exception_names.contains(&pkg.name.as_str()) { |
| continue; |
| } |
| let license = match &pkg.license { |
| Some(license) => license, |
| None => { |
| check.error(format!( |
| "dependency `{}` in workspace `{workspace}` does not define a license expression", |
| pkg.id |
| )); |
| continue; |
| } |
| }; |
| if !LICENSES.contains(&license.as_str()) && !LICENSES_TOOLS.contains(&license.as_str()) { |
| check.error(format!( |
| "invalid license `{}` for package `{}` in workspace `{workspace}`", |
| license, pkg.id |
| )); |
| } |
| } |
| } |
| |
| fn check_runtime_no_duplicate_dependencies(metadata: &Metadata, check: &mut RunningCheck) { |
| let mut seen_pkgs = HashSet::new(); |
| for pkg in &metadata.packages { |
| if pkg.source.is_none() { |
| continue; |
| } |
| |
| // Skip the `wasi` crate here which the standard library explicitly |
| // depends on two version of (one for the `wasm32-wasip1` target and |
| // another for the `wasm32-wasip2` target). |
| if pkg.name.to_string() != "wasi" && !seen_pkgs.insert(&*pkg.name) { |
| check.error(format!( |
| "duplicate package `{}` is not allowed for the standard library", |
| pkg.name |
| )); |
| } |
| } |
| } |
| |
| fn check_runtime_no_proc_macros(metadata: &Metadata, check: &mut RunningCheck) { |
| for pkg in &metadata.packages { |
| if pkg.targets.iter().any(|target| target.is_proc_macro()) { |
| check.error(format!( |
| "proc macro `{}` is not allowed as standard library dependency.\n\ |
| Using proc macros in the standard library would break cross-compilation \ |
| as proc-macros don't get shipped for the host tuple.", |
| pkg.name |
| )); |
| } |
| } |
| } |
| |
| /// Checks the dependency of `restricted_dependency_crates` at the given path. Changes `bad` to |
| /// `true` if a check failed. |
| /// |
| /// Specifically, this checks that the dependencies are on the `permitted_dependencies`. |
| fn check_permitted_dependencies( |
| metadata: &Metadata, |
| descr: &str, |
| permitted_dependencies: &[&'static str], |
| restricted_dependency_crates: &[&'static str], |
| permitted_location: ListLocation, |
| check: &mut RunningCheck, |
| ) { |
| let mut has_permitted_dep_error = false; |
| let mut deps = HashSet::new(); |
| for to_check in restricted_dependency_crates { |
| let to_check = pkg_from_name(metadata, to_check); |
| deps_of(metadata, &to_check.id, &mut deps); |
| } |
| |
| // Check that the PERMITTED_DEPENDENCIES does not have unused entries. |
| for permitted in permitted_dependencies { |
| fn compare(pkg: &Package, permitted: &str) -> bool { |
| if let Some((name, version)) = permitted.split_once("@") { |
| let Ok(version) = Version::parse(version) else { |
| return false; |
| }; |
| *pkg.name == name && pkg.version == version |
| } else { |
| *pkg.name == permitted |
| } |
| } |
| if !deps.iter().any(|dep_id| compare(pkg_from_id(metadata, dep_id), permitted)) { |
| check.error(format!( |
| "could not find allowed package `{permitted}`\n\ |
| Remove from PERMITTED_DEPENDENCIES list if it is no longer used.", |
| )); |
| has_permitted_dep_error = true; |
| } |
| } |
| |
| // Get in a convenient form. |
| let permitted_dependencies: HashMap<_, _> = permitted_dependencies |
| .iter() |
| .map(|s| { |
| if let Some((name, version)) = s.split_once('@') { |
| (name, Version::parse(version).ok()) |
| } else { |
| (*s, None) |
| } |
| }) |
| .collect(); |
| |
| for dep in deps { |
| let dep = pkg_from_id(metadata, dep); |
| // If this path is in-tree, we don't require it to be explicitly permitted. |
| if dep.source.is_some() { |
| let is_eq = if let Some(version) = permitted_dependencies.get(dep.name.as_str()) { |
| if let Some(version) = version { version == &dep.version } else { true } |
| } else { |
| false |
| }; |
| if !is_eq { |
| check.error(format!("Dependency for {descr} not explicitly permitted: {}", dep.id)); |
| has_permitted_dep_error = true; |
| } |
| } |
| } |
| |
| if has_permitted_dep_error { |
| eprintln!("Go to `{}:{}` for the list.", permitted_location.path, permitted_location.line); |
| } |
| } |
| |
| /// Finds a package with the given name. |
| fn pkg_from_name<'a>(metadata: &'a Metadata, name: &'static str) -> &'a Package { |
| let mut i = metadata.packages.iter().filter(|p| *p.name == name); |
| let result = |
| i.next().unwrap_or_else(|| panic!("could not find package `{name}` in package list")); |
| assert!(i.next().is_none(), "more than one package found for `{name}`"); |
| result |
| } |
| |
| fn pkg_from_id<'a>(metadata: &'a Metadata, id: &PackageId) -> &'a Package { |
| metadata.packages.iter().find(|p| &p.id == id).unwrap() |
| } |
| |
| /// Recursively find all dependencies. |
| fn deps_of<'a>(metadata: &'a Metadata, pkg_id: &'a PackageId, result: &mut HashSet<&'a PackageId>) { |
| if !result.insert(pkg_id) { |
| return; |
| } |
| let node = metadata |
| .resolve |
| .as_ref() |
| .unwrap() |
| .nodes |
| .iter() |
| .find(|n| &n.id == pkg_id) |
| .unwrap_or_else(|| panic!("could not find `{pkg_id}` in resolve")); |
| for dep in &node.deps { |
| deps_of(metadata, &dep.pkg, result); |
| } |
| } |