mirror of
				https://github.com/rust-lang/rust.git
				synced 2025-10-31 04:57:19 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			250 lines
		
	
	
		
			8.1 KiB
		
	
	
	
		
			Rust
		
	
	
	
	
	
			
		
		
	
	
			250 lines
		
	
	
		
			8.1 KiB
		
	
	
	
		
			Rust
		
	
	
	
	
	
| //! The main parser interface.
 | |
| 
 | |
| // tidy-alphabetical-start
 | |
| #![allow(internal_features)]
 | |
| #![allow(rustc::diagnostic_outside_of_impl)]
 | |
| #![allow(rustc::untranslatable_diagnostic)]
 | |
| #![feature(array_windows)]
 | |
| #![feature(assert_matches)]
 | |
| #![feature(box_patterns)]
 | |
| #![feature(debug_closure_helpers)]
 | |
| #![feature(if_let_guard)]
 | |
| #![feature(iter_intersperse)]
 | |
| #![feature(let_chains)]
 | |
| #![feature(string_from_utf8_lossy_owned)]
 | |
| // tidy-alphabetical-end
 | |
| 
 | |
| use std::path::{Path, PathBuf};
 | |
| use std::str::Utf8Error;
 | |
| use std::sync::Arc;
 | |
| 
 | |
| use rustc_ast as ast;
 | |
| use rustc_ast::tokenstream::TokenStream;
 | |
| use rustc_ast::{AttrItem, Attribute, MetaItemInner, token};
 | |
| use rustc_ast_pretty::pprust;
 | |
| use rustc_errors::{Diag, EmissionGuarantee, FatalError, PResult, pluralize};
 | |
| use rustc_session::parse::ParseSess;
 | |
| use rustc_span::source_map::SourceMap;
 | |
| use rustc_span::{FileName, SourceFile, Span};
 | |
| pub use unicode_normalization::UNICODE_VERSION as UNICODE_NORMALIZATION_VERSION;
 | |
| 
 | |
| pub const MACRO_ARGUMENTS: Option<&str> = Some("macro arguments");
 | |
| 
 | |
| #[macro_use]
 | |
| pub mod parser;
 | |
| use parser::{Parser, make_unclosed_delims_error};
 | |
| pub mod lexer;
 | |
| pub mod validate_attr;
 | |
| 
 | |
| mod errors;
 | |
| 
 | |
| rustc_fluent_macro::fluent_messages! { "../messages.ftl" }
 | |
| 
 | |
| // Unwrap the result if `Ok`, otherwise emit the diagnostics and abort.
 | |
| pub fn unwrap_or_emit_fatal<T>(expr: Result<T, Vec<Diag<'_>>>) -> T {
 | |
|     match expr {
 | |
|         Ok(expr) => expr,
 | |
|         Err(errs) => {
 | |
|             for err in errs {
 | |
|                 err.emit();
 | |
|             }
 | |
|             FatalError.raise()
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| /// Creates a new parser from a source string. On failure, the errors must be consumed via
 | |
| /// `unwrap_or_emit_fatal`, `emit`, `cancel`, etc., otherwise a panic will occur when they are
 | |
| /// dropped.
 | |
| pub fn new_parser_from_source_str(
 | |
|     psess: &ParseSess,
 | |
|     name: FileName,
 | |
|     source: String,
 | |
| ) -> Result<Parser<'_>, Vec<Diag<'_>>> {
 | |
|     let source_file = psess.source_map().new_source_file(name, source);
 | |
|     new_parser_from_source_file(psess, source_file)
 | |
| }
 | |
| 
 | |
| /// Creates a new parser from a filename. On failure, the errors must be consumed via
 | |
| /// `unwrap_or_emit_fatal`, `emit`, `cancel`, etc., otherwise a panic will occur when they are
 | |
| /// dropped.
 | |
| ///
 | |
| /// If a span is given, that is used on an error as the source of the problem.
 | |
| pub fn new_parser_from_file<'a>(
 | |
|     psess: &'a ParseSess,
 | |
|     path: &Path,
 | |
|     sp: Option<Span>,
 | |
| ) -> Result<Parser<'a>, Vec<Diag<'a>>> {
 | |
|     let sm = psess.source_map();
 | |
|     let source_file = sm.load_file(path).unwrap_or_else(|e| {
 | |
|         let msg = format!("couldn't read `{}`: {}", path.display(), e);
 | |
|         let mut err = psess.dcx().struct_fatal(msg);
 | |
|         if let Ok(contents) = std::fs::read(path)
 | |
|             && let Err(utf8err) = String::from_utf8(contents.clone())
 | |
|         {
 | |
|             utf8_error(
 | |
|                 sm,
 | |
|                 &path.display().to_string(),
 | |
|                 sp,
 | |
|                 &mut err,
 | |
|                 utf8err.utf8_error(),
 | |
|                 &contents,
 | |
|             );
 | |
|         }
 | |
|         if let Some(sp) = sp {
 | |
|             err.span(sp);
 | |
|         }
 | |
|         err.emit();
 | |
|     });
 | |
|     new_parser_from_source_file(psess, source_file)
 | |
| }
 | |
| 
 | |
| pub fn utf8_error<E: EmissionGuarantee>(
 | |
|     sm: &SourceMap,
 | |
|     path: &str,
 | |
|     sp: Option<Span>,
 | |
|     err: &mut Diag<'_, E>,
 | |
|     utf8err: Utf8Error,
 | |
|     contents: &[u8],
 | |
| ) {
 | |
|     // The file exists, but it wasn't valid UTF-8.
 | |
|     let start = utf8err.valid_up_to();
 | |
|     let note = format!("invalid utf-8 at byte `{start}`");
 | |
|     let msg = if let Some(len) = utf8err.error_len() {
 | |
|         format!(
 | |
|             "byte{s} `{bytes}` {are} not valid utf-8",
 | |
|             bytes = if len == 1 {
 | |
|                 format!("{:?}", contents[start])
 | |
|             } else {
 | |
|                 format!("{:?}", &contents[start..start + len])
 | |
|             },
 | |
|             s = pluralize!(len),
 | |
|             are = if len == 1 { "is" } else { "are" },
 | |
|         )
 | |
|     } else {
 | |
|         note.clone()
 | |
|     };
 | |
|     let contents = String::from_utf8_lossy(contents).to_string();
 | |
|     let source = sm.new_source_file(PathBuf::from(path).into(), contents);
 | |
|     let span = Span::with_root_ctxt(
 | |
|         source.normalized_byte_pos(start as u32),
 | |
|         source.normalized_byte_pos(start as u32),
 | |
|     );
 | |
|     if span.is_dummy() {
 | |
|         err.note(note);
 | |
|     } else {
 | |
|         if sp.is_some() {
 | |
|             err.span_note(span, msg);
 | |
|         } else {
 | |
|             err.span(span);
 | |
|             err.span_label(span, msg);
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| /// Given a session and a `source_file`, return a parser. Returns any buffered errors from lexing
 | |
| /// the initial token stream.
 | |
| fn new_parser_from_source_file(
 | |
|     psess: &ParseSess,
 | |
|     source_file: Arc<SourceFile>,
 | |
| ) -> Result<Parser<'_>, Vec<Diag<'_>>> {
 | |
|     let end_pos = source_file.end_position();
 | |
|     let stream = source_file_to_stream(psess, source_file, None)?;
 | |
|     let mut parser = Parser::new(psess, stream, None);
 | |
|     if parser.token == token::Eof {
 | |
|         parser.token.span = Span::new(end_pos, end_pos, parser.token.span.ctxt(), None);
 | |
|     }
 | |
|     Ok(parser)
 | |
| }
 | |
| 
 | |
| pub fn source_str_to_stream(
 | |
|     psess: &ParseSess,
 | |
|     name: FileName,
 | |
|     source: String,
 | |
|     override_span: Option<Span>,
 | |
| ) -> Result<TokenStream, Vec<Diag<'_>>> {
 | |
|     let source_file = psess.source_map().new_source_file(name, source);
 | |
|     source_file_to_stream(psess, source_file, override_span)
 | |
| }
 | |
| 
 | |
| /// Given a source file, produces a sequence of token trees. Returns any buffered errors from
 | |
| /// parsing the token stream.
 | |
| fn source_file_to_stream<'psess>(
 | |
|     psess: &'psess ParseSess,
 | |
|     source_file: Arc<SourceFile>,
 | |
|     override_span: Option<Span>,
 | |
| ) -> Result<TokenStream, Vec<Diag<'psess>>> {
 | |
|     let src = source_file.src.as_ref().unwrap_or_else(|| {
 | |
|         psess.dcx().bug(format!(
 | |
|             "cannot lex `source_file` without source: {}",
 | |
|             psess.source_map().filename_for_diagnostics(&source_file.name)
 | |
|         ));
 | |
|     });
 | |
| 
 | |
|     lexer::lex_token_trees(psess, src.as_str(), source_file.start_pos, override_span)
 | |
| }
 | |
| 
 | |
| /// Runs the given subparser `f` on the tokens of the given `attr`'s item.
 | |
| pub fn parse_in<'a, T>(
 | |
|     psess: &'a ParseSess,
 | |
|     tts: TokenStream,
 | |
|     name: &'static str,
 | |
|     mut f: impl FnMut(&mut Parser<'a>) -> PResult<'a, T>,
 | |
| ) -> PResult<'a, T> {
 | |
|     let mut parser = Parser::new(psess, tts, Some(name));
 | |
|     let result = f(&mut parser)?;
 | |
|     if parser.token != token::Eof {
 | |
|         parser.unexpected()?;
 | |
|     }
 | |
|     Ok(result)
 | |
| }
 | |
| 
 | |
| pub fn fake_token_stream_for_item(psess: &ParseSess, item: &ast::Item) -> TokenStream {
 | |
|     let source = pprust::item_to_string(item);
 | |
|     let filename = FileName::macro_expansion_source_code(&source);
 | |
|     unwrap_or_emit_fatal(source_str_to_stream(psess, filename, source, Some(item.span)))
 | |
| }
 | |
| 
 | |
| pub fn fake_token_stream_for_crate(psess: &ParseSess, krate: &ast::Crate) -> TokenStream {
 | |
|     let source = pprust::crate_to_string_for_macros(krate);
 | |
|     let filename = FileName::macro_expansion_source_code(&source);
 | |
|     unwrap_or_emit_fatal(source_str_to_stream(
 | |
|         psess,
 | |
|         filename,
 | |
|         source,
 | |
|         Some(krate.spans.inner_span),
 | |
|     ))
 | |
| }
 | |
| 
 | |
| pub fn parse_cfg_attr(
 | |
|     cfg_attr: &Attribute,
 | |
|     psess: &ParseSess,
 | |
| ) -> Option<(MetaItemInner, Vec<(AttrItem, Span)>)> {
 | |
|     const CFG_ATTR_GRAMMAR_HELP: &str = "#[cfg_attr(condition, attribute, other_attribute, ...)]";
 | |
|     const CFG_ATTR_NOTE_REF: &str = "for more information, visit \
 | |
|         <https://doc.rust-lang.org/reference/conditional-compilation.html#the-cfg_attr-attribute>";
 | |
| 
 | |
|     match cfg_attr.get_normal_item().args {
 | |
|         ast::AttrArgs::Delimited(ast::DelimArgs { dspan, delim, ref tokens })
 | |
|             if !tokens.is_empty() =>
 | |
|         {
 | |
|             crate::validate_attr::check_cfg_attr_bad_delim(psess, dspan, delim);
 | |
|             match parse_in(psess, tokens.clone(), "`cfg_attr` input", |p| p.parse_cfg_attr()) {
 | |
|                 Ok(r) => return Some(r),
 | |
|                 Err(e) => {
 | |
|                     e.with_help(format!("the valid syntax is `{CFG_ATTR_GRAMMAR_HELP}`"))
 | |
|                         .with_note(CFG_ATTR_NOTE_REF)
 | |
|                         .emit();
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
|         _ => {
 | |
|             psess.dcx().emit_err(errors::MalformedCfgAttr {
 | |
|                 span: cfg_attr.span,
 | |
|                 sugg: CFG_ATTR_GRAMMAR_HELP,
 | |
|             });
 | |
|         }
 | |
|     }
 | |
|     None
 | |
| }
 | 
