@@ -8,18 +8,18 @@ use crate::reimplementation::tokenisation::processing::escape_processing::{
88} ;
99use crate :: tokens_common:: { NumericBase , Origin } ;
1010
11- use super :: tokens_matching:: { Nonterminal , TokenKindMatch } ;
11+ use super :: tokens_matching:: { Nonterminal , TokenisationMatch } ;
1212
1313mod escape_processing;
1414
1515/// Converts a match to a fine-grained token, or rejects the match.
1616///
17- /// This is the "Processing a match" stage of extracting a fine-grained token .
17+ /// This implements "Processing a tokenisation nonterminal match" .
1818///
1919/// If the match is accepted, returns a fine-grained token.
2020///
21- /// If the match is rejected , distinguishes rejection from "model error".
22- pub fn process ( match_data : & TokenKindMatch ) -> Result < FineToken , Error > {
21+ /// If the match is not accepted , distinguishes rejection from "model error".
22+ pub fn process ( match_data : & TokenisationMatch ) -> Result < FineToken , Error > {
2323 let token_data = match match_data. matched_nonterminal {
2424 Nonterminal :: Whitespace => process_whitespace ( match_data) ?,
2525 Nonterminal :: Line_comment => process_line_comment ( match_data) ?,
@@ -54,7 +54,7 @@ pub fn process(match_data: &TokenKindMatch) -> Result<FineToken, Error> {
5454 match_data. matched_nonterminal
5555 ) ) ) ;
5656 }
57- _ => return model_error ( "unhandled token-kind nonterminal" ) ,
57+ _ => return model_error ( "unhandled tokenisation nonterminal" ) ,
5858 } ;
5959 Ok ( FineToken {
6060 data : token_data,
@@ -97,7 +97,7 @@ impl From<escape_processing::Error> for Error {
9797 }
9898}
9999
100- impl TokenKindMatch {
100+ impl TokenisationMatch {
101101 /// Returns the characters consumed by the specified subsidiary nonterminal, or None if that
102102 /// nonterminal did not participate in the match.
103103 ///
@@ -138,11 +138,11 @@ impl TokenKindMatch {
138138 }
139139}
140140
141- fn process_whitespace ( _m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
141+ fn process_whitespace ( _m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
142142 Ok ( FineTokenData :: Whitespace )
143143}
144144
145- fn process_line_comment ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
145+ fn process_line_comment ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
146146 let comment_content = m. consumed ( Nonterminal :: LINE_COMMENT_CONTENT ) ?;
147147 let ( style, body) = match comment_content. chars ( ) {
148148 [ '/' , '/' , ..] => ( CommentStyle :: NonDoc , & [ ] as & [ char ] ) ,
@@ -159,7 +159,7 @@ fn process_line_comment(m: &TokenKindMatch) -> Result<FineTokenData, Error> {
159159 } )
160160}
161161
162- fn process_block_comment ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
162+ fn process_block_comment ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
163163 let comment_content = m. consumed_by_first_participating ( Nonterminal :: BLOCK_COMMENT_CONTENT ) ?;
164164 let ( style, body) = match comment_content. chars ( ) {
165165 [ '*' , '*' , ..] => ( CommentStyle :: NonDoc , & [ ] as & [ char ] ) ,
@@ -176,7 +176,7 @@ fn process_block_comment(m: &TokenKindMatch) -> Result<FineTokenData, Error> {
176176 } )
177177}
178178
179- fn process_character_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
179+ fn process_character_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
180180 use LiteralComponent :: * ;
181181 let single_quoted_content = m. consumed ( Nonterminal :: SINGLE_QUOTED_CONTENT ) ?;
182182 let single_escape_interpretation =
@@ -209,7 +209,7 @@ fn process_character_literal(m: &TokenKindMatch) -> Result<FineTokenData, Error>
209209 } )
210210}
211211
212- fn process_byte_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
212+ fn process_byte_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
213213 use LiteralComponent :: * ;
214214 let single_quoted_content = m. consumed ( Nonterminal :: SINGLE_QUOTED_CONTENT ) ?;
215215 let single_escape_interpretation =
@@ -246,7 +246,7 @@ fn process_byte_literal(m: &TokenKindMatch) -> Result<FineTokenData, Error> {
246246 } )
247247}
248248
249- fn process_string_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
249+ fn process_string_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
250250 use LiteralComponent :: * ;
251251 let double_quoted_content = m. consumed ( Nonterminal :: DOUBLE_QUOTED_CONTENT ) ?;
252252 let escape_interpretation = match try_escape_interpretation ( double_quoted_content) ? {
@@ -282,7 +282,7 @@ fn process_string_literal(m: &TokenKindMatch) -> Result<FineTokenData, Error> {
282282 } )
283283}
284284
285- fn process_byte_string_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
285+ fn process_byte_string_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
286286 use LiteralComponent :: * ;
287287 let double_quoted_content = m. consumed ( Nonterminal :: DOUBLE_QUOTED_CONTENT ) ?;
288288 let escape_interpretation = match try_escape_interpretation ( double_quoted_content) ? {
@@ -321,7 +321,7 @@ fn process_byte_string_literal(m: &TokenKindMatch) -> Result<FineTokenData, Erro
321321 } )
322322}
323323
324- fn process_c_string_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
324+ fn process_c_string_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
325325 use LiteralComponent :: * ;
326326 let double_quoted_content = m. consumed ( Nonterminal :: DOUBLE_QUOTED_CONTENT ) ?;
327327 let escape_interpretation = match try_escape_interpretation ( double_quoted_content) ? {
@@ -377,7 +377,7 @@ fn process_c_string_literal(m: &TokenKindMatch) -> Result<FineTokenData, Error>
377377 } )
378378}
379379
380- fn process_raw_string_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
380+ fn process_raw_string_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
381381 let raw_double_quoted_content = m. consumed ( Nonterminal :: RAW_DOUBLE_QUOTED_CONTENT ) ?. clone ( ) ;
382382 if raw_double_quoted_content. contains ( '\u{000d}' ) {
383383 return rejected ( "CR non-escape" ) ;
@@ -392,7 +392,7 @@ fn process_raw_string_literal(m: &TokenKindMatch) -> Result<FineTokenData, Error
392392 } )
393393}
394394
395- fn process_raw_byte_string_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
395+ fn process_raw_byte_string_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
396396 let raw_double_quoted_content = m. consumed ( Nonterminal :: RAW_DOUBLE_QUOTED_CONTENT ) ?;
397397 if raw_double_quoted_content. scalar_values ( ) . any ( |n| n > 127 ) {
398398 return rejected ( "non-ASCII character" ) ;
@@ -414,7 +414,7 @@ fn process_raw_byte_string_literal(m: &TokenKindMatch) -> Result<FineTokenData,
414414 } )
415415}
416416
417- fn process_raw_c_string_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
417+ fn process_raw_c_string_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
418418 let raw_double_quoted_content = m. consumed ( Nonterminal :: RAW_DOUBLE_QUOTED_CONTENT ) ?;
419419 if raw_double_quoted_content. contains ( '\u{000d}' ) {
420420 return rejected ( "CR in raw content" ) ;
@@ -433,7 +433,7 @@ fn process_raw_c_string_literal(m: &TokenKindMatch) -> Result<FineTokenData, Err
433433 } )
434434}
435435
436- fn process_float_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
436+ fn process_float_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
437437 let body = match (
438438 m. maybe_consumed ( Nonterminal :: FLOAT_BODY_WITH_EXPONENT ) ?,
439439 m. maybe_consumed ( Nonterminal :: FLOAT_BODY_WITHOUT_EXPONENT ) ?,
@@ -453,7 +453,7 @@ fn process_float_literal(m: &TokenKindMatch) -> Result<FineTokenData, Error> {
453453 } )
454454}
455455
456- fn process_integer_literal ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
456+ fn process_integer_literal ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
457457 let base = match (
458458 m. maybe_consumed ( Nonterminal :: INTEGER_BINARY_LITERAL ) ?,
459459 m. maybe_consumed ( Nonterminal :: INTEGER_OCTAL_LITERAL ) ?,
@@ -504,7 +504,7 @@ fn process_integer_literal(m: &TokenKindMatch) -> Result<FineTokenData, Error> {
504504 } )
505505}
506506
507- fn process_raw_lifetime_or_label ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
507+ fn process_raw_lifetime_or_label ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
508508 let name = m. consumed ( Nonterminal :: IDENT ) ?. clone ( ) ;
509509 let s = name. to_string ( ) ;
510510 if s == "_" || s == "crate" || s == "self" || s == "super" || s == "Self" {
@@ -513,12 +513,12 @@ fn process_raw_lifetime_or_label(m: &TokenKindMatch) -> Result<FineTokenData, Er
513513 Ok ( FineTokenData :: RawLifetimeOrLabel { name } )
514514}
515515
516- fn process_lifetime_or_label ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
516+ fn process_lifetime_or_label ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
517517 let name = m. consumed ( Nonterminal :: IDENT ) ?. clone ( ) ;
518518 Ok ( FineTokenData :: LifetimeOrLabel { name } )
519519}
520520
521- fn process_raw_ident ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
521+ fn process_raw_ident ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
522522 let represented_ident = m. consumed ( Nonterminal :: IDENT ) ?. nfc ( ) ;
523523 let s = represented_ident. to_string ( ) ;
524524 if s == "_" || s == "crate" || s == "self" || s == "super" || s == "Self" {
@@ -527,13 +527,13 @@ fn process_raw_ident(m: &TokenKindMatch) -> Result<FineTokenData, Error> {
527527 Ok ( FineTokenData :: RawIdent { represented_ident } )
528528}
529529
530- fn process_ident ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
530+ fn process_ident ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
531531 Ok ( FineTokenData :: Ident {
532532 represented_ident : m. consumed ( Nonterminal :: IDENT ) ?. nfc ( ) ,
533533 } )
534534}
535535
536- fn process_punctuation ( m : & TokenKindMatch ) -> Result < FineTokenData , Error > {
536+ fn process_punctuation ( m : & TokenisationMatch ) -> Result < FineTokenData , Error > {
537537 let mark = match m. consumed . chars ( ) {
538538 [ c] => * c,
539539 _ => return rejected ( "impossible Punctuation match" ) ,
0 commit comments