• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 #![allow(
2     clippy::assertions_on_result_states,
3     clippy::items_after_statements,
4     clippy::needless_pass_by_value,
5     clippy::needless_raw_string_hashes,
6     clippy::non_ascii_literal,
7     clippy::octal_escapes
8 )]
9 
10 use proc_macro2::{Ident, Literal, Punct, Spacing, Span, TokenStream, TokenTree};
11 use std::ffi::CStr;
12 use std::iter;
13 use std::str::{self, FromStr};
14 
15 #[test]
idents()16 fn idents() {
17     assert_eq!(
18         Ident::new("String", Span::call_site()).to_string(),
19         "String"
20     );
21     assert_eq!(Ident::new("fn", Span::call_site()).to_string(), "fn");
22     assert_eq!(Ident::new("_", Span::call_site()).to_string(), "_");
23 }
24 
25 #[test]
raw_idents()26 fn raw_idents() {
27     assert_eq!(
28         Ident::new_raw("String", Span::call_site()).to_string(),
29         "r#String"
30     );
31     assert_eq!(Ident::new_raw("fn", Span::call_site()).to_string(), "r#fn");
32 }
33 
34 #[test]
35 #[should_panic(expected = "`r#_` cannot be a raw identifier")]
ident_raw_underscore()36 fn ident_raw_underscore() {
37     Ident::new_raw("_", Span::call_site());
38 }
39 
40 #[test]
41 #[should_panic(expected = "`r#super` cannot be a raw identifier")]
ident_raw_reserved()42 fn ident_raw_reserved() {
43     Ident::new_raw("super", Span::call_site());
44 }
45 
46 #[test]
47 #[should_panic(expected = "Ident is not allowed to be empty; use Option<Ident>")]
ident_empty()48 fn ident_empty() {
49     Ident::new("", Span::call_site());
50 }
51 
52 #[test]
53 #[should_panic(expected = "Ident cannot be a number; use Literal instead")]
ident_number()54 fn ident_number() {
55     Ident::new("255", Span::call_site());
56 }
57 
58 #[test]
59 #[should_panic(expected = "\"a#\" is not a valid Ident")]
ident_invalid()60 fn ident_invalid() {
61     Ident::new("a#", Span::call_site());
62 }
63 
64 #[test]
65 #[should_panic(expected = "not a valid Ident")]
raw_ident_empty()66 fn raw_ident_empty() {
67     Ident::new("r#", Span::call_site());
68 }
69 
70 #[test]
71 #[should_panic(expected = "not a valid Ident")]
raw_ident_number()72 fn raw_ident_number() {
73     Ident::new("r#255", Span::call_site());
74 }
75 
76 #[test]
77 #[should_panic(expected = "\"r#a#\" is not a valid Ident")]
raw_ident_invalid()78 fn raw_ident_invalid() {
79     Ident::new("r#a#", Span::call_site());
80 }
81 
82 #[test]
83 #[should_panic(expected = "not a valid Ident")]
lifetime_empty()84 fn lifetime_empty() {
85     Ident::new("'", Span::call_site());
86 }
87 
88 #[test]
89 #[should_panic(expected = "not a valid Ident")]
lifetime_number()90 fn lifetime_number() {
91     Ident::new("'255", Span::call_site());
92 }
93 
94 #[test]
95 #[should_panic(expected = r#""'a#" is not a valid Ident"#)]
lifetime_invalid()96 fn lifetime_invalid() {
97     Ident::new("'a#", Span::call_site());
98 }
99 
100 #[test]
literal_string()101 fn literal_string() {
102     #[track_caller]
103     fn assert(literal: Literal, expected: &str) {
104         assert_eq!(literal.to_string(), expected.trim());
105     }
106 
107     assert(Literal::string(""), r#"  ""  "#);
108     assert(Literal::string("aA"), r#"  "aA"  "#);
109     assert(Literal::string("\t"), r#"  "\t"  "#);
110     assert(Literal::string("❤"), r#"  "❤"  "#);
111     assert(Literal::string("'"), r#"  "'"  "#);
112     assert(Literal::string("\""), r#"  "\""  "#);
113     assert(Literal::string("\0"), r#"  "\0"  "#);
114     assert(Literal::string("\u{1}"), r#"  "\u{1}"  "#);
115     assert(
116         Literal::string("a\00b\07c\08d\0e\0"),
117         r#"  "a\x000b\x007c\08d\0e\0"  "#,
118     );
119 
120     "\"\\\r\n    x\"".parse::<TokenStream>().unwrap();
121     "\"\\\r\n  \rx\"".parse::<TokenStream>().unwrap_err();
122 }
123 
124 #[test]
literal_raw_string()125 fn literal_raw_string() {
126     "r\"\r\n\"".parse::<TokenStream>().unwrap();
127 
128     fn raw_string_literal_with_hashes(n: usize) -> String {
129         let mut literal = String::new();
130         literal.push('r');
131         literal.extend(iter::repeat('#').take(n));
132         literal.push('"');
133         literal.push('"');
134         literal.extend(iter::repeat('#').take(n));
135         literal
136     }
137 
138     raw_string_literal_with_hashes(255)
139         .parse::<TokenStream>()
140         .unwrap();
141 
142     // https://github.com/rust-lang/rust/pull/95251
143     raw_string_literal_with_hashes(256)
144         .parse::<TokenStream>()
145         .unwrap_err();
146 }
147 
148 #[test]
literal_byte_character()149 fn literal_byte_character() {
150     #[track_caller]
151     fn assert(literal: Literal, expected: &str) {
152         assert_eq!(literal.to_string(), expected.trim());
153     }
154 
155     assert(Literal::byte_character(b'a'), r#"  b'a'  "#);
156     assert(Literal::byte_character(b'\0'), r#"  b'\0'  "#);
157     assert(Literal::byte_character(b'\t'), r#"  b'\t'  "#);
158     assert(Literal::byte_character(b'\n'), r#"  b'\n'  "#);
159     assert(Literal::byte_character(b'\r'), r#"  b'\r'  "#);
160     assert(Literal::byte_character(b'\''), r#"  b'\''  "#);
161     assert(Literal::byte_character(b'\\'), r#"  b'\\'  "#);
162     assert(Literal::byte_character(b'\x1f'), r#"  b'\x1F'  "#);
163     assert(Literal::byte_character(b'"'), r#"  b'"'  "#);
164 }
165 
166 #[test]
literal_byte_string()167 fn literal_byte_string() {
168     #[track_caller]
169     fn assert(literal: Literal, expected: &str) {
170         assert_eq!(literal.to_string(), expected.trim());
171     }
172 
173     assert(Literal::byte_string(b""), r#"  b""  "#);
174     assert(Literal::byte_string(b"\0"), r#"  b"\0"  "#);
175     assert(Literal::byte_string(b"\t"), r#"  b"\t"  "#);
176     assert(Literal::byte_string(b"\n"), r#"  b"\n"  "#);
177     assert(Literal::byte_string(b"\r"), r#"  b"\r"  "#);
178     assert(Literal::byte_string(b"\""), r#"  b"\""  "#);
179     assert(Literal::byte_string(b"\\"), r#"  b"\\"  "#);
180     assert(Literal::byte_string(b"\x1f"), r#"  b"\x1F"  "#);
181     assert(Literal::byte_string(b"'"), r#"  b"'"  "#);
182     assert(
183         Literal::byte_string(b"a\00b\07c\08d\0e\0"),
184         r#"  b"a\x000b\x007c\08d\0e\0"  "#,
185     );
186 
187     "b\"\\\r\n    x\"".parse::<TokenStream>().unwrap();
188     "b\"\\\r\n  \rx\"".parse::<TokenStream>().unwrap_err();
189     "b\"\\\r\n  \u{a0}x\"".parse::<TokenStream>().unwrap_err();
190     "br\"\u{a0}\"".parse::<TokenStream>().unwrap_err();
191 }
192 
193 #[test]
literal_c_string()194 fn literal_c_string() {
195     #[track_caller]
196     fn assert(literal: Literal, expected: &str) {
197         assert_eq!(literal.to_string(), expected.trim());
198     }
199 
200     assert(Literal::c_string(<&CStr>::default()), r#"  c""  "#);
201     assert(
202         Literal::c_string(CStr::from_bytes_with_nul(b"aA\0").unwrap()),
203         r#"  c"aA"  "#,
204     );
205     assert(
206         Literal::c_string(CStr::from_bytes_with_nul(b"aA\0").unwrap()),
207         r#"  c"aA"  "#,
208     );
209     assert(
210         Literal::c_string(CStr::from_bytes_with_nul(b"\t\0").unwrap()),
211         r#"  c"\t"  "#,
212     );
213     assert(
214         Literal::c_string(CStr::from_bytes_with_nul(b"\xE2\x9D\xA4\0").unwrap()),
215         r#"  c"❤"  "#,
216     );
217     assert(
218         Literal::c_string(CStr::from_bytes_with_nul(b"'\0").unwrap()),
219         r#"  c"'"  "#,
220     );
221     assert(
222         Literal::c_string(CStr::from_bytes_with_nul(b"\"\0").unwrap()),
223         r#"  c"\""  "#,
224     );
225     assert(
226         Literal::c_string(CStr::from_bytes_with_nul(b"\x7F\xFF\xFE\xCC\xB3\0").unwrap()),
227         r#"  c"\u{7f}\xFF\xFE\u{333}"  "#,
228     );
229 
230     let strings = r###"
231         c"hello\x80我叫\u{1F980}"  // from the RFC
232         cr"\"
233         cr##"Hello "world"!"##
234         c"\t\n\r\"\\"
235     "###;
236 
237     let mut tokens = strings.parse::<TokenStream>().unwrap().into_iter();
238 
239     for expected in &[
240         r#"c"hello\x80我叫\u{1F980}""#,
241         r#"cr"\""#,
242         r###"cr##"Hello "world"!"##"###,
243         r#"c"\t\n\r\"\\""#,
244     ] {
245         match tokens.next().unwrap() {
246             TokenTree::Literal(literal) => {
247                 assert_eq!(literal.to_string(), *expected);
248             }
249             unexpected => panic!("unexpected token: {:?}", unexpected),
250         }
251     }
252 
253     if let Some(unexpected) = tokens.next() {
254         panic!("unexpected token: {:?}", unexpected);
255     }
256 
257     for invalid in &[r#"c"\0""#, r#"c"\x00""#, r#"c"\u{0}""#, "c\"\0\""] {
258         if let Ok(unexpected) = invalid.parse::<TokenStream>() {
259             panic!("unexpected token: {:?}", unexpected);
260         }
261     }
262 }
263 
264 #[test]
literal_character()265 fn literal_character() {
266     #[track_caller]
267     fn assert(literal: Literal, expected: &str) {
268         assert_eq!(literal.to_string(), expected.trim());
269     }
270 
271     assert(Literal::character('a'), r#"  'a'  "#);
272     assert(Literal::character('\t'), r#"  '\t'  "#);
273     assert(Literal::character('❤'), r#"  '❤'  "#);
274     assert(Literal::character('\''), r#"  '\''  "#);
275     assert(Literal::character('"'), r#"  '"'  "#);
276     assert(Literal::character('\0'), r#"  '\0'  "#);
277     assert(Literal::character('\u{1}'), r#"  '\u{1}'  "#);
278 }
279 
280 #[test]
literal_integer()281 fn literal_integer() {
282     #[track_caller]
283     fn assert(literal: Literal, expected: &str) {
284         assert_eq!(literal.to_string(), expected);
285     }
286 
287     assert(Literal::u8_suffixed(10), "10u8");
288     assert(Literal::u16_suffixed(10), "10u16");
289     assert(Literal::u32_suffixed(10), "10u32");
290     assert(Literal::u64_suffixed(10), "10u64");
291     assert(Literal::u128_suffixed(10), "10u128");
292     assert(Literal::usize_suffixed(10), "10usize");
293 
294     assert(Literal::i8_suffixed(10), "10i8");
295     assert(Literal::i16_suffixed(10), "10i16");
296     assert(Literal::i32_suffixed(10), "10i32");
297     assert(Literal::i64_suffixed(10), "10i64");
298     assert(Literal::i128_suffixed(10), "10i128");
299     assert(Literal::isize_suffixed(10), "10isize");
300 
301     assert(Literal::u8_unsuffixed(10), "10");
302     assert(Literal::u16_unsuffixed(10), "10");
303     assert(Literal::u32_unsuffixed(10), "10");
304     assert(Literal::u64_unsuffixed(10), "10");
305     assert(Literal::u128_unsuffixed(10), "10");
306     assert(Literal::usize_unsuffixed(10), "10");
307 
308     assert(Literal::i8_unsuffixed(10), "10");
309     assert(Literal::i16_unsuffixed(10), "10");
310     assert(Literal::i32_unsuffixed(10), "10");
311     assert(Literal::i64_unsuffixed(10), "10");
312     assert(Literal::i128_unsuffixed(10), "10");
313     assert(Literal::isize_unsuffixed(10), "10");
314 
315     assert(Literal::i32_suffixed(-10), "-10i32");
316     assert(Literal::i32_unsuffixed(-10), "-10");
317 }
318 
319 #[test]
literal_float()320 fn literal_float() {
321     #[track_caller]
322     fn assert(literal: Literal, expected: &str) {
323         assert_eq!(literal.to_string(), expected);
324     }
325 
326     assert(Literal::f32_suffixed(10.0), "10f32");
327     assert(Literal::f32_suffixed(-10.0), "-10f32");
328     assert(Literal::f64_suffixed(10.0), "10f64");
329     assert(Literal::f64_suffixed(-10.0), "-10f64");
330 
331     assert(Literal::f32_unsuffixed(10.0), "10.0");
332     assert(Literal::f32_unsuffixed(-10.0), "-10.0");
333     assert(Literal::f64_unsuffixed(10.0), "10.0");
334     assert(Literal::f64_unsuffixed(-10.0), "-10.0");
335 
336     assert(
337         Literal::f64_unsuffixed(1e100),
338         "10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000.0",
339     );
340 }
341 
342 #[test]
literal_suffix()343 fn literal_suffix() {
344     fn token_count(p: &str) -> usize {
345         p.parse::<TokenStream>().unwrap().into_iter().count()
346     }
347 
348     assert_eq!(token_count("999u256"), 1);
349     assert_eq!(token_count("999r#u256"), 3);
350     assert_eq!(token_count("1."), 1);
351     assert_eq!(token_count("1.f32"), 3);
352     assert_eq!(token_count("1.0_0"), 1);
353     assert_eq!(token_count("1._0"), 3);
354     assert_eq!(token_count("1._m"), 3);
355     assert_eq!(token_count("\"\"s"), 1);
356     assert_eq!(token_count("r\"\"r"), 1);
357     assert_eq!(token_count("r#\"\"#r"), 1);
358     assert_eq!(token_count("b\"\"b"), 1);
359     assert_eq!(token_count("br\"\"br"), 1);
360     assert_eq!(token_count("br#\"\"#br"), 1);
361     assert_eq!(token_count("c\"\"c"), 1);
362     assert_eq!(token_count("cr\"\"cr"), 1);
363     assert_eq!(token_count("cr#\"\"#cr"), 1);
364     assert_eq!(token_count("'c'c"), 1);
365     assert_eq!(token_count("b'b'b"), 1);
366     assert_eq!(token_count("0E"), 1);
367     assert_eq!(token_count("0o0A"), 1);
368     assert_eq!(token_count("0E--0"), 4);
369     assert_eq!(token_count("0.0ECMA"), 1);
370 }
371 
372 #[test]
literal_iter_negative()373 fn literal_iter_negative() {
374     let negative_literal = Literal::i32_suffixed(-3);
375     let tokens = TokenStream::from(TokenTree::Literal(negative_literal));
376     let mut iter = tokens.into_iter();
377     match iter.next().unwrap() {
378         TokenTree::Punct(punct) => {
379             assert_eq!(punct.as_char(), '-');
380             assert_eq!(punct.spacing(), Spacing::Alone);
381         }
382         unexpected => panic!("unexpected token {:?}", unexpected),
383     }
384     match iter.next().unwrap() {
385         TokenTree::Literal(literal) => {
386             assert_eq!(literal.to_string(), "3i32");
387         }
388         unexpected => panic!("unexpected token {:?}", unexpected),
389     }
390     assert!(iter.next().is_none());
391 }
392 
393 #[test]
literal_parse()394 fn literal_parse() {
395     assert!("1".parse::<Literal>().is_ok());
396     assert!("-1".parse::<Literal>().is_ok());
397     assert!("-1u12".parse::<Literal>().is_ok());
398     assert!("1.0".parse::<Literal>().is_ok());
399     assert!("-1.0".parse::<Literal>().is_ok());
400     assert!("-1.0f12".parse::<Literal>().is_ok());
401     assert!("'a'".parse::<Literal>().is_ok());
402     assert!("\"\n\"".parse::<Literal>().is_ok());
403     assert!("0 1".parse::<Literal>().is_err());
404     assert!(" 0".parse::<Literal>().is_err());
405     assert!("0 ".parse::<Literal>().is_err());
406     assert!("/* comment */0".parse::<Literal>().is_err());
407     assert!("0/* comment */".parse::<Literal>().is_err());
408     assert!("0// comment".parse::<Literal>().is_err());
409     assert!("- 1".parse::<Literal>().is_err());
410     assert!("- 1.0".parse::<Literal>().is_err());
411     assert!("-\"\"".parse::<Literal>().is_err());
412 }
413 
414 #[test]
literal_span()415 fn literal_span() {
416     let positive = "0.1".parse::<Literal>().unwrap();
417     let negative = "-0.1".parse::<Literal>().unwrap();
418     let subspan = positive.subspan(1..2);
419 
420     #[cfg(not(span_locations))]
421     {
422         let _ = negative;
423         assert!(subspan.is_none());
424     }
425 
426     #[cfg(span_locations)]
427     {
428         assert_eq!(positive.span().start().column, 0);
429         assert_eq!(positive.span().end().column, 3);
430         assert_eq!(negative.span().start().column, 0);
431         assert_eq!(negative.span().end().column, 4);
432         assert_eq!(subspan.unwrap().source_text().unwrap(), ".");
433     }
434 
435     assert!(positive.subspan(1..4).is_none());
436 }
437 
438 #[cfg(span_locations)]
439 #[test]
source_text()440 fn source_text() {
441     let input = "    �� a z    ";
442     let mut tokens = input
443         .parse::<proc_macro2::TokenStream>()
444         .unwrap()
445         .into_iter();
446 
447     let first = tokens.next().unwrap();
448     assert_eq!("��", first.span().source_text().unwrap());
449 
450     let second = tokens.next().unwrap();
451     let third = tokens.next().unwrap();
452     assert_eq!("z", third.span().source_text().unwrap());
453     assert_eq!("a", second.span().source_text().unwrap());
454 }
455 
456 #[test]
roundtrip()457 fn roundtrip() {
458     fn roundtrip(p: &str) {
459         println!("parse: {}", p);
460         let s = p.parse::<TokenStream>().unwrap().to_string();
461         println!("first: {}", s);
462         let s2 = s.parse::<TokenStream>().unwrap().to_string();
463         assert_eq!(s, s2);
464     }
465     roundtrip("a");
466     roundtrip("<<");
467     roundtrip("<<=");
468     roundtrip(
469         "
470         1
471         1.0
472         1f32
473         2f64
474         1usize
475         4isize
476         4e10
477         1_000
478         1_0i32
479         8u8
480         9
481         0
482         0xffffffffffffffffffffffffffffffff
483         1x
484         1u80
485         1f320
486     ",
487     );
488     roundtrip("'a");
489     roundtrip("'_");
490     roundtrip("'static");
491     roundtrip(r"'\u{10__FFFF}'");
492     roundtrip("\"\\u{10_F0FF__}foo\\u{1_0_0_0__}\"");
493 }
494 
495 #[test]
fail()496 fn fail() {
497     fn fail(p: &str) {
498         if let Ok(s) = p.parse::<TokenStream>() {
499             panic!("should have failed to parse: {}\n{:#?}", p, s);
500         }
501     }
502     fail("' static");
503     fail("r#1");
504     fail("r#_");
505     fail("\"\\u{0000000}\""); // overlong unicode escape (rust allows at most 6 hex digits)
506     fail("\"\\u{999999}\""); // outside of valid range of char
507     fail("\"\\u{_0}\""); // leading underscore
508     fail("\"\\u{}\""); // empty
509     fail("b\"\r\""); // bare carriage return in byte string
510     fail("r\"\r\""); // bare carriage return in raw string
511     fail("\"\\\r  \""); // backslash carriage return
512     fail("'aa'aa");
513     fail("br##\"\"#");
514     fail("cr##\"\"#");
515     fail("\"\\\n\u{85}\r\"");
516 }
517 
518 #[cfg(span_locations)]
519 #[test]
span_test()520 fn span_test() {
521     check_spans(
522         "\
523 /// This is a document comment
524 testing 123
525 {
526   testing 234
527 }",
528         &[
529             (1, 0, 1, 30),  // #
530             (1, 0, 1, 30),  // [ ... ]
531             (1, 0, 1, 30),  // doc
532             (1, 0, 1, 30),  // =
533             (1, 0, 1, 30),  // "This is..."
534             (2, 0, 2, 7),   // testing
535             (2, 8, 2, 11),  // 123
536             (3, 0, 5, 1),   // { ... }
537             (4, 2, 4, 9),   // testing
538             (4, 10, 4, 13), // 234
539         ],
540     );
541 }
542 
543 #[cfg(procmacro2_semver_exempt)]
544 #[test]
default_span()545 fn default_span() {
546     let start = Span::call_site().start();
547     assert_eq!(start.line, 1);
548     assert_eq!(start.column, 0);
549     let end = Span::call_site().end();
550     assert_eq!(end.line, 1);
551     assert_eq!(end.column, 0);
552     let source_file = Span::call_site().source_file();
553     assert_eq!(source_file.path().to_string_lossy(), "<unspecified>");
554     assert!(!source_file.is_real());
555 }
556 
557 #[cfg(procmacro2_semver_exempt)]
558 #[test]
span_join()559 fn span_join() {
560     let source1 = "aaa\nbbb"
561         .parse::<TokenStream>()
562         .unwrap()
563         .into_iter()
564         .collect::<Vec<_>>();
565     let source2 = "ccc\nddd"
566         .parse::<TokenStream>()
567         .unwrap()
568         .into_iter()
569         .collect::<Vec<_>>();
570 
571     assert!(source1[0].span().source_file() != source2[0].span().source_file());
572     assert_eq!(
573         source1[0].span().source_file(),
574         source1[1].span().source_file()
575     );
576 
577     let joined1 = source1[0].span().join(source1[1].span());
578     let joined2 = source1[0].span().join(source2[0].span());
579     assert!(joined1.is_some());
580     assert!(joined2.is_none());
581 
582     let start = joined1.unwrap().start();
583     let end = joined1.unwrap().end();
584     assert_eq!(start.line, 1);
585     assert_eq!(start.column, 0);
586     assert_eq!(end.line, 2);
587     assert_eq!(end.column, 3);
588 
589     assert_eq!(
590         joined1.unwrap().source_file(),
591         source1[0].span().source_file()
592     );
593 }
594 
595 #[test]
no_panic()596 fn no_panic() {
597     let s = str::from_utf8(b"b\'\xc2\x86  \x00\x00\x00^\"").unwrap();
598     assert!(s.parse::<TokenStream>().is_err());
599 }
600 
601 #[test]
punct_before_comment()602 fn punct_before_comment() {
603     let mut tts = TokenStream::from_str("~// comment").unwrap().into_iter();
604     match tts.next().unwrap() {
605         TokenTree::Punct(tt) => {
606             assert_eq!(tt.as_char(), '~');
607             assert_eq!(tt.spacing(), Spacing::Alone);
608         }
609         wrong => panic!("wrong token {:?}", wrong),
610     }
611 }
612 
613 #[test]
joint_last_token()614 fn joint_last_token() {
615     // This test verifies that we match the behavior of libproc_macro *not* in
616     // the range nightly-2020-09-06 through nightly-2020-09-10, in which this
617     // behavior was temporarily broken.
618     // See https://github.com/rust-lang/rust/issues/76399
619 
620     let joint_punct = Punct::new(':', Spacing::Joint);
621     let stream = TokenStream::from(TokenTree::Punct(joint_punct));
622     let punct = match stream.into_iter().next().unwrap() {
623         TokenTree::Punct(punct) => punct,
624         _ => unreachable!(),
625     };
626     assert_eq!(punct.spacing(), Spacing::Joint);
627 }
628 
629 #[test]
raw_identifier()630 fn raw_identifier() {
631     let mut tts = TokenStream::from_str("r#dyn").unwrap().into_iter();
632     match tts.next().unwrap() {
633         TokenTree::Ident(raw) => assert_eq!("r#dyn", raw.to_string()),
634         wrong => panic!("wrong token {:?}", wrong),
635     }
636     assert!(tts.next().is_none());
637 }
638 
639 #[test]
test_debug_ident()640 fn test_debug_ident() {
641     let ident = Ident::new("proc_macro", Span::call_site());
642 
643     #[cfg(not(span_locations))]
644     let expected = "Ident(proc_macro)";
645 
646     #[cfg(span_locations)]
647     let expected = "Ident { sym: proc_macro }";
648 
649     assert_eq!(expected, format!("{:?}", ident));
650 }
651 
652 #[test]
test_debug_tokenstream()653 fn test_debug_tokenstream() {
654     let tts = TokenStream::from_str("[a + 1]").unwrap();
655 
656     #[cfg(not(span_locations))]
657     let expected = "\
658 TokenStream [
659     Group {
660         delimiter: Bracket,
661         stream: TokenStream [
662             Ident {
663                 sym: a,
664             },
665             Punct {
666                 char: '+',
667                 spacing: Alone,
668             },
669             Literal {
670                 lit: 1,
671             },
672         ],
673     },
674 ]\
675     ";
676 
677     #[cfg(not(span_locations))]
678     let expected_before_trailing_commas = "\
679 TokenStream [
680     Group {
681         delimiter: Bracket,
682         stream: TokenStream [
683             Ident {
684                 sym: a
685             },
686             Punct {
687                 char: '+',
688                 spacing: Alone
689             },
690             Literal {
691                 lit: 1
692             }
693         ]
694     }
695 ]\
696     ";
697 
698     #[cfg(span_locations)]
699     let expected = "\
700 TokenStream [
701     Group {
702         delimiter: Bracket,
703         stream: TokenStream [
704             Ident {
705                 sym: a,
706                 span: bytes(2..3),
707             },
708             Punct {
709                 char: '+',
710                 spacing: Alone,
711                 span: bytes(4..5),
712             },
713             Literal {
714                 lit: 1,
715                 span: bytes(6..7),
716             },
717         ],
718         span: bytes(1..8),
719     },
720 ]\
721     ";
722 
723     #[cfg(span_locations)]
724     let expected_before_trailing_commas = "\
725 TokenStream [
726     Group {
727         delimiter: Bracket,
728         stream: TokenStream [
729             Ident {
730                 sym: a,
731                 span: bytes(2..3)
732             },
733             Punct {
734                 char: '+',
735                 spacing: Alone,
736                 span: bytes(4..5)
737             },
738             Literal {
739                 lit: 1,
740                 span: bytes(6..7)
741             }
742         ],
743         span: bytes(1..8)
744     }
745 ]\
746     ";
747 
748     let actual = format!("{:#?}", tts);
749     if actual.ends_with(",\n]") {
750         assert_eq!(expected, actual);
751     } else {
752         assert_eq!(expected_before_trailing_commas, actual);
753     }
754 }
755 
756 #[test]
default_tokenstream_is_empty()757 fn default_tokenstream_is_empty() {
758     let default_token_stream = <TokenStream as Default>::default();
759 
760     assert!(default_token_stream.is_empty());
761 }
762 
763 #[test]
tokenstream_size_hint()764 fn tokenstream_size_hint() {
765     let tokens = "a b (c d) e".parse::<TokenStream>().unwrap();
766 
767     assert_eq!(tokens.into_iter().size_hint(), (4, Some(4)));
768 }
769 
770 #[test]
tuple_indexing()771 fn tuple_indexing() {
772     // This behavior may change depending on https://github.com/rust-lang/rust/pull/71322
773     let mut tokens = "tuple.0.0".parse::<TokenStream>().unwrap().into_iter();
774     assert_eq!("tuple", tokens.next().unwrap().to_string());
775     assert_eq!(".", tokens.next().unwrap().to_string());
776     assert_eq!("0.0", tokens.next().unwrap().to_string());
777     assert!(tokens.next().is_none());
778 }
779 
780 #[cfg(span_locations)]
781 #[test]
non_ascii_tokens()782 fn non_ascii_tokens() {
783     check_spans("// abc", &[]);
784     check_spans("// ábc", &[]);
785     check_spans("// abc x", &[]);
786     check_spans("// ábc x", &[]);
787     check_spans("/* abc */ x", &[(1, 10, 1, 11)]);
788     check_spans("/* ábc */ x", &[(1, 10, 1, 11)]);
789     check_spans("/* ab\nc */ x", &[(2, 5, 2, 6)]);
790     check_spans("/* áb\nc */ x", &[(2, 5, 2, 6)]);
791     check_spans("/*** abc */ x", &[(1, 12, 1, 13)]);
792     check_spans("/*** ábc */ x", &[(1, 12, 1, 13)]);
793     check_spans(r#""abc""#, &[(1, 0, 1, 5)]);
794     check_spans(r#""ábc""#, &[(1, 0, 1, 5)]);
795     check_spans(r##"r#"abc"#"##, &[(1, 0, 1, 8)]);
796     check_spans(r##"r#"ábc"#"##, &[(1, 0, 1, 8)]);
797     check_spans("r#\"a\nc\"#", &[(1, 0, 2, 3)]);
798     check_spans("r#\"á\nc\"#", &[(1, 0, 2, 3)]);
799     check_spans("'a'", &[(1, 0, 1, 3)]);
800     check_spans("'á'", &[(1, 0, 1, 3)]);
801     check_spans("//! abc", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]);
802     check_spans("//! ábc", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]);
803     check_spans("//! abc\n", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]);
804     check_spans("//! ábc\n", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]);
805     check_spans("/*! abc */", &[(1, 0, 1, 10), (1, 0, 1, 10), (1, 0, 1, 10)]);
806     check_spans("/*! ábc */", &[(1, 0, 1, 10), (1, 0, 1, 10), (1, 0, 1, 10)]);
807     check_spans("/*! a\nc */", &[(1, 0, 2, 4), (1, 0, 2, 4), (1, 0, 2, 4)]);
808     check_spans("/*! á\nc */", &[(1, 0, 2, 4), (1, 0, 2, 4), (1, 0, 2, 4)]);
809     check_spans("abc", &[(1, 0, 1, 3)]);
810     check_spans("ábc", &[(1, 0, 1, 3)]);
811     check_spans("ábć", &[(1, 0, 1, 3)]);
812     check_spans("abc// foo", &[(1, 0, 1, 3)]);
813     check_spans("ábc// foo", &[(1, 0, 1, 3)]);
814     check_spans("ábć// foo", &[(1, 0, 1, 3)]);
815     check_spans("b\"a\\\n c\"", &[(1, 0, 2, 3)]);
816 }
817 
818 #[cfg(span_locations)]
check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)])819 fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) {
820     let ts = p.parse::<TokenStream>().unwrap();
821     check_spans_internal(ts, &mut lines);
822     assert!(lines.is_empty(), "leftover ranges: {:?}", lines);
823 }
824 
825 #[cfg(span_locations)]
check_spans_internal(ts: TokenStream, lines: &mut &[(usize, usize, usize, usize)])826 fn check_spans_internal(ts: TokenStream, lines: &mut &[(usize, usize, usize, usize)]) {
827     for i in ts {
828         if let Some((&(sline, scol, eline, ecol), rest)) = lines.split_first() {
829             *lines = rest;
830 
831             let start = i.span().start();
832             assert_eq!(start.line, sline, "sline did not match for {}", i);
833             assert_eq!(start.column, scol, "scol did not match for {}", i);
834 
835             let end = i.span().end();
836             assert_eq!(end.line, eline, "eline did not match for {}", i);
837             assert_eq!(end.column, ecol, "ecol did not match for {}", i);
838 
839             if let TokenTree::Group(g) = i {
840                 check_spans_internal(g.stream().clone(), lines);
841             }
842         }
843     }
844 }
845 
846 #[test]
whitespace()847 fn whitespace() {
848     // space, horizontal tab, vertical tab, form feed, carriage return, line
849     // feed, non-breaking space, left-to-right mark, right-to-left mark
850     let various_spaces = " \t\u{b}\u{c}\r\n\u{a0}\u{200e}\u{200f}";
851     let tokens = various_spaces.parse::<TokenStream>().unwrap();
852     assert_eq!(tokens.into_iter().count(), 0);
853 
854     let lone_carriage_returns = " \r \r\r\n ";
855     lone_carriage_returns.parse::<TokenStream>().unwrap();
856 }
857 
858 #[test]
byte_order_mark()859 fn byte_order_mark() {
860     let string = "\u{feff}foo";
861     let tokens = string.parse::<TokenStream>().unwrap();
862     match tokens.into_iter().next().unwrap() {
863         TokenTree::Ident(ident) => assert_eq!(ident, "foo"),
864         _ => unreachable!(),
865     }
866 
867     let string = "foo\u{feff}";
868     string.parse::<TokenStream>().unwrap_err();
869 }
870 
871 #[cfg(span_locations)]
create_span() -> proc_macro2::Span872 fn create_span() -> proc_macro2::Span {
873     let tts: TokenStream = "1".parse().unwrap();
874     match tts.into_iter().next().unwrap() {
875         TokenTree::Literal(literal) => literal.span(),
876         _ => unreachable!(),
877     }
878 }
879 
880 #[cfg(span_locations)]
881 #[test]
test_invalidate_current_thread_spans()882 fn test_invalidate_current_thread_spans() {
883     let actual = format!("{:#?}", create_span());
884     assert_eq!(actual, "bytes(1..2)");
885     let actual = format!("{:#?}", create_span());
886     assert_eq!(actual, "bytes(3..4)");
887 
888     proc_macro2::extra::invalidate_current_thread_spans();
889 
890     let actual = format!("{:#?}", create_span());
891     // Test that span offsets have been reset after the call
892     // to invalidate_current_thread_spans()
893     assert_eq!(actual, "bytes(1..2)");
894 }
895 
896 #[cfg(span_locations)]
897 #[test]
898 #[should_panic(expected = "Invalid span with no related FileInfo!")]
test_use_span_after_invalidation()899 fn test_use_span_after_invalidation() {
900     let span = create_span();
901 
902     proc_macro2::extra::invalidate_current_thread_spans();
903 
904     span.source_text();
905 }
906