@@ -213,11 +213,14 @@ mod tests {
213213 #[ test]
214214 fn test_single_line_comment ( ) {
215215 let mut trie = TokenTrie :: new ( ) ;
216- trie. insert ( b"//" , TokenMatch {
217- token_type : TokenType :: LineComment ,
218- close : None ,
219- advance : 0 ,
220- } ) ;
216+ trie. insert (
217+ b"//" ,
218+ TokenMatch {
219+ token_type : TokenType :: LineComment ,
220+ close : None ,
221+ advance : 0 ,
222+ } ,
223+ ) ;
221224 let m = trie. match_at ( b"// comment" , 0 ) . unwrap ( ) ;
222225 assert_eq ! ( m. token_type, TokenType :: LineComment ) ;
223226 assert_eq ! ( m. advance, 2 ) ;
@@ -227,11 +230,14 @@ mod tests {
227230 #[ test]
228231 fn test_block_comment ( ) {
229232 let mut trie = TokenTrie :: new ( ) ;
230- trie. insert ( b"/*" , TokenMatch {
231- token_type : TokenType :: BlockCommentStart ,
232- close : Some ( b"*/" . to_vec ( ) ) ,
233- advance : 0 ,
234- } ) ;
233+ trie. insert (
234+ b"/*" ,
235+ TokenMatch {
236+ token_type : TokenType :: BlockCommentStart ,
237+ close : Some ( b"*/" . to_vec ( ) ) ,
238+ advance : 0 ,
239+ } ,
240+ ) ;
235241 let m = trie. match_at ( b"/* block */" , 0 ) . unwrap ( ) ;
236242 assert_eq ! ( m. token_type, TokenType :: BlockCommentStart ) ;
237243 assert_eq ! ( m. advance, 2 ) ;
@@ -241,11 +247,14 @@ mod tests {
241247 #[ test]
242248 fn test_no_match_at_wrong_position ( ) {
243249 let mut trie = TokenTrie :: new ( ) ;
244- trie. insert ( b"//" , TokenMatch {
245- token_type : TokenType :: LineComment ,
246- close : None ,
247- advance : 0 ,
248- } ) ;
250+ trie. insert (
251+ b"//" ,
252+ TokenMatch {
253+ token_type : TokenType :: LineComment ,
254+ close : None ,
255+ advance : 0 ,
256+ } ,
257+ ) ;
249258 assert_eq ! ( trie. match_at( b"x // y" , 0 ) , None ) ;
250259 let m = trie. match_at ( b"x // y" , 2 ) . unwrap ( ) ;
251260 assert_eq ! ( m. token_type, TokenType :: LineComment ) ;
@@ -254,11 +263,14 @@ mod tests {
254263 #[ test]
255264 fn test_string_delimiter ( ) {
256265 let mut trie = TokenTrie :: new ( ) ;
257- trie. insert ( b"\" " , TokenMatch {
258- token_type : TokenType :: StringDelimiter ,
259- close : Some ( b"\" " . to_vec ( ) ) ,
260- advance : 0 ,
261- } ) ;
266+ trie. insert (
267+ b"\" " ,
268+ TokenMatch {
269+ token_type : TokenType :: StringDelimiter ,
270+ close : Some ( b"\" " . to_vec ( ) ) ,
271+ advance : 0 ,
272+ } ,
273+ ) ;
262274 let m = trie. match_at ( b"\" hello\" " , 0 ) . unwrap ( ) ;
263275 assert_eq ! ( m. token_type, TokenType :: StringDelimiter ) ;
264276 assert_eq ! ( m. close. as_deref( ) , Some ( b"\" " . as_slice( ) ) ) ;
@@ -267,16 +279,22 @@ mod tests {
267279 #[ test]
268280 fn test_process_mask_filters_correctly ( ) {
269281 let mut trie = TokenTrie :: new ( ) ;
270- trie. insert ( b"//" , TokenMatch {
271- token_type : TokenType :: LineComment ,
272- close : None ,
273- advance : 0 ,
274- } ) ;
275- trie. insert ( b"\" " , TokenMatch {
276- token_type : TokenType :: StringDelimiter ,
277- close : Some ( b"\" " . to_vec ( ) ) ,
278- advance : 0 ,
279- } ) ;
282+ trie. insert (
283+ b"//" ,
284+ TokenMatch {
285+ token_type : TokenType :: LineComment ,
286+ close : None ,
287+ advance : 0 ,
288+ } ,
289+ ) ;
290+ trie. insert (
291+ b"\" " ,
292+ TokenMatch {
293+ token_type : TokenType :: StringDelimiter ,
294+ close : Some ( b"\" " . to_vec ( ) ) ,
295+ advance : 0 ,
296+ } ,
297+ ) ;
280298 let mask = trie. process_mask ( ) ;
281299 assert ! ( should_process( b'/' , mask) ) ;
282300 assert ! ( should_process( b'"' , mask) ) ;
@@ -287,16 +305,22 @@ mod tests {
287305 #[ test]
288306 fn test_longer_match_wins ( ) {
289307 let mut trie = TokenTrie :: new ( ) ;
290- trie. insert ( b"\" " , TokenMatch {
291- token_type : TokenType :: StringDelimiter ,
292- close : Some ( b"\" " . to_vec ( ) ) ,
293- advance : 0 ,
294- } ) ;
295- trie. insert ( b"\" \" \" " , TokenMatch {
296- token_type : TokenType :: DocStringDelimiter ,
297- close : Some ( b"\" \" \" " . to_vec ( ) ) ,
298- advance : 0 ,
299- } ) ;
308+ trie. insert (
309+ b"\" " ,
310+ TokenMatch {
311+ token_type : TokenType :: StringDelimiter ,
312+ close : Some ( b"\" " . to_vec ( ) ) ,
313+ advance : 0 ,
314+ } ,
315+ ) ;
316+ trie. insert (
317+ b"\" \" \" " ,
318+ TokenMatch {
319+ token_type : TokenType :: DocStringDelimiter ,
320+ close : Some ( b"\" \" \" " . to_vec ( ) ) ,
321+ advance : 0 ,
322+ } ,
323+ ) ;
300324 let m = trie. match_at ( b"\" \" \" hello\" \" \" " , 0 ) . unwrap ( ) ;
301325 assert_eq ! ( m. token_type, TokenType :: DocStringDelimiter ) ;
302326 assert_eq ! ( m. advance, 3 ) ;
0 commit comments