mirror of
				https://github.com/alecthomas/chroma.git
				synced 2025-10-30 23:57:49 +02:00 
			
		
		
		
	
		
			
				
	
	
		
			53 lines
		
	
	
		
			1.1 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			53 lines
		
	
	
		
			1.1 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| package chroma
 | |
| 
 | |
| import (
 | |
| 	"testing"
 | |
| 
 | |
| 	"github.com/alecthomas/assert"
 | |
| )
 | |
| 
 | |
| func TestTokenTypeClassifiers(t *testing.T) {
 | |
| 	assert.True(t, GenericDeleted.InCategory(Generic))
 | |
| 	assert.True(t, LiteralStringBacktick.InSubCategory(String))
 | |
| 	assert.Equal(t, LiteralStringBacktick.String(), "LiteralStringBacktick")
 | |
| }
 | |
| 
 | |
| func TestSimpleLexer(t *testing.T) {
 | |
| 	lexer, err := NewLexer(
 | |
| 		&Config{
 | |
| 			Name:      "INI",
 | |
| 			Aliases:   []string{"ini", "cfg"},
 | |
| 			Filenames: []string{"*.ini", "*.cfg"},
 | |
| 		},
 | |
| 		map[string][]Rule{
 | |
| 			"root": {
 | |
| 				{`\s+`, Whitespace, nil},
 | |
| 				{`;.*?$`, Comment, nil},
 | |
| 				{`\[.*?\]$`, Keyword, nil},
 | |
| 				{`(.*?)(\s*)(=)(\s*)(.*?)$`, ByGroups(Name, Whitespace, Operator, Whitespace, String), nil},
 | |
| 			},
 | |
| 		},
 | |
| 	)
 | |
| 	assert.NoError(t, err)
 | |
| 	actual, err := Tokenise(lexer, nil, `
 | |
| 	; this is a comment
 | |
| 	[section]
 | |
| 	a = 10
 | |
| `)
 | |
| 	assert.NoError(t, err)
 | |
| 	expected := []Token{
 | |
| 		{Whitespace, "\n\t"},
 | |
| 		{Comment, "; this is a comment"},
 | |
| 		{Whitespace, "\n\t"},
 | |
| 		{Keyword, "[section]"},
 | |
| 		{Whitespace, "\n\t"},
 | |
| 		{Name, "a"},
 | |
| 		{Whitespace, " "},
 | |
| 		{Operator, "="},
 | |
| 		{Whitespace, " "},
 | |
| 		{LiteralString, "10"},
 | |
| 		{Whitespace, "\n"},
 | |
| 	}
 | |
| 	assert.Equal(t, expected, actual)
 | |
| }
 |