1
0
mirror of https://github.com/alecthomas/chroma.git synced 2025-01-28 03:29:41 +02:00

Add RemappingLexer.

This pattern is used heavily in Pygments for certain lexers,
particularly Lisp variants. See #43.
This commit is contained in:
Alec Thomas 2017-09-27 11:28:51 +10:00
parent c99eebc024
commit 102aa61ac3
2 changed files with 105 additions and 0 deletions

76
remap.go Normal file
View File

@ -0,0 +1,76 @@
package chroma
type remappingLexer struct {
lexer Lexer
mapper func(*Token) []*Token
}
// RemappingLexer remaps a token to a set of, potentially empty, tokens.
func RemappingLexer(lexer Lexer, mapper func(*Token) []*Token) Lexer {
return &remappingLexer{lexer, mapper}
}
func (r *remappingLexer) Config() *Config {
return r.lexer.Config()
}
func (r *remappingLexer) Tokenise(options *TokeniseOptions, text string) (Iterator, error) {
it, err := r.lexer.Tokenise(options, text)
if err != nil {
return nil, err
}
buffer := []*Token{}
return func() *Token {
for {
if len(buffer) > 0 {
t := buffer[0]
buffer = buffer[1:]
return t
}
t := it()
if t == nil {
return t
}
buffer = r.mapper(t)
}
}, nil
}
type TypeMapping struct {
From TokenType
To TokenType
}
type TypeRemappingMap map[TypeMapping][]string
// TypeRemappingLexer remaps types of tokens coming from a parent Lexer.
//
// eg. Map "defvaralias" tokens of type NameVariable to NameFunction:
//
// mapping := TypeRemappingMap{
// {NameVariable, NameFunction}: {"defvaralias"},
// }
// lexer = TypeRemappingLexer(lexer, mapping)
func TypeRemappingLexer(lexer Lexer, mapping TypeRemappingMap) Lexer {
// Lookup table for fast remapping.
lut := map[TokenType]map[string]TokenType{}
for rt, kl := range mapping {
km, ok := lut[rt.From]
if !ok {
km = map[string]TokenType{}
lut[rt.From] = km
}
for _, k := range kl {
km[k] = rt.To
}
}
return RemappingLexer(lexer, func(t *Token) []*Token {
if k, ok := lut[t.Type]; ok {
if tt, ok := k[t.Value]; ok {
t = t.Clone()
t.Type = tt
}
}
return []*Token{t}
})
}

29
remap_test.go Normal file
View File

@ -0,0 +1,29 @@
package chroma
import (
"testing"
"github.com/stretchr/testify/assert"
)
func TestRemappingLexer(t *testing.T) {
var lexer Lexer = MustNewLexer(nil, Rules{
"root": {
{`\s+`, Whitespace, nil},
{`\w+`, Name, nil},
},
})
lexer = TypeRemappingLexer(lexer, TypeRemappingMap{
{Name, Keyword}: {"if", "else"},
})
it, err := lexer.Tokenise(nil, `if true then print else end`)
assert.NoError(t, err)
expected := []*Token{
{Keyword, "if"}, {TextWhitespace, " "}, {Name, "true"}, {TextWhitespace, " "}, {Name, "then"},
{TextWhitespace, " "}, {Name, "print"}, {TextWhitespace, " "}, {Keyword, "else"},
{TextWhitespace, " "}, {Name, "end"},
}
actual := it.Tokens()
assert.Equal(t, expected, actual)
}