Documentation ¶
Overview ¶
Package participle constructs parsers from definitions in struct tags and parses directly into those structs. The approach is philosophically similar to how other marshallers work in Go, "unmarshalling" an instance of a grammar into a struct.
The supported annotation syntax is:
- `@<expr>` Capture expression into the field.
- `@@` Recursively capture using the fields own type.
- `<identifier>` Match named lexer token.
- `( ... )` Group.
- `"..."` Match the literal (note that the lexer must emit tokens matching this literal exactly).
- `"...":<identifier>` Match the literal, specifying the exact lexer token type to match.
- `<expr> <expr> ...` Match expressions.
- `<expr> | <expr>` Match one of the alternatives.
The following modifiers can be used after any expression:
- `*` Expression can match zero or more times.
- `+` Expression must match one or more times.
- `?` Expression can match zero or once.
- `!` Require a non-empty match (this is useful with a sequence of optional matches eg. `("a"? "b"? "c"?)!`).
Supported but deprecated:
- `{ ... }` Match 0 or more times (**DEPRECATED** - prefer `( ... )*`).
- `[ ... ]` Optional (**DEPRECATED** - prefer `( ... )?`).
Here's an example of an EBNF grammar.
type Group struct { Expression *Expression `"(" @@ ")"` } type Option struct { Expression *Expression `"[" @@ "]"` } type Repetition struct { Expression *Expression `"{" @@ "}"` } type Literal struct { Start string `@String` // lexer.Lexer token "String" End string `("…" @String)?` } type Term struct { Name string ` @Ident` Literal *Literal `| @@` Group *Group `| @@` Option *Option `| @@` Repetition *Repetition `| @@` } type Sequence struct { Terms []*Term `@@+` } type Expression struct { Alternatives []*Sequence `@@ ("|" @@)*` } type Expressions []*Expression type Production struct { Name string `@Ident "="` Expressions Expressions `@@+ "."` } type EBNF struct { Productions []*Production `@@*` }
Index ¶
- Variables
- func AnnotateError(pos lexer.Position, err error) error
- func ErrorWithTokenf(tok lexer.Token, format string, args ...interface{}) error
- func Errorf(pos lexer.Position, format string, args ...interface{}) error
- func Wrapf(pos lexer.Position, err error, format string, args ...interface{}) error
- type Capture
- type Error
- type Mapper
- type Option
- type ParseOption
- type Parseable
- type Parser
- func (p *Parser) Lex(r io.Reader) ([]lexer.Token, error)
- func (p *Parser) Lexer() lexer.Definition
- func (p *Parser) Parse(r io.Reader, v interface{}, options ...ParseOption) (err error)
- func (p *Parser) ParseBytes(b []byte, v interface{}, options ...ParseOption) error
- func (p *Parser) ParseFromLexer(lex *lexer.PeekingLexer, v interface{}, options ...ParseOption) error
- func (p *Parser) ParseString(s string, v interface{}, options ...ParseOption) error
- func (p *Parser) String() string
- type UnexpectedTokenError
Constants ¶
This section is empty.
Variables ¶
var ( // MaxIterations limits the number of elements capturable by {}. MaxIterations = 1000000 // NextMatch should be returned by Parseable.Parse() method implementations to indicate // that the node did not match and that other matches should be attempted, if appropriate. NextMatch = errors.New("no match") // nolint: golint )
var DropToken = errors.New("drop token") // nolint: golint
DropToken can be returned by a Mapper to remove a token from the stream.
Functions ¶
func AnnotateError ¶ added in v0.4.0
AnnotateError wraps an existing error with a position.
If the existing error is a lexer.Error or participle.Error it will be returned unmodified.
func ErrorWithTokenf ¶ added in v0.4.2
ErrorWithTokenf creats a new Error with the given token as context.
Types ¶
type Capture ¶
Capture can be implemented by fields in order to transform captured tokens into field values.
type Error ¶
type Error interface { error // Unadorned message. Message() string // Closest token to error location. Token() lexer.Token }
Error represents an error while parsing.
The error will contain positional information if available.
type Mapper ¶
Mapper function for mutating tokens before being applied to the AST.
If the Mapper func returns an error of DropToken, the token will be removed from the stream.
type Option ¶
An Option to modify the behaviour of the Parser.
func CaseInsensitive ¶ added in v0.2.0
CaseInsensitive allows the specified token types to be matched case-insensitively.
func Lexer ¶
func Lexer(def lexer.Definition) Option
Lexer is an Option that sets the lexer to use with the given grammar.
func Map ¶
Map is an Option that configures the Parser to apply a mapping function to each Token from the lexer.
This can be useful to eg. upper-case all tokens of a certain type, or dequote strings.
"symbols" specifies the token symbols that the Mapper will be applied to. If empty, all tokens will be mapped.
func Unquote ¶
Unquote applies strconv.Unquote() to tokens of the given types.
Tokens of type "String" will be unquoted if no other types are provided.
func Upper ¶
Upper is an Option that upper-cases all tokens of the given type. Useful for case normalisation.
func UseLookahead ¶
UseLookahead allows branch lookahead up to "n" tokens.
If parsing cannot be disambiguated before "n" tokens of lookahead, parsing will fail.
Note that increasing lookahead has a minor performance impact, but also reduces the accuracy of error reporting.
type ParseOption ¶ added in v0.4.0
type ParseOption func(p *parseContext)
ParseOption modifies how an individual parse is applied.
func AllowTrailing ¶ added in v0.4.0
func AllowTrailing(ok bool) ParseOption
AllowTrailing tokens without erroring.
That is, do not error if a full parse completes but additional tokens remain.
type Parseable ¶
type Parseable interface { // Parse into the receiver. // // Should return NextMatch if no tokens matched and parsing should continue. // Nil should be returned if parsing was successful. Parse(lex *lexer.PeekingLexer) error }
The Parseable interface can be implemented by any element in the grammar to provide custom parsing.
type Parser ¶
type Parser struct {
// contains filtered or unexported fields
}
A Parser for a particular grammar and lexer.
func Build ¶
Build constructs a parser for the given grammar.
If "Lexer()" is not provided as an option, a default lexer based on text/scanner will be used. This scans typical Go- like tokens.
See documentation for details
func (*Parser) Lexer ¶ added in v0.4.2
func (p *Parser) Lexer() lexer.Definition
Lexer returns the parser's builtin lexer.
func (*Parser) Parse ¶
func (p *Parser) Parse(r io.Reader, v interface{}, options ...ParseOption) (err error)
Parse from r into grammar v which must be of the same type as the grammar passed to participle.Build().
This may return a participle.Error.
func (*Parser) ParseBytes ¶
func (p *Parser) ParseBytes(b []byte, v interface{}, options ...ParseOption) error
ParseBytes is a convenience around Parse().
This may return a participle.Error.
func (*Parser) ParseFromLexer ¶ added in v0.4.0
func (p *Parser) ParseFromLexer(lex *lexer.PeekingLexer, v interface{}, options ...ParseOption) error
ParseFromLexer into grammar v which must be of the same type as the grammar passed to participle.Build().
This may return a participle.Error.
func (*Parser) ParseString ¶
func (p *Parser) ParseString(s string, v interface{}, options ...ParseOption) error
ParseString is a convenience around Parse().
This may return a participle.Error.
type UnexpectedTokenError ¶ added in v0.4.0
UnexpectedTokenError is returned by Parse when an unexpected token is encountered.
This is useful for composing parsers in order to detect when a sub-parser has terminated.
func (UnexpectedTokenError) Error ¶ added in v0.4.0
func (u UnexpectedTokenError) Error() string
func (UnexpectedTokenError) Message ¶ added in v0.4.2
func (u UnexpectedTokenError) Message() string
func (UnexpectedTokenError) Token ¶ added in v0.4.2
func (u UnexpectedTokenError) Token() lexer.Token
Source Files ¶
Directories ¶
Path | Synopsis |
---|---|
_examples
module
|
|
Package lexer defines interfaces and implementations used by Participle to perform lexing.
|
Package lexer defines interfaces and implementations used by Participle to perform lexing. |
ebnf
Package ebnf is an EBNF lexer for Participle.
|
Package ebnf is an EBNF lexer for Participle. |
ebnf/internal
Package internal is a library for EBNF grammars.
|
Package internal is a library for EBNF grammars. |
regex
Package regex provides a regex based lexer using a readable list of named patterns.
|
Package regex provides a regex based lexer using a readable list of named patterns. |
stateful
Package stateful defines a nested stateful lexer.
|
Package stateful defines a nested stateful lexer. |