// Copyright 2015 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package lex
import (
"text/scanner"
"cmd/internal/src"
)
// A Slice reads from a slice of Tokens.
type Slice struct {
tokens []Token
base *src.PosBase
line int
pos int
}
func NewSlice(base *src.PosBase, line int, tokens []Token) *Slice {
return &Slice{
tokens: tokens,
base: base,
line: line,
pos: -1, // Next will advance to zero.
}
}
func (s *Slice) Next() ScanToken {
s.pos++
if s.pos >= len(s.tokens) {
return scanner.EOF
}
return s.tokens[s.pos].ScanToken
}
func (s *Slice) Text() string {
return s.tokens[s.pos].text
}
func (s *Slice) File() string {
return s.base.Filename()
}
func (s *Slice) Base() *src.PosBase {
return s.base
}
func (s *Slice) SetBase(base *src.PosBase) {
// Cannot happen because we only have slices of already-scanned text,
// but be prepared.
s.base = base
}
func (s *Slice) Line() int {
return s.line
}
func (s *Slice) Col() int {
// TODO: Col is only called when defining a macro and all it cares about is increasing
// position to discover whether there is a blank before the parenthesis.
// We only get here if defining a macro inside a macro.
// This imperfect implementation means we cannot tell the difference between
// #define A #define B(x) x
// and
// #define A #define B (x) x
// The first has definition of B has an argument, the second doesn't. Because we let
// text/scanner strip the blanks for us, this is extremely rare, hard to fix, and not worth it.
return s.pos
}
func (s *Slice) Close() {
}