Source file src/cmd/asm/internal/lex/lex.go

     1  // Copyright 2015 The Go Authors. All rights reserved.
     2  // Use of this source code is governed by a BSD-style
     3  // license that can be found in the LICENSE file.
     4  
     5  // Package lex implements lexical analysis for the assembler.
     6  package lex
     7  
     8  import (
     9  	"fmt"
    10  	"log"
    11  	"os"
    12  	"strings"
    13  	"text/scanner"
    14  
    15  	"cmd/internal/src"
    16  )
    17  
    18  // A ScanToken represents an input item. It is a simple wrapping of rune, as
    19  // returned by text/scanner.Scanner, plus a couple of extra values.
    20  type ScanToken rune
    21  
    22  const (
    23  	// Asm defines some two-character lexemes. We make up
    24  	// a rune/ScanToken value for them - ugly but simple.
    25  	LSH          ScanToken = -1000 - iota // << Left shift.
    26  	RSH                                   // >> Logical right shift.
    27  	ARR                                   // -> Used on ARM for shift type 3, arithmetic right shift.
    28  	ROT                                   // @> Used on ARM for shift type 4, rotate right.
    29  	Include                               // included file started here
    30  	BuildComment                          // //go:build or +build comment
    31  	macroName                             // name of macro that should not be expanded
    32  )
    33  
    34  // IsRegisterShift reports whether the token is one of the ARM register shift operators.
    35  func IsRegisterShift(r ScanToken) bool {
    36  	return ROT <= r && r <= LSH // Order looks backwards because these are negative.
    37  }
    38  
    39  func (t ScanToken) String() string {
    40  	switch t {
    41  	case scanner.EOF:
    42  		return "EOF"
    43  	case scanner.Ident:
    44  		return "identifier"
    45  	case scanner.Int:
    46  		return "integer constant"
    47  	case scanner.Float:
    48  		return "float constant"
    49  	case scanner.Char:
    50  		return "rune constant"
    51  	case scanner.String:
    52  		return "string constant"
    53  	case scanner.RawString:
    54  		return "raw string constant"
    55  	case scanner.Comment:
    56  		return "comment"
    57  	default:
    58  		return fmt.Sprintf("%q", rune(t))
    59  	}
    60  }
    61  
    62  // NewLexer returns a lexer for the named file and the given link context.
    63  func NewLexer(name string) TokenReader {
    64  	input := NewInput(name)
    65  	fd, err := os.Open(name)
    66  	if err != nil {
    67  		log.Fatalf("%s\n", err)
    68  	}
    69  	input.Push(NewTokenizer(name, fd, fd))
    70  	return input
    71  }
    72  
    73  // The other files in this directory each contain an implementation of TokenReader.
    74  
    75  // A TokenReader is like a reader, but returns lex tokens of type Token. It also can tell you what
    76  // the text of the most recently returned token is, and where it was found.
    77  // The underlying scanner elides all spaces except newline, so the input looks like a stream of
    78  // Tokens; original spacing is lost but we don't need it.
    79  type TokenReader interface {
    80  	// Next returns the next token.
    81  	Next() ScanToken
    82  	// The following methods all refer to the most recent token returned by Next.
    83  	// Text returns the original string representation of the token.
    84  	Text() string
    85  	// File reports the source file name of the token.
    86  	File() string
    87  	// Base reports the position base of the token.
    88  	Base() *src.PosBase
    89  	// SetBase sets the position base.
    90  	SetBase(*src.PosBase)
    91  	// Line reports the source line number of the token.
    92  	Line() int
    93  	// Col reports the source column number of the token.
    94  	Col() int
    95  	// Close does any teardown required.
    96  	Close()
    97  }
    98  
    99  // A Token is a scan token plus its string value.
   100  // A macro is stored as a sequence of Tokens with spaces stripped.
   101  type Token struct {
   102  	ScanToken
   103  	text string
   104  }
   105  
   106  // Make returns a Token with the given rune (ScanToken) and text representation.
   107  func Make(token ScanToken, text string) Token {
   108  	// Substitute the substitutes for . and /.
   109  	text = strings.ReplaceAll(text, "\u00B7", ".")
   110  	text = strings.ReplaceAll(text, "\u2215", "/")
   111  	return Token{ScanToken: token, text: text}
   112  }
   113  
   114  func (l Token) String() string {
   115  	return l.text
   116  }
   117  
   118  // A Macro represents the definition of a #defined macro.
   119  type Macro struct {
   120  	name   string   // The #define name.
   121  	args   []string // Formal arguments.
   122  	tokens []Token  // Body of macro.
   123  }
   124  
   125  // Tokenize turns a string into a list of Tokens; used to parse the -D flag and in tests.
   126  func Tokenize(str string) []Token {
   127  	t := NewTokenizer("command line", strings.NewReader(str), nil)
   128  	var tokens []Token
   129  	for {
   130  		tok := t.Next()
   131  		if tok == scanner.EOF {
   132  			break
   133  		}
   134  		tokens = append(tokens, Make(tok, t.Text()))
   135  	}
   136  	return tokens
   137  }
   138  

View as plain text