From 2dc98e0cf4f4d27bf53d941dc9b81871008ecb0b Mon Sep 17 00:00:00 2001 From: brettlangdon Date: Sun, 13 Sep 2015 22:29:28 -0400 Subject: [PATCH] use scanner instead of tokenizer --- main.go | 9 +++------ 1 file changed, 3 insertions(+), 6 deletions(-) diff --git a/main.go b/main.go index 897d9fe..6b95bd4 100644 --- a/main.go +++ b/main.go @@ -4,17 +4,14 @@ import ( "fmt" "os" - "github.com/brettlangdon/gython/parser" + "github.com/brettlangdon/gython/scanner" "github.com/brettlangdon/gython/token" ) func main() { - tokenizer, err := parser.TokenizerFromFileName(os.Args[1]) - if err != nil { - panic(err) - } + tokenizer := scanner.NewScanner(os.Stdin) for { - tok := tokenizer.Next() + tok := tokenizer.NextToken() tokenRange := fmt.Sprintf("%d,%d-%d,%d:", tok.LineStart, tok.ColumnStart, tok.LineEnd, tok.ColumnEnd) literalRep := fmt.Sprintf("%#v", tok.Literal) fmt.Printf("%-20s%-15s%15s\n", tokenRange, tok.String(), literalRep)