This commit is contained in:
Marvin Blum
2015-12-15 23:19:41 +01:00
parent f1f564d89b
commit 69e684a230
6 changed files with 227 additions and 226 deletions

View File

@@ -1,34 +1,34 @@
package main package main
import ( import (
"parser"
"tokenizer"
"fmt" "fmt"
"io/ioutil" "io/ioutil"
"path/filepath"
"os" "os"
"parser"
"path/filepath"
"strings" "strings"
"tokenizer"
) )
const ( const (
version = "1.1.1" version = "1.1.1"
extension = ".asl" extension = ".asl"
sqfextension = ".sqf" sqfextension = ".sqf"
PathSeparator = string(os.PathSeparator) PathSeparator = string(os.PathSeparator)
) )
type ASLFile struct { type ASLFile struct {
in string in string
out string out string
newname string newname string
} }
var ( var (
recursive bool = false recursive bool = false
pretty bool = false pretty bool = false
exit bool = false exit bool = false
aslFiles []ASLFile aslFiles []ASLFile
inDir string inDir string
) )
func usage() { func usage() {
@@ -42,109 +42,110 @@ func usage() {
} }
func flags(flag string) bool { func flags(flag string) bool {
flag = strings.ToLower(flag) flag = strings.ToLower(flag)
if flag[0] == '-' { if flag[0] == '-' {
if flag == "-v" { if flag == "-v" {
fmt.Println("asl version "+version) fmt.Println("asl version " + version)
exit = true exit = true
} else if flag == "-r" { } else if flag == "-r" {
recursive = true recursive = true
} else if flag == "-pretty" { } else if flag == "-pretty" {
pretty = true pretty = true
} else if flag == "--help" { } else if flag == "--help" {
usage() usage()
exit = true exit = true
} }
return true return true
} }
return false return false
} }
func readAslFiles(path string) { func readAslFiles(path string) {
dir, err := ioutil.ReadDir(path) dir, err := ioutil.ReadDir(path)
if err != nil {
fmt.Println("Error reading in directory!")
return
}
for i := 0; i < len(dir); i++ {
name := dir[i].Name()
if dir[i].IsDir() && recursive {
readAslFiles(filepath.FromSlash(path+PathSeparator+name))
continue
}
if !dir[i].IsDir() && strings.ToLower(filepath.Ext(name)) == extension { if err != nil {
in := filepath.FromSlash(path+PathSeparator+dir[i].Name()) fmt.Println("Error reading in directory!")
out := filepath.FromSlash("./"+path[len(inDir):len(path)]) return
newname := name[:len(name)-len(filepath.Ext(name))] }
file := ASLFile{in, out, newname} for i := 0; i < len(dir); i++ {
aslFiles = append(aslFiles, file) name := dir[i].Name()
}
} if dir[i].IsDir() && recursive {
readAslFiles(filepath.FromSlash(path + PathSeparator + name))
continue
}
if !dir[i].IsDir() && strings.ToLower(filepath.Ext(name)) == extension {
in := filepath.FromSlash(path + PathSeparator + dir[i].Name())
out := filepath.FromSlash("./" + path[len(inDir):len(path)])
newname := name[:len(name)-len(filepath.Ext(name))]
file := ASLFile{in, out, newname}
aslFiles = append(aslFiles, file)
}
}
} }
func compile(path string) { func compile(path string) {
for i := 0; i < len(aslFiles); i++ { for i := 0; i < len(aslFiles); i++ {
out := filepath.FromSlash(path+PathSeparator+aslFiles[i].out+PathSeparator+aslFiles[i].newname+sqfextension) out := filepath.FromSlash(path + PathSeparator + aslFiles[i].out + PathSeparator + aslFiles[i].newname + sqfextension)
fmt.Println(aslFiles[i].in+" -> "+out) fmt.Println(aslFiles[i].in + " -> " + out)
code, err := ioutil.ReadFile(aslFiles[i].in) code, err := ioutil.ReadFile(aslFiles[i].in)
if err != nil { if err != nil {
fmt.Println("Error reading file: "+aslFiles[i].in) fmt.Println("Error reading file: " + aslFiles[i].in)
continue continue
} }
token := tokenizer.Tokenize(code, false) token := tokenizer.Tokenize(code, false)
compiler := parser.Compiler{} compiler := parser.Compiler{}
sqf := compiler.Parse(token, pretty) sqf := compiler.Parse(token, pretty)
os.MkdirAll(filepath.FromSlash(path+PathSeparator+aslFiles[i].out), 0777) os.MkdirAll(filepath.FromSlash(path+PathSeparator+aslFiles[i].out), 0777)
err = ioutil.WriteFile(out, []byte(sqf), 0666) err = ioutil.WriteFile(out, []byte(sqf), 0666)
if err != nil { if err != nil {
fmt.Println("Error writing file: "+aslFiles[i].out) fmt.Println("Error writing file: " + aslFiles[i].out)
fmt.Println(err) fmt.Println(err)
} }
} }
} }
func main() { func main() {
args := os.Args args := os.Args
// flags // flags
if len(args) < 2 { if len(args) < 2 {
usage() usage()
return return
} }
var i int var i int
for i = 1; i < len(args) && flags(args[i]); i++ {} for i = 1; i < len(args) && flags(args[i]); i++ {
if exit {
return
} }
if exit {
return
}
// in/out parameter // in/out parameter
out := "" out := ""
if i < len(args) { if i < len(args) {
inDir = args[i] inDir = args[i]
i++ i++
} else { } else {
return return
} }
if i < len(args) { if i < len(args) {
out = args[i] out = args[i]
} }
readAslFiles(inDir) readAslFiles(inDir)
compile(out) compile(out)
} }

View File

@@ -10,7 +10,7 @@ const new_line = "\r\n"
// and writes SQF code into desired location. // and writes SQF code into desired location.
func (c *Compiler) Parse(token []tokenizer.Token, prettyPrinting bool) string { func (c *Compiler) Parse(token []tokenizer.Token, prettyPrinting bool) string {
if !c.initParser(token, prettyPrinting) { if !c.initParser(token, prettyPrinting) {
return "" return ""
} }
for c.tokenIndex < len(token) { for c.tokenIndex < len(token) {
@@ -21,9 +21,9 @@ func (c *Compiler) Parse(token []tokenizer.Token, prettyPrinting bool) string {
} }
func (c *Compiler) parseBlock() { func (c *Compiler) parseBlock() {
if c.get().Preprocessor { if c.get().Preprocessor {
c.parsePreprocessor() c.parsePreprocessor()
} else if c.accept("var") { } else if c.accept("var") {
c.parseVar() c.parseVar()
} else if c.accept("if") { } else if c.accept("if") {
c.parseIf() c.parseIf()
@@ -42,9 +42,9 @@ func (c *Compiler) parseBlock() {
} else if c.accept("try") { } else if c.accept("try") {
c.parseTryCatch() c.parseTryCatch()
} else if c.accept("exitwith") { } else if c.accept("exitwith") {
c.parseExitWith() c.parseExitWith()
} else if c.accept("waituntil") { } else if c.accept("waituntil") {
c.parseWaitUntil() c.parseWaitUntil()
} else if c.accept("case") || c.accept("default") { } else if c.accept("case") || c.accept("default") {
return return
} else { } else {
@@ -57,9 +57,9 @@ func (c *Compiler) parseBlock() {
} }
func (c *Compiler) parsePreprocessor() { func (c *Compiler) parsePreprocessor() {
// we definitely want a new line before and after // we definitely want a new line before and after
c.appendOut(new_line+c.get().Token+new_line, false) c.appendOut(new_line+c.get().Token+new_line, false)
c.next() c.next()
} }
func (c *Compiler) parseVar() { func (c *Compiler) parseVar() {
@@ -78,7 +78,7 @@ func (c *Compiler) parseVar() {
} }
func (c *Compiler) parseArray(out bool) string { func (c *Compiler) parseArray(out bool) string {
output := "" output := ""
c.expect("[") c.expect("[")
output += "[" output += "["
@@ -87,17 +87,17 @@ func (c *Compiler) parseArray(out bool) string {
for c.accept(",") { for c.accept(",") {
c.next() c.next()
output += ","+c.parseExpression(false) output += "," + c.parseExpression(false)
} }
} }
c.expect("]") c.expect("]")
output += "]" output += "]"
if out { if out {
c.appendOut(output, false) c.appendOut(output, false)
} }
return output return output
} }
@@ -231,20 +231,20 @@ func (c *Compiler) parseFunctionParameter() {
if c.accept("{") { if c.accept("{") {
return return
} }
c.appendOut("params [", false) c.appendOut("params [", false)
for !c.accept(")") { for !c.accept(")") {
name := c.get().Token name := c.get().Token
c.next() c.next()
if c.accept("=") { if c.accept("=") {
c.next() c.next()
value := c.get().Token value := c.get().Token
c.next() c.next()
c.appendOut("[\""+name+"\","+value+"]", false) c.appendOut("[\""+name+"\","+value+"]", false)
} else { } else {
c.appendOut("\""+name+"\"", false) c.appendOut("\""+name+"\"", false)
} }
if !c.accept(")") { if !c.accept(")") {
@@ -252,7 +252,7 @@ func (c *Compiler) parseFunctionParameter() {
c.appendOut(",", false) c.appendOut(",", false)
} }
} }
c.appendOut("];", true) c.appendOut("];", true)
} }
@@ -279,47 +279,47 @@ func (c *Compiler) parseTryCatch() {
} }
func (c *Compiler) parseExitWith() { func (c *Compiler) parseExitWith() {
c.expect("exitwith") c.expect("exitwith")
c.expect("{") c.expect("{")
c.appendOut("if (true) exitWith {", true) c.appendOut("if (true) exitWith {", true)
c.parseBlock() c.parseBlock()
c.expect("}") c.expect("}")
c.appendOut("};", true) c.appendOut("};", true)
} }
func (c *Compiler) parseWaitUntil() { func (c *Compiler) parseWaitUntil() {
c.expect("waituntil") c.expect("waituntil")
c.expect("(") c.expect("(")
c.appendOut("waitUntil {", false) c.appendOut("waitUntil {", false)
c.parseExpression(true) c.parseExpression(true)
if c.accept(";") { if c.accept(";") {
c.next() c.next()
c.appendOut(";", false) c.appendOut(";", false)
c.parseExpression(true) c.parseExpression(true)
} }
c.expect(")") c.expect(")")
c.expect(";") c.expect(";")
c.appendOut("};", true) c.appendOut("};", true)
} }
func (c *Compiler) parseInlineCode() string { func (c *Compiler) parseInlineCode() string {
c.expect("code") c.expect("code")
c.expect("(") c.expect("(")
code := c.get().Token code := c.get().Token
c.next() c.next()
output := "{}" output := "{}"
if len(code) > 2 { if len(code) > 2 {
compiler := Compiler{} compiler := Compiler{}
output = "{"+compiler.Parse(tokenizer.Tokenize([]byte(code[1:len(code)-1]), true), false)+"}" output = "{" + compiler.Parse(tokenizer.Tokenize([]byte(code[1:len(code)-1]), true), false) + "}"
} }
c.expect(")") c.expect(")")
return output return output
} }
// Everything that does not start with a keyword. // Everything that does not start with a keyword.
@@ -459,19 +459,19 @@ func (c *Compiler) parseIdentifier() string {
output := "" output := ""
if c.accept("code") { if c.accept("code") {
output += c.parseInlineCode() output += c.parseInlineCode()
} else if c.seek("(") && !c.accept("!") && !c.accept("-") { } else if c.seek("(") && !c.accept("!") && !c.accept("-") {
name := c.get().Token name := c.get().Token
c.next() c.next()
output = "(" + c.parseFunctionCall(false, name) + ")" output = "(" + c.parseFunctionCall(false, name) + ")"
} else if c.accept("[") { } else if c.accept("[") {
output += c.parseArray(false) output += c.parseArray(false)
} else if c.seek("[") { } else if c.seek("[") {
output += "("+c.get().Token output += "(" + c.get().Token
c.next() c.next()
c.expect("[") c.expect("[")
output += " select ("+c.parseExpression(false)+"))" output += " select (" + c.parseExpression(false) + "))"
c.expect("]") c.expect("]")
} else if c.accept("!") || c.accept("-") { } else if c.accept("!") || c.accept("-") {
output = c.get().Token output = c.get().Token
c.next() c.next()

View File

@@ -1,16 +1,16 @@
package parser package parser
import ( import (
"strconv" "strconv"
"tokenizer" "tokenizer"
) )
type Compiler struct { type Compiler struct {
tokens []tokenizer.Token tokens []tokenizer.Token
tokenIndex int tokenIndex int
out string out string
offset int offset int
pretty bool pretty bool
} }
// Initilizes the parser. // Initilizes the parser.
@@ -24,7 +24,7 @@ func (c *Compiler) initParser(token []tokenizer.Token, prettyPrinting bool) bool
c.out = "" c.out = ""
c.offset = 0 c.offset = 0
c.pretty = prettyPrinting c.pretty = prettyPrinting
return true return true
} }
@@ -38,7 +38,7 @@ func (c *Compiler) accept(token string) bool {
// Throws if current token does not match expected one. // Throws if current token does not match expected one.
func (c *Compiler) expect(token string) { func (c *Compiler) expect(token string) {
if !c.tokenEqual(token, c.get()) { if !c.tokenEqual(token, c.get()) {
panic("Parse error, expected '" + token + "' but was '" + c.get().Token + "' in line "+strconv.Itoa(c.get().Line)+" at "+strconv.Itoa(c.get().Column)) panic("Parse error, expected '" + token + "' but was '" + c.get().Token + "' in line " + strconv.Itoa(c.get().Line) + " at " + strconv.Itoa(c.get().Column))
} }
c.next() c.next()

View File

@@ -1,10 +1,10 @@
package parser_test package parser_test
import ( import (
"tokenizer"
"parser"
"io/ioutil" "io/ioutil"
"parser"
"testing" "testing"
"tokenizer"
) )
func TestParserDeclaration(t *testing.T) { func TestParserDeclaration(t *testing.T) {
@@ -50,9 +50,9 @@ func TestParserForeach(t *testing.T) {
} }
func TestParserSwitch(t *testing.T) { func TestParserSwitch(t *testing.T) {
got := getCompiled(t, "test/tokenizer_switch.asl") got := getCompiled(t, "test/tokenizer_switch.asl")
want := "switch (x) do {\r\ncase 1:\r\n{\r\nx = 1;\r\n};\r\ncase 2:\r\n{\r\nx = 2;\r\n};\r\ndefault:\r\n{\r\nx = 3;\r\n};\r\n};\r\n" want := "switch (x) do {\r\ncase 1:\r\n{\r\nx = 1;\r\n};\r\ncase 2:\r\n{\r\nx = 2;\r\n};\r\ndefault:\r\n{\r\nx = 3;\r\n};\r\n};\r\n"
equal(t, got, want) equal(t, got, want)
} }

View File

@@ -5,10 +5,10 @@ import (
) )
type Token struct { type Token struct {
Token string Token string
Preprocessor bool Preprocessor bool
Line int Line int
Column int Column int
} }
var delimiter = []byte{ var delimiter = []byte{
@@ -59,10 +59,10 @@ var new_line = []byte{'\r', '\n'}
// Tokenizes the given byte array into syntax tokens, // Tokenizes the given byte array into syntax tokens,
// which can be parsed later. // which can be parsed later.
func Tokenize(code []byte, doStripSlashes bool) []Token { func Tokenize(code []byte, doStripSlashes bool) []Token {
if doStripSlashes { if doStripSlashes {
code = stripSlashes(code); code = stripSlashes(code)
} }
code = removeComments(code) code = removeComments(code)
tokens := make([]Token, 0) tokens := make([]Token, 0)
token, mask, isstring, line, column := "", false, false, 0, 0 token, mask, isstring, line, column := "", false, false, 0, 0
@@ -70,10 +70,10 @@ func Tokenize(code []byte, doStripSlashes bool) []Token {
for i := 0; i < len(code); i++ { for i := 0; i < len(code); i++ {
c := code[i] c := code[i]
column++ column++
if byteArrayContains(new_line, c) { if byteArrayContains(new_line, c) {
line++ line++
column = 0 column = 0
} }
// string masks (backslash) // string masks (backslash)
@@ -95,8 +95,8 @@ func Tokenize(code []byte, doStripSlashes bool) []Token {
} else { } else {
// preprocessor, delimeter, keyword or variable/expression // preprocessor, delimeter, keyword or variable/expression
if c == preprocessor { if c == preprocessor {
tokens = append(tokens, preprocessorLine(code, &i, line, column)) tokens = append(tokens, preprocessorLine(code, &i, line, column))
token = "" token = ""
} else if byteArrayContains(delimiter, c) { } else if byteArrayContains(delimiter, c) {
if token != "" { if token != "" {
tokens = append(tokens, Token{token, false, line, column}) tokens = append(tokens, Token{token, false, line, column})
@@ -121,23 +121,23 @@ func Tokenize(code []byte, doStripSlashes bool) []Token {
// Removes slashes from input code. // Removes slashes from input code.
// This is used for the "code" keyword for correct strings in resulting code. // This is used for the "code" keyword for correct strings in resulting code.
func stripSlashes(code []byte) []byte { func stripSlashes(code []byte) []byte {
newcode := make([]byte, len(code)) newcode := make([]byte, len(code))
j, mask := 0, false j, mask := 0, false
for i := 0; i < len(code); i++ { for i := 0; i < len(code); i++ {
c := code[i] c := code[i]
if c == '\\' && !mask { if c == '\\' && !mask {
mask = true mask = true
continue continue
} }
newcode[j] = code[i] newcode[j] = code[i]
mask = false mask = false
j++ j++
} }
return newcode return newcode
} }
// Removes all comments from input byte array. // Removes all comments from input byte array.
@@ -180,31 +180,31 @@ func removeComments(code []byte) []byte {
// Reads preprocessor command until end of line // Reads preprocessor command until end of line
func preprocessorLine(code []byte, i *int, lineNr, column int) Token { func preprocessorLine(code []byte, i *int, lineNr, column int) Token {
c := byte('0') c := byte('0')
var line string var line string
for *i < len(code) { for *i < len(code) {
c = code[*i] c = code[*i]
if byteArrayContains(new_line, c) { if byteArrayContains(new_line, c) {
break break
} }
line += string(c) line += string(c)
(*i)++ (*i)++
} }
// read all new line characters (\r and \n) // read all new line characters (\r and \n)
c = code[*i] c = code[*i]
for byteArrayContains(new_line, c) { for byteArrayContains(new_line, c) {
(*i)++ (*i)++
c = code[*i] c = code[*i]
} }
(*i)-- // for will count up 1, so subtract it here (*i)-- // for will count up 1, so subtract it here
return Token{line, true, lineNr, column} return Token{line, true, lineNr, column}
} }
// Returns the next character in code starting at i. // Returns the next character in code starting at i.

View File

@@ -1,9 +1,9 @@
package tokenizer_test package tokenizer_test
import ( import (
"tokenizer"
"io/ioutil" "io/ioutil"
"testing" "testing"
"tokenizer"
) )
func TestTokenizerVar(t *testing.T) { func TestTokenizerVar(t *testing.T) {
@@ -96,9 +96,9 @@ func TestTokenizerPreprocessor(t *testing.T) {
func TestTokenizerMask(t *testing.T) { func TestTokenizerMask(t *testing.T) {
got := getTokens(t, "test/tokenizer_mask.asl") got := getTokens(t, "test/tokenizer_mask.asl")
//var y = code("var z = \"Hello \\"World\\"\";"); //var y = code("var z = \"Hello \\"World\\"\";");
want := []string{"var", "x", "=", "\"Hello \\\"World\\\"\"", ";", want := []string{"var", "x", "=", "\"Hello \\\"World\\\"\"", ";",
"var", "y", "=", "code", "(", "\"var z = \\\"Hello \\\\\"World\\\\\"\\\";\"", ")", ";"} "var", "y", "=", "code", "(", "\"var z = \\\"Hello \\\\\"World\\\\\"\\\";\"", ")", ";"}
compareLength(t, &got, &want) compareLength(t, &got, &want)
compareTokens(t, &got, &want) compareTokens(t, &got, &want)