diff options
Diffstat (limited to 'tests/lexer')
-rw-r--r-- | tests/lexer/buildfile | 7 | ||||
-rw-r--r-- | tests/lexer/driver.cxx | 179 |
2 files changed, 0 insertions, 186 deletions
diff --git a/tests/lexer/buildfile b/tests/lexer/buildfile deleted file mode 100644 index 9a17904..0000000 --- a/tests/lexer/buildfile +++ /dev/null @@ -1,7 +0,0 @@ -# file : tests/lexer/buildfile -# copyright : Copyright (c) 2014-2016 Code Synthesis Ltd -# license : MIT; see accompanying LICENSE file - -import libs = libbutl%lib{butl} - -exe{driver}: cxx{driver ../../build/{lexer diagnostics utility}} $libs diff --git a/tests/lexer/driver.cxx b/tests/lexer/driver.cxx deleted file mode 100644 index 5575956..0000000 --- a/tests/lexer/driver.cxx +++ /dev/null @@ -1,179 +0,0 @@ -// file : tests/build/lexer/driver.cxx -*- C++ -*- -// copyright : Copyright (c) 2014-2016 Code Synthesis Ltd -// license : MIT; see accompanying LICENSE file - -#include <string> -#include <vector> -#include <cassert> -#include <sstream> -#include <iostream> - -#include <build/token> -#include <build/lexer> - -using namespace std; -using namespace build; - -using tokens = vector<string>; - -static tokens -lex (const char*); - -ostream& -operator<< (ostream&, const tokens&); - -int -main () -{ - ostream cnull (nullptr); - diag_stream = &cnull; - - // Whitespaces. - // - assert (lex ("") == tokens ({""})); - assert (lex ("\n") == tokens ({""})); - assert (lex ("\n\n") == tokens ({""})); - assert (lex (" \t \n") == tokens ({""})); - assert (lex ("#comment") == tokens ({""})); - assert (lex (" #comment") == tokens ({""})); - assert (lex ("#comment\n") == tokens ({""})); - assert (lex ("#comment\\\n") == tokens ({""})); - assert (lex ("#comment 1\n#comment2") == tokens ({""})); - - // Punctuation. - // - assert (lex (": \n { }") == tokens ({":", "\n", "{", "}", ""})); - - // Names. - // - assert (lex ("foo") == tokens ({"foo", ""})); - assert (lex ("foo.bar") == tokens ({"foo.bar", ""})); - - // Escaping. - // - assert (lex (" \\\n") == tokens ({""})); - assert (lex ("\\\nfoo") == tokens ({"foo", ""})); - assert (lex (" \\ foo") == tokens ({" foo", ""})); - assert (lex ("fo\\ o\\:") == tokens ({"fo o:", ""})); - assert (lex ("foo\\\nbar") == tokens ({"foobar", ""})); - assert (lex ("foo \\\nbar") == tokens ({"foo", "bar", ""})); - assert (lex ("\\'foo") == tokens ({"'foo", ""})); - - assert (lex (" \\") == tokens ({"<lexer error>"})); - assert (lex (" foo\\") == tokens ({"<lexer error>"})); - - - // Quoting ''. - // - assert (lex ("''") == tokens ({"", ""})); - assert (lex ("'foo'") == tokens ({"foo", ""})); - assert (lex ("'foo bar'") == tokens ({"foo bar", ""})); - assert (lex ("'foo 'bar") == tokens ({"foo bar", ""})); - assert (lex ("foo' bar'") == tokens ({"foo bar", ""})); - assert (lex ("'foo ''bar'") == tokens ({"foo bar", ""})); - assert (lex ("foo' 'bar") == tokens ({"foo bar", ""})); - assert (lex ("'foo\nbar'") == tokens ({"foo\nbar", ""})); - assert (lex ("'#:${}()=+\n'") == tokens ({"#:${}()=+\n", ""})); - assert (lex ("'\"'") == tokens ({"\"", ""})); - assert (lex ("'\\'") == tokens ({"\\", ""})); - - assert (lex ("'foo bar") == tokens ({"<lexer error>"})); - - // Quoting "". - // - assert (lex ("\"\"") == tokens ({"", ""})); - assert (lex ("\"foo\"") == tokens ({"foo", ""})); - assert (lex ("\"foo bar\"") == tokens ({"foo bar", ""})); - assert (lex ("\"foo \"bar") == tokens ({"foo bar", ""})); - assert (lex ("foo\" bar\"") == tokens ({"foo bar", ""})); - assert (lex ("\"foo \"\"bar\"") == tokens ({"foo bar", ""})); - assert (lex ("foo\" \"bar") == tokens ({"foo bar", ""})); - assert (lex ("\"foo\nbar\"") == tokens ({"foo\nbar", ""})); - assert (lex ("\"#:{})=+\n\"") == tokens ({"#:{})=+\n", ""})); - assert (lex ("\"'\"") == tokens ({"'", ""})); - assert (lex ("\"\\\"") == tokens ({"\\", ""})); - - assert (lex ("\"$\"") == tokens ({"", "$", "", ""})); - assert (lex ("\"foo$bar\"") == tokens ({"foo", "$", "bar", ""})); - assert (lex ("foo\"$\"bar") == tokens ({"foo", "$", "bar", ""})); - assert (lex ("f\"oo$ba\"r") == tokens ({"foo", "$", "bar", ""})); - - assert (lex ("\"foo bar") == tokens ({"<lexer error>"})); - assert (lex ("\"foo $") == tokens ({"foo ", "$", "<lexer error>"})); - assert (lex ("\"foo $bar") == tokens ({"foo ", "$", "<lexer error>"})); - - // Combinations. - // - assert (lex ("foo: bar") == tokens ({"foo", ":", "bar", ""})); - assert (lex ("\n \nfoo: bar") == tokens ({"foo", ":", "bar", ""})); - assert (lex ("foo: bar\n") == tokens ({"foo", ":", "bar", "\n", ""})); - assert (lex ("foo: bar#comment") == tokens ({"foo", ":", "bar", ""})); - assert (lex ("exe{foo}: obj{bar}") == - tokens ({"exe", "{", "foo", "}", ":", "obj", "{", "bar", "}", ""})); - assert (lex ("foo: bar\nbaz: biz") == - tokens ({"foo", ":", "bar", "\n", "baz", ":", "biz", ""})); - assert (lex ("foo: bar#comment\nbaz: biz") == - tokens ({"foo", ":", "bar", "\n", "baz", ":", "biz", ""})); - assert (lex ("foo:#comment \\\nbar") == - tokens ({"foo", ":", "\n", "bar", ""})); -} - -static tokens -lex (const char* s) -{ - tokens r; - istringstream is (s); - - is.exceptions (istream::failbit | istream::badbit); - lexer l (is, ""); - - try - { - for (token t (l.next ());; t = l.next ()) - { - string v; - - switch (t.type) - { - case token_type::eos: v = ""; break; - case token_type::newline: v = "\n"; break; - case token_type::pair_separator: v = l.pair_separator (); break; - case token_type::colon: v = ":"; break; - case token_type::lcbrace: v = "{"; break; - case token_type::rcbrace: v = "}"; break; - case token_type::equal: v = "="; break; - case token_type::plus_equal: v = "+="; break; - case token_type::dollar: v = "$"; break; - case token_type::lparen: v = "("; break; - case token_type::rparen: v = ")"; break; - case token_type::name: v = t.value.c_str (); break; - } - - // cerr << t.line () << ':' << t.column () << ':' << v << endl; - - r.push_back (move (v)); - - if (t.type == token_type::eos) - break; - } - } - catch (const failed&) - { - r.push_back ("<lexer error>"); - } - catch (const io_error&) - { - r.push_back ("<io error>"); - } - - return r; -} - -ostream& -operator<< (ostream& os, const tokens& ts) -{ - for (const string& t: ts) - os << '"' << t << '"' << ' '; - - return os; -} |