aboutsummaryrefslogtreecommitdiff
path: root/tests/lexer/driver.cxx
diff options
context:
space:
mode:
authorBoris Kolpackov <boris@codesynthesis.com>2016-12-01 12:07:18 +0200
committerBoris Kolpackov <boris@codesynthesis.com>2016-12-01 12:07:18 +0200
commit0ed86c76239d4f2904ea4ae1a77902a9e0db2a6d (patch)
tree5e8e1d43a0140f5739b21be7cbe120ebb9360d4f /tests/lexer/driver.cxx
parent12fc75188cc1f0a8c4c443c04e7a022131198c36 (diff)
Move old tests to old-tests/
Diffstat (limited to 'tests/lexer/driver.cxx')
-rw-r--r--tests/lexer/driver.cxx179
1 files changed, 0 insertions, 179 deletions
diff --git a/tests/lexer/driver.cxx b/tests/lexer/driver.cxx
deleted file mode 100644
index 5575956..0000000
--- a/tests/lexer/driver.cxx
+++ /dev/null
@@ -1,179 +0,0 @@
-// file : tests/build/lexer/driver.cxx -*- C++ -*-
-// copyright : Copyright (c) 2014-2016 Code Synthesis Ltd
-// license : MIT; see accompanying LICENSE file
-
-#include <string>
-#include <vector>
-#include <cassert>
-#include <sstream>
-#include <iostream>
-
-#include <build/token>
-#include <build/lexer>
-
-using namespace std;
-using namespace build;
-
-using tokens = vector<string>;
-
-static tokens
-lex (const char*);
-
-ostream&
-operator<< (ostream&, const tokens&);
-
-int
-main ()
-{
- ostream cnull (nullptr);
- diag_stream = &cnull;
-
- // Whitespaces.
- //
- assert (lex ("") == tokens ({""}));
- assert (lex ("\n") == tokens ({""}));
- assert (lex ("\n\n") == tokens ({""}));
- assert (lex (" \t \n") == tokens ({""}));
- assert (lex ("#comment") == tokens ({""}));
- assert (lex (" #comment") == tokens ({""}));
- assert (lex ("#comment\n") == tokens ({""}));
- assert (lex ("#comment\\\n") == tokens ({""}));
- assert (lex ("#comment 1\n#comment2") == tokens ({""}));
-
- // Punctuation.
- //
- assert (lex (": \n { }") == tokens ({":", "\n", "{", "}", ""}));
-
- // Names.
- //
- assert (lex ("foo") == tokens ({"foo", ""}));
- assert (lex ("foo.bar") == tokens ({"foo.bar", ""}));
-
- // Escaping.
- //
- assert (lex (" \\\n") == tokens ({""}));
- assert (lex ("\\\nfoo") == tokens ({"foo", ""}));
- assert (lex (" \\ foo") == tokens ({" foo", ""}));
- assert (lex ("fo\\ o\\:") == tokens ({"fo o:", ""}));
- assert (lex ("foo\\\nbar") == tokens ({"foobar", ""}));
- assert (lex ("foo \\\nbar") == tokens ({"foo", "bar", ""}));
- assert (lex ("\\'foo") == tokens ({"'foo", ""}));
-
- assert (lex (" \\") == tokens ({"<lexer error>"}));
- assert (lex (" foo\\") == tokens ({"<lexer error>"}));
-
-
- // Quoting ''.
- //
- assert (lex ("''") == tokens ({"", ""}));
- assert (lex ("'foo'") == tokens ({"foo", ""}));
- assert (lex ("'foo bar'") == tokens ({"foo bar", ""}));
- assert (lex ("'foo 'bar") == tokens ({"foo bar", ""}));
- assert (lex ("foo' bar'") == tokens ({"foo bar", ""}));
- assert (lex ("'foo ''bar'") == tokens ({"foo bar", ""}));
- assert (lex ("foo' 'bar") == tokens ({"foo bar", ""}));
- assert (lex ("'foo\nbar'") == tokens ({"foo\nbar", ""}));
- assert (lex ("'#:${}()=+\n'") == tokens ({"#:${}()=+\n", ""}));
- assert (lex ("'\"'") == tokens ({"\"", ""}));
- assert (lex ("'\\'") == tokens ({"\\", ""}));
-
- assert (lex ("'foo bar") == tokens ({"<lexer error>"}));
-
- // Quoting "".
- //
- assert (lex ("\"\"") == tokens ({"", ""}));
- assert (lex ("\"foo\"") == tokens ({"foo", ""}));
- assert (lex ("\"foo bar\"") == tokens ({"foo bar", ""}));
- assert (lex ("\"foo \"bar") == tokens ({"foo bar", ""}));
- assert (lex ("foo\" bar\"") == tokens ({"foo bar", ""}));
- assert (lex ("\"foo \"\"bar\"") == tokens ({"foo bar", ""}));
- assert (lex ("foo\" \"bar") == tokens ({"foo bar", ""}));
- assert (lex ("\"foo\nbar\"") == tokens ({"foo\nbar", ""}));
- assert (lex ("\"#:{})=+\n\"") == tokens ({"#:{})=+\n", ""}));
- assert (lex ("\"'\"") == tokens ({"'", ""}));
- assert (lex ("\"\\\"") == tokens ({"\\", ""}));
-
- assert (lex ("\"$\"") == tokens ({"", "$", "", ""}));
- assert (lex ("\"foo$bar\"") == tokens ({"foo", "$", "bar", ""}));
- assert (lex ("foo\"$\"bar") == tokens ({"foo", "$", "bar", ""}));
- assert (lex ("f\"oo$ba\"r") == tokens ({"foo", "$", "bar", ""}));
-
- assert (lex ("\"foo bar") == tokens ({"<lexer error>"}));
- assert (lex ("\"foo $") == tokens ({"foo ", "$", "<lexer error>"}));
- assert (lex ("\"foo $bar") == tokens ({"foo ", "$", "<lexer error>"}));
-
- // Combinations.
- //
- assert (lex ("foo: bar") == tokens ({"foo", ":", "bar", ""}));
- assert (lex ("\n \nfoo: bar") == tokens ({"foo", ":", "bar", ""}));
- assert (lex ("foo: bar\n") == tokens ({"foo", ":", "bar", "\n", ""}));
- assert (lex ("foo: bar#comment") == tokens ({"foo", ":", "bar", ""}));
- assert (lex ("exe{foo}: obj{bar}") ==
- tokens ({"exe", "{", "foo", "}", ":", "obj", "{", "bar", "}", ""}));
- assert (lex ("foo: bar\nbaz: biz") ==
- tokens ({"foo", ":", "bar", "\n", "baz", ":", "biz", ""}));
- assert (lex ("foo: bar#comment\nbaz: biz") ==
- tokens ({"foo", ":", "bar", "\n", "baz", ":", "biz", ""}));
- assert (lex ("foo:#comment \\\nbar") ==
- tokens ({"foo", ":", "\n", "bar", ""}));
-}
-
-static tokens
-lex (const char* s)
-{
- tokens r;
- istringstream is (s);
-
- is.exceptions (istream::failbit | istream::badbit);
- lexer l (is, "");
-
- try
- {
- for (token t (l.next ());; t = l.next ())
- {
- string v;
-
- switch (t.type)
- {
- case token_type::eos: v = ""; break;
- case token_type::newline: v = "\n"; break;
- case token_type::pair_separator: v = l.pair_separator (); break;
- case token_type::colon: v = ":"; break;
- case token_type::lcbrace: v = "{"; break;
- case token_type::rcbrace: v = "}"; break;
- case token_type::equal: v = "="; break;
- case token_type::plus_equal: v = "+="; break;
- case token_type::dollar: v = "$"; break;
- case token_type::lparen: v = "("; break;
- case token_type::rparen: v = ")"; break;
- case token_type::name: v = t.value.c_str (); break;
- }
-
- // cerr << t.line () << ':' << t.column () << ':' << v << endl;
-
- r.push_back (move (v));
-
- if (t.type == token_type::eos)
- break;
- }
- }
- catch (const failed&)
- {
- r.push_back ("<lexer error>");
- }
- catch (const io_error&)
- {
- r.push_back ("<io error>");
- }
-
- return r;
-}
-
-ostream&
-operator<< (ostream& os, const tokens& ts)
-{
- for (const string& t: ts)
- os << '"' << t << '"' << ' ';
-
- return os;
-}