diff options
Diffstat (limited to 'test/core')
-rw-r--r-- | test/core/utils/CodeTokenizerTest.cpp | 30 | ||||
-rw-r--r-- | test/core/utils/TokenizerTest.cpp | 34 |
2 files changed, 64 insertions, 0 deletions
diff --git a/test/core/utils/CodeTokenizerTest.cpp b/test/core/utils/CodeTokenizerTest.cpp new file mode 100644 index 0000000..d0f9a17 --- /dev/null +++ b/test/core/utils/CodeTokenizerTest.cpp @@ -0,0 +1,30 @@ +/* + Ousía + Copyright (C) 2014, 2015 Benjamin Paaßen, Andreas Stöckel + + This program is free software: you can redistribute it and/or modify + it under the terms of the GNU General Public License as published by + the Free Software Foundation, either version 3 of the License, or + (at your option) any later version. + + This program is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + GNU General Public License for more details. + + You should have received a copy of the GNU General Public License + along with this program. If not, see <http://www.gnu.org/licenses/>. +*/ + +#include <gtest/gtest.h> + +#include <core/utils/CodeTokenizer.hpp> + +namespace ousia { +namespace utils { +TEST(CodeTokenizer, testTokenizer) +{ + +} +} +} diff --git a/test/core/utils/TokenizerTest.cpp b/test/core/utils/TokenizerTest.cpp index f441fd8..ba06c33 100644 --- a/test/core/utils/TokenizerTest.cpp +++ b/test/core/utils/TokenizerTest.cpp @@ -18,6 +18,8 @@ #include <gtest/gtest.h> +#include <core/utils/BufferedCharReader.hpp> + #include <core/utils/Tokenizer.hpp> namespace ousia { @@ -59,5 +61,37 @@ TEST(TokenTreeNode, testConstructor) ASSERT_EQ(4, abd.tokenId); ASSERT_EQ(0, abd.children.size()); } + +TEST(Tokenizer, testTokenization) +{ + TokenTreeNode root{{{"/", 1}, {"/*", 2}, {"*/", 3}}}; + + BufferedCharReader reader; + reader.feed("Test/Test /* Block Comment */"); + // 12345678901234567890123456789 + // 0 1 2 + + std::vector<Token> expected = { + {TOKEN_TEXT, "Test", 1, 1, 5, 1}, + {1, "/", 5, 1, 6, 1}, + {TOKEN_TEXT, "Test ", 6, 1, 11, 1}, + {2, "/*", 11, 1, 13, 1}, + {TOKEN_TEXT, " Block Comment ", 13, 1, 28, 1}, + {3, "*/", 28, 1, 30, 1}}; + + Tokenizer tokenizer{reader, root}; + + Token t; + for (auto &te : expected) { + ASSERT_TRUE(tokenizer.next(t)); + ASSERT_EQ(te.tokenId, t.tokenId); + ASSERT_EQ(te.content, t.content); + ASSERT_EQ(te.startColumn, t.startColumn); + ASSERT_EQ(te.startLine, t.startLine); + ASSERT_EQ(te.endColumn, t.endColumn); + ASSERT_EQ(te.endLine, t.endLine); + } + ASSERT_FALSE(tokenizer.next(t)); +} } } |