Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(465)

Unified Diff: Source/core/css/parser/CSSTokenizerTest.cpp

Issue 646893003: CSS Tokenizer: Comprehensive unit tests (Closed) Base URL: https://chromium.googlesource.com/chromium/blink.git@master
Patch Set: git cl try Created 6 years, 2 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « Source/core/css/parser/CSSParserToken.cpp ('k') | no next file » | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: Source/core/css/parser/CSSTokenizerTest.cpp
diff --git a/Source/core/css/parser/CSSTokenizerTest.cpp b/Source/core/css/parser/CSSTokenizerTest.cpp
index 6b414561f118eb34978dd0bac95159c796d2e74e..f4757883568eb05c765b97a3437af9d4bb1ddd57 100644
--- a/Source/core/css/parser/CSSTokenizerTest.cpp
+++ b/Source/core/css/parser/CSSTokenizerTest.cpp
@@ -6,15 +6,295 @@
#include "core/css/parser/CSSTokenizer.h"
#include "core/css/parser/MediaQueryBlockWatcher.h"
-#include "wtf/PassOwnPtr.h"
#include <gtest/gtest.h>
namespace blink {
-typedef struct {
- const char* input;
- const char* output;
-} TestCase;
+// This let's us see the line numbers of failing tests
+#define TEST_TOKENS(string, ...) { \
+ String s = string; \
+ SCOPED_TRACE(s.ascii().data()); \
+ testTokens(string, __VA_ARGS__); \
+}
+
+void compareTokens(const CSSParserToken& expected, const CSSParserToken& actual)
+{
+ ASSERT_EQ(expected.type(), actual.type());
+ switch (expected.type()) {
+ case DelimiterToken:
+ ASSERT_EQ(expected.delimiter(), actual.delimiter());
+ break;
+ case IdentToken:
+ case FunctionToken:
+ case StringToken:
+ ASSERT_EQ(expected.value(), actual.value());
+ break;
+ case DimensionToken:
+ ASSERT_EQ(expected.value(), actual.value());
+ // fallthrough
+ case NumberToken:
+ case PercentageToken:
+ ASSERT_EQ(expected.numericValueType(), actual.numericValueType());
+ ASSERT_DOUBLE_EQ(expected.numericValue(), actual.numericValue());
+ break;
+ default:
+ break;
+ }
+}
+
+void testTokens(const String& string, const CSSParserToken& token1, const CSSParserToken& token2 = CSSParserToken(EOFToken), const CSSParserToken& token3 = CSSParserToken(EOFToken))
+{
+ Vector<CSSParserToken> expectedTokens;
+ expectedTokens.append(token1);
+ if (token2.type() != EOFToken) {
+ expectedTokens.append(token2);
+ if (token3.type() != EOFToken)
+ expectedTokens.append(token3);
+ }
+
+ Vector<CSSParserToken> actualTokens;
+ CSSTokenizer::tokenize(string, actualTokens);
+ ASSERT_FALSE(actualTokens.isEmpty());
+ ASSERT_EQ(EOFToken, actualTokens.last().type());
+ actualTokens.removeLast();
+
+ ASSERT_EQ(expectedTokens.size(), actualTokens.size());
+ for (size_t i = 0; i < expectedTokens.size(); ++i)
+ compareTokens(expectedTokens[i], actualTokens[i]);
+}
+
+static CSSParserToken ident(const String& string) { return CSSParserToken(IdentToken, string); }
+static CSSParserToken string(const String& string) { return CSSParserToken(StringToken, string); }
+static CSSParserToken function(const String& string) { return CSSParserToken(FunctionToken, string); }
+static CSSParserToken delim(char c) { return CSSParserToken(DelimiterToken, c); }
+
+static CSSParserToken number(NumericValueType type, double value)
+{
+ return CSSParserToken(NumberToken, value, type);
+}
+
+static CSSParserToken dimension(NumericValueType type, double value, const String& string)
+{
+ CSSParserToken token = number(type, value);
+ token.convertToDimensionWithUnit(string);
+ return token;
+}
+
+static CSSParserToken percentage(NumericValueType type, double value)
+{
+ CSSParserToken token = number(type, value);
+ token.convertToPercentage();
+ return token;
+}
+
+DEFINE_STATIC_LOCAL(CSSParserToken, whitespace, (WhitespaceToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, colon, (ColonToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, semicolon, (SemicolonToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, comma, (CommaToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, leftParenthesis, (LeftParenthesisToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, rightParenthesis, (RightParenthesisToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, leftBracket, (LeftBracketToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, rightBracket, (RightBracketToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, leftBrace, (LeftBraceToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, rightBrace, (RightBraceToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, badString, (BadStringToken));
+DEFINE_STATIC_LOCAL(CSSParserToken, comment, (CommentToken));
+
+String fromUChar32(UChar32 c)
+{
+ StringBuilder input;
+ input.append(c);
+ return input.toString();
+}
+
+TEST(CSSTokenizerTest, SingleCharacterTokens)
+{
+ TEST_TOKENS("(", leftParenthesis);
+ TEST_TOKENS(")", rightParenthesis);
+ TEST_TOKENS("[", leftBracket);
+ TEST_TOKENS("]", rightBracket);
+ TEST_TOKENS(",", comma);
+ TEST_TOKENS(":", colon);
+ TEST_TOKENS(";", semicolon);
+ TEST_TOKENS(")[", rightParenthesis, leftBracket);
+ TEST_TOKENS("[)", leftBracket, rightParenthesis);
+ TEST_TOKENS("{}", leftBrace, rightBrace);
+ TEST_TOKENS(",,", comma, comma);
+}
+
+TEST(CSSTokenizerTest, DelimiterToken)
+{
+ TEST_TOKENS("*", delim('*'));
+ TEST_TOKENS("%", delim('%'));
+ TEST_TOKENS("~", delim('~'));
+ TEST_TOKENS("&", delim('&'));
+ TEST_TOKENS("\x7f", delim('\x7f'));
+ TEST_TOKENS("\1", delim('\x1'));
+}
+
+TEST(CSSTokenizerTest, WhitespaceTokens)
+{
+ TEST_TOKENS(" ", whitespace);
+ TEST_TOKENS("\n\rS", whitespace, ident("S"));
+ TEST_TOKENS(" *", whitespace, delim('*'));
+ TEST_TOKENS("\r\n\f\t2", whitespace, number(IntegerValueType, 2));
+}
+
+TEST(CSSTokenizerTest, Escapes)
+{
+ TEST_TOKENS("hel\\6Co", ident("hello"));
+ TEST_TOKENS("\\26 B", ident("&B"));
+ TEST_TOKENS("'hel\\6c o'", string("hello"));
+ TEST_TOKENS("'spac\\65\r\ns'", string("spaces"));
+ TEST_TOKENS("spac\\65\r\ns", ident("spaces"));
+ TEST_TOKENS("spac\\65\n\rs", ident("space"), whitespace, ident("s"));
+ TEST_TOKENS("sp\\61\tc\\65\fs", ident("spaces"));
+ TEST_TOKENS("hel\\6c o", ident("hell"), whitespace, ident("o"));
+ TEST_TOKENS("test\\\n", ident("test"), delim('\\'), whitespace);
+ TEST_TOKENS("eof\\", ident("eof"), delim('\\'));
+ TEST_TOKENS("test\\D799", ident("test" + fromUChar32(0xD799)));
+ TEST_TOKENS("\\E000", ident(fromUChar32(0xE000)));
+ TEST_TOKENS("te\\s\\t", ident("test"));
+ TEST_TOKENS("spaces\\ in\\\tident", ident("spaces in\tident"));
+ TEST_TOKENS("\\.\\,\\:\\!", ident(".,:!"));
+ // FIXME: We don't correctly return replacement characters
+ // String replacement = fromUChar32(0xFFFD);
+ // TEST_TOKENS("null\\0", ident("null" + replacement));
+ // TEST_TOKENS("null\\0000", ident("null" + replacement));
+ // TEST_TOKENS("large\\110000", ident("large" + replacement));
+ // TEST_TOKENS("surrogate\\D800", ident("surrogate" + replacement));
+ // TEST_TOKENS("surrogate\\0DABC", ident("surrogate" + replacement));
+ // TEST_TOKENS("\\00DFFFsurrogate", ident(replacement + "surrogate"));
+ // FIXME: We don't correctly return supplementary plane characters
+ // TEST_TOKENS("\\10fFfF", ident(fromUChar32(0x10ffff) + "0"));
+ // TEST_TOKENS("\\10000000", ident(fromUChar32(0x100000) + "000"));
+ // FIXME: We don't correctly match newlines (normally handled in preprocessing)
+ // TEST_TOKENS("\\\r", delim('\\'), whitespace);
+ // TEST_TOKENS("\\\f", delim('\\'), whitespace);
+ // TEST_TOKENS("\\\r\n", delim('\\'), whitespace);
+}
+
+TEST(CSSTokenizerTest, IdentToken)
+{
+ TEST_TOKENS("simple-ident", ident("simple-ident"));
+ TEST_TOKENS("testing123", ident("testing123"));
+ TEST_TOKENS("hello!", ident("hello"), delim('!'));
+ TEST_TOKENS("world\5", ident("world"), delim('\5'));
+ TEST_TOKENS("_under score", ident("_under"), whitespace, ident("score"));
+ TEST_TOKENS("-_underscore", ident("-_underscore"));
+ TEST_TOKENS("-text", ident("-text"));
+ TEST_TOKENS("-\\6d", ident("-m"));
+ TEST_TOKENS(fromUChar32(0x2003), ident(fromUChar32(0x2003))); // em-space
+ TEST_TOKENS(fromUChar32(0xA0), ident(fromUChar32(0xA0))); // non-breaking space
+ TEST_TOKENS(fromUChar32(0x1234), ident(fromUChar32(0x1234)));
+ TEST_TOKENS(fromUChar32(0x12345), ident(fromUChar32(0x12345)));
+ // FIXME: These are idents in the editor's draft
+ // TEST_TOKENS("--abc", ident("--abc"));
+ // TEST_TOKENS("--", ident("--"));
+ // TEST_TOKENS("---", ident("---"));
+ // FIXME: Preprocessing is supposed to replace U+0000 with U+FFFD
+ // TEST_TOKENS("\0", ident(fromUChar32(0xFFFD)));
+}
+
+TEST(CSSTokenizerTest, FunctionToken)
+{
+ TEST_TOKENS("scale(2)", function("scale"), number(IntegerValueType, 2), rightParenthesis);
+ TEST_TOKENS("foo-bar\\ baz(", function("foo-bar baz"));
+ TEST_TOKENS("fun\\(ction(", function("fun(ction"));
+ TEST_TOKENS("-foo(", function("-foo"));
+}
+
+TEST(CSSTokenizerTest, StringToken)
+{
+ TEST_TOKENS("'text'", string("text"));
+ TEST_TOKENS("\"text\"", string("text"));
+ TEST_TOKENS("'testing, 123!'", string("testing, 123!"));
+ TEST_TOKENS("'es\\'ca\\\"pe'", string("es'ca\"pe"));
+ TEST_TOKENS("'\"quotes\"'", string("\"quotes\""));
+ TEST_TOKENS("\"'quotes'\"", string("'quotes'"));
+ TEST_TOKENS("\"mismatch'", string("mismatch'"));
+ TEST_TOKENS("'text\5\t\13'", string("text\5\t\13"));
+ TEST_TOKENS("\"end on eof", string("end on eof"));
+ TEST_TOKENS("'esca\\\nped'", string("escaped"));
+ TEST_TOKENS("\"esc\\\faped\"", string("escaped"));
+ TEST_TOKENS("'new\\\rline'", string("newline"));
+ TEST_TOKENS("'bad\nstring", badString, whitespace, ident("string"));
+ TEST_TOKENS("'bad\rstring", badString, whitespace, ident("string"));
+ TEST_TOKENS("'bad\r\nstring", badString, whitespace, ident("string"));
+ TEST_TOKENS("'bad\fstring", badString, whitespace, ident("string"));
+ // FIXME: Preprocessing is supposed to replace U+0000 with U+FFFD
+ // TEST_TOKENS("'\0'", string(fromUChar32(0xFFFD)));
+ // FIXME: We don't correctly match newlines (normally handled in preprocessing)
+ // TEST_TOKENS("\"new\\\r\nline\"", string("newline"));
+}
+
+TEST(CSSTokenizerTest, NumberToken)
+{
+ TEST_TOKENS("10", number(IntegerValueType, 10));
+ TEST_TOKENS("12.0", number(NumberValueType, 12));
+ TEST_TOKENS("+45.6", number(NumberValueType, 45.6));
+ TEST_TOKENS("-7", number(IntegerValueType, -7));
+ TEST_TOKENS("010", number(IntegerValueType, 10));
+ TEST_TOKENS("10e0", number(NumberValueType, 10));
+ TEST_TOKENS("12e3", number(NumberValueType, 12000));
+ TEST_TOKENS("3e+1", number(NumberValueType, 30));
+ TEST_TOKENS("12E-1", number(NumberValueType, 1.2));
+ TEST_TOKENS(".7", number(NumberValueType, 0.7));
+ TEST_TOKENS("-.3", number(NumberValueType, -0.3));
+ TEST_TOKENS("+637.54e-2", number(NumberValueType, 6.3754));
+ TEST_TOKENS("-12.34E+2", number(NumberValueType, -1234));
+
+ TEST_TOKENS("+ 5", delim('+'), whitespace, number(IntegerValueType, 5));
+ TEST_TOKENS("--11", delim('-'), number(IntegerValueType, -11));
+ TEST_TOKENS("-+12", delim('-'), number(IntegerValueType, 12));
+ TEST_TOKENS("+-21", delim('+'), number(IntegerValueType, -21));
+ TEST_TOKENS("++22", delim('+'), number(IntegerValueType, 22));
+ TEST_TOKENS("13.", number(IntegerValueType, 13), delim('.'));
+ TEST_TOKENS("1.e2", number(IntegerValueType, 1), delim('.'), ident("e2"));
+ TEST_TOKENS("2e3.5", number(NumberValueType, 2000), number(NumberValueType, 0.5));
+ TEST_TOKENS("2e3.", number(NumberValueType, 2000), delim('.'));
+}
+
+TEST(CSSTokenizerTest, DimensionToken)
+{
+ TEST_TOKENS("10px", dimension(IntegerValueType, 10, "px"));
+ TEST_TOKENS("12.0em", dimension(NumberValueType, 12, "em"));
+ TEST_TOKENS("-12.0em", dimension(NumberValueType, -12, "em"));
+ TEST_TOKENS("+45.6__qem", dimension(NumberValueType, 45.6, "__qem"));
+ TEST_TOKENS("5e", dimension(IntegerValueType, 5, "e"));
+ TEST_TOKENS("5px-2px", dimension(IntegerValueType, 5, "px-2px"));
+ TEST_TOKENS("5e-", dimension(IntegerValueType, 5, "e-"));
+ TEST_TOKENS("5\\ ", dimension(IntegerValueType, 5, " "));
+ TEST_TOKENS("40\\70\\78", dimension(IntegerValueType, 40, "px"));
+ TEST_TOKENS("4e3e2", dimension(NumberValueType, 4000, "e2"));
+ TEST_TOKENS("0x10px", dimension(IntegerValueType, 0, "x1px"));
+ TEST_TOKENS("4unit ", dimension(IntegerValueType, 4, "unit"), whitespace);
+ TEST_TOKENS("5e+", dimension(IntegerValueType, 5, "e"), delim('+'));
+ TEST_TOKENS("2e.5", dimension(IntegerValueType, 2, "e"), number(NumberValueType, 0.5));
+ TEST_TOKENS("2e+.5", dimension(IntegerValueType, 2, "e"), number(NumberValueType, 0.5));
+}
+
+TEST(CSSTokenizerTest, PercentageToken)
+{
+ TEST_TOKENS("10%", percentage(IntegerValueType, 10));
+ TEST_TOKENS("+12.0%", percentage(NumberValueType, 12));
+ TEST_TOKENS("-48.99%", percentage(NumberValueType, -48.99));
+ TEST_TOKENS("6e-1%", percentage(NumberValueType, 0.6));
+ TEST_TOKENS("5%%", percentage(IntegerValueType, 5), delim('%'));
+}
+
+TEST(CSSTokenizerTest, CommentToken)
+{
+ TEST_TOKENS("/*comment*/", comment);
+ TEST_TOKENS("/**\\2f**/", comment);
+ TEST_TOKENS("/**y*a*y**/", comment);
+ TEST_TOKENS("/* \n :) \n */", comment);
+ TEST_TOKENS("/*/*/", comment);
+ TEST_TOKENS("/**/*", comment, delim('*'));
+ // FIXME: Should an EOF-terminated comment get a token?
+ // TEST_TOKENS("/******", comment);
+}
+
typedef struct {
const char* input;
@@ -22,64 +302,6 @@ typedef struct {
const unsigned finalLevel;
} BlockTestCase;
-TEST(CSSTokenizerTest, Basic)
-{
- TestCase testCases[] = {
- { "(max-width: 50px)", "(max-width: 50px)" },
- { "(max-width: 1e+2px)", "(max-width: 100.000000px)" },
- { "(max-width: 1e2px)", "(max-width: 100.000000px)" },
- { "(max-width: 1000e-1px)", "(max-width: 100.000000px)" },
- { "(max-width: 50\\70\\78)", "(max-width: 50px)" },
- { "(max-width: /* comment */50px)", "(max-width: 50px)" },
- { "(max-width: /** *commen*t */60px)", "(max-width: 60px)" },
- { "(max-width: /** *commen*t **/70px)", "(max-width: 70px)" },
- { "(max-width: /** *commen*t **//**/80px)", "(max-width: 80px)" },
- { "(max-width: /*/ **/90px)", "(max-width: 90px)" },
- { "(max-width: /*/ **/*100px)", "(max-width: '*'100px)" },
- { "(max-width: 110px/*)", "(max-width: 110px" },
- { "(max-width: 120px)/*", "(max-width: 120px)" },
- { "(max-width: 130px)/**", "(max-width: 130px)" },
- { "(max-width: /***/140px)/**/", "(max-width: 140px)" },
- { "(max-width: '40px')", "(max-width: 40px)" },
- { "(max-width: '40px", "(max-width: 40px" },
- { "(max-width: '40px\n", "(max-width: " },
- { "(max-width: '40px\\", "(max-width: 40px" },
- { "(max-width: '40px\\\n", "(max-width: 40px" },
- { "(max-width: '40px\\\n')", "(max-width: 40px)" },
- { "(max-width: '40\\70\\78')", "(max-width: 40px)" },
- { "(max-width: '40\\\npx')", "(max-width: 40px)" },
- { "(max-aspect-ratio: 5)", "(max-aspect-ratio: 5)" },
- { "(max-aspect-ratio: +5)", "(max-aspect-ratio: 5)" },
- { "(max-aspect-ratio: -5)", "(max-aspect-ratio: -5)" },
- { "(max-aspect-ratio: -+5)", "(max-aspect-ratio: '-'5)" },
- { "(max-aspect-ratio: +-5)", "(max-aspect-ratio: '+'-5)" },
- { "(max-aspect-ratio: +bla5)", "(max-aspect-ratio: '+'bla5)" },
- { "(max-aspect-ratio: +5bla)", "(max-aspect-ratio: 5other)" },
- { "(max-aspect-ratio: -bla)", "(max-aspect-ratio: -bla)" },
- { "(max-aspect-ratio: --bla)", "(max-aspect-ratio: '-'-bla)" },
- { "5e0", "5.000000" },
- { "5.0", "5.000000" },
- { "5.", "5'.'" },
- { "5.0e-1", "0.500000" },
- { "5.e-1", "5'.'e-1" },
- { "hel\\6co", "hello" },
- { "wor\\6c d", "world" },
- { "wor\\6c\r\nd wor\\6c\n\rd", "world worl d" },
- { "cod\\65point esca\\70\fe \\74\test", "codepoint escape test" },
- { "esca\\70\f\te \\74 \nest", "escap e t est" },
- { 0, 0 } // Do not remove the terminator line.
- };
-
- for (int i = 0; testCases[i].input; ++i) {
- Vector<CSSParserToken> tokens;
- CSSTokenizer::tokenize(testCases[i].input, tokens);
- StringBuilder output;
- for (size_t j = 0; j < tokens.size(); ++j)
- output.append(tokens[j].textForUnitTests());
- ASSERT_STREQ(testCases[i].output, output.toString().ascii().data());
- }
-}
-
TEST(CSSTokenizerBlockTest, Basic)
{
BlockTestCase testCases[] = {
@@ -128,56 +350,4 @@ TEST(CSSTokenizerBlockTest, Basic)
}
}
-void testToken(UChar c, CSSParserTokenType tokenType)
-{
- Vector<CSSParserToken> tokens;
- StringBuilder input;
- input.append(c);
- CSSTokenizer::tokenize(input.toString(), tokens);
- ASSERT_EQ(tokens[0].type(), tokenType);
-}
-
-TEST(CSSTokenizerCodepointsTest, Basic)
-{
- for (UChar c = 0; c <= 1000; ++c) {
- if (isASCIIDigit(c))
- testToken(c, NumberToken);
- else if (isASCIIAlpha(c))
- testToken(c, IdentToken);
- else if (c == '_')
- testToken(c, IdentToken);
- else if (c == '\r' || c == ' ' || c == '\n' || c == '\t' || c == '\f')
- testToken(c, WhitespaceToken);
- else if (c == '(')
- testToken(c, LeftParenthesisToken);
- else if (c == ')')
- testToken(c, RightParenthesisToken);
- else if (c == '[')
- testToken(c, LeftBracketToken);
- else if (c == ']')
- testToken(c, RightBracketToken);
- else if (c == '{')
- testToken(c, LeftBraceToken);
- else if (c == '}')
- testToken(c, RightBraceToken);
- else if (c == '.' || c == '+' || c == '-' || c == '/' || c == '\\')
- testToken(c, DelimiterToken);
- else if (c == '\'' || c == '"')
- testToken(c, StringToken);
- else if (c == ',')
- testToken(c, CommaToken);
- else if (c == ':')
- testToken(c, ColonToken);
- else if (c == ';')
- testToken(c, SemicolonToken);
- else if (!c)
- testToken(c, EOFToken);
- else if (c > SCHAR_MAX)
- testToken(c, IdentToken);
- else
- testToken(c, DelimiterToken);
- }
- testToken(USHRT_MAX, IdentToken);
-}
-
} // namespace
« no previous file with comments | « Source/core/css/parser/CSSParserToken.cpp ('k') | no next file » | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698