mirror of
				https://github.com/LadybirdBrowser/ladybird.git
				synced 2025-11-03 23:00:58 +00:00 
			
		
		
		
	This is required for the `<urange>` type, and custom properties, to work correctly, as both need to know exactly what the original text was.
		
			
				
	
	
		
			1346 lines
		
	
	
	
		
			47 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
			
		
		
	
	
			1346 lines
		
	
	
	
		
			47 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
/*
 | 
						||
 * Copyright (c) 2020-2022, the SerenityOS developers.
 | 
						||
 * Copyright (c) 2021-2023, Sam Atkins <atkinssj@serenityos.org>
 | 
						||
 *
 | 
						||
 * SPDX-License-Identifier: BSD-2-Clause
 | 
						||
 */
 | 
						||
 | 
						||
#include <AK/CharacterTypes.h>
 | 
						||
#include <AK/Debug.h>
 | 
						||
#include <AK/FloatingPointStringConversions.h>
 | 
						||
#include <AK/SourceLocation.h>
 | 
						||
#include <AK/Vector.h>
 | 
						||
#include <LibTextCodec/Decoder.h>
 | 
						||
#include <LibWeb/CSS/Parser/Tokenizer.h>
 | 
						||
#include <LibWeb/Infra/Strings.h>
 | 
						||
 | 
						||
namespace Web::CSS::Parser {
 | 
						||
 | 
						||
// U+FFFD REPLACEMENT CHARACTER (<28>)
 | 
						||
#define REPLACEMENT_CHARACTER 0xFFFD
 | 
						||
static constexpr u32 TOKENIZER_EOF = 0xFFFFFFFF;
 | 
						||
 | 
						||
static inline void log_parse_error(SourceLocation const& location = SourceLocation::current())
 | 
						||
{
 | 
						||
    dbgln_if(CSS_TOKENIZER_DEBUG, "Parse error (css tokenization) {} ", location);
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_eof(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == TOKENIZER_EOF;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_quotation_mark(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x22;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_greater_than_maximum_allowed_code_point(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point > 0x10FFFF;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_low_line(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x5F;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#ident-start-code-point
 | 
						||
static inline bool is_ident_start_code_point(u32 code_point)
 | 
						||
{
 | 
						||
    // FIXME: We use !is_ascii() for "non-ASCII code point" in the spec, but it's not quite right -
 | 
						||
    //        it treats EOF as a valid! The spec also lacks a definition of code point. For now, the
 | 
						||
    //        !is_eof() check is a hack, but it should work.
 | 
						||
    return !is_eof(code_point) && (is_ascii_alpha(code_point) || !is_ascii(code_point) || is_low_line(code_point));
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_hyphen_minus(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x2D;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#ident-code-point
 | 
						||
static inline bool is_ident_code_point(u32 code_point)
 | 
						||
{
 | 
						||
    return is_ident_start_code_point(code_point) || is_ascii_digit(code_point) || is_hyphen_minus(code_point);
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_non_printable(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point <= 0x8 || code_point == 0xB || (code_point >= 0xE && code_point <= 0x1F) || code_point == 0x7F;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_number_sign(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x23;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_reverse_solidus(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x5C;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_apostrophe(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x27;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_left_paren(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x28;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_right_paren(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x29;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_plus_sign(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x2B;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_comma(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x2C;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_full_stop(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x2E;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_newline(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0xA;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_asterisk(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x2A;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_solidus(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x2F;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_colon(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x3A;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_semicolon(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x3B;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_less_than_sign(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x3C;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_greater_than_sign(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x3E;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_at(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x40;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_open_square_bracket(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x5B;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_closed_square_bracket(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x5D;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_open_curly_bracket(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x7B;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_closed_curly_bracket(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x7D;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_whitespace(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x9 || code_point == 0xA || code_point == 0x20;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_percent(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x25;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_exclamation_mark(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x21;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_e(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x65;
 | 
						||
}
 | 
						||
 | 
						||
static inline bool is_E(u32 code_point)
 | 
						||
{
 | 
						||
    return code_point == 0x45;
 | 
						||
}
 | 
						||
 | 
						||
ErrorOr<Vector<Token>> Tokenizer::tokenize(StringView input, StringView encoding)
 | 
						||
{
 | 
						||
    // https://www.w3.org/TR/css-syntax-3/#css-filter-code-points
 | 
						||
    auto filter_code_points = [](StringView input, auto encoding) -> ErrorOr<String> {
 | 
						||
        auto decoder = TextCodec::decoder_for(encoding);
 | 
						||
        VERIFY(decoder.has_value());
 | 
						||
 | 
						||
        StringBuilder builder { input.length() };
 | 
						||
        bool last_was_carriage_return = false;
 | 
						||
 | 
						||
        // To filter code points from a stream of (unfiltered) code points input:
 | 
						||
        TRY(decoder->process(input, [&builder, &last_was_carriage_return](u32 code_point) -> ErrorOr<void> {
 | 
						||
            // Replace any U+000D CARRIAGE RETURN (CR) code points,
 | 
						||
            // U+000C FORM FEED (FF) code points,
 | 
						||
            // or pairs of U+000D CARRIAGE RETURN (CR) followed by U+000A LINE FEED (LF)
 | 
						||
            // in input by a single U+000A LINE FEED (LF) code point.
 | 
						||
            if (code_point == '\r') {
 | 
						||
                if (last_was_carriage_return) {
 | 
						||
                    TRY(builder.try_append('\n'));
 | 
						||
                } else {
 | 
						||
                    last_was_carriage_return = true;
 | 
						||
                }
 | 
						||
            } else {
 | 
						||
                if (last_was_carriage_return)
 | 
						||
                    TRY(builder.try_append('\n'));
 | 
						||
 | 
						||
                if (code_point == '\n') {
 | 
						||
                    if (!last_was_carriage_return)
 | 
						||
                        TRY(builder.try_append('\n'));
 | 
						||
 | 
						||
                } else if (code_point == '\f') {
 | 
						||
                    TRY(builder.try_append('\n'));
 | 
						||
                    // Replace any U+0000 NULL or surrogate code points in input with U+FFFD REPLACEMENT CHARACTER (<28>).
 | 
						||
                } else if (code_point == 0x00 || (code_point >= 0xD800 && code_point <= 0xDFFF)) {
 | 
						||
                    TRY(builder.try_append_code_point(REPLACEMENT_CHARACTER));
 | 
						||
                } else {
 | 
						||
                    TRY(builder.try_append_code_point(code_point));
 | 
						||
                }
 | 
						||
 | 
						||
                last_was_carriage_return = false;
 | 
						||
            }
 | 
						||
            return {};
 | 
						||
        }));
 | 
						||
        return builder.to_string();
 | 
						||
    };
 | 
						||
 | 
						||
    Tokenizer tokenizer { TRY(filter_code_points(input, encoding)) };
 | 
						||
    return tokenizer.tokenize();
 | 
						||
}
 | 
						||
 | 
						||
Tokenizer::Tokenizer(String decoded_input)
 | 
						||
    : m_decoded_input(move(decoded_input))
 | 
						||
    , m_utf8_view(m_decoded_input)
 | 
						||
    , m_utf8_iterator(m_utf8_view.begin())
 | 
						||
{
 | 
						||
}
 | 
						||
 | 
						||
ErrorOr<Vector<Token>> Tokenizer::tokenize()
 | 
						||
{
 | 
						||
    Vector<Token> tokens;
 | 
						||
    for (;;) {
 | 
						||
        auto token_start = m_position;
 | 
						||
        auto token = TRY(consume_a_token());
 | 
						||
        token.m_start_position = token_start;
 | 
						||
        token.m_end_position = m_position;
 | 
						||
        TRY(tokens.try_append(token));
 | 
						||
 | 
						||
        if (token.is(Token::Type::EndOfFile)) {
 | 
						||
            return tokens;
 | 
						||
        }
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
u32 Tokenizer::next_code_point()
 | 
						||
{
 | 
						||
    if (m_utf8_iterator == m_utf8_view.end())
 | 
						||
        return TOKENIZER_EOF;
 | 
						||
    m_prev_utf8_iterator = m_utf8_iterator;
 | 
						||
    ++m_utf8_iterator;
 | 
						||
    auto code_point = *m_prev_utf8_iterator;
 | 
						||
 | 
						||
    m_prev_position = m_position;
 | 
						||
    if (is_newline(code_point)) {
 | 
						||
        m_position.line++;
 | 
						||
        m_position.column = 0;
 | 
						||
    } else {
 | 
						||
        m_position.column++;
 | 
						||
    }
 | 
						||
 | 
						||
    dbgln_if(CSS_TOKENIZER_DEBUG, "(Tokenizer) Next code_point: {:d}", code_point);
 | 
						||
    return code_point;
 | 
						||
}
 | 
						||
 | 
						||
u32 Tokenizer::peek_code_point(size_t offset) const
 | 
						||
{
 | 
						||
    auto it = m_utf8_iterator;
 | 
						||
    for (size_t i = 0; i < offset && it != m_utf8_view.end(); ++i)
 | 
						||
        ++it;
 | 
						||
    if (it == m_utf8_view.end())
 | 
						||
        return TOKENIZER_EOF;
 | 
						||
    dbgln_if(CSS_TOKENIZER_DEBUG, "(Tokenizer) Peek code_point: {:d}", *m_prev_utf8_iterator);
 | 
						||
    return *it;
 | 
						||
}
 | 
						||
 | 
						||
U32Twin Tokenizer::peek_twin() const
 | 
						||
{
 | 
						||
    U32Twin values { TOKENIZER_EOF, TOKENIZER_EOF };
 | 
						||
    auto it = m_utf8_iterator;
 | 
						||
    for (size_t i = 0; i < 2 && it != m_utf8_view.end(); ++i) {
 | 
						||
        values.set(i, *it);
 | 
						||
        ++it;
 | 
						||
    }
 | 
						||
    dbgln_if(CSS_TOKENIZER_DEBUG, "(Tokenizer) Peek twin: {:d},{:d}", values.first, values.second);
 | 
						||
    return values;
 | 
						||
}
 | 
						||
 | 
						||
U32Triplet Tokenizer::peek_triplet() const
 | 
						||
{
 | 
						||
    U32Triplet values { TOKENIZER_EOF, TOKENIZER_EOF, TOKENIZER_EOF };
 | 
						||
    auto it = m_utf8_iterator;
 | 
						||
    for (size_t i = 0; i < 3 && it != m_utf8_view.end(); ++i) {
 | 
						||
        values.set(i, *it);
 | 
						||
        ++it;
 | 
						||
    }
 | 
						||
    dbgln_if(CSS_TOKENIZER_DEBUG, "(Tokenizer) Peek triplet: {:d},{:d},{:d}", values.first, values.second, values.third);
 | 
						||
    return values;
 | 
						||
}
 | 
						||
 | 
						||
U32Twin Tokenizer::start_of_input_stream_twin()
 | 
						||
{
 | 
						||
    U32Twin twin;
 | 
						||
    // FIXME: Reconsuming just to read the current code point again is weird.
 | 
						||
    reconsume_current_input_code_point();
 | 
						||
    twin.first = next_code_point();
 | 
						||
    twin.second = peek_code_point();
 | 
						||
 | 
						||
    return twin;
 | 
						||
}
 | 
						||
 | 
						||
U32Triplet Tokenizer::start_of_input_stream_triplet()
 | 
						||
{
 | 
						||
    U32Triplet triplet;
 | 
						||
    // FIXME: Reconsuming just to read the current code point again is weird.
 | 
						||
    reconsume_current_input_code_point();
 | 
						||
    triplet.first = next_code_point();
 | 
						||
    auto next_two = peek_twin();
 | 
						||
    triplet.second = next_two.first;
 | 
						||
    triplet.third = next_two.second;
 | 
						||
 | 
						||
    return triplet;
 | 
						||
}
 | 
						||
 | 
						||
Token Tokenizer::create_new_token(Token::Type type)
 | 
						||
{
 | 
						||
    Token token = {};
 | 
						||
    token.m_type = type;
 | 
						||
    return token;
 | 
						||
}
 | 
						||
 | 
						||
Token Tokenizer::create_eof_token()
 | 
						||
{
 | 
						||
    return create_new_token(Token::Type::EndOfFile);
 | 
						||
}
 | 
						||
 | 
						||
Token Tokenizer::create_value_token(Token::Type type, FlyString&& value, String&& representation)
 | 
						||
{
 | 
						||
    auto token = create_new_token(type);
 | 
						||
    token.m_value = move(value);
 | 
						||
    token.m_representation = move(representation);
 | 
						||
    return token;
 | 
						||
}
 | 
						||
 | 
						||
Token Tokenizer::create_value_token(Token::Type type, u32 value, String&& representation)
 | 
						||
{
 | 
						||
    auto token = create_new_token(type);
 | 
						||
    token.m_value = String::from_code_point(value);
 | 
						||
    token.m_representation = move(representation);
 | 
						||
    return token;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-escaped-code-point
 | 
						||
u32 Tokenizer::consume_escaped_code_point()
 | 
						||
{
 | 
						||
    // This section describes how to consume an escaped code point.
 | 
						||
    // It assumes that the U+005C REVERSE SOLIDUS (\) has already been consumed and that the next
 | 
						||
    // input code point has already been verified to be part of a valid escape.
 | 
						||
    // It will return a code point.
 | 
						||
 | 
						||
    // Consume the next input code point.
 | 
						||
    auto input = next_code_point();
 | 
						||
 | 
						||
    // hex digit
 | 
						||
    if (is_ascii_hex_digit(input)) {
 | 
						||
        // Consume as many hex digits as possible, but no more than 5.
 | 
						||
        // Note that this means 1-6 hex digits have been consumed in total.
 | 
						||
        StringBuilder builder;
 | 
						||
        builder.append_code_point(input);
 | 
						||
 | 
						||
        size_t counter = 0;
 | 
						||
        while (is_ascii_hex_digit(peek_code_point()) && counter++ < 5) {
 | 
						||
            builder.append_code_point(next_code_point());
 | 
						||
        }
 | 
						||
 | 
						||
        // If the next input code point is whitespace, consume it as well.
 | 
						||
        if (is_whitespace(peek_code_point())) {
 | 
						||
            (void)next_code_point();
 | 
						||
        }
 | 
						||
 | 
						||
        // Interpret the hex digits as a hexadecimal number.
 | 
						||
        auto unhexed = AK::StringUtils::convert_to_uint_from_hex<u32>(builder.string_view()).value_or(0);
 | 
						||
        // If this number is zero, or is for a surrogate, or is greater than the maximum allowed
 | 
						||
        // code point, return U+FFFD REPLACEMENT CHARACTER (<28>).
 | 
						||
        if (unhexed == 0 || is_unicode_surrogate(unhexed) || is_greater_than_maximum_allowed_code_point(unhexed)) {
 | 
						||
            return REPLACEMENT_CHARACTER;
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, return the code point with that value.
 | 
						||
        return unhexed;
 | 
						||
    }
 | 
						||
 | 
						||
    // EOF
 | 
						||
    if (is_eof(input)) {
 | 
						||
        // This is a parse error. Return U+FFFD REPLACEMENT CHARACTER (<28>).
 | 
						||
        log_parse_error();
 | 
						||
        return REPLACEMENT_CHARACTER;
 | 
						||
    }
 | 
						||
 | 
						||
    // anything else
 | 
						||
    // Return the current input code point.
 | 
						||
    return input;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-ident-like-token
 | 
						||
ErrorOr<Token> Tokenizer::consume_an_ident_like_token()
 | 
						||
{
 | 
						||
    // This section describes how to consume an ident-like token from a stream of code points.
 | 
						||
    // It returns an <ident-token>, <function-token>, <url-token>, or <bad-url-token>.
 | 
						||
 | 
						||
    // Consume an ident sequence, and let string be the result.
 | 
						||
    auto start_byte_offset = current_byte_offset();
 | 
						||
    auto string = TRY(consume_an_ident_sequence());
 | 
						||
 | 
						||
    // If string’s value is an ASCII case-insensitive match for "url", and the next input code
 | 
						||
    // point is U+0028 LEFT PARENTHESIS ((), consume it.
 | 
						||
    if (Infra::is_ascii_case_insensitive_match(string, "url"sv) && is_left_paren(peek_code_point())) {
 | 
						||
        (void)next_code_point();
 | 
						||
 | 
						||
        // While the next two input code points are whitespace, consume the next input code point.
 | 
						||
        for (;;) {
 | 
						||
            auto maybe_whitespace = peek_twin();
 | 
						||
            if (!(is_whitespace(maybe_whitespace.first) && is_whitespace(maybe_whitespace.second))) {
 | 
						||
                break;
 | 
						||
            }
 | 
						||
 | 
						||
            (void)next_code_point();
 | 
						||
        }
 | 
						||
 | 
						||
        // If the next one or two input code points are U+0022 QUOTATION MARK ("), U+0027 APOSTROPHE ('),
 | 
						||
        // or whitespace followed by U+0022 QUOTATION MARK (") or U+0027 APOSTROPHE ('), then create a
 | 
						||
        // <function-token> with its value set to string and return it.
 | 
						||
        auto next_two = peek_twin();
 | 
						||
        if (is_quotation_mark(next_two.first) || is_apostrophe(next_two.first) || (is_whitespace(next_two.first) && (is_quotation_mark(next_two.second) || is_apostrophe(next_two.second)))) {
 | 
						||
            return create_value_token(Token::Type::Function, move(string), TRY(input_since(start_byte_offset)));
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, consume a url token, and return it.
 | 
						||
        return consume_a_url_token();
 | 
						||
    }
 | 
						||
 | 
						||
    // Otherwise, if the next input code point is U+0028 LEFT PARENTHESIS ((), consume it.
 | 
						||
    if (is_left_paren(peek_code_point())) {
 | 
						||
        (void)next_code_point();
 | 
						||
 | 
						||
        // Create a <function-token> with its value set to string and return it.
 | 
						||
        return create_value_token(Token::Type::Function, move(string), TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // Otherwise, create an <ident-token> with its value set to string and return it.
 | 
						||
    return create_value_token(Token::Type::Ident, move(string), TRY(input_since(start_byte_offset)));
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-number
 | 
						||
Number Tokenizer::consume_a_number()
 | 
						||
{
 | 
						||
    // This section describes how to consume a number from a stream of code points.
 | 
						||
    // It returns a numeric value, and a type which is either "integer" or "number".
 | 
						||
    //
 | 
						||
    // Note: This algorithm does not do the verification of the first few code points
 | 
						||
    // that are necessary to ensure a number can be obtained from the stream. Ensure
 | 
						||
    // that the stream starts with a number before calling this algorithm.
 | 
						||
 | 
						||
    // Execute the following steps in order:
 | 
						||
 | 
						||
    // 1. Initially set type to "integer". Let repr be the empty string.
 | 
						||
    StringBuilder repr;
 | 
						||
    Number::Type type = Number::Type::Integer;
 | 
						||
 | 
						||
    // 2. If the next input code point is U+002B PLUS SIGN (+) or U+002D HYPHEN-MINUS (-),
 | 
						||
    // consume it and append it to repr.
 | 
						||
    bool has_explicit_sign = false;
 | 
						||
    auto next_input = peek_code_point();
 | 
						||
    if (is_plus_sign(next_input) || is_hyphen_minus(next_input)) {
 | 
						||
        has_explicit_sign = true;
 | 
						||
        repr.append_code_point(next_code_point());
 | 
						||
    }
 | 
						||
 | 
						||
    // 3. While the next input code point is a digit, consume it and append it to repr.
 | 
						||
    for (;;) {
 | 
						||
        auto digits = peek_code_point();
 | 
						||
        if (!is_ascii_digit(digits))
 | 
						||
            break;
 | 
						||
 | 
						||
        repr.append_code_point(next_code_point());
 | 
						||
    }
 | 
						||
 | 
						||
    // 4. If the next 2 input code points are U+002E FULL STOP (.) followed by a digit, then:
 | 
						||
    auto maybe_number = peek_twin();
 | 
						||
    if (is_full_stop(maybe_number.first) && is_ascii_digit(maybe_number.second)) {
 | 
						||
        // 1. Consume them.
 | 
						||
        // 2. Append them to repr.
 | 
						||
        repr.append_code_point(next_code_point());
 | 
						||
        repr.append_code_point(next_code_point());
 | 
						||
 | 
						||
        // 3. Set type to "number".
 | 
						||
        type = Number::Type::Number;
 | 
						||
 | 
						||
        // 4. While the next input code point is a digit, consume it and append it to repr.
 | 
						||
        for (;;) {
 | 
						||
            auto digit = peek_code_point();
 | 
						||
            if (!is_ascii_digit(digit))
 | 
						||
                break;
 | 
						||
 | 
						||
            repr.append_code_point(next_code_point());
 | 
						||
        }
 | 
						||
    }
 | 
						||
 | 
						||
    // 5. If the next 2 or 3 input code points are U+0045 LATIN CAPITAL LETTER E (E) or
 | 
						||
    // U+0065 LATIN SMALL LETTER E (e), optionally followed by U+002D HYPHEN-MINUS (-)
 | 
						||
    // or U+002B PLUS SIGN (+), followed by a digit, then:
 | 
						||
    auto maybe_exp = peek_triplet();
 | 
						||
    if ((is_E(maybe_exp.first) || is_e(maybe_exp.first))
 | 
						||
        && (((is_plus_sign(maybe_exp.second) || is_hyphen_minus(maybe_exp.second)) && is_ascii_digit(maybe_exp.third))
 | 
						||
            || (is_ascii_digit(maybe_exp.second)))) {
 | 
						||
        // 1. Consume them.
 | 
						||
        // 2. Append them to repr.
 | 
						||
        if (is_plus_sign(maybe_exp.second) || is_hyphen_minus(maybe_exp.second)) {
 | 
						||
            if (is_ascii_digit(maybe_exp.third)) {
 | 
						||
                repr.append_code_point(next_code_point());
 | 
						||
                repr.append_code_point(next_code_point());
 | 
						||
                repr.append_code_point(next_code_point());
 | 
						||
            }
 | 
						||
        } else if (is_ascii_digit(maybe_exp.second)) {
 | 
						||
            repr.append_code_point(next_code_point());
 | 
						||
            repr.append_code_point(next_code_point());
 | 
						||
        }
 | 
						||
 | 
						||
        // 3. Set type to "number".
 | 
						||
        type = Number::Type::Number;
 | 
						||
 | 
						||
        // 4. While the next input code point is a digit, consume it and append it to repr.
 | 
						||
        for (;;) {
 | 
						||
            auto digits = peek_code_point();
 | 
						||
            if (!is_ascii_digit(digits))
 | 
						||
                break;
 | 
						||
 | 
						||
            repr.append_code_point(next_code_point());
 | 
						||
        }
 | 
						||
    }
 | 
						||
 | 
						||
    // 6. Convert repr to a number, and set the value to the returned value.
 | 
						||
    auto value = convert_a_string_to_a_number(repr.string_view());
 | 
						||
 | 
						||
    // 7. Return value and type.
 | 
						||
    if (type == Number::Type::Integer && has_explicit_sign)
 | 
						||
        return Number { Number::Type::IntegerWithExplicitSign, value };
 | 
						||
    return Number { type, value };
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#convert-string-to-number
 | 
						||
float Tokenizer::convert_a_string_to_a_number(StringView string)
 | 
						||
{
 | 
						||
    // FIXME: We already found the whole part, fraction part and exponent during
 | 
						||
    //        validation, we could probably skip
 | 
						||
    return string.to_float(AK::TrimWhitespace::No).release_value();
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-name
 | 
						||
ErrorOr<FlyString> Tokenizer::consume_an_ident_sequence()
 | 
						||
{
 | 
						||
    // This section describes how to consume an ident sequence from a stream of code points.
 | 
						||
    // It returns a string containing the largest name that can be formed from adjacent
 | 
						||
    // code points in the stream, starting from the first.
 | 
						||
    //
 | 
						||
    // Note: This algorithm does not do the verification of the first few code points that
 | 
						||
    // are necessary to ensure the returned code points would constitute an <ident-token>.
 | 
						||
    // If that is the intended use, ensure that the stream starts with an ident sequence before
 | 
						||
    // calling this algorithm.
 | 
						||
 | 
						||
    // Let result initially be an empty string.
 | 
						||
    StringBuilder result;
 | 
						||
 | 
						||
    // Repeatedly consume the next input code point from the stream:
 | 
						||
    for (;;) {
 | 
						||
        auto input = next_code_point();
 | 
						||
 | 
						||
        if (is_eof(input))
 | 
						||
            break;
 | 
						||
 | 
						||
        // name code point
 | 
						||
        if (is_ident_code_point(input)) {
 | 
						||
            // Append the code point to result.
 | 
						||
            TRY(result.try_append_code_point(input));
 | 
						||
            continue;
 | 
						||
        }
 | 
						||
 | 
						||
        // the stream starts with a valid escape
 | 
						||
        if (is_valid_escape_sequence(start_of_input_stream_twin())) {
 | 
						||
            // Consume an escaped code point. Append the returned code point to result.
 | 
						||
            TRY(result.try_append_code_point(consume_escaped_code_point()));
 | 
						||
            continue;
 | 
						||
        }
 | 
						||
 | 
						||
        // anything else
 | 
						||
        // Reconsume the current input code point. Return result.
 | 
						||
        reconsume_current_input_code_point();
 | 
						||
        break;
 | 
						||
    }
 | 
						||
 | 
						||
    return result.to_fly_string();
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-url-token
 | 
						||
ErrorOr<Token> Tokenizer::consume_a_url_token()
 | 
						||
{
 | 
						||
    // This section describes how to consume a url token from a stream of code points.
 | 
						||
    // It returns either a <url-token> or a <bad-url-token>.
 | 
						||
    //
 | 
						||
    // Note: This algorithm assumes that the initial "url(" has already been consumed.
 | 
						||
    // This algorithm also assumes that it’s being called to consume an "unquoted" value,
 | 
						||
    // like url(foo). A quoted value, like url("foo"), is parsed as a <function-token>.
 | 
						||
    // Consume an ident-like token automatically handles this distinction; this algorithm
 | 
						||
    // shouldn’t be called directly otherwise.
 | 
						||
 | 
						||
    // 1. Initially create a <url-token> with its value set to the empty string.
 | 
						||
    auto start_byte_offset = current_byte_offset();
 | 
						||
    auto token = create_new_token(Token::Type::Url);
 | 
						||
    StringBuilder builder;
 | 
						||
 | 
						||
    // 2. Consume as much whitespace as possible.
 | 
						||
    consume_as_much_whitespace_as_possible();
 | 
						||
 | 
						||
    auto make_token = [&]() -> ErrorOr<Token> {
 | 
						||
        token.m_value = TRY(FlyString::from_utf8(builder.string_view()));
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    };
 | 
						||
 | 
						||
    // 3. Repeatedly consume the next input code point from the stream:
 | 
						||
    for (;;) {
 | 
						||
        auto input = next_code_point();
 | 
						||
 | 
						||
        // U+0029 RIGHT PARENTHESIS ())
 | 
						||
        if (is_right_paren(input)) {
 | 
						||
            // Return the <url-token>.
 | 
						||
            return make_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // EOF
 | 
						||
        if (is_eof(input)) {
 | 
						||
            // This is a parse error. Return the <url-token>.
 | 
						||
            log_parse_error();
 | 
						||
            return make_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // whitespace
 | 
						||
        if (is_whitespace(input)) {
 | 
						||
            // Consume as much whitespace as possible.
 | 
						||
            consume_as_much_whitespace_as_possible();
 | 
						||
 | 
						||
            // If the next input code point is U+0029 RIGHT PARENTHESIS ()) or EOF, consume it
 | 
						||
            // and return the <url-token> (if EOF was encountered, this is a parse error);
 | 
						||
            input = peek_code_point();
 | 
						||
 | 
						||
            if (is_right_paren(input)) {
 | 
						||
                (void)next_code_point();
 | 
						||
                return make_token();
 | 
						||
            }
 | 
						||
 | 
						||
            if (is_eof(input)) {
 | 
						||
                (void)next_code_point();
 | 
						||
                log_parse_error();
 | 
						||
                return make_token();
 | 
						||
            }
 | 
						||
 | 
						||
            // otherwise, consume the remnants of a bad url, create a <bad-url-token>, and return it.
 | 
						||
            consume_the_remnants_of_a_bad_url();
 | 
						||
            auto bad_url_token = create_new_token(Token::Type::BadUrl);
 | 
						||
            bad_url_token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
            return bad_url_token;
 | 
						||
        }
 | 
						||
 | 
						||
        // U+0022 QUOTATION MARK (")
 | 
						||
        // U+0027 APOSTROPHE (')
 | 
						||
        // U+0028 LEFT PARENTHESIS (()
 | 
						||
        // non-printable code point
 | 
						||
        if (is_quotation_mark(input) || is_apostrophe(input) || is_left_paren(input) || is_non_printable(input)) {
 | 
						||
            // This is a parse error. Consume the remnants of a bad url, create a <bad-url-token>, and return it.
 | 
						||
            log_parse_error();
 | 
						||
            consume_the_remnants_of_a_bad_url();
 | 
						||
            auto bad_url_token = create_new_token(Token::Type::BadUrl);
 | 
						||
            bad_url_token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
            return bad_url_token;
 | 
						||
        }
 | 
						||
 | 
						||
        // U+005C REVERSE SOLIDUS (\)
 | 
						||
        if (is_reverse_solidus(input)) {
 | 
						||
            // If the stream starts with a valid escape,
 | 
						||
            if (is_valid_escape_sequence(start_of_input_stream_twin())) {
 | 
						||
                // consume an escaped code point and append the returned code point to the <url-token>’s value.
 | 
						||
                builder.append_code_point(consume_escaped_code_point());
 | 
						||
                continue;
 | 
						||
            } else {
 | 
						||
                // Otherwise, this is a parse error.
 | 
						||
                log_parse_error();
 | 
						||
                // Consume the remnants of a bad url, create a <bad-url-token>, and return it.
 | 
						||
                consume_the_remnants_of_a_bad_url();
 | 
						||
                auto bad_url_token = create_new_token(Token::Type::BadUrl);
 | 
						||
                bad_url_token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
                return bad_url_token;
 | 
						||
            }
 | 
						||
        }
 | 
						||
 | 
						||
        // anything else
 | 
						||
        // Append the current input code point to the <url-token>’s value.
 | 
						||
        builder.append_code_point(input);
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-remnants-of-bad-url
 | 
						||
void Tokenizer::consume_the_remnants_of_a_bad_url()
 | 
						||
{
 | 
						||
    // This section describes how to consume the remnants of a bad url from a stream of code points,
 | 
						||
    // "cleaning up" after the tokenizer realizes that it’s in the middle of a <bad-url-token> rather
 | 
						||
    // than a <url-token>. It returns nothing; its sole use is to consume enough of the input stream
 | 
						||
    // to reach a recovery point where normal tokenizing can resume.
 | 
						||
 | 
						||
    // Repeatedly consume the next input code point from the stream:
 | 
						||
    for (;;) {
 | 
						||
        auto input = next_code_point();
 | 
						||
 | 
						||
        // U+0029 RIGHT PARENTHESIS ())
 | 
						||
        // EOF
 | 
						||
        if (is_eof(input) || is_right_paren(input)) {
 | 
						||
            // Return.
 | 
						||
            return;
 | 
						||
        }
 | 
						||
 | 
						||
        // the input stream starts with a valid escape
 | 
						||
        if (is_valid_escape_sequence(start_of_input_stream_twin())) {
 | 
						||
            // Consume an escaped code point.
 | 
						||
            // This allows an escaped right parenthesis ("\)") to be encountered without ending
 | 
						||
            // the <bad-url-token>. This is otherwise identical to the "anything else" clause.
 | 
						||
            (void)consume_escaped_code_point();
 | 
						||
        }
 | 
						||
 | 
						||
        // anything else
 | 
						||
        // Do nothing.
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
void Tokenizer::consume_as_much_whitespace_as_possible()
 | 
						||
{
 | 
						||
    while (is_whitespace(peek_code_point())) {
 | 
						||
        (void)next_code_point();
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
void Tokenizer::reconsume_current_input_code_point()
 | 
						||
{
 | 
						||
    m_utf8_iterator = m_prev_utf8_iterator;
 | 
						||
    m_position = m_prev_position;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-numeric-token
 | 
						||
ErrorOr<Token> Tokenizer::consume_a_numeric_token()
 | 
						||
{
 | 
						||
    // This section describes how to consume a numeric token from a stream of code points.
 | 
						||
    // It returns either a <number-token>, <percentage-token>, or <dimension-token>.
 | 
						||
 | 
						||
    auto start_byte_offset = current_byte_offset();
 | 
						||
 | 
						||
    // Consume a number and let number be the result.
 | 
						||
    auto number = consume_a_number();
 | 
						||
 | 
						||
    // If the next 3 input code points would start an ident sequence, then:
 | 
						||
    if (would_start_an_ident_sequence(peek_triplet())) {
 | 
						||
        // 1. Create a <dimension-token> with the same value and type flag as number,
 | 
						||
        //    and a unit set initially to the empty string.
 | 
						||
        auto token = create_new_token(Token::Type::Dimension);
 | 
						||
        token.m_number_value = number;
 | 
						||
 | 
						||
        // 2. Consume an ident sequence. Set the <dimension-token>’s unit to the returned value.
 | 
						||
        auto unit = TRY(consume_an_ident_sequence());
 | 
						||
        VERIFY(!unit.is_empty());
 | 
						||
        // NOTE: We intentionally store this in the `value`, to save space.
 | 
						||
        token.m_value = move(unit);
 | 
						||
 | 
						||
        // 3. Return the <dimension-token>.
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // Otherwise, if the next input code point is U+0025 PERCENTAGE SIGN (%), consume it.
 | 
						||
    if (is_percent(peek_code_point())) {
 | 
						||
        (void)next_code_point();
 | 
						||
 | 
						||
        // Create a <percentage-token> with the same value as number, and return it.
 | 
						||
        auto token = create_new_token(Token::Type::Percentage);
 | 
						||
        token.m_number_value = number;
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // Otherwise, create a <number-token> with the same value and type flag as number, and return it.
 | 
						||
    auto token = create_new_token(Token::Type::Number);
 | 
						||
    token.m_number_value = number;
 | 
						||
    token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
    return token;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#starts-with-a-number
 | 
						||
bool Tokenizer::would_start_a_number(U32Triplet values)
 | 
						||
{
 | 
						||
    // This section describes how to check if three code points would start a number.
 | 
						||
    // The algorithm described here can be called explicitly with three code points,
 | 
						||
    // or can be called with the input stream itself. In the latter case, the three
 | 
						||
    // code points in question are the current input code point and the next two input
 | 
						||
    // code points, in that order.
 | 
						||
    //
 | 
						||
    // Note: This algorithm will not consume any additional code points.
 | 
						||
 | 
						||
    // Look at the first code point:
 | 
						||
 | 
						||
    // U+002B PLUS SIGN (+)
 | 
						||
    // U+002D HYPHEN-MINUS (-)
 | 
						||
    if (is_plus_sign(values.first) || is_hyphen_minus(values.first)) {
 | 
						||
        // If the second code point is a digit, return true.
 | 
						||
        if (is_ascii_digit(values.second))
 | 
						||
            return true;
 | 
						||
 | 
						||
        // Otherwise, if the second code point is a U+002E FULL STOP (.) and the third
 | 
						||
        // code point is a digit, return true.
 | 
						||
        if (is_full_stop(values.second) && is_ascii_digit(values.third))
 | 
						||
            return true;
 | 
						||
 | 
						||
        // Otherwise, return false.
 | 
						||
        return false;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+002E FULL STOP (.)
 | 
						||
    if (is_full_stop(values.first))
 | 
						||
        // If the second code point is a digit, return true. Otherwise, return false.
 | 
						||
        return is_ascii_digit(values.second);
 | 
						||
 | 
						||
    // digit
 | 
						||
    if (is_ascii_digit(values.first))
 | 
						||
        // Return true.
 | 
						||
        return true;
 | 
						||
 | 
						||
    // anything else
 | 
						||
    // Return false.
 | 
						||
    return false;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#starts-with-a-valid-escape
 | 
						||
bool Tokenizer::is_valid_escape_sequence(U32Twin values)
 | 
						||
{
 | 
						||
    // This section describes how to check if two code points are a valid escape.
 | 
						||
    // The algorithm described here can be called explicitly with two code points,
 | 
						||
    // or can be called with the input stream itself. In the latter case, the two
 | 
						||
    // code points in question are the current input code point and the next input
 | 
						||
    // code point, in that order.
 | 
						||
    //
 | 
						||
    // Note: This algorithm will not consume any additional code point.
 | 
						||
 | 
						||
    // If the first code point is not U+005C REVERSE SOLIDUS (\), return false.
 | 
						||
    if (!is_reverse_solidus(values.first))
 | 
						||
        return false;
 | 
						||
 | 
						||
    // Otherwise, if the second code point is a newline, return false.
 | 
						||
    if (is_newline(values.second))
 | 
						||
        return false;
 | 
						||
 | 
						||
    // Otherwise, return true.
 | 
						||
    return true;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#would-start-an-identifier
 | 
						||
bool Tokenizer::would_start_an_ident_sequence(U32Triplet values)
 | 
						||
{
 | 
						||
    // This section describes how to check if three code points would start an ident sequence.
 | 
						||
    // The algorithm described here can be called explicitly with three code points, or
 | 
						||
    // can be called with the input stream itself. In the latter case, the three code
 | 
						||
    // points in question are the current input code point and the next two input code
 | 
						||
    // points, in that order.
 | 
						||
    //
 | 
						||
    // Note: This algorithm will not consume any additional code points.
 | 
						||
 | 
						||
    // Look at the first code point:
 | 
						||
 | 
						||
    // U+002D HYPHEN-MINUS
 | 
						||
    if (is_hyphen_minus(values.first)) {
 | 
						||
        // If the second code point is a name-start code point or a U+002D HYPHEN-MINUS,
 | 
						||
        // or the second and third code points are a valid escape, return true.
 | 
						||
        if (is_ident_start_code_point(values.second) || is_hyphen_minus(values.second) || is_valid_escape_sequence(values.to_twin_23()))
 | 
						||
            return true;
 | 
						||
        // Otherwise, return false.
 | 
						||
        return false;
 | 
						||
    }
 | 
						||
 | 
						||
    // name-start code point
 | 
						||
    if (is_ident_start_code_point(values.first)) {
 | 
						||
        // Return true.
 | 
						||
        return true;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+005C REVERSE SOLIDUS (\)
 | 
						||
    if (is_reverse_solidus(values.first)) {
 | 
						||
        // If the first and second code points are a valid escape, return true.
 | 
						||
        if (is_valid_escape_sequence(values.to_twin_12()))
 | 
						||
            return true;
 | 
						||
        // Otherwise, return false.
 | 
						||
        return false;
 | 
						||
    }
 | 
						||
 | 
						||
    // anything else
 | 
						||
    // Return false.
 | 
						||
    return false;
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-string-token
 | 
						||
ErrorOr<Token> Tokenizer::consume_string_token(u32 ending_code_point)
 | 
						||
{
 | 
						||
    // This section describes how to consume a string token from a stream of code points.
 | 
						||
    // It returns either a <string-token> or <bad-string-token>.
 | 
						||
    //
 | 
						||
    // This algorithm may be called with an ending code point, which denotes the code point
 | 
						||
    // that ends the string. If an ending code point is not specified, the current input
 | 
						||
    // code point is used.
 | 
						||
 | 
						||
    // Initially create a <string-token> with its value set to the empty string.
 | 
						||
    auto start_byte_offset = current_byte_offset();
 | 
						||
    auto token = create_new_token(Token::Type::String);
 | 
						||
    StringBuilder builder;
 | 
						||
 | 
						||
    auto make_token = [&]() -> ErrorOr<Token> {
 | 
						||
        token.m_value = TRY(FlyString::from_utf8(builder.string_view()));
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    };
 | 
						||
 | 
						||
    // Repeatedly consume the next input code point from the stream:
 | 
						||
    for (;;) {
 | 
						||
        auto input = next_code_point();
 | 
						||
 | 
						||
        // ending code point
 | 
						||
        if (input == ending_code_point)
 | 
						||
            return make_token();
 | 
						||
 | 
						||
        // EOF
 | 
						||
        if (is_eof(input)) {
 | 
						||
            // This is a parse error. Return the <string-token>.
 | 
						||
            log_parse_error();
 | 
						||
            return make_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // newline
 | 
						||
        if (is_newline(input)) {
 | 
						||
            // This is a parse error. Reconsume the current input code point, create a
 | 
						||
            // <bad-string-token>, and return it.
 | 
						||
            reconsume_current_input_code_point();
 | 
						||
            auto bad_string_token = create_new_token(Token::Type::BadString);
 | 
						||
            bad_string_token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
            return bad_string_token;
 | 
						||
        }
 | 
						||
 | 
						||
        // U+005C REVERSE SOLIDUS (\)
 | 
						||
        if (is_reverse_solidus(input)) {
 | 
						||
            // If the next input code point is EOF, do nothing.
 | 
						||
            auto next_input = peek_code_point();
 | 
						||
            if (is_eof(next_input))
 | 
						||
                continue;
 | 
						||
 | 
						||
            // Otherwise, if the next input code point is a newline, consume it.
 | 
						||
            if (is_newline(next_input)) {
 | 
						||
                (void)next_code_point();
 | 
						||
                continue;
 | 
						||
            }
 | 
						||
 | 
						||
            // Otherwise, (the stream starts with a valid escape) consume an escaped code
 | 
						||
            // point and append the returned code point to the <string-token>’s value.
 | 
						||
            auto escaped = consume_escaped_code_point();
 | 
						||
            builder.append_code_point(escaped);
 | 
						||
            continue;
 | 
						||
        }
 | 
						||
 | 
						||
        // anything else
 | 
						||
        // Append the current input code point to the <string-token>’s value.
 | 
						||
        builder.append_code_point(input);
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-comment
 | 
						||
void Tokenizer::consume_comments()
 | 
						||
{
 | 
						||
    // This section describes how to consume comments from a stream of code points.
 | 
						||
    // It returns nothing.
 | 
						||
 | 
						||
start:
 | 
						||
    // If the next two input code point are U+002F SOLIDUS (/) followed by a U+002A ASTERISK (*),
 | 
						||
    // consume them and all following code points up to and including the first U+002A ASTERISK (*)
 | 
						||
    // followed by a U+002F SOLIDUS (/), or up to an EOF code point. Return to the start of this step.
 | 
						||
    //
 | 
						||
    // If the preceding paragraph ended by consuming an EOF code point, this is a parse error.
 | 
						||
    //
 | 
						||
    // Return nothing.
 | 
						||
    auto twin = peek_twin();
 | 
						||
    if (!(is_solidus(twin.first) && is_asterisk(twin.second)))
 | 
						||
        return;
 | 
						||
 | 
						||
    (void)next_code_point();
 | 
						||
    (void)next_code_point();
 | 
						||
 | 
						||
    for (;;) {
 | 
						||
        auto twin_inner = peek_twin();
 | 
						||
        if (is_eof(twin_inner.first) || is_eof(twin_inner.second)) {
 | 
						||
            log_parse_error();
 | 
						||
            return;
 | 
						||
        }
 | 
						||
 | 
						||
        if (is_asterisk(twin_inner.first) && is_solidus(twin_inner.second)) {
 | 
						||
            (void)next_code_point();
 | 
						||
            (void)next_code_point();
 | 
						||
            goto start;
 | 
						||
        }
 | 
						||
 | 
						||
        (void)next_code_point();
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
// https://www.w3.org/TR/css-syntax-3/#consume-token
 | 
						||
ErrorOr<Token> Tokenizer::consume_a_token()
 | 
						||
{
 | 
						||
    // This section describes how to consume a token from a stream of code points.
 | 
						||
    // It will return a single token of any type.
 | 
						||
 | 
						||
    // Consume comments.
 | 
						||
    consume_comments();
 | 
						||
 | 
						||
    // Consume the next input code point.
 | 
						||
    auto start_byte_offset = current_byte_offset();
 | 
						||
    auto input = next_code_point();
 | 
						||
 | 
						||
    // whitespace
 | 
						||
    if (is_whitespace(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is whitespace");
 | 
						||
        // Consume as much whitespace as possible. Return a <whitespace-token>.
 | 
						||
        consume_as_much_whitespace_as_possible();
 | 
						||
        auto token = create_new_token(Token::Type::Whitespace);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+0022 QUOTATION MARK (")
 | 
						||
    if (is_quotation_mark(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is quotation mark");
 | 
						||
        // Consume a string token and return it.
 | 
						||
        return consume_string_token(input);
 | 
						||
    }
 | 
						||
 | 
						||
    // U+0023 NUMBER SIGN (#)
 | 
						||
    if (is_number_sign(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is number sign");
 | 
						||
 | 
						||
        // If the next input code point is an ident code point or the next two input code points
 | 
						||
        // are a valid escape, then:
 | 
						||
        auto next_input = peek_code_point();
 | 
						||
        auto maybe_escape = peek_twin();
 | 
						||
 | 
						||
        if (is_ident_code_point(next_input) || is_valid_escape_sequence(maybe_escape)) {
 | 
						||
            // 1. Create a <hash-token>.
 | 
						||
            auto token = create_new_token(Token::Type::Hash);
 | 
						||
 | 
						||
            // 2. If the next 3 input code points would start an ident sequence, set the <hash-token>’s
 | 
						||
            //    type flag to "id".
 | 
						||
            if (would_start_an_ident_sequence(peek_triplet()))
 | 
						||
                token.m_hash_type = Token::HashType::Id;
 | 
						||
 | 
						||
            // 3. Consume an ident sequence, and set the <hash-token>’s value to the returned string.
 | 
						||
            auto name = TRY(consume_an_ident_sequence());
 | 
						||
            token.m_value = move(name);
 | 
						||
 | 
						||
            // 4. Return the <hash-token>.
 | 
						||
            token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
            return token;
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, return a <delim-token> with its value set to the current input code point.
 | 
						||
        return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // U+0027 APOSTROPHE (')
 | 
						||
    if (is_apostrophe(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is apostrophe");
 | 
						||
        // Consume a string token and return it.
 | 
						||
        return consume_string_token(input);
 | 
						||
    }
 | 
						||
 | 
						||
    // U+0028 LEFT PARENTHESIS (()
 | 
						||
    if (is_left_paren(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is left paren");
 | 
						||
        // Return a <(-token>.
 | 
						||
        Token token = create_new_token(Token::Type::OpenParen);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+0029 RIGHT PARENTHESIS ())
 | 
						||
    if (is_right_paren(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is right paren");
 | 
						||
        // Return a <)-token>.
 | 
						||
        Token token = create_new_token(Token::Type::CloseParen);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+002B PLUS SIGN (+)
 | 
						||
    if (is_plus_sign(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is plus sign");
 | 
						||
        // If the input stream starts with a number, reconsume the current input code point,
 | 
						||
        // consume a numeric token and return it.
 | 
						||
        if (would_start_a_number(start_of_input_stream_triplet())) {
 | 
						||
            reconsume_current_input_code_point();
 | 
						||
            return consume_a_numeric_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, return a <delim-token> with its value set to the current input code point.
 | 
						||
        return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // U+002C COMMA (,)
 | 
						||
    if (is_comma(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is comma");
 | 
						||
        // Return a <comma-token>.
 | 
						||
        Token token = create_new_token(Token::Type::Comma);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+002D HYPHEN-MINUS (-)
 | 
						||
    if (is_hyphen_minus(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is hyphen minus");
 | 
						||
        // If the input stream starts with a number, reconsume the current input code point,
 | 
						||
        // consume a numeric token, and return it.
 | 
						||
        if (would_start_a_number(start_of_input_stream_triplet())) {
 | 
						||
            reconsume_current_input_code_point();
 | 
						||
            return consume_a_numeric_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, if the next 2 input code points are U+002D HYPHEN-MINUS U+003E
 | 
						||
        // GREATER-THAN SIGN (->), consume them and return a <CDC-token>.
 | 
						||
        auto next_twin = peek_twin();
 | 
						||
        if (is_hyphen_minus(next_twin.first) && is_greater_than_sign(next_twin.second)) {
 | 
						||
            (void)next_code_point();
 | 
						||
            (void)next_code_point();
 | 
						||
 | 
						||
            Token token = create_new_token(Token::Type::CDC);
 | 
						||
            token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
            return token;
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, if the input stream starts with an identifier, reconsume the current
 | 
						||
        // input code point, consume an ident-like token, and return it.
 | 
						||
        if (would_start_an_ident_sequence(start_of_input_stream_triplet())) {
 | 
						||
            reconsume_current_input_code_point();
 | 
						||
            return consume_an_ident_like_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, return a <delim-token> with its value set to the current input code point.
 | 
						||
        return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // U+002E FULL STOP (.)
 | 
						||
    if (is_full_stop(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is full stop");
 | 
						||
        // If the input stream starts with a number, reconsume the current input code point,
 | 
						||
        // consume a numeric token, and return it.
 | 
						||
        if (would_start_a_number(start_of_input_stream_triplet())) {
 | 
						||
            reconsume_current_input_code_point();
 | 
						||
            return consume_a_numeric_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, return a <delim-token> with its value set to the current input code point.
 | 
						||
        return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // U+003A COLON (:)
 | 
						||
    if (is_colon(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is colon");
 | 
						||
        // Return a <colon-token>.
 | 
						||
        Token token = create_new_token(Token::Type::Colon);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+003B SEMICOLON (;)
 | 
						||
    if (is_semicolon(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is semicolon");
 | 
						||
        // Return a <semicolon-token>.
 | 
						||
        Token token = create_new_token(Token::Type::Semicolon);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+003C LESS-THAN SIGN (<)
 | 
						||
    if (is_less_than_sign(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is less than");
 | 
						||
        // If the next 3 input code points are U+0021 EXCLAMATION MARK U+002D HYPHEN-MINUS
 | 
						||
        // U+002D HYPHEN-MINUS (!--), consume them and return a <CDO-token>.
 | 
						||
        auto maybe_cdo = peek_triplet();
 | 
						||
        if (is_exclamation_mark(maybe_cdo.first) && is_hyphen_minus(maybe_cdo.second) && is_hyphen_minus(maybe_cdo.third)) {
 | 
						||
            (void)next_code_point();
 | 
						||
            (void)next_code_point();
 | 
						||
            (void)next_code_point();
 | 
						||
 | 
						||
            Token token = create_new_token(Token::Type::CDO);
 | 
						||
            token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
            return token;
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, return a <delim-token> with its value set to the current input code point.
 | 
						||
        return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // U+0040 COMMERCIAL AT (@)
 | 
						||
    if (is_at(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is at");
 | 
						||
        // If the next 3 input code points would start an ident sequence, consume an ident sequence, create
 | 
						||
        // an <at-keyword-token> with its value set to the returned value, and return it.
 | 
						||
        if (would_start_an_ident_sequence(peek_triplet())) {
 | 
						||
            auto name = TRY(consume_an_ident_sequence());
 | 
						||
            return create_value_token(Token::Type::AtKeyword, move(name), TRY(input_since(start_byte_offset)));
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, return a <delim-token> with its value set to the current input code point.
 | 
						||
        return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // U+005B LEFT SQUARE BRACKET ([)
 | 
						||
    if (is_open_square_bracket(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is open square");
 | 
						||
        // Return a <[-token>.
 | 
						||
        Token token = create_new_token(Token::Type::OpenSquare);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+005C REVERSE SOLIDUS (\)
 | 
						||
    if (is_reverse_solidus(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is reverse solidus");
 | 
						||
        // If the input stream starts with a valid escape, reconsume the current input code point,
 | 
						||
        // consume an ident-like token, and return it.
 | 
						||
        if (is_valid_escape_sequence(start_of_input_stream_twin())) {
 | 
						||
            reconsume_current_input_code_point();
 | 
						||
            return consume_an_ident_like_token();
 | 
						||
        }
 | 
						||
 | 
						||
        // Otherwise, this is a parse error. Return a <delim-token> with its value set to the
 | 
						||
        // current input code point.
 | 
						||
        log_parse_error();
 | 
						||
        return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
    }
 | 
						||
 | 
						||
    // U+005D RIGHT SQUARE BRACKET (])
 | 
						||
    if (is_closed_square_bracket(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is closed square");
 | 
						||
        // Return a <]-token>.
 | 
						||
        Token token = create_new_token(Token::Type::CloseSquare);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+007B LEFT CURLY BRACKET ({)
 | 
						||
    if (is_open_curly_bracket(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is open curly");
 | 
						||
        // Return a <{-token>.
 | 
						||
        Token token = create_new_token(Token::Type::OpenCurly);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // U+007D RIGHT CURLY BRACKET (})
 | 
						||
    if (is_closed_curly_bracket(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is closed curly");
 | 
						||
        // Return a <}-token>.
 | 
						||
        Token token = create_new_token(Token::Type::CloseCurly);
 | 
						||
        token.m_representation = TRY(input_since(start_byte_offset));
 | 
						||
        return token;
 | 
						||
    }
 | 
						||
 | 
						||
    // digit
 | 
						||
    if (is_ascii_digit(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is digit");
 | 
						||
        // Reconsume the current input code point, consume a numeric token, and return it.
 | 
						||
        reconsume_current_input_code_point();
 | 
						||
        return consume_a_numeric_token();
 | 
						||
    }
 | 
						||
 | 
						||
    // name-start code point
 | 
						||
    if (is_ident_start_code_point(input)) {
 | 
						||
        dbgln_if(CSS_TOKENIZER_DEBUG, "is name start");
 | 
						||
        // Reconsume the current input code point, consume an ident-like token, and return it.
 | 
						||
        reconsume_current_input_code_point();
 | 
						||
        return consume_an_ident_like_token();
 | 
						||
    }
 | 
						||
 | 
						||
    // EOF
 | 
						||
    if (is_eof(input)) {
 | 
						||
        // Return an <EOF-token>.
 | 
						||
        return create_eof_token();
 | 
						||
    }
 | 
						||
 | 
						||
    // anything else
 | 
						||
    dbgln_if(CSS_TOKENIZER_DEBUG, "is delimiter");
 | 
						||
    // Return a <delim-token> with its value set to the current input code point.
 | 
						||
    return create_value_token(Token::Type::Delim, input, TRY(input_since(start_byte_offset)));
 | 
						||
}
 | 
						||
 | 
						||
size_t Tokenizer::current_byte_offset() const
 | 
						||
{
 | 
						||
    return m_utf8_iterator.ptr() - m_utf8_view.bytes();
 | 
						||
}
 | 
						||
 | 
						||
ErrorOr<String> Tokenizer::input_since(size_t offset) const
 | 
						||
{
 | 
						||
    return m_decoded_input.substring_from_byte_offset_with_shared_superstring(offset, current_byte_offset() - offset);
 | 
						||
}
 | 
						||
 | 
						||
}
 |