blob: 5a3473a4b083ca3b5f69652e2b2455fb00de468c [file] [log] [blame]
//===-- ResourceScriptToken.cpp ---------------------------------*- C++-*-===//
// The LLVM Compiler Infrastructure
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
// This file implements an interface defined in ResourceScriptToken.h.
// In particular, it defines an .rc script tokenizer.
#include "ResourceScriptToken.h"
#include "llvm/Support/raw_ostream.h"
#include <algorithm>
#include <cassert>
#include <cctype>
#include <cstdlib>
#include <utility>
using namespace llvm;
using Kind = RCToken::Kind;
// Checks if Representation is a correct description of an RC integer.
// It should be a 32-bit unsigned integer, either decimal, octal (0[0-7]+),
// or hexadecimal (0x[0-9a-f]+). It might be followed by a single 'L'
// character (that is the difference between our representation and
// StringRef's one). If Representation is correct, 'true' is returned and
// the return value is put back in Num.
static bool rcGetAsInteger(StringRef Representation, uint32_t &Num) {
size_t Length = Representation.size();
if (Length == 0)
return false;
// Strip the last 'L' if unnecessary.
if (std::toupper(Representation.back()) == 'L')
Representation = Representation.drop_back(1);
return !Representation.getAsInteger<uint32_t>(0, Num);
RCToken::RCToken(RCToken::Kind RCTokenKind, StringRef Value)
: TokenKind(RCTokenKind), TokenValue(Value) {}
uint32_t RCToken::intValue() const {
assert(TokenKind == Kind::Int);
// We assume that the token already is a correct integer (checked by
// rcGetAsInteger).
uint32_t Result;
bool IsSuccess = rcGetAsInteger(TokenValue, Result);
(void)IsSuccess; // Silence the compiler warning when -DNDEBUG flag is on.
return Result;
bool RCToken::isLongInt() const {
return TokenKind == Kind::Int && std::toupper(TokenValue.back()) == 'L';
StringRef RCToken::value() const { return TokenValue; }
Kind RCToken::kind() const { return TokenKind; }
bool RCToken::isBinaryOp() const {
switch (TokenKind) {
case Kind::Plus:
case Kind::Minus:
case Kind::Pipe:
case Kind::Amp:
return true;
return false;
static Error getStringError(const Twine &message) {
return make_error<StringError>("Error parsing file: " + message,
namespace {
class Tokenizer {
Tokenizer(StringRef Input) : Data(Input), DataLength(Input.size()) {}
Expected<std::vector<RCToken>> run();
// All 'advancing' methods return boolean values; if they're equal to false,
// the stream has ended or failed.
bool advance(size_t Amount = 1);
bool skipWhitespaces();
// Consumes a token. If any problem occurred, a non-empty Error is returned.
Error consumeToken(const Kind TokenKind);
// Check if tokenizer is about to read FollowingChars.
bool willNowRead(StringRef FollowingChars) const;
// Check if tokenizer can start reading an identifier at current position.
// The original tool did non specify the rules to determine what is a correct
// identifier. We assume they should follow the C convention:
// [a-zA-Z_][a-zA-Z0-9_]*.
bool canStartIdentifier() const;
// Check if tokenizer can continue reading an identifier.
bool canContinueIdentifier() const;
// Check if tokenizer can start reading an integer.
// A correct integer always starts with a 0-9 digit,
// can contain characters 0-9A-Fa-f (digits),
// Ll (marking the integer is 32-bit), Xx (marking the representation
// is hexadecimal). As some kind of separator should come after the
// integer, we can consume the integer until a non-alphanumeric
// character.
bool canStartInt() const;
bool canContinueInt() const;
bool canStartString() const;
// Check if tokenizer can start reading a single line comment (e.g. a comment
// that begins with '//')
bool canStartLineComment() const;
// Check if tokenizer can start or finish reading a block comment (e.g. a
// comment that begins with '/*' and ends with '*/')
bool canStartBlockComment() const;
// Throw away all remaining characters on the current line.
void skipCurrentLine();
bool streamEof() const;
// Classify the token that is about to be read from the current position.
Kind classifyCurrentToken() const;
// Process the Kind::Identifier token - check if it is
// an identifier describing a block start or end.
void processIdentifier(RCToken &token) const;
StringRef Data;
size_t DataLength, Pos;
void Tokenizer::skipCurrentLine() {
Pos = Data.find_first_of("\r\n", Pos);
Pos = Data.find_first_not_of("\r\n", Pos);
if (Pos == StringRef::npos)
Pos = DataLength;
Expected<std::vector<RCToken>> Tokenizer::run() {
Pos = 0;
std::vector<RCToken> Result;
// Consume an optional UTF-8 Byte Order Mark.
if (willNowRead("\xef\xbb\xbf"))
while (!streamEof()) {
if (!skipWhitespaces())
Kind TokenKind = classifyCurrentToken();
if (TokenKind == Kind::Invalid)
return getStringError("Invalid token found at position " + Twine(Pos));
const size_t TokenStart = Pos;
if (Error TokenError = consumeToken(TokenKind))
return std::move(TokenError);
// Comments are just deleted, don't bother saving them.
if (TokenKind == Kind::LineComment || TokenKind == Kind::StartComment)
RCToken Token(TokenKind, Data.take_front(Pos).drop_front(TokenStart));
if (TokenKind == Kind::Identifier) {
} else if (TokenKind == Kind::Int) {
uint32_t TokenInt;
if (!rcGetAsInteger(Token.value(), TokenInt)) {
// The integer has incorrect format or cannot be represented in
// a 32-bit integer.
return getStringError("Integer invalid or too large: " +
return Result;
bool Tokenizer::advance(size_t Amount) {
Pos += Amount;
return !streamEof();
bool Tokenizer::skipWhitespaces() {
while (!streamEof() && std::isspace(Data[Pos]))
return !streamEof();
Error Tokenizer::consumeToken(const Kind TokenKind) {
switch (TokenKind) {
// One-character token consumption.
#define TOKEN(Name)
#define SHORT_TOKEN(Name, Ch) case Kind::Name:
#include "ResourceScriptTokenList.h"
#undef TOKEN
return Error::success();
case Kind::LineComment:
return Error::success();
case Kind::StartComment: {
auto EndPos = Data.find("*/", Pos);
if (EndPos == StringRef::npos)
return getStringError(
"Unclosed multi-line comment beginning at position " + Twine(Pos));
advance(EndPos - Pos);
return Error::success();
case Kind::Identifier:
while (!streamEof() && canContinueIdentifier())
return Error::success();
case Kind::Int:
while (!streamEof() && canContinueInt())
return Error::success();
case Kind::String:
// Consume the preceding 'L', if there is any.
if (std::toupper(Data[Pos]) == 'L')
// Consume the double-quote.
// Consume the characters until the end of the file, line or string.
while (true) {
if (streamEof()) {
return getStringError("Unterminated string literal.");
} else if (Data[Pos] == '"') {
// Consume the ending double-quote.
// However, if another '"' follows this double-quote, the string didn't
// end and we just included '"' into the string.
if (!willNowRead("\""))
return Error::success();
} else if (Data[Pos] == '\n') {
return getStringError("String literal not terminated in the line.");
case Kind::Invalid:
assert(false && "Cannot consume an invalid token.");
llvm_unreachable("Unknown RCToken::Kind");
bool Tokenizer::willNowRead(StringRef FollowingChars) const {
return Data.drop_front(Pos).startswith(FollowingChars);
bool Tokenizer::canStartIdentifier() const {
const char CurChar = Data[Pos];
return std::isalpha(CurChar) || CurChar == '_';
bool Tokenizer::canContinueIdentifier() const {
const char CurChar = Data[Pos];
return std::isalnum(CurChar) || CurChar == '_';
bool Tokenizer::canStartInt() const {
return std::isdigit(Data[Pos]);
bool Tokenizer::canStartBlockComment() const {
return Data.drop_front(Pos).startswith("/*");
bool Tokenizer::canStartLineComment() const {
return Data.drop_front(Pos).startswith("//");
bool Tokenizer::canContinueInt() const {
return std::isalnum(Data[Pos]);
bool Tokenizer::canStartString() const {
return willNowRead("\"") || willNowRead("L\"") || willNowRead("l\"");
bool Tokenizer::streamEof() const { return Pos == DataLength; }
Kind Tokenizer::classifyCurrentToken() const {
if (canStartBlockComment())
return Kind::StartComment;
if (canStartLineComment())
return Kind::LineComment;
if (canStartInt())
return Kind::Int;
if (canStartString())
return Kind::String;
// BEGIN and END are at this point of lexing recognized as identifiers.
if (canStartIdentifier())
return Kind::Identifier;
const char CurChar = Data[Pos];
switch (CurChar) {
// One-character token classification.
#define TOKEN(Name)
#define SHORT_TOKEN(Name, Ch) \
case Ch: \
return Kind::Name;
#include "ResourceScriptTokenList.h"
#undef TOKEN
return Kind::Invalid;
void Tokenizer::processIdentifier(RCToken &Token) const {
assert(Token.kind() == Kind::Identifier);
StringRef Name = Token.value();
if (Name.equals_lower("begin"))
Token = RCToken(Kind::BlockBegin, Name);
else if (Name.equals_lower("end"))
Token = RCToken(Kind::BlockEnd, Name);
} // anonymous namespace
namespace llvm {
Expected<std::vector<RCToken>> tokenizeRC(StringRef Input) {
return Tokenizer(Input).run();
} // namespace llvm