Files
qt-creator/tests/manual/plain-cplusplus/Preprocessor.cpp
hjk 2931a499e6 Long live the king!
Change-Id: I2b72b34c0cfeafc8bdbaf49b83ff723544f2b6e2
Reviewed-by: Daniel Teske <daniel.teske@nokia.com>
2012-01-26 19:55:36 +01:00

346 lines
9.0 KiB
C++

/**************************************************************************
**
** This file is part of Qt Creator
**
** Copyright (c) 2012 Nokia Corporation and/or its subsidiary(-ies).
**
** Contact: Nokia Corporation (qt-info@nokia.com)
**
**
** GNU Lesser General Public License Usage
**
** This file may be used under the terms of the GNU Lesser General Public
** License version 2.1 as published by the Free Software Foundation and
** appearing in the file LICENSE.LGPL included in the packaging of this file.
** Please review the following information to ensure the GNU Lesser General
** Public License version 2.1 requirements will be met:
** http://www.gnu.org/licenses/old-licenses/lgpl-2.1.html.
**
** In addition, as a special exception, Nokia gives you certain additional
** rights. These rights are described in the Nokia Qt LGPL Exception
** version 1.1, included in the file LGPL_EXCEPTION.txt in this package.
**
** Other Usage
**
** Alternatively, this file may be used in accordance with the terms and
** conditions contained in a signed written agreement between you and Nokia.
**
** If you have questions regarding the use of this file, please contact
** Nokia at qt-info@nokia.com.
**
**************************************************************************/
#include "Preprocessor.h"
#include "Lexer.h"
#include <list>
#include <iostream>
#include <cassert>
using namespace CPlusPlus;
std::ostream &operator << (std::ostream &out, const StringRef &s)
{
out.write(s.text(), s.size());
return out;
}
struct Preprocessor::TokenBuffer
{
std::list<Token> tokens;
const Macro *macro;
TokenBuffer *next;
template <typename _Iterator>
TokenBuffer(_Iterator firstToken, _Iterator lastToken, const Macro *macro, TokenBuffer *next)
: tokens(firstToken, lastToken), macro(macro), next(next) {}
};
Lexer *Preprocessor::switchLexer(Lexer *lex)
{
Lexer *previousLexer = _lexer;
_lexer = lex;
return previousLexer;
}
StringRef Preprocessor::switchSource(const StringRef &source)
{
StringRef previousSource = _source;
_source = source;
return previousSource;
}
const Preprocessor::Macro *Preprocessor::resolveMacro(const StringRef &name) const
{
std::map<StringRef, Macro>::const_iterator it = macros.find(name);
if (it != macros.end()) {
const Macro *m = &it->second;
for (TokenBuffer *r = _tokenBuffer; r; r = r->next) {
if (r->macro == m)
return 0;
}
return m;
}
return 0;
}
void Preprocessor::collectActualArguments(Token *tk, std::vector<std::vector<Token> > *actuals)
{
lex(tk);
assert(tk->is(T_LPAREN));
lex(tk);
std::vector<Token> tokens;
scanActualArgument(tk, &tokens);
actuals->push_back(tokens);
while (tk->is(T_COMMA)) {
lex(tk);
std::vector<Token> tokens;
scanActualArgument(tk, &tokens);
actuals->push_back(tokens);
}
assert(tk->is(T_RPAREN));
lex(tk);
}
void Preprocessor::scanActualArgument(Token *tk, std::vector<Token> *tokens)
{
int count = 0;
while (tk->isNot(T_EOF_SYMBOL)) {
if (tk->is(T_LPAREN))
++count;
else if (tk->is(T_RPAREN)) {
if (! count)
break;
--count;
}
else if (! count && tk->is(T_COMMA))
break;
tokens->push_back(*tk);
lex(tk);
}
}
void Preprocessor::lex(Token *tk)
{
_Lagain:
if (_tokenBuffer) {
if (_tokenBuffer->tokens.empty()) {
TokenBuffer *r = _tokenBuffer;
_tokenBuffer = _tokenBuffer->next;
delete r;
goto _Lagain;
}
*tk = _tokenBuffer->tokens.front();
_tokenBuffer->tokens.pop_front();
} else {
_lexer->scan(tk);
}
_Lclassify:
if (! inPreprocessorDirective) {
if (tk->newline() && tk->is(T_POUND)) {
handlePreprocessorDirective(tk);
goto _Lclassify;
} else if (tk->is(T_IDENTIFIER)) {
const StringRef id = asStringRef(*tk);
if (const Macro *macro = resolveMacro(id)) {
std::vector<Token> body = macro->body;
if (macro->isFunctionLike) {
std::vector<std::vector<Token> > actuals;
collectActualArguments(tk, &actuals);
std::vector<Token> expanded;
for (size_t i = 0; i < body.size(); ++i) {
const Token &token = body[i];
if (token.isNot(T_IDENTIFIER))
expanded.push_back(token);
else {
const StringRef id = asStringRef(token);
size_t j = 0;
for (; j < macro->formals.size(); ++j) {
if (macro->formals[j] == id) {
expanded.insert(expanded.end(), actuals[j].begin(), actuals[j].end());
break;
}
}
if (j == macro->formals.size())
expanded.push_back(token);
}
}
const Token currentTokenBuffer[] = { *tk };
_tokenBuffer = new TokenBuffer(currentTokenBuffer, currentTokenBuffer + 1,
/*macro */ 0, _tokenBuffer);
body = expanded;
}
_tokenBuffer = new TokenBuffer(body.begin(), body.end(),
macro, _tokenBuffer);
goto _Lagain;
}
}
}
}
void Preprocessor::handlePreprocessorDirective(Token *tk)
{
inPreprocessorDirective = true;
lex(tk); // scan the directive
if (tk->newline() && ! tk->joined())
return; // nothing to do.
const StringRef ppDefine("define", 6);
if (tk->is(T_IDENTIFIER)) {
const StringRef directive = asStringRef(*tk);
if (directive == ppDefine)
handleDefineDirective(tk);
else
skipPreprocesorDirective(tk);
}
inPreprocessorDirective = false;
}
bool Preprocessor::isValidToken(const Token &tk) const
{
if (tk.isNot(T_EOF_SYMBOL) && (! tk.newline() || tk.joined()))
return true;
return false;
}
void Preprocessor::handleDefineDirective(Token *tk)
{
lex(tk);
if (tk->is(T_IDENTIFIER)) {
const StringRef macroName = asStringRef(*tk);
Macro macro;
lex(tk);
if (isValidToken(*tk) && tk->is(T_LPAREN) && ! tk->whitespace()) {
macro.isFunctionLike = true;
lex(tk); // skip `('
if (isValidToken(*tk) && tk->is(T_IDENTIFIER)) {
macro.formals.push_back(asStringRef(*tk));
lex(tk);
while (isValidToken(*tk) && tk->is(T_COMMA)) {
lex(tk);
if (isValidToken(*tk) && tk->is(T_IDENTIFIER)) {
macro.formals.push_back(asStringRef(*tk));
lex(tk);
}
}
}
if (isValidToken(*tk) && tk->is(T_RPAREN))
lex(tk); // skip `)'
}
while (isValidToken(*tk)) {
macro.body.push_back(*tk);
lex(tk);
}
macros.insert(std::make_pair(macroName, macro));
} else {
skipPreprocesorDirective(tk);
}
}
void Preprocessor::skipPreprocesorDirective(Token *tk)
{
do {
lex(tk);
} while (isValidToken(*tk));
}
StringRef Preprocessor::asStringRef(const Token &tk) const
{ return StringRef(_source.begin() + tk.begin(), tk.length()); }
Preprocessor::Preprocessor(std::ostream &out)
: out(out), _lexer(0), inPreprocessorDirective(false)
{ }
void Preprocessor::operator()(const char *source, unsigned size, const StringRef &currentFileName)
{
_currentFileName = currentFileName;
run(source, size);
}
void Preprocessor::run(const char *source, unsigned size)
{
_tokenBuffer = 0;
const StringRef previousSource = switchSource(StringRef(source, size));
Lexer thisLexer(source, source + size);
thisLexer.setScanKeywords(false);
Lexer *previousLexer = switchLexer(&thisLexer);
inPreprocessorDirective = false;
Token tk;
unsigned lineno = 0;
do {
lex(&tk);
if (lineno != tk.lineno) {
if (lineno > tk.lineno || tk.lineno - lineno > 3)
out << std::endl << "#line " << tk.lineno << " \"" << _currentFileName << "\"" << std::endl;
else {
for (unsigned i = lineno; i < tk.lineno; ++i)
out << std::endl;
}
lineno = tk.lineno;
} else {
if (tk.newline())
out << std::endl;
if (tk.whitespace())
out << ' ';
}
out << asStringRef(tk);
lineno = tk.lineno;
} while (tk.isNot(T_EOF_SYMBOL));
out << std::endl;
switchLexer(previousLexer);
switchSource(previousSource);
}