250 lines
7.7 KiB
C++
250 lines
7.7 KiB
C++
/*
|
|
Open Asset Import Library (assimp)
|
|
----------------------------------------------------------------------
|
|
|
|
Copyright (c) 2006-2017, assimp team
|
|
|
|
All rights reserved.
|
|
|
|
Redistribution and use of this software in source and binary forms,
|
|
with or without modification, are permitted provided that the
|
|
following conditions are met:
|
|
|
|
* Redistributions of source code must retain the above
|
|
copyright notice, this list of conditions and the
|
|
following disclaimer.
|
|
|
|
* Redistributions in binary form must reproduce the above
|
|
copyright notice, this list of conditions and the
|
|
following disclaimer in the documentation and/or other
|
|
materials provided with the distribution.
|
|
|
|
* Neither the name of the assimp team, nor the names of its
|
|
contributors may be used to endorse or promote products
|
|
derived from this software without specific prior
|
|
written permission of the assimp team.
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
----------------------------------------------------------------------
|
|
*/
|
|
|
|
/** @file FBXTokenizer.cpp
|
|
* @brief Implementation of the FBX broadphase lexer
|
|
*/
|
|
|
|
#ifndef ASSIMP_BUILD_NO_FBX_IMPORTER
|
|
|
|
// tab width for logging columns
|
|
#define ASSIMP_FBX_TAB_WIDTH 4
|
|
|
|
#include "ParsingUtils.h"
|
|
|
|
#include "FBXTokenizer.h"
|
|
#include "FBXUtil.h"
|
|
#include "Exceptional.h"
|
|
|
|
namespace Assimp {
|
|
namespace FBX {
|
|
|
|
// ------------------------------------------------------------------------------------------------
|
|
Token::Token(const char* sbegin, const char* send, TokenType type, unsigned int line, unsigned int column)
|
|
:
|
|
#ifdef DEBUG
|
|
contents(sbegin, static_cast<size_t>(send-sbegin)),
|
|
#endif
|
|
sbegin(sbegin)
|
|
, send(send)
|
|
, type(type)
|
|
, line(line)
|
|
, column(column)
|
|
{
|
|
ai_assert(sbegin);
|
|
ai_assert(send);
|
|
|
|
// tokens must be of non-zero length
|
|
ai_assert(static_cast<size_t>(send-sbegin) > 0);
|
|
}
|
|
|
|
|
|
// ------------------------------------------------------------------------------------------------
|
|
Token::~Token()
|
|
{
|
|
}
|
|
|
|
|
|
namespace {
|
|
|
|
// ------------------------------------------------------------------------------------------------
|
|
// signal tokenization error, this is always unrecoverable. Throws DeadlyImportError.
|
|
AI_WONT_RETURN void TokenizeError(const std::string& message, unsigned int line, unsigned int column) AI_WONT_RETURN_SUFFIX;
|
|
AI_WONT_RETURN void TokenizeError(const std::string& message, unsigned int line, unsigned int column)
|
|
{
|
|
throw DeadlyImportError(Util::AddLineAndColumn("FBX-Tokenize",message,line,column));
|
|
}
|
|
|
|
|
|
// process a potential data token up to 'cur', adding it to 'output_tokens'.
|
|
// ------------------------------------------------------------------------------------------------
|
|
void ProcessDataToken( TokenList& output_tokens, const char*& start, const char*& end,
|
|
unsigned int line,
|
|
unsigned int column,
|
|
TokenType type = TokenType_DATA,
|
|
bool must_have_token = false)
|
|
{
|
|
if (start && end) {
|
|
// sanity check:
|
|
// tokens should have no whitespace outside quoted text and [start,end] should
|
|
// properly delimit the valid range.
|
|
bool in_double_quotes = false;
|
|
for (const char* c = start; c != end + 1; ++c) {
|
|
if (*c == '\"') {
|
|
in_double_quotes = !in_double_quotes;
|
|
}
|
|
|
|
if (!in_double_quotes && IsSpaceOrNewLine(*c)) {
|
|
TokenizeError("unexpected whitespace in token", line, column);
|
|
}
|
|
}
|
|
|
|
if (in_double_quotes) {
|
|
TokenizeError("non-terminated double quotes", line, column);
|
|
}
|
|
|
|
output_tokens.push_back(new_Token(start,end + 1,type,line,column));
|
|
}
|
|
else if (must_have_token) {
|
|
TokenizeError("unexpected character, expected data token", line, column);
|
|
}
|
|
|
|
start = end = NULL;
|
|
}
|
|
|
|
}
|
|
|
|
// ------------------------------------------------------------------------------------------------
|
|
void Tokenize(TokenList& output_tokens, const char* input)
|
|
{
|
|
ai_assert(input);
|
|
|
|
// line and column numbers numbers are one-based
|
|
unsigned int line = 1;
|
|
unsigned int column = 1;
|
|
|
|
bool comment = false;
|
|
bool in_double_quotes = false;
|
|
bool pending_data_token = false;
|
|
|
|
const char* token_begin = NULL, *token_end = NULL;
|
|
for (const char* cur = input;*cur;column += (*cur == '\t' ? ASSIMP_FBX_TAB_WIDTH : 1), ++cur) {
|
|
const char c = *cur;
|
|
|
|
if (IsLineEnd(c)) {
|
|
comment = false;
|
|
|
|
column = 0;
|
|
++line;
|
|
}
|
|
|
|
if(comment) {
|
|
continue;
|
|
}
|
|
|
|
if(in_double_quotes) {
|
|
if (c == '\"') {
|
|
in_double_quotes = false;
|
|
token_end = cur;
|
|
|
|
ProcessDataToken(output_tokens,token_begin,token_end,line,column);
|
|
pending_data_token = false;
|
|
}
|
|
continue;
|
|
}
|
|
|
|
switch(c)
|
|
{
|
|
case '\"':
|
|
if (token_begin) {
|
|
TokenizeError("unexpected double-quote", line, column);
|
|
}
|
|
token_begin = cur;
|
|
in_double_quotes = true;
|
|
continue;
|
|
|
|
case ';':
|
|
ProcessDataToken(output_tokens,token_begin,token_end,line,column);
|
|
comment = true;
|
|
continue;
|
|
|
|
case '{':
|
|
ProcessDataToken(output_tokens,token_begin,token_end, line, column);
|
|
output_tokens.push_back(new_Token(cur,cur+1,TokenType_OPEN_BRACKET,line,column));
|
|
continue;
|
|
|
|
case '}':
|
|
ProcessDataToken(output_tokens,token_begin,token_end,line,column);
|
|
output_tokens.push_back(new_Token(cur,cur+1,TokenType_CLOSE_BRACKET,line,column));
|
|
continue;
|
|
|
|
case ',':
|
|
if (pending_data_token) {
|
|
ProcessDataToken(output_tokens,token_begin,token_end,line,column,TokenType_DATA,true);
|
|
}
|
|
output_tokens.push_back(new_Token(cur,cur+1,TokenType_COMMA,line,column));
|
|
continue;
|
|
|
|
case ':':
|
|
if (pending_data_token) {
|
|
ProcessDataToken(output_tokens,token_begin,token_end,line,column,TokenType_KEY,true);
|
|
}
|
|
else {
|
|
TokenizeError("unexpected colon", line, column);
|
|
}
|
|
continue;
|
|
}
|
|
|
|
if (IsSpaceOrNewLine(c)) {
|
|
|
|
if (token_begin) {
|
|
// peek ahead and check if the next token is a colon in which
|
|
// case this counts as KEY token.
|
|
TokenType type = TokenType_DATA;
|
|
for (const char* peek = cur; *peek && IsSpaceOrNewLine(*peek); ++peek) {
|
|
if (*peek == ':') {
|
|
type = TokenType_KEY;
|
|
cur = peek;
|
|
break;
|
|
}
|
|
}
|
|
|
|
ProcessDataToken(output_tokens,token_begin,token_end,line,column,type);
|
|
}
|
|
|
|
pending_data_token = false;
|
|
}
|
|
else {
|
|
token_end = cur;
|
|
if (!token_begin) {
|
|
token_begin = cur;
|
|
}
|
|
|
|
pending_data_token = true;
|
|
}
|
|
}
|
|
}
|
|
|
|
} // !FBX
|
|
} // !Assimp
|
|
|
|
#endif
|