| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517 |
- #pragma once
- #include "common.h"
- #include "error.h"
- #include "str.h"
- namespace pkpy{
- typedef uint8_t TokenIndex;
- constexpr const char* kTokens[] = {
- "is not", "not in", "yield from",
- "@eof", "@eol", "@sof",
- "@id", "@num", "@str", "@fstr",
- "@indent", "@dedent",
- /*****************************************/
- "+", "+=", "-", "-=", // (INPLACE_OP - 1) can get '=' removed
- "*", "*=", "/", "/=", "//", "//=", "%", "%=",
- "&", "&=", "|", "|=", "^", "^=",
- "<<", "<<=", ">>", ">>=",
- /*****************************************/
- ".", ",", ":", ";", "#", "(", ")", "[", "]", "{", "}",
- "**", "=", ">", "<", "...", "->", "?", "@", "==", "!=", ">=", "<=",
- /** KW_BEGIN **/
- "class", "import", "as", "def", "lambda", "pass", "del", "from", "with", "yield",
- "None", "in", "is", "and", "or", "not", "True", "False", "global", "try", "except", "finally",
- "goto", "label", // extended keywords, not available in cpython
- "while", "for", "if", "elif", "else", "break", "continue", "return", "assert", "raise"
- };
- using TokenValue = std::variant<std::monostate, i64, f64, Str>;
- const TokenIndex kTokenCount = sizeof(kTokens) / sizeof(kTokens[0]);
- constexpr TokenIndex TK(const char token[]) {
- for(int k=0; k<kTokenCount; k++){
- const char* i = kTokens[k];
- const char* j = token;
- while(*i && *j && *i == *j) { i++; j++;}
- if(*i == *j) return k;
- }
- #ifdef __GNUC__
- // for old version of gcc, it is not smart enough to ignore FATAL_ERROR()
- // so we must do a normal return
- return 255;
- #else
- FATAL_ERROR();
- #endif
- }
- #define TK_STR(t) kTokens[t]
- const std::map<std::string_view, TokenIndex> kTokenKwMap = [](){
- std::map<std::string_view, TokenIndex> map;
- for(int k=TK("class"); k<kTokenCount; k++) map[kTokens[k]] = k;
- return map;
- }();
- struct Token{
- TokenIndex type;
- const char* start;
- int length;
- int line;
- TokenValue value;
- Str str() const { return Str(start, length);}
- std::string_view sv() const { return std::string_view(start, length);}
- std::string info() const {
- std::stringstream ss;
- ss << line << ": " << TK_STR(type) << " '" << (
- sv()=="\n" ? "\\n" : sv()
- ) << "'";
- return ss.str();
- }
- };
- // https://docs.python.org/3/reference/expressions.html#operator-precedence
- enum Precedence {
- PREC_NONE,
- PREC_TUPLE, // ,
- PREC_LAMBDA, // lambda
- PREC_TERNARY, // ?:
- PREC_LOGICAL_OR, // or
- PREC_LOGICAL_AND, // and
- PREC_LOGICAL_NOT, // not
- PREC_EQUALITY, // == !=
- PREC_TEST, // in / is / is not / not in
- PREC_COMPARISION, // < > <= >=
- PREC_BITWISE_OR, // |
- PREC_BITWISE_XOR, // ^
- PREC_BITWISE_AND, // &
- PREC_BITWISE_SHIFT, // << >>
- PREC_TERM, // + -
- PREC_FACTOR, // * / % //
- PREC_UNARY, // - not
- PREC_EXPONENT, // **
- PREC_CALL, // ()
- PREC_SUBSCRIPT, // []
- PREC_ATTRIB, // .index
- PREC_PRIMARY,
- };
- enum StringType { NORMAL_STRING, RAW_STRING, F_STRING };
- struct Lexer {
- shared_ptr<SourceData> src;
- const char* token_start;
- const char* curr_char;
- int current_line = 1;
- std::vector<Token> nexts;
- stack<int> indents;
- int brackets_level = 0;
- bool used = false;
- char peekchar() const{ return *curr_char; }
- bool match_n_chars(int n, char c0){
- const char* c = curr_char;
- for(int i=0; i<n; i++){
- if(*c == '\0') return false;
- if(*c != c0) return false;
- c++;
- }
- for(int i=0; i<n; i++) eatchar_include_newline();
- return true;
- }
- int eat_spaces(){
- int count = 0;
- while (true) {
- switch (peekchar()) {
- case ' ' : count+=1; break;
- case '\t': count+=4; break;
- default: return count;
- }
- eatchar();
- }
- }
- bool eat_indentation(){
- if(brackets_level > 0) return true;
- int spaces = eat_spaces();
- if(peekchar() == '#') skip_line_comment();
- if(peekchar() == '\0' || peekchar() == '\n') return true;
- // https://docs.python.org/3/reference/lexical_analysis.html#indentation
- if(spaces > indents.top()){
- indents.push(spaces);
- nexts.push_back(Token{TK("@indent"), token_start, 0, current_line});
- } else if(spaces < indents.top()){
- while(spaces < indents.top()){
- indents.pop();
- nexts.push_back(Token{TK("@dedent"), token_start, 0, current_line});
- }
- if(spaces != indents.top()){
- return false;
- }
- }
- return true;
- }
- char eatchar() {
- char c = peekchar();
- if(c == '\n') throw std::runtime_error("eatchar() cannot consume a newline");
- curr_char++;
- return c;
- }
- char eatchar_include_newline() {
- char c = peekchar();
- curr_char++;
- if (c == '\n'){
- current_line++;
- src->line_starts.push_back(curr_char);
- }
- return c;
- }
- int eat_name() {
- curr_char--;
- while(true){
- unsigned char c = peekchar();
- int u8bytes = utf8len(c, true);
- if(u8bytes == 0) return 1;
- if(u8bytes == 1){
- if(isalpha(c) || c=='_' || isdigit(c)) {
- curr_char++;
- continue;
- }else{
- break;
- }
- }
- // handle multibyte char
- std::string u8str(curr_char, u8bytes);
- if(u8str.size() != u8bytes) return 2;
- uint32_t value = 0;
- for(int k=0; k < u8bytes; k++){
- uint8_t b = u8str[k];
- if(k==0){
- if(u8bytes == 2) value = (b & 0b00011111) << 6;
- else if(u8bytes == 3) value = (b & 0b00001111) << 12;
- else if(u8bytes == 4) value = (b & 0b00000111) << 18;
- }else{
- value |= (b & 0b00111111) << (6*(u8bytes-k-1));
- }
- }
- if(is_unicode_Lo_char(value)) curr_char += u8bytes;
- else break;
- }
- int length = (int)(curr_char - token_start);
- if(length == 0) return 3;
- std::string_view name(token_start, length);
- if(src->mode == JSON_MODE){
- if(name == "true"){
- add_token(TK("True"));
- } else if(name == "false"){
- add_token(TK("False"));
- } else if(name == "null"){
- add_token(TK("None"));
- } else {
- return 4;
- }
- return 0;
- }
- if(kTokenKwMap.count(name)){
- if(name == "not"){
- if(strncmp(curr_char, " in", 3) == 0){
- curr_char += 3;
- add_token(TK("not in"));
- return 0;
- }
- }else if(name == "is"){
- if(strncmp(curr_char, " not", 4) == 0){
- curr_char += 4;
- add_token(TK("is not"));
- return 0;
- }
- }else if(name == "yield"){
- if(strncmp(curr_char, " from", 5) == 0){
- curr_char += 5;
- add_token(TK("yield from"));
- return 0;
- }
- }
- add_token(kTokenKwMap.at(name));
- } else {
- add_token(TK("@id"));
- }
- return 0;
- }
- void skip_line_comment() {
- char c;
- while ((c = peekchar()) != '\0') {
- if (c == '\n') return;
- eatchar();
- }
- }
-
- bool matchchar(char c) {
- if (peekchar() != c) return false;
- eatchar_include_newline();
- return true;
- }
- void add_token(TokenIndex type, TokenValue value={}) {
- switch(type){
- case TK("{"): case TK("["): case TK("("): brackets_level++; break;
- case TK(")"): case TK("]"): case TK("}"): brackets_level--; break;
- }
- nexts.push_back( Token{
- type,
- token_start,
- (int)(curr_char - token_start),
- current_line - ((type == TK("@eol")) ? 1 : 0),
- value
- });
- }
- void add_token_2(char c, TokenIndex one, TokenIndex two) {
- if (matchchar(c)) add_token(two);
- else add_token(one);
- }
- Str eat_string_until(char quote, bool raw) {
- bool quote3 = match_n_chars(2, quote);
- std::vector<char> buff;
- while (true) {
- char c = eatchar_include_newline();
- if (c == quote){
- if(quote3 && !match_n_chars(2, quote)){
- buff.push_back(c);
- continue;
- }
- break;
- }
- if (c == '\0'){
- if(quote3 && src->mode == REPL_MODE){
- throw NeedMoreLines(false);
- }
- SyntaxError("EOL while scanning string literal");
- }
- if (c == '\n'){
- if(!quote3) SyntaxError("EOL while scanning string literal");
- else{
- buff.push_back(c);
- continue;
- }
- }
- if (!raw && c == '\\') {
- switch (eatchar_include_newline()) {
- case '"': buff.push_back('"'); break;
- case '\'': buff.push_back('\''); break;
- case '\\': buff.push_back('\\'); break;
- case 'n': buff.push_back('\n'); break;
- case 'r': buff.push_back('\r'); break;
- case 't': buff.push_back('\t'); break;
- default: SyntaxError("invalid escape char");
- }
- } else {
- buff.push_back(c);
- }
- }
- return Str(buff.data(), buff.size());
- }
- void eat_string(char quote, StringType type) {
- Str s = eat_string_until(quote, type == RAW_STRING);
- if(type == F_STRING){
- add_token(TK("@fstr"), s);
- }else{
- add_token(TK("@str"), s);
- }
- }
- void eat_number() {
- static const std::regex pattern("^(0x)?[0-9a-fA-F]+(\\.[0-9]+)?");
- std::smatch m;
- const char* i = token_start;
- while(*i != '\n' && *i != '\0') i++;
- std::string s = std::string(token_start, i);
- try{
- if (std::regex_search(s, m, pattern)) {
- // here is m.length()-1, since the first char was eaten by lex_token()
- for(int j=0; j<m.length()-1; j++) eatchar();
- int base = 10;
- size_t size;
- if (m[1].matched) base = 16;
- if (m[2].matched) {
- if(base == 16) SyntaxError("hex literal should not contain a dot");
- add_token(TK("@num"), Number::stof(m[0], &size));
- } else {
- add_token(TK("@num"), Number::stoi(m[0], &size, base));
- }
- if (size != m.length()) FATAL_ERROR();
- }
- }catch(std::exception& _){
- SyntaxError("invalid number literal");
- }
- }
- bool lex_one_token() {
- while (peekchar() != '\0') {
- token_start = curr_char;
- char c = eatchar_include_newline();
- switch (c) {
- case '\'': case '"': eat_string(c, NORMAL_STRING); return true;
- case '#': skip_line_comment(); break;
- case '{': add_token(TK("{")); return true;
- case '}': add_token(TK("}")); return true;
- case ',': add_token(TK(",")); return true;
- case ':': add_token(TK(":")); return true;
- case ';': add_token(TK(";")); return true;
- case '(': add_token(TK("(")); return true;
- case ')': add_token(TK(")")); return true;
- case '[': add_token(TK("[")); return true;
- case ']': add_token(TK("]")); return true;
- case '@': add_token(TK("@")); return true;
- case '%': add_token_2('=', TK("%"), TK("%=")); return true;
- case '&': add_token_2('=', TK("&"), TK("&=")); return true;
- case '|': add_token_2('=', TK("|"), TK("|=")); return true;
- case '^': add_token_2('=', TK("^"), TK("^=")); return true;
- case '?': add_token(TK("?")); return true;
- case '.': {
- if(matchchar('.')) {
- if(matchchar('.')) {
- add_token(TK("..."));
- } else {
- SyntaxError("invalid token '..'");
- }
- } else {
- add_token(TK("."));
- }
- return true;
- }
- case '=': add_token_2('=', TK("="), TK("==")); return true;
- case '+': add_token_2('=', TK("+"), TK("+=")); return true;
- case '>': {
- if(matchchar('=')) add_token(TK(">="));
- else if(matchchar('>')) add_token_2('=', TK(">>"), TK(">>="));
- else add_token(TK(">"));
- return true;
- }
- case '<': {
- if(matchchar('=')) add_token(TK("<="));
- else if(matchchar('<')) add_token_2('=', TK("<<"), TK("<<="));
- else add_token(TK("<"));
- return true;
- }
- case '-': {
- if(matchchar('=')) add_token(TK("-="));
- else if(matchchar('>')) add_token(TK("->"));
- else add_token(TK("-"));
- return true;
- }
- case '!':
- if(matchchar('=')) add_token(TK("!="));
- else SyntaxError("expected '=' after '!'");
- break;
- case '*':
- if (matchchar('*')) {
- add_token(TK("**")); // '**'
- } else {
- add_token_2('=', TK("*"), TK("*="));
- }
- return true;
- case '/':
- if(matchchar('/')) {
- add_token_2('=', TK("//"), TK("//="));
- } else {
- add_token_2('=', TK("/"), TK("/="));
- }
- return true;
- case ' ': case '\t': eat_spaces(); break;
- case '\n': {
- add_token(TK("@eol"));
- if(!eat_indentation()) IndentationError("unindent does not match any outer indentation level");
- return true;
- }
- default: {
- if(c == 'f'){
- if(matchchar('\'')) {eat_string('\'', F_STRING); return true;}
- if(matchchar('"')) {eat_string('"', F_STRING); return true;}
- }else if(c == 'r'){
- if(matchchar('\'')) {eat_string('\'', RAW_STRING); return true;}
- if(matchchar('"')) {eat_string('"', RAW_STRING); return true;}
- }
- if (c >= '0' && c <= '9') {
- eat_number();
- return true;
- }
- switch (eat_name())
- {
- case 0: break;
- case 1: SyntaxError("invalid char: " + std::string(1, c));
- case 2: SyntaxError("invalid utf8 sequence: " + std::string(1, c));
- case 3: SyntaxError("@id contains invalid char"); break;
- case 4: SyntaxError("invalid JSON token"); break;
- default: FATAL_ERROR();
- }
- return true;
- }
- }
- }
- token_start = curr_char;
- while(indents.size() > 1){
- indents.pop();
- add_token(TK("@dedent"));
- return true;
- }
- add_token(TK("@eof"));
- return false;
- }
- /***** Error Reporter *****/
- void throw_err(Str type, Str msg){
- int lineno = current_line;
- const char* cursor = curr_char;
- if(peekchar() == '\n'){
- lineno--;
- cursor--;
- }
- throw_err(type, msg, lineno, cursor);
- }
- void throw_err(Str type, Str msg, int lineno, const char* cursor){
- auto e = Exception("SyntaxError", msg);
- e.st_push(src->snapshot(lineno, cursor));
- throw e;
- }
- void SyntaxError(Str msg){ throw_err("SyntaxError", msg); }
- void SyntaxError(){ throw_err("SyntaxError", "invalid syntax"); }
- void IndentationError(Str msg){ throw_err("IndentationError", msg); }
- Lexer(shared_ptr<SourceData> src) {
- this->src = src;
- this->token_start = src->source.c_str();
- this->curr_char = src->source.c_str();
- this->nexts.push_back(Token{TK("@sof"), token_start, 0, current_line});
- this->indents.push(0);
- }
- std::vector<Token> run() {
- if(used) FATAL_ERROR();
- used = true;
- while (lex_one_token());
- return std::move(nexts);
- }
- };
- } // namespace pkpy
|