From a1d03a7d5e369e350a7c559e7209efd0d61be3b8 Mon Sep 17 00:00:00 2001 From: DCiliberto Date: Wed, 25 Mar 2020 22:04:08 -0700 Subject: [PATCH 1/6] skeleton for interfaces in parser.dart --- lib/src/parser/parser.dart | 66 ++++++++++++++++++++++++++++++++++++++ 1 file changed, 66 insertions(+) diff --git a/lib/src/parser/parser.dart b/lib/src/parser/parser.dart index 5ed3327..f0fc8fb 100644 --- a/lib/src/parser/parser.dart +++ b/lib/src/parser/parser.dart @@ -9,7 +9,20 @@ import '../../parser.dart'; import '../../parser.dart'; import '../../parser.dart'; import '../../parser.dart'; +import '../interfaces/interface.dart'; +import '../tokens/variants/identifier_token.dart'; +import '../tokens/variants/identifier_token.dart'; +import '../tokens/variants/identifier_token.dart'; import '../tokens/variants/integer_literal_token.dart'; +import '../tokens/variants/keyword_token.dart'; +import '../tokens/variants/keyword_token.dart'; +import '../tokens/variants/operator_or_punctuator_token.dart'; +import '../tokens/variants/operator_or_punctuator_token.dart'; +import '../tokens/variants/operator_or_punctuator_token.dart'; +import '../tokens/variants/operator_or_punctuator_token.dart'; +import '../tokens/variants/operator_or_punctuator_token.dart'; +import '../tokens/variants/operator_or_punctuator_token.dart'; +import '../tokens/variants/operator_or_punctuator_token.dart'; import '../tokens/variants/operator_or_punctuator_token.dart'; class Parser { @@ -56,6 +69,59 @@ class Parser { } + ParseResult _parseInterface(final int startPos){ + int currPos = startPos; + if(_tokens[currPos] is KeywordToken){ + if(_tokens[currPos].value == 'new' || _tokens[currPos].value == 'public' + ||_tokens[currPos].value == 'protected'||_tokens[currPos].value == 'internal' + ||_tokens[currPos].value == 'private'){ + //make helper function for this + } + if(_tokens[currPos] is KeywordToken && _tokens[currPos].value == 'interface'){ + if(_tokens[currPos] is IdentifierToken){ + //param list + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ + if(true){ + //check for params here maybe make helper + } + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '>'){ + + } + } + //end param list + //start interface base + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ':'){ + if(_tokens[currPos] is IdentifierToken){ + + } + } + //end interface base + //start type param constraint + //end type param constraint + //start interface body + if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '{'){ + + //fuck dude idk, maybe make a helper method, + //close bracket + if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '}'){ + + } + } + //end interfacebody + //check if this shit is actually allowed + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ';'){ + + } + + } + } + } + else{ + return null; + } + + } + ParseResult parseExp(final int startPos) { // if (_tokens[startPos] is IdentifierToken) { From 51fc0a62dfdb45bdfeed630ea2ccab1817c68127 Mon Sep 17 00:00:00 2001 From: DCiliberto Date: Fri, 27 Mar 2020 17:16:09 -0700 Subject: [PATCH 2/6] nested if for parsing interfaces and classes, ',' now lexes as OperatororPunctuatorToken --- lib/src/classes/classes.dart | 5 + lib/src/lexer/lexer.dart | 3 +- lib/src/parser/parser.dart | 191 +++++++++++++++++++++++++++++++++-- packages-microsoft-prod.deb | Bin 0 -> 3120 bytes 4 files changed, 187 insertions(+), 12 deletions(-) create mode 100644 packages-microsoft-prod.deb diff --git a/lib/src/classes/classes.dart b/lib/src/classes/classes.dart index e69de29..b2b70df 100644 --- a/lib/src/classes/classes.dart +++ b/lib/src/classes/classes.dart @@ -0,0 +1,5 @@ +import 'package:cs2dart/src/abstractions/abstraction.dart'; + +abstract class Class implements Abstraction { + +} \ No newline at end of file diff --git a/lib/src/lexer/lexer.dart b/lib/src/lexer/lexer.dart index f3fa7f4..d4f28c4 100644 --- a/lib/src/lexer/lexer.dart +++ b/lib/src/lexer/lexer.dart @@ -246,6 +246,7 @@ class Lexer { char == 41 || // right parenthesis char == 42 || // asterisk char == 43 || // plus sign + char == 44 || //comma char == 45 || // hyphen-minus char == 46 || // full stop, period char == 47 || // slash @@ -261,7 +262,7 @@ class Lexer { char == 123 || // left curly bracket char == 124 || // vertical bar char == 125 || // right curly bracket - char == 126) { // tilde + char == 126 ) { // tilde return true; } return false; diff --git a/lib/src/parser/parser.dart b/lib/src/parser/parser.dart index f0fc8fb..d94c68a 100644 --- a/lib/src/parser/parser.dart +++ b/lib/src/parser/parser.dart @@ -10,6 +10,7 @@ import '../../parser.dart'; import '../../parser.dart'; import '../../parser.dart'; import '../interfaces/interface.dart'; +import '../classes/classes.dart'; import '../tokens/variants/identifier_token.dart'; import '../tokens/variants/identifier_token.dart'; import '../tokens/variants/identifier_token.dart'; @@ -69,16 +70,179 @@ class Parser { } - ParseResult _parseInterface(final int startPos){ - int currPos = startPos; + //dump this into parser_assist later + bool _isInterfaceModifier(KeywordToken input){//modifie to check for repeats + if (input.value == 'new' || input.value == 'public' + ||input.value == 'protected'||input.value == 'internal' + || input.value == 'private' ){ + return true; + } + else{ + return false; + } + + } + //end interface helper functions + + //start class helper functions + bool _isClassModifier(KeywordToken input){//modify to check for repeats + if (input.value == 'new' || input.value == 'public' + ||input.value == 'protected'||input.value == 'internal' + || input.value == 'private' || input.value == 'abstract' || input.value == 'static' ){ + return true; + } + else{ + return false; + } + + } + //end class helper functions + + ParseResult _parseClass(final int startPos){ + var currPos = startPos; if(_tokens[currPos] is KeywordToken){ - if(_tokens[currPos].value == 'new' || _tokens[currPos].value == 'public' - ||_tokens[currPos].value == 'protected'||_tokens[currPos].value == 'internal' - ||_tokens[currPos].value == 'private'){ - //make helper function for this + currPos++; + //while loop to parse modifiers + while(_tokens[currPos] is KeywordToken && _isClassModifier(_tokens[currPos])) + { + currPos++; + }//end loop + + //start interface parsing + if(_tokens[currPos] is KeywordToken && _tokens[currPos].value == 'class'){ + currPos++; + if(_tokens[currPos] is IdentifierToken){//parsing identifier + currPos++; + //param list + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ + currPos++; + while(){//refactor loop vars for here, or make helper function + + //check for params here maybe makebool helper + currPos++; + } + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '>'){ + currPos++; + //start type param constraint + while(_tokens[currPos] is IdentifierToken && _tokens[currPos].value == 'where'){ + currPos++; + if(_tokens[currPos] is IdentifierToken){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ':'){ + currPos++; + var exitLoop = false; + var ateConstructorConstraint = false; + while (!exitLoop){ + if(_tokens[currPos] is KeywordToken && _tokens[currPos].value == 'new' && !ateConstructorConstraint){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '('){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ')'){ + currPos++; + ateConstructorConstraint = true; + } + else{ + return null; + } + } + else{ + return null; + } + } + else if(_tokens[currPos] is IdentifierToken){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ + currPos++; + if(_tokens[currPos] is IdentifierToken){ + currPos++; + if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '>'){ + currPos++; + } + } + } + } + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ',' ){ + currPos++; + } + else{ + currPos++; + exitLoop = true; + } + } + } + else{ + return null; + } + } + else{ + return null; + } + } + //end type param constraint + } + else{ + return null; + } + } + //end param list + //start class base + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ':'){ + currPos++; + + if(_tokens[currPos] is IdentifierToken){ + currPos++; + + } + } + //end class base + + //start class body + if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '{'){ + currPos++; + + //fuck dude idk, maybe make a helper method, + //close bracket + if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '}'){ + currPos++; + + } + } + //end class body + //ending semicolon, optional + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ';'){ + currPos++; + + } } + else{ + return null; + } + // final return + return null; + } + //end identifier parsing + else{ + return null; + } + } + else{ + return null; + } + + } + + ParseResult _parseInterface(final int startPos){ + var currPos = startPos; + if(_tokens[currPos] is KeywordToken){ + //while loop to parse modifiers + while(_tokens[currPos] is KeywordToken && _isInterfaceModifier(_tokens[currPos])) + { + + }//end loop + + //start interface parsing if(_tokens[currPos] is KeywordToken && _tokens[currPos].value == 'interface'){ - if(_tokens[currPos] is IdentifierToken){ + if(_tokens[currPos] is IdentifierToken){//parsing identifier //param list if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ if(true){ @@ -92,7 +256,7 @@ class Parser { //start interface base if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ':'){ if(_tokens[currPos] is IdentifierToken){ - + } } //end interface base @@ -108,12 +272,17 @@ class Parser { } } //end interfacebody - //check if this shit is actually allowed + //ending semicolon, optional if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ';'){ - } - + } } + // final return + return null; + } + //end identifier parsing + else{ + return null; } } else{ diff --git a/packages-microsoft-prod.deb b/packages-microsoft-prod.deb new file mode 100644 index 0000000000000000000000000000000000000000..a9114940be156b9c453b205f820596ee1ff271a9 GIT binary patch literal 3120 zcmai!cQo6L9>#f<65guPP`gB>#E242?NpU2rEiB(gi@<^?3bufs$w;b+5~BBYSb!f zMXL7Rdo-%l9*N(5?|sj?_n&*ueg1gPbI$jC&vVZ6$LF%DwdYd@b&#!{jia@zjE$qK zwWq(FsHnW?fAn|$+YyS2|2q8b|BPsmf~-8~shcat)9tw|#@bWXK0xmO{=*dzN`E(? zK_ahN8CV$xo-iC|LL=EG{haL}%vh33Gc@tzD_x@I z{f&WQAV`wp^PHUjhG>Ym%ZWB2b}kw>)URf%0!wSG=S>rLfv9QVG)3EkpX_(5ZXd?p z$p#_&ZY-?gmvbl8WVTQ#p#8UW5aizj2Bj))oa-xC@HI?uLuAA6}fg-4YImYzI;kHPzF!d9a&WeQ*}!+Eva*|w|BamMpr=6YW*8_ZM~0nzt0W5=~AxxwuKXJ zF!tZcGN<9{VWUH4hgwF{h%J}52fVYTxWKy)XOX2Z zzlt3$71Kfx3w{2V`Hj88_W4!HQ$|R}6Ncas#0n2zx_ovq43YzW*k+#g`5<;1R5&ak z^Omr`z`@N)Cc;}QpTCst6=0vXe%xbjP=5uAw=A8U^z43-$vaR~bvH@81~ynIgyCp6 z$f?F5Gs{>iSLl7DZhgw!a%z&htZW1J3zF@~>QdknM)we)Rdiw)fLuBPT+3E~jN@7x zx!mBD6N{LJ5EysSNS2ifr0W&EIYUVv2D!T-Qg zmF3~(oTtv-(Fgm*P>ef0_Xk%X@qr63D63$?dsNpOMR@vd}?NTOeLaf&b|g&edy>M5de+-i65RiiG@E)mVn z>^jQ0&&l*>)kVW7=WVp^}6>Ib4c_tI);@qmZ2t8*?!XG6n2wQ96 z3e^=<+jGK>E&qK2sp5ku=Ft*tvigKAq5R;QCResrM*IAO+(C@gSWI;Im8+EYO5KH} zieg{MldCc|#^)2h5wo%!oBhw*SH!y$B5lFnTwT-hS5RF?EjJjuwlUyv|IRm%PNuJm zuw`h_HDL(l2l!kCbMY$if!Ah4r-xj_QKZ1 z42?-Y6~)W|Wnci=T4SvLjmp0;`2T1NM<5jb3oTRFjJO~3k>9xU>)bxqAJFVn92M{E zrFW(%1S$n~#~0+)LG;HGe{@X$9+fjhL}|UhXe_At>w`)9orr+Ss1L2IPm3KGFJa7X zb@MFvT>!vPtAyNCuJXLZ}Hut1lh5PCZfP!S{>w_tTx2I@aZEU<_++AhW&q zdMmfGq#DiMuM{4S@yCo>!QL{_4O+14rR{i#78jS|z1tUL&IWpMhBe332z{3ejUf6Q zcUi-8n#xLpZS6Tc>4W1pyLn&D%(hiOOHWP=(GD--dr0tstNh%?J(b{2-*;WOLPJxI zsS|^t8zb{R1w#7`YjGXC2JUW+fo0TU2|G)(B*#@jJ7QR)6=4>(MY-cG66|a3kmnng z`zznVYeuN%BO=haHlLmenb+J_H8GAjcG zB~EW;(MXv6jrEH4>Rb!x)cUhaCf@tU!^Aq5FEDe&>!1h1V@@Z1OXnYiiSdAp-|>^( ztE87Ge!oRX?Ll1~9R$qFRQo;`Lx$g|^}=>$i4wZh`NxH*&5WTfYq zpd`Y6$qeke_&MZym5&V0SK`82xfke4TZaT6F8|erASW zw#`O9=bubFk`~6~T)z@s%=H8bSXY;`!$)Le%l@Mm`hWO;hZbI-QT3WW`=3_FA|n@q z7CKlazZ(7!I})Cd=0bYSPxyKYNUG)X{^(k5dGpyxMx6W6=$S8C`Xun?Qrcwyq4yBw zo5v@+oPPZH+npcEaz$xrm%=t-IIEjuu72^bJGeTo0LNaP_mg_|Ek)};I}6|6c>at# zP5;r@(aY>uu`eSbWzox-#ZP()ca-Chp= z7;st;{faBLeP%#J`OE?*{u<$tx}nptxKffv-#PhCM;ySDE%_GOK8og<9%a_#+-`cp z5=L_a-RNtwU_@V^*a^fW0Lh+K6Y4-0Ivy|x1^pgL5X)5OAZW@Mq~ z>xOHe^n(KWfCODFpcm3GU==p6_;QYpKCAWt;E-VKn?2J|qQQkrgpg4A>J?`e_mZMD zfmuZi3G2+(>8v6ke=KY<4A{V@9$`%&Z(khGLNHn7zu8=2iBa0ZC$N_2D21o+zHl^p z#&gTMCgpZ~179NA$6foir%Hov7m}}Ma53OcoP(cK%nv_#!#PzQjwccEQmwkJuljW* zB^wSyFOkzo{YI+K#JU3E3Cp1r-A{hYfE2OmQ7A4Bkmu0Xq$xAI>}@s=2RQQn1%@~J zny1Qa(}PXE8_gx#fHPf#j{q0=Yc+;7;CI)6B|zs?B#Rxu2GKJqrzqqW_78>;;LEXw z2_Vvdkvo-Uvw*$I7Y_t8))xQ}8`JtjIXruw=xJ)`Q55*0W~t4LD#vJPcxc8XBTp!m z6_mrKKoIEU@l)K<(tCCtH5kts9UmaHl>fWixFhemT6VKzdAU@@4XW->l+dj=P~x&0 zHsl;fT-TmUl`3-^o^2bemM^D8WxbFM|82FW6%RFpk0yV40=*D{ZjA5RBP&9g4Ze3v zu$ZNvzE3@!onT%%sj6f?qCV7_%8>??6$2+mv0_F*@|jiKh>Uq`@+gUBeXC_2Y*6@m zdYioBcZlQ=D4mm9AdJ?$|0bx?rE+vn&iu7=LShVlpv_Tlm3Oi<7=yLa+q>_310CBu zVrg0Pw4$`D4s%MANVfh0Q@+=A_dsXeGu8C__+@Dda&{aNFhf&oJz8Bk-hK-Qi^vj$ t4wdSlxN4}X;G1{WLV0BZ Date: Fri, 27 Mar 2020 18:17:20 -0700 Subject: [PATCH 3/6] keyword tests for lexer --- lib/src/parser/parser.dart | 124 +++++++++++++++-- test/lexer_tests/lexer_test.dart | 221 ++++++++++++++++++++++++++++++- 2 files changed, 334 insertions(+), 11 deletions(-) diff --git a/lib/src/parser/parser.dart b/lib/src/parser/parser.dart index d94c68a..e0da77f 100644 --- a/lib/src/parser/parser.dart +++ b/lib/src/parser/parser.dart @@ -158,7 +158,17 @@ class Parser { if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '>'){ currPos++; } + else{ + return null; + } } + else{ + + return null; + } + } + else{ + return null; } } if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ',' ){ @@ -193,6 +203,9 @@ class Parser { currPos++; } + else{ + return null; + } } //end class base @@ -206,6 +219,9 @@ class Parser { currPos++; } + else{ + return null; + } } //end class body //ending semicolon, optional @@ -234,49 +250,138 @@ class Parser { ParseResult _parseInterface(final int startPos){ var currPos = startPos; if(_tokens[currPos] is KeywordToken){ + currPos++; //while loop to parse modifiers while(_tokens[currPos] is KeywordToken && _isInterfaceModifier(_tokens[currPos])) { - + currPos++; }//end loop //start interface parsing if(_tokens[currPos] is KeywordToken && _tokens[currPos].value == 'interface'){ + currPos++; if(_tokens[currPos] is IdentifierToken){//parsing identifier + currPos++; //param list if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ - if(true){ - //check for params here maybe make helper + currPos++; + while(){//refactor loop vars for here, or make helper function + + //check for params here maybe makebool helper + currPos++; } if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '>'){ + currPos++; + //start type param constraint + while(_tokens[currPos] is IdentifierToken && _tokens[currPos].value == 'where'){ + currPos++; + if(_tokens[currPos] is IdentifierToken){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ':'){ + currPos++; + var exitLoop = false; + var ateConstructorConstraint = false; + while (!exitLoop){ + if(_tokens[currPos] is KeywordToken && _tokens[currPos].value == 'new' && !ateConstructorConstraint){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '('){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ')'){ + currPos++; + ateConstructorConstraint = true; + } + else{ + return null; + } + } + else{ + return null; + } + } + else if(_tokens[currPos] is IdentifierToken){ + currPos++; + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ + currPos++; + if(_tokens[currPos] is IdentifierToken){ + currPos++; + if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '>'){ + currPos++; + } + else{ + return null; + } + } + else{ + return null; + } + } + else{ + return null; + } + } + if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ',' ){ + currPos++; + } + else{ + currPos++; + exitLoop = true; + } + } + } + else{ + return null; + } + } + else{ + return null; + } + } + //end type param constraint } + else{ + return null; + } } //end param list - //start interface base + //start class base if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ':'){ + currPos++; + if(_tokens[currPos] is IdentifierToken){ + currPos++; } + else{ + return null; + } } - //end interface base - //start type param constraint - //end type param constraint - //start interface body + //end class base + + //start class body if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '{'){ + currPos++; //fuck dude idk, maybe make a helper method, //close bracket if (_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '}'){ + currPos++; } + else{ + return null; + } } - //end interfacebody + //end class body //ending semicolon, optional if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == ';'){ + currPos++; } } + else{ + return null; + } // final return return null; } @@ -289,6 +394,7 @@ class Parser { return null; } + } ParseResult parseExp(final int startPos) { diff --git a/test/lexer_tests/lexer_test.dart b/test/lexer_tests/lexer_test.dart index e210627..648d66a 100644 --- a/test/lexer_tests/lexer_test.dart +++ b/test/lexer_tests/lexer_test.dart @@ -10,8 +10,90 @@ void main() { expect(TokenType.identifier, token.type); }); - test("The 'return' utf16 test", () { - var token = Lexer('return').nextToken(); + //keyword tests + + test("The 'as' utf16 test", () { + var token = Lexer('as').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'bool' utf16 test", () { + var token = Lexer('bool').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'break' utf16 test", () { + var token = Lexer('break').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'byte' utf16 test", () { + var token = Lexer('byte').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'case' utf16 test", () { + var token = Lexer('case').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'char' utf16 test", () { + var token = Lexer('char').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'class' utf16 test", () { + var token = Lexer('class').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'const' utf16 test", () { + var token = Lexer('const').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'continue' utf16 test", () { + var token = Lexer('continue').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'decimal' utf16 test", () { + var token = Lexer('decimal').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'do' utf16 test", () { + var token = Lexer('do').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'double' utf16 test", () { + var token = Lexer('double').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'else' utf16 test", () { + var token = Lexer('else').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'false' utf16 test", () { + var token = Lexer('false').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'float' utf16 test", () { + var token = Lexer('float').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'for' utf16 test", () { + var token = Lexer('for').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'foreach' utf16 test", () { + var token = Lexer('foreach').nextToken(); expect(TokenType.keyword, token.type); }); @@ -20,11 +102,146 @@ void main() { expect(TokenType.keyword, token.type); }); + test("The 'in' utf16 test", () { + var token = Lexer('in').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'intbreak' utf16 test", () { + var token = Lexer('int').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'interface' utf16 test", () { + var token = Lexer('interface').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'internal' utf16 test", () { + var token = Lexer('internal').nextToken(); + expect(TokenType.keyword, token.type); + }); + test("The 'protected' utf16 test", () { var token = Lexer('protected').nextToken(); expect(TokenType.keyword, token.type); }); + test("The 'is' utf16 test", () { + var token = Lexer('is').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'long' utf16 test", () { + var token = Lexer('long').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'new' utf16 test", () { + var token = Lexer('new').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'null' utf16 test", () { + var token = Lexer('break').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'object' utf16 test", () { + var token = Lexer('object').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'private' utf16 test", () { + var token = Lexer('private').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'public' utf16 test", () { + var token = Lexer('protected').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'private' utf16 test", () { + var token = Lexer('public').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'readonly' utf16 test", () { + var token = Lexer('readonly').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'return' utf16 test", () { + var token = Lexer('return').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'sbyte' utf16 test", () { + var token = Lexer('sbyte').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'short' utf16 test", () { + var token = Lexer('short').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'static' utf16 test", () { + var token = Lexer('break').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'string' utf16 test", () { + var token = Lexer('string').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'switch' utf16 test", () { + var token = Lexer('switch').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'this' utf16 test", () { + var token = Lexer('this').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'true' utf16 test", () { + var token = Lexer('true').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'typeof' utf16 test", () { + var token = Lexer('typeof').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'uint' utf16 test", () { + var token = Lexer('uint').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'ulong' utf16 test", () { + var token = Lexer('ulong').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'ushort' utf16 test", () { + var token = Lexer('ushort').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'void' utf16 test", () { + var token = Lexer('void').nextToken(); + expect(TokenType.keyword, token.type); + }); + + test("The 'while' utf16 test", () { + var token = Lexer('while').nextToken(); + expect(TokenType.keyword, token.type); + }); + //end keyword tests test('Simple assignment test', () { var line = Lexer('boolean x = !false;'); var list = line.lexify(); From fbeac3fa82b2bf1ead7f321d0a6fcc9d10fcdb00 Mon Sep 17 00:00:00 2001 From: DCiliberto Date: Fri, 27 Mar 2020 18:47:04 -0700 Subject: [PATCH 4/6] identifier tests, keyword tests, realliteraltests, fixed realliteral type suffixes --- lib/src/lexer/lexer.dart | 31 +++--- lib/src/parser/parser.dart | 4 +- test/lexer_tests/lexer_test.dart | 178 +++++++++++++++++++++++-------- 3 files changed, 152 insertions(+), 61 deletions(-) diff --git a/lib/src/lexer/lexer.dart b/lib/src/lexer/lexer.dart index d4f28c4..3bc2c93 100644 --- a/lib/src/lexer/lexer.dart +++ b/lib/src/lexer/lexer.dart @@ -548,15 +548,19 @@ class Lexer { _next(); if (_isSign(_current)) { _next(); - if (_isRealTypeSuffix(_current)) { - _next(); - } - } - if (_isRealTypeSuffix(_current)) { - _next(); + // if (_isRealTypeSuffix(_current)) { + // _next(); + // } } + // if (_isRealTypeSuffix(_current)) { + // _next(); + // } } + } + if (_isRealTypeSuffix(_current)) { + _next(); + } // form: XXX'.'XXX end = _position; chunk = _text.substring(start, end); @@ -570,15 +574,18 @@ class Lexer { _next(); if (_isSign(_current)) { _next(); - if (_isRealTypeSuffix(_current)) { - _next(); - } - } - if (_isRealTypeSuffix(_current)) { - _next(); + // if (_isRealTypeSuffix(_current)) { + // _next(); + // } } + // if (_isRealTypeSuffix(_current)) { + // _next(); + // } } } + if (_isRealTypeSuffix(_current)) { + _next(); + } // form: '.'XXX end = _position; chunk = _text.substring(start, end); diff --git a/lib/src/parser/parser.dart b/lib/src/parser/parser.dart index e0da77f..cc55e57 100644 --- a/lib/src/parser/parser.dart +++ b/lib/src/parser/parser.dart @@ -116,7 +116,7 @@ class Parser { //param list if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ currPos++; - while(){//refactor loop vars for here, or make helper function + while(true){//refactor loop vars for here, or make helper function //check for params here maybe makebool helper currPos++; @@ -265,7 +265,7 @@ class Parser { //param list if(_tokens[currPos] is OperatorOrPunctuatorToken && _tokens[currPos].value == '<'){ currPos++; - while(){//refactor loop vars for here, or make helper function + while(true){//refactor loop vars for here, or make helper function //check for params here maybe makebool helper currPos++; diff --git a/test/lexer_tests/lexer_test.dart b/test/lexer_tests/lexer_test.dart index 648d66a..9a55a2e 100644 --- a/test/lexer_tests/lexer_test.dart +++ b/test/lexer_tests/lexer_test.dart @@ -5,243 +5,327 @@ import 'package:test/test.dart'; void main() { - test("The 'hello' utf16 test", () { + //identifier tests + test('basic identifier Test', () { var token = Lexer('hello').nextToken(); expect(TokenType.identifier, token.type); }); + test('undersocre identifier Test', () { + var token = Lexer('_4ello').nextToken(); + expect(TokenType.identifier, token.type); + }); + //end identifier tests //keyword tests - test("The 'as' utf16 test", () { + test("The 'as' keyword test", () { var token = Lexer('as').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'bool' utf16 test", () { + test("The 'bool' keyword test", () { var token = Lexer('bool').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'break' utf16 test", () { + test("The 'break' keyword test", () { var token = Lexer('break').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'byte' utf16 test", () { + test("The 'byte' keyword test", () { var token = Lexer('byte').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'case' utf16 test", () { + test("The 'case' keyword test", () { var token = Lexer('case').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'char' utf16 test", () { + test("The 'char' keyword test", () { var token = Lexer('char').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'class' utf16 test", () { + test("The 'class' keyword test", () { var token = Lexer('class').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'const' utf16 test", () { + test("The 'const' keyword test", () { var token = Lexer('const').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'continue' utf16 test", () { + test("The 'continue' keyword test", () { var token = Lexer('continue').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'decimal' utf16 test", () { + test("The 'decimal' keyword test", () { var token = Lexer('decimal').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'do' utf16 test", () { + test("The 'do' keyword test", () { var token = Lexer('do').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'double' utf16 test", () { + test("The 'double' keyword test", () { var token = Lexer('double').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'else' utf16 test", () { + test("The 'else' keyword test", () { var token = Lexer('else').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'false' utf16 test", () { + test("The 'false' keyword test", () { var token = Lexer('false').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'float' utf16 test", () { + test("The 'float' keyword test", () { var token = Lexer('float').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'for' utf16 test", () { + test("The 'for' keyword test", () { var token = Lexer('for').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'foreach' utf16 test", () { + test("The 'foreach' keyword test", () { var token = Lexer('foreach').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'if' utf16 test", () { + test("The 'if' keyword test", () { var token = Lexer('if').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'in' utf16 test", () { + test("The 'in' keyword test", () { var token = Lexer('in').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'intbreak' utf16 test", () { + test("The 'intbreak' keyword test", () { var token = Lexer('int').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'interface' utf16 test", () { + test("The 'interface' keyword test", () { var token = Lexer('interface').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'internal' utf16 test", () { + test("The 'internal' keyword test", () { var token = Lexer('internal').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'protected' utf16 test", () { + test("The 'protected' keyword test", () { var token = Lexer('protected').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'is' utf16 test", () { + test("The 'is' keyword test", () { var token = Lexer('is').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'long' utf16 test", () { + test("The 'long' keyword test", () { var token = Lexer('long').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'new' utf16 test", () { + test("The 'new' keyword test", () { var token = Lexer('new').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'null' utf16 test", () { + test("The 'null' keyword test", () { var token = Lexer('break').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'object' utf16 test", () { + test("The 'object' keyword test", () { var token = Lexer('object').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'private' utf16 test", () { + test("The 'private' keyword test", () { var token = Lexer('private').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'public' utf16 test", () { + test("The 'public' keyword test", () { var token = Lexer('protected').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'private' utf16 test", () { + test("The 'private' keyword test", () { var token = Lexer('public').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'readonly' utf16 test", () { + test("The 'readonly' keyword test", () { var token = Lexer('readonly').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'return' utf16 test", () { + test("The 'return' keyword test", () { var token = Lexer('return').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'sbyte' utf16 test", () { + test("The 'sbyte' keyword test", () { var token = Lexer('sbyte').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'short' utf16 test", () { + test("The 'short' keyword test", () { var token = Lexer('short').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'static' utf16 test", () { + test("The 'static' keyword test", () { var token = Lexer('break').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'string' utf16 test", () { + test("The 'string' keyword test", () { var token = Lexer('string').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'switch' utf16 test", () { + test("The 'switch' keyword test", () { var token = Lexer('switch').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'this' utf16 test", () { + test("The 'this' keyword test", () { var token = Lexer('this').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'true' utf16 test", () { + test("The 'true' keyword test", () { var token = Lexer('true').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'typeof' utf16 test", () { + test("The 'typeof' keyword test", () { var token = Lexer('typeof').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'uint' utf16 test", () { + test("The 'uint' keyword test", () { var token = Lexer('uint').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'ulong' utf16 test", () { + test("The 'ulong' keyword test", () { var token = Lexer('ulong').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'ushort' utf16 test", () { + test("The 'ushort' keyword test", () { var token = Lexer('ushort').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'void' utf16 test", () { + test("The 'void' keyword test", () { var token = Lexer('void').nextToken(); expect(TokenType.keyword, token.type); }); - test("The 'while' utf16 test", () { + test("The 'while' keyword test", () { var token = Lexer('while').nextToken(); expect(TokenType.keyword, token.type); }); //end keyword tests + + //begin real literal tests + test('basic realLiteral Test', () { + var token = Lexer('1').nextToken(); + expect(TokenType.realLiteral, token.type); + }); + + test('decimal realLiteral Test', () { + var token = Lexer('1.1').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test('decimal exponent realLiteral Test', () { + var token = Lexer('1.1e10').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test('decimal exponent with sign realLiteral Test', () { + var token = Lexer('1.1e-10').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test('decimal exponent big E realLiteral Test', () { + var token = Lexer('1.1E10').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test('decimal exponent big E with sign realLiteral Test', () { + var token = Lexer('1.1E-10').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test("type suffix 'F' realLiteral Test", () {//fails check if need to implement + var token = Lexer('1F').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test("type suffix 'f' realLiteral Test", () {//fails check if need to implement + var token = Lexer('1f').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test("type suffix 'D' realLiteral Test", () {//fails check if need to implement + var token = Lexer('1D').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test("type suffix 'd' realLiteral Test", () {//fails check if need to implement + var token = Lexer('1d').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test("type suffix 'M' realLiteral Test", () {//fails check if need to implement + var token = Lexer('1M').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test("type suffix 'm' realLiteral Test", () {//fails check if need to implement + var token = Lexer('1m').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test('type suffix with decimal realLiteral Test', () {//fails check if need to implement + var token = Lexer('1.1f').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test('type suffix with decimal 2 realLiteral Test', () {//fails check if need to implement + var token = Lexer('.1f').nextToken(); + expect(TokenType.identifier, token.type); + }); + + test('type suffix with exponent realLiteral Test', () {//fails check if need to implement + var token = Lexer('1.1e1f').nextToken(); + expect(TokenType.identifier, token.type); + }); + + //end real literal tests test('Simple assignment test', () { var line = Lexer('boolean x = !false;'); var list = line.lexify(); From 5e7c6c7ae60cdfa36a905733fd8cdb2e6ea4cc2d Mon Sep 17 00:00:00 2001 From: DCiliberto Date: Fri, 27 Mar 2020 21:23:47 -0700 Subject: [PATCH 5/6] Bug Huntin --- lib/src/lexer/lexer.dart | 17 +- test/lexer_tests/lexer_test.dart | 291 +++++++++++++++++++++++++++++-- 2 files changed, 290 insertions(+), 18 deletions(-) diff --git a/lib/src/lexer/lexer.dart b/lib/src/lexer/lexer.dart index 3bc2c93..1a98f29 100644 --- a/lib/src/lexer/lexer.dart +++ b/lib/src/lexer/lexer.dart @@ -239,6 +239,7 @@ class Lexer { bool _isOperatorOrPunctuator(int char) { // single chars + print(new String.fromCharCode(char)); if (char == 33 || // Exclamation mark char == 37 || // percent sign char == 38 || // ampersand @@ -246,9 +247,9 @@ class Lexer { char == 41 || // right parenthesis char == 42 || // asterisk char == 43 || // plus sign - char == 44 || //comma + char == 44 || // comma char == 45 || // hyphen-minus - char == 46 || // full stop, period + char == 0 || // full stop, period char == 47 || // slash char == 58 || // colon char == 59 || // semicolon @@ -263,7 +264,9 @@ class Lexer { char == 124 || // vertical bar char == 125 || // right curly bracket char == 126 ) { // tilde + print('here'); return true; + } return false; } @@ -494,6 +497,10 @@ class Lexer { // go through on all of these different types of characters if (_isOperatorOrPunctuator(_current)) { + // if(_current == 0){ + // _next(); + // return OperatorOrPunctuatorToken('.'); + // } _next(); if (_isDoubleOp(_current)) { _next(); @@ -568,7 +575,10 @@ class Lexer { } else if (_isDecimalPoint(_current)) { _next(); //copied from while loop above - while (lexer_assist.isDecimalDigit(_current)) { + if(lexer_assist.isDecimalDigit(_current)) + { + _next(); + while (lexer_assist.isDecimalDigit(_current)) { _next(); if (_isExponentPart(_current)) { _next(); @@ -590,6 +600,7 @@ class Lexer { end = _position; chunk = _text.substring(start, end); return RealLiteralToken(chunk); + } } else { // not recognized // reset position diff --git a/test/lexer_tests/lexer_test.dart b/test/lexer_tests/lexer_test.dart index 9a55a2e..b4fccf8 100644 --- a/test/lexer_tests/lexer_test.dart +++ b/test/lexer_tests/lexer_test.dart @@ -257,75 +257,336 @@ void main() { test('decimal realLiteral Test', () { var token = Lexer('1.1').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test('decimal exponent realLiteral Test', () { var token = Lexer('1.1e10').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test('decimal exponent with sign realLiteral Test', () { var token = Lexer('1.1e-10').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test('decimal exponent big E realLiteral Test', () { var token = Lexer('1.1E10').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test('decimal exponent big E with sign realLiteral Test', () { var token = Lexer('1.1E-10').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test("type suffix 'F' realLiteral Test", () {//fails check if need to implement var token = Lexer('1F').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test("type suffix 'f' realLiteral Test", () {//fails check if need to implement var token = Lexer('1f').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test("type suffix 'D' realLiteral Test", () {//fails check if need to implement var token = Lexer('1D').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test("type suffix 'd' realLiteral Test", () {//fails check if need to implement var token = Lexer('1d').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test("type suffix 'M' realLiteral Test", () {//fails check if need to implement var token = Lexer('1M').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test("type suffix 'm' realLiteral Test", () {//fails check if need to implement var token = Lexer('1m').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test('type suffix with decimal realLiteral Test', () {//fails check if need to implement var token = Lexer('1.1f').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test('type suffix with decimal 2 realLiteral Test', () {//fails check if need to implement var token = Lexer('.1f').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); test('type suffix with exponent realLiteral Test', () {//fails check if need to implement var token = Lexer('1.1e1f').nextToken(); - expect(TokenType.identifier, token.type); + expect(TokenType.realLiteral, token.type); }); - //end real literal tests + + //begin intergerliteral tests ***come back to this + + //end intergerliteral tests + + //begin interpolated stringliteral tests + test('basic interpolated string literal test', () {//fails check if need to implement + var token = Lexer('\$\"hi\"').nextToken(); + expect(TokenType.interpolatedStringLiteral, token.type); + }); + + test('basic interpolated string literal test', () {//fails check if need to implement + var token = Lexer('\$\"hi\"').nextToken(); + expect(TokenType.interpolatedStringLiteral, token.type); + }); + + test('interpolated string literal with brakets test', () {//fails check if need to implement + var token = Lexer('\$\"hi{hello}\"').nextToken(); + expect(TokenType.interpolatedStringLiteral, token.type); + }); + //end interpolated string literal tests + + //begin string literal tests + test('basic string literal test', () {//fails check if need to implement + var token = Lexer('\"hello\"').nextToken(); + expect(TokenType.stringLiteral, token.type); + }); + //end string literal tests + + //begin character literal tests + test('basic character literal test', () {//fails check if need to implement + var token = Lexer("\'a\'").nextToken(); + expect(TokenType.characterLiteral, token.type); + }); + //end character literal tests + + //begin operator or punctuator tests + test('{ operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('{').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('} operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('}').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('[ operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('[').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('] operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer(']').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('( operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('(').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test(') operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer(')').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('. operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('.').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test(', operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer(',').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test(': operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer(':').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('; operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer(';').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('+ operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('+').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('- operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('-').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('* operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('*').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('/ operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('/').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('% operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('%').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('& operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('&').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('| operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('|').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('^ operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('^').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('! operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('!').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('~ operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('~').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('< operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('<').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('> operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('>').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('? operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('?').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('?? operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('??').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test(':: operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('::').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('++ operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('++').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('-- operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('--').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('&& operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('&&').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('|| operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('||').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('-> operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('->').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('== operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('==').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('!= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('!=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('<= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('<=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('>= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('>=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('+= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('+=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('-= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('-=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('/= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('/=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('&= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('%=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('&= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('&=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('|= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('|=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('^= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('^=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('<< operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('<<').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('<<= operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('<<=').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + + test('=> operatorOrPunctuator test', () {//fails check if need to implement + var token = Lexer('=>').nextToken(); + expect(TokenType.operatorOrPunctuator, token.type); + }); + //end operator or punctuator tests test('Simple assignment test', () { var line = Lexer('boolean x = !false;'); var list = line.lexify(); From 0032868234799af3645b803616b90cbb8fc91352 Mon Sep 17 00:00:00 2001 From: DCiliberto Date: Fri, 27 Mar 2020 22:32:56 -0700 Subject: [PATCH 6/6] fixed stupid fucking self impossed bug --- lib/src/lexer/lexer.dart | 18 +++++++++--------- 1 file changed, 9 insertions(+), 9 deletions(-) diff --git a/lib/src/lexer/lexer.dart b/lib/src/lexer/lexer.dart index 1a98f29..beb44b7 100644 --- a/lib/src/lexer/lexer.dart +++ b/lib/src/lexer/lexer.dart @@ -239,7 +239,6 @@ class Lexer { bool _isOperatorOrPunctuator(int char) { // single chars - print(new String.fromCharCode(char)); if (char == 33 || // Exclamation mark char == 37 || // percent sign char == 38 || // ampersand @@ -249,7 +248,7 @@ class Lexer { char == 43 || // plus sign char == 44 || // comma char == 45 || // hyphen-minus - char == 0 || // full stop, period + char == 46 || // full stop, period char == 47 || // slash char == 58 || // colon char == 59 || // semicolon @@ -264,7 +263,6 @@ class Lexer { char == 124 || // vertical bar char == 125 || // right curly bracket char == 126 ) { // tilde - print('here'); return true; } @@ -573,6 +571,7 @@ class Lexer { chunk = _text.substring(start, end); return RealLiteralToken(chunk); } else if (_isDecimalPoint(_current)) { + _next(); //copied from while loop above if(lexer_assist.isDecimalDigit(_current)) @@ -584,13 +583,7 @@ class Lexer { _next(); if (_isSign(_current)) { _next(); - // if (_isRealTypeSuffix(_current)) { - // _next(); - // } } - // if (_isRealTypeSuffix(_current)) { - // _next(); - // } } } if (_isRealTypeSuffix(_current)) { @@ -601,10 +594,17 @@ class Lexer { chunk = _text.substring(start, end); return RealLiteralToken(chunk); } + else{ + //decrement next + + _position = start; + return null; + } } else { // not recognized // reset position _position = start; + return null; } }