refeat: ILexer接口适配 (#38)
Co-authored-by: Huaps <1183155719@qq.com> Co-authored-by: duqoo <92306417+duqoo@users.noreply.github.com> Reviewed-on: PostGuard/Canon#38
This commit is contained in:
parent
d631a28703
commit
4b6635796c
|
@ -7,12 +7,7 @@ namespace Canon.Core.Abstractions;
|
||||||
/// </summary>
|
/// </summary>
|
||||||
public interface ISourceReader
|
public interface ISourceReader
|
||||||
{
|
{
|
||||||
/// <summary>
|
public char Current { get; }
|
||||||
/// 尝试读取下一个字符
|
|
||||||
/// </summary>
|
|
||||||
/// <param name="c">读取到的字符</param>
|
|
||||||
/// <returns>是否成功读取</returns>
|
|
||||||
public bool TryReadChar([NotNullWhen(true)] out char? c);
|
|
||||||
|
|
||||||
/// <summary>
|
/// <summary>
|
||||||
/// 源文件名称
|
/// 源文件名称
|
||||||
|
@ -28,4 +23,23 @@ public interface ISourceReader
|
||||||
/// 当前读取字符的列号
|
/// 当前读取字符的列号
|
||||||
/// </summary>
|
/// </summary>
|
||||||
public uint Pos { get; }
|
public uint Pos { get; }
|
||||||
|
|
||||||
|
/// <summary>
|
||||||
|
/// 回退一个字符
|
||||||
|
/// </summary>
|
||||||
|
/// <returns>回退是否成功</returns>
|
||||||
|
public bool Retract();
|
||||||
|
|
||||||
|
/// <summary>
|
||||||
|
/// 前进一个字符
|
||||||
|
/// </summary>
|
||||||
|
/// <returns></returns>
|
||||||
|
public bool MoveNext();
|
||||||
|
|
||||||
|
/// <summary>
|
||||||
|
/// 读取下一个字符但是移进
|
||||||
|
/// </summary>
|
||||||
|
/// <param name="c">读取到的下一个字符</param>
|
||||||
|
/// <returns>是否能够读取下一个字符</returns>
|
||||||
|
public bool TryPeekChar([NotNullWhen(true)] out char? c);
|
||||||
}
|
}
|
||||||
|
|
|
@ -9,7 +9,6 @@ public enum SemanticTokenType
|
||||||
Identifier,
|
Identifier,
|
||||||
Character,
|
Character,
|
||||||
Empty,
|
Empty,
|
||||||
Error, // 加了一个错误token
|
|
||||||
/// <summary>
|
/// <summary>
|
||||||
/// 语法分析中的栈底符号
|
/// 语法分析中的栈底符号
|
||||||
/// </summary>
|
/// </summary>
|
||||||
|
@ -90,10 +89,15 @@ public enum NumberType
|
||||||
|
|
||||||
public enum StateType
|
public enum StateType
|
||||||
{
|
{
|
||||||
|
Start,
|
||||||
|
Comment,
|
||||||
Word,
|
Word,
|
||||||
Digit,
|
Num,
|
||||||
Delimiter,
|
Delimiter,
|
||||||
Operator
|
Operator,
|
||||||
|
BreakPoint,
|
||||||
|
Unknown,
|
||||||
|
Done
|
||||||
}
|
}
|
||||||
|
|
||||||
public enum BasicIdType
|
public enum BasicIdType
|
||||||
|
|
83
Canon.Core/LexicalParser/LexRules.cs
Normal file
83
Canon.Core/LexicalParser/LexRules.cs
Normal file
|
@ -0,0 +1,83 @@
|
||||||
|
namespace Canon.Core.LexicalParser;
|
||||||
|
|
||||||
|
public static class LexRules
|
||||||
|
{
|
||||||
|
// 保留关键字
|
||||||
|
private static readonly string[] _keywords =
|
||||||
|
[
|
||||||
|
"Program", "Const", "Var", "Procedure",
|
||||||
|
"Function", "Begin", "End", "Array",
|
||||||
|
"Of", "If", "Then", "Else",
|
||||||
|
"For", "To", "Do", "Integer",
|
||||||
|
"Real", "Boolean", "Character", "Divide",
|
||||||
|
"Not", "Mod", "And", "Or"
|
||||||
|
];
|
||||||
|
|
||||||
|
private static readonly string[] _delimiter = [";", ",", ":", ".", "(", ")", "[", "]", "'", "\"", ".."];
|
||||||
|
|
||||||
|
private static readonly string[] _operator = ["=", "<>", "<", "<=", ">", ">=", "+", "-", "*", "/", ":="];
|
||||||
|
|
||||||
|
// 判断字符
|
||||||
|
public static bool IsDigit(char _ch) {
|
||||||
|
if (_ch >= '0' && _ch <= '9') return true;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static bool IsHexDigit(char _ch)
|
||||||
|
{
|
||||||
|
if ((_ch >= '0' && _ch <= '9') || (_ch<= 'F' && _ch >= 'A')) return true;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static bool IsLetter(char _ch) {
|
||||||
|
if ((_ch >= 'A' && _ch <= 'Z') || (_ch >= 'a' && _ch <= 'z' || _ch == '_')) {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static bool IsKeyword(string tokenString)
|
||||||
|
{
|
||||||
|
|
||||||
|
foreach (var t in _keywords)
|
||||||
|
{
|
||||||
|
if (string.Equals(tokenString, t, StringComparison.OrdinalIgnoreCase)) return true;
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
public static bool IsDelimiter(char ch)
|
||||||
|
{
|
||||||
|
foreach (var delimiter in _delimiter)
|
||||||
|
{
|
||||||
|
if (delimiter.Contains(ch))
|
||||||
|
{
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static bool IsOperator(char ch)
|
||||||
|
{
|
||||||
|
foreach (var o in _operator)
|
||||||
|
{
|
||||||
|
if (o.Contains(ch))
|
||||||
|
{
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static bool IsBreakPoint(char ch)
|
||||||
|
{
|
||||||
|
if (ch == ' ' || ch == '\n' || ch == '\t' || ch == '\r')
|
||||||
|
{
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
95
Canon.Core/LexicalParser/LexemeFactory.cs
Normal file
95
Canon.Core/LexicalParser/LexemeFactory.cs
Normal file
|
@ -0,0 +1,95 @@
|
||||||
|
using Canon.Core.Enums;
|
||||||
|
|
||||||
|
namespace Canon.Core.LexicalParser;
|
||||||
|
|
||||||
|
public static class LexemeFactory
|
||||||
|
{
|
||||||
|
|
||||||
|
public static SemanticToken MakeToken(SemanticTokenType tokenType,string literal,uint _line,uint _chPos)
|
||||||
|
{
|
||||||
|
SemanticToken? token;
|
||||||
|
switch (tokenType)
|
||||||
|
{
|
||||||
|
case SemanticTokenType.Character:
|
||||||
|
CharacterSemanticToken characterSemanticToken = new CharacterSemanticToken()
|
||||||
|
{
|
||||||
|
LinePos = _line, CharacterPos = _chPos, LiteralValue = literal,
|
||||||
|
};
|
||||||
|
token = characterSemanticToken;
|
||||||
|
break;
|
||||||
|
case SemanticTokenType.Identifier:
|
||||||
|
IdentifierSemanticToken identifierSemanticToken = new IdentifierSemanticToken()
|
||||||
|
{
|
||||||
|
LinePos = _line, CharacterPos = _chPos, LiteralValue = literal,
|
||||||
|
};
|
||||||
|
token = identifierSemanticToken;
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
throw new ArgumentOutOfRangeException(nameof(tokenType), tokenType, null);
|
||||||
|
}
|
||||||
|
|
||||||
|
return token;
|
||||||
|
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
public static KeywordSemanticToken MakeToken(KeywordType keywordType,string literal,uint _line,uint _chPos)
|
||||||
|
{
|
||||||
|
KeywordSemanticToken keywordSemanticToken = new KeywordSemanticToken
|
||||||
|
{
|
||||||
|
LinePos = _line,
|
||||||
|
CharacterPos = _chPos,
|
||||||
|
LiteralValue = literal,
|
||||||
|
KeywordType = keywordType
|
||||||
|
};
|
||||||
|
return keywordSemanticToken;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static DelimiterSemanticToken MakeToken(DelimiterType delimiterType,string literal,uint _line,uint _chPos)
|
||||||
|
{
|
||||||
|
DelimiterSemanticToken delimiterSemanticToken = new DelimiterSemanticToken()
|
||||||
|
{
|
||||||
|
LinePos = _line,
|
||||||
|
CharacterPos = _chPos,
|
||||||
|
LiteralValue = literal,
|
||||||
|
DelimiterType = delimiterType
|
||||||
|
};
|
||||||
|
return delimiterSemanticToken;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static NumberSemanticToken MakeToken(NumberType numberType,string literal,uint _line,uint _chPos)
|
||||||
|
{
|
||||||
|
string temp = literal;
|
||||||
|
string result;
|
||||||
|
if (numberType == NumberType.Hex)
|
||||||
|
{
|
||||||
|
result = string.Concat("0x", temp.AsSpan(1, temp.Length - 1));
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
result = temp;
|
||||||
|
}
|
||||||
|
|
||||||
|
NumberSemanticToken numberSemanticToken = new NumberSemanticToken()
|
||||||
|
{
|
||||||
|
LinePos = _line,
|
||||||
|
CharacterPos = _chPos,
|
||||||
|
LiteralValue = result,
|
||||||
|
NumberType = numberType
|
||||||
|
};
|
||||||
|
return numberSemanticToken;
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
public static OperatorSemanticToken MakeToken(OperatorType operatorType,string literal,uint _line,uint _chPos)
|
||||||
|
{
|
||||||
|
OperatorSemanticToken operatorSemanticToken = new OperatorSemanticToken()
|
||||||
|
{
|
||||||
|
LinePos = _line,
|
||||||
|
CharacterPos = _chPos,
|
||||||
|
LiteralValue = literal,
|
||||||
|
OperatorType = operatorType
|
||||||
|
};
|
||||||
|
return operatorSemanticToken;
|
||||||
|
}
|
||||||
|
}
|
File diff suppressed because it is too large
Load Diff
|
@ -339,18 +339,4 @@ public class EndSemanticToken : SemanticToken
|
||||||
public override SemanticTokenType TokenType => SemanticTokenType.End;
|
public override SemanticTokenType TokenType => SemanticTokenType.End;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// <summary>
|
|
||||||
/// 错误类型记号
|
|
||||||
/// </summary>
|
|
||||||
public class ErrorSemanticToken : SemanticToken
|
|
||||||
{
|
|
||||||
public override SemanticTokenType TokenType => SemanticTokenType.Error;
|
|
||||||
|
|
||||||
public static bool TryParse(uint linePos, uint characterPos, LinkedListNode<char> now,
|
|
||||||
out IdentifierSemanticToken? token)
|
|
||||||
{
|
|
||||||
token = null;
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
|
@ -3,12 +3,14 @@ using Canon.Core.CodeGenerators;
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
using Canon.Core.SyntaxNodes;
|
using Canon.Core.SyntaxNodes;
|
||||||
using Canon.Tests.GeneratedParserTests;
|
using Canon.Tests.GeneratedParserTests;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
|
||||||
namespace Canon.Tests.CCodeGeneratorTests;
|
namespace Canon.Tests.CCodeGeneratorTests;
|
||||||
|
|
||||||
public class BasicTests
|
public class BasicTests
|
||||||
{
|
{
|
||||||
private readonly IGrammarParser _parser = GeneratedGrammarParser.Instance;
|
private readonly IGrammarParser _parser = GeneratedGrammarParser.Instance;
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
|
|
||||||
[Fact]
|
[Fact]
|
||||||
public void ProgramStructTest()
|
public void ProgramStructTest()
|
||||||
|
@ -21,9 +23,7 @@ public class BasicTests
|
||||||
end.
|
end.
|
||||||
""";
|
""";
|
||||||
|
|
||||||
Lexer lexer = new(program);
|
IEnumerable<SemanticToken> tokens = _lexer.Tokenize(new StringSourceReader(program));
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
|
||||||
tokens.Add(SemanticToken.End);
|
|
||||||
|
|
||||||
ProgramStruct root = _parser.Analyse(tokens);
|
ProgramStruct root = _parser.Analyse(tokens);
|
||||||
root.GenerateCCode(builder);
|
root.GenerateCCode(builder);
|
||||||
|
|
|
@ -4,12 +4,14 @@ using Canon.Core.GrammarParser;
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
using Canon.Core.SyntaxNodes;
|
using Canon.Core.SyntaxNodes;
|
||||||
using Canon.Tests.GeneratedParserTests;
|
using Canon.Tests.GeneratedParserTests;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
|
||||||
namespace Canon.Tests.GrammarParserTests;
|
namespace Canon.Tests.GrammarParserTests;
|
||||||
|
|
||||||
public class PascalGrammarTests
|
public class PascalGrammarTests
|
||||||
{
|
{
|
||||||
private readonly IGrammarParser _parser = GeneratedGrammarParser.Instance;
|
private readonly IGrammarParser _parser = GeneratedGrammarParser.Instance;
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
|
|
||||||
[Fact]
|
[Fact]
|
||||||
public void DoNothingTest()
|
public void DoNothingTest()
|
||||||
|
@ -20,9 +22,7 @@ public class PascalGrammarTests
|
||||||
end.
|
end.
|
||||||
""";
|
""";
|
||||||
|
|
||||||
Lexer lexer = new(program);
|
IEnumerable<SemanticToken> tokens = _lexer.Tokenize(new StringSourceReader(program));
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
|
||||||
tokens.Add(SemanticToken.End);
|
|
||||||
|
|
||||||
ProgramStruct root = _parser.Analyse(tokens);
|
ProgramStruct root = _parser.Analyse(tokens);
|
||||||
Assert.Equal("DoNothing", root.Head.ProgramName.LiteralValue);
|
Assert.Equal("DoNothing", root.Head.ProgramName.LiteralValue);
|
||||||
|
@ -39,10 +39,7 @@ public class PascalGrammarTests
|
||||||
a := 1 + 1
|
a := 1 + 1
|
||||||
end.
|
end.
|
||||||
""";
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokens = _lexer.Tokenize(new StringSourceReader(program));
|
||||||
Lexer lexer = new(program);
|
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
|
||||||
tokens.Add(SemanticToken.End);
|
|
||||||
|
|
||||||
ProgramStruct root = _parser.Analyse(tokens);
|
ProgramStruct root = _parser.Analyse(tokens);
|
||||||
Assert.Equal("Add", root.Head.ProgramName.LiteralValue);
|
Assert.Equal("Add", root.Head.ProgramName.LiteralValue);
|
||||||
|
@ -59,10 +56,7 @@ public class PascalGrammarTests
|
||||||
writeln( str, ret );
|
writeln( str, ret );
|
||||||
end.
|
end.
|
||||||
""";
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokens = _lexer.Tokenize(new StringSourceReader(program));
|
||||||
Lexer lexer = new(program);
|
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
|
||||||
tokens.Add(SemanticToken.End);
|
|
||||||
|
|
||||||
ProgramStruct root = _parser.Analyse(tokens);
|
ProgramStruct root = _parser.Analyse(tokens);
|
||||||
Assert.Equal("exFunction", root.Head.ProgramName.LiteralValue);
|
Assert.Equal("exFunction", root.Head.ProgramName.LiteralValue);
|
||||||
|
@ -79,10 +73,7 @@ public class PascalGrammarTests
|
||||||
begin
|
begin
|
||||||
end.
|
end.
|
||||||
""";
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokens = _lexer.Tokenize(new StringSourceReader(program));
|
||||||
Lexer lexer = new(program);
|
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
|
||||||
tokens.Add(SemanticToken.End);
|
|
||||||
|
|
||||||
ProgramStruct root = _parser.Analyse(tokens);
|
ProgramStruct root = _parser.Analyse(tokens);
|
||||||
Assert.Equal("main", root.Head.ProgramName.LiteralValue);
|
Assert.Equal("main", root.Head.ProgramName.LiteralValue);
|
||||||
|
|
|
@ -2,13 +2,15 @@
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
using Xunit.Abstractions;
|
using Xunit.Abstractions;
|
||||||
using Canon.Core.Exceptions;
|
using Canon.Core.Exceptions;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
|
||||||
namespace Canon.Tests.LexicalParserTests
|
namespace Canon.Tests.LexicalParserTests
|
||||||
{
|
{
|
||||||
public class CharacterTypeTests
|
public class CharacterTypeTests
|
||||||
{
|
{
|
||||||
private readonly ITestOutputHelper _testOutputHelper;
|
private readonly ITestOutputHelper _testOutputHelper;
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
public CharacterTypeTests(ITestOutputHelper testOutputHelper)
|
public CharacterTypeTests(ITestOutputHelper testOutputHelper)
|
||||||
{
|
{
|
||||||
_testOutputHelper = testOutputHelper;
|
_testOutputHelper = testOutputHelper;
|
||||||
|
@ -20,16 +22,15 @@ namespace Canon.Tests.LexicalParserTests
|
||||||
|
|
||||||
public void TestCharacterType(string input, string? expectedResult)
|
public void TestCharacterType(string input, string? expectedResult)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(input));
|
||||||
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
if (expectedResult == null)
|
if (expectedResult == null)
|
||||||
{
|
{
|
||||||
Assert.Throws<LexemeException>(() => lexer.Tokenize());
|
Assert.Throws<LexemeException>(() => tokens);
|
||||||
}
|
}
|
||||||
else
|
else
|
||||||
{
|
{
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
|
||||||
_testOutputHelper.WriteLine(tokens[0].LiteralValue);
|
_testOutputHelper.WriteLine(tokens[0].LiteralValue);
|
||||||
Assert.Single(tokens);
|
|
||||||
Assert.Equal(SemanticTokenType.Character, tokens[0].TokenType);
|
Assert.Equal(SemanticTokenType.Character, tokens[0].TokenType);
|
||||||
Assert.Equal(expectedResult, tokens[0].LiteralValue);
|
Assert.Equal(expectedResult, tokens[0].LiteralValue);
|
||||||
}
|
}
|
||||||
|
@ -43,8 +44,8 @@ namespace Canon.Tests.LexicalParserTests
|
||||||
//[InlineData("\"x\'", 1, 3, LexemeException.LexemeErrorType.UnclosedStringLiteral)]
|
//[InlineData("\"x\'", 1, 3, LexemeException.LexemeErrorType.UnclosedStringLiteral)]
|
||||||
public void TestParseCharacterError(string input, uint expectedLine, uint expectedCharPosition, LexemeErrorType expectedErrorType)
|
public void TestParseCharacterError(string input, uint expectedLine, uint expectedCharPosition, LexemeErrorType expectedErrorType)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
|
||||||
var ex = Assert.Throws<LexemeException>(() => lexer.Tokenize());
|
var ex = Assert.Throws<LexemeException>(() => _lexer.Tokenize(new StringSourceReader(input)).ToList());
|
||||||
_testOutputHelper.WriteLine(ex.ToString());
|
_testOutputHelper.WriteLine(ex.ToString());
|
||||||
Assert.Equal(expectedErrorType, ex.ErrorType);
|
Assert.Equal(expectedErrorType, ex.ErrorType);
|
||||||
Assert.Equal(expectedLine, ex.Line);
|
Assert.Equal(expectedLine, ex.Line);
|
||||||
|
|
|
@ -1,10 +1,13 @@
|
||||||
using Canon.Core.Enums;
|
using Canon.Core.Enums;
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
namespace Canon.Tests.LexicalParserTests;
|
namespace Canon.Tests.LexicalParserTests;
|
||||||
|
|
||||||
public class DelimiterTests
|
public class DelimiterTests
|
||||||
{
|
{
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
|
|
||||||
[Theory]
|
[Theory]
|
||||||
[InlineData(",123", DelimiterType.Comma)]
|
[InlineData(",123", DelimiterType.Comma)]
|
||||||
// [InlineData(".123", DelimiterType.Period)]
|
// [InlineData(".123", DelimiterType.Period)]
|
||||||
|
@ -16,8 +19,8 @@ public class DelimiterTests
|
||||||
[InlineData("]asd", DelimiterType.RightSquareBracket)]
|
[InlineData("]asd", DelimiterType.RightSquareBracket)]
|
||||||
public void SmokeTest(string input, DelimiterType type)
|
public void SmokeTest(string input, DelimiterType type)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(input));
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
|
||||||
SemanticToken token = tokens[0];
|
SemanticToken token = tokens[0];
|
||||||
Assert.Equal(SemanticTokenType.Delimiter, token.TokenType);
|
Assert.Equal(SemanticTokenType.Delimiter, token.TokenType);
|
||||||
|
|
|
@ -2,11 +2,14 @@
|
||||||
using Canon.Core.Exceptions;
|
using Canon.Core.Exceptions;
|
||||||
using Xunit.Abstractions;
|
using Xunit.Abstractions;
|
||||||
using Canon.Core.Enums;
|
using Canon.Core.Enums;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
|
||||||
namespace Canon.Tests.LexicalParserTests
|
namespace Canon.Tests.LexicalParserTests
|
||||||
{
|
{
|
||||||
public class ErrorSingleTests
|
public class ErrorSingleTests
|
||||||
{
|
{
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
private readonly ITestOutputHelper _testOutputHelper;
|
private readonly ITestOutputHelper _testOutputHelper;
|
||||||
public ErrorSingleTests(ITestOutputHelper testOutputHelper)
|
public ErrorSingleTests(ITestOutputHelper testOutputHelper)
|
||||||
{
|
{
|
||||||
|
@ -20,9 +23,7 @@ namespace Canon.Tests.LexicalParserTests
|
||||||
[InlineData("identifier_with_special_chars@#",1, 30, LexemeErrorType.UnknownCharacterOrString)]
|
[InlineData("identifier_with_special_chars@#",1, 30, LexemeErrorType.UnknownCharacterOrString)]
|
||||||
public void TestUnknownCharacterError(string pascalProgram, uint expectedLine, uint expectedCharPosition, LexemeErrorType expectedErrorType)
|
public void TestUnknownCharacterError(string pascalProgram, uint expectedLine, uint expectedCharPosition, LexemeErrorType expectedErrorType)
|
||||||
{
|
{
|
||||||
var lexer = new Lexer(pascalProgram);
|
var ex = Assert.Throws<LexemeException>(() => _lexer.Tokenize(new StringSourceReader(pascalProgram)).ToList());
|
||||||
|
|
||||||
var ex = Assert.Throws<LexemeException>(() => lexer.Tokenize());
|
|
||||||
_testOutputHelper.WriteLine(ex.ToString());
|
_testOutputHelper.WriteLine(ex.ToString());
|
||||||
Assert.Equal(expectedErrorType, ex.ErrorType);
|
Assert.Equal(expectedErrorType, ex.ErrorType);
|
||||||
Assert.Equal(expectedLine, ex.Line);
|
Assert.Equal(expectedLine, ex.Line);
|
||||||
|
|
|
@ -1,10 +1,13 @@
|
||||||
using Canon.Core.Enums;
|
using Canon.Core.Enums;
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
namespace Canon.Tests.LexicalParserTests
|
namespace Canon.Tests.LexicalParserTests
|
||||||
{
|
{
|
||||||
public class IdentifierTests
|
public class IdentifierTests
|
||||||
{
|
{
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
|
|
||||||
[Theory]
|
[Theory]
|
||||||
[InlineData("identifier", true)]
|
[InlineData("identifier", true)]
|
||||||
[InlineData("_identifier", true)]
|
[InlineData("_identifier", true)]
|
||||||
|
@ -14,10 +17,9 @@ namespace Canon.Tests.LexicalParserTests
|
||||||
[InlineData("andand", true)]
|
[InlineData("andand", true)]
|
||||||
public void TestParseIdentifier(string input, bool expectedResult)
|
public void TestParseIdentifier(string input, bool expectedResult)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(input));
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
|
||||||
Assert.Single(tokens);
|
|
||||||
Assert.Equal(expectedResult, tokens.FirstOrDefault()?.TokenType == SemanticTokenType.Identifier);
|
Assert.Equal(expectedResult, tokens.FirstOrDefault()?.TokenType == SemanticTokenType.Identifier);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,10 +1,14 @@
|
||||||
using Canon.Core.Enums;
|
using Canon.Core.Enums;
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
|
|
||||||
namespace Canon.Tests.LexicalParserTests;
|
namespace Canon.Tests.LexicalParserTests;
|
||||||
|
|
||||||
public class KeywordTypeTests
|
public class KeywordTypeTests
|
||||||
{
|
{
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
|
|
||||||
[Theory]
|
[Theory]
|
||||||
[InlineData("program", KeywordType.Program)]
|
[InlineData("program", KeywordType.Program)]
|
||||||
[InlineData("const", KeywordType.Const)]
|
[InlineData("const", KeywordType.Const)]
|
||||||
|
@ -24,8 +28,8 @@ public class KeywordTypeTests
|
||||||
[InlineData("DO", KeywordType.Do)]
|
[InlineData("DO", KeywordType.Do)]
|
||||||
public void SmokeTest(string input, KeywordType type)
|
public void SmokeTest(string input, KeywordType type)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(input));
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
|
||||||
SemanticToken token = tokens[0];
|
SemanticToken token = tokens[0];
|
||||||
Assert.Equal(SemanticTokenType.Keyword, token.TokenType);
|
Assert.Equal(SemanticTokenType.Keyword, token.TokenType);
|
||||||
|
|
|
@ -3,12 +3,15 @@ using Canon.Core.Enums;
|
||||||
using Canon.Core.Exceptions;
|
using Canon.Core.Exceptions;
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
using Xunit.Abstractions;
|
using Xunit.Abstractions;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
|
|
||||||
namespace Canon.Tests.LexicalParserTests;
|
namespace Canon.Tests.LexicalParserTests;
|
||||||
|
|
||||||
public class LexicalFileTests
|
public class LexicalFileTests
|
||||||
{
|
{
|
||||||
private readonly ITestOutputHelper _testOutputHelper;
|
private readonly ITestOutputHelper _testOutputHelper;
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
|
|
||||||
public LexicalFileTests(ITestOutputHelper testOutputHelper)
|
public LexicalFileTests(ITestOutputHelper testOutputHelper)
|
||||||
{
|
{
|
||||||
|
@ -126,14 +129,16 @@ public class LexicalFileTests
|
||||||
}
|
}
|
||||||
: token).ToList();
|
: token).ToList();
|
||||||
|
|
||||||
var lexer = new Lexer(pascalProgram);
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(pascalProgram));
|
||||||
var actualTokens = lexer.Tokenize();
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
|
||||||
|
var actualTokens = tokens;
|
||||||
for (int i = 0; i < expectedTokens.Count; i++)
|
for (int i = 0; i < expectedTokens.Count; i++)
|
||||||
{
|
{
|
||||||
_testOutputHelper.WriteLine($"Expect: {expectedTokens[i]}");
|
_testOutputHelper.WriteLine($"Expect: {expectedTokens[i]}");
|
||||||
_testOutputHelper.WriteLine($"Actual: {actualTokens[i]}");
|
_testOutputHelper.WriteLine($"Actual: {actualTokens[i]}");
|
||||||
_testOutputHelper.WriteLine("----");
|
_testOutputHelper.WriteLine("----");
|
||||||
Assert.Equal(expectedTokens[i], actualTokens[i]);
|
// Assert.Equal(expectedTokens[i], actualTokens[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
Assert.Equal(expectedTokens, actualTokens);
|
Assert.Equal(expectedTokens, actualTokens);
|
||||||
|
@ -284,8 +289,7 @@ public class LexicalFileTests
|
||||||
x := 42;
|
x := 42;
|
||||||
end.
|
end.
|
||||||
""";
|
""";
|
||||||
var lexer = new Lexer(pascalProgram);
|
var ex = Assert.Throws<LexemeException>(() => _lexer.Tokenize(new StringSourceReader(pascalProgram)).ToList());
|
||||||
var ex = Assert.Throws<LexemeException>(() => lexer.Tokenize());
|
|
||||||
//打印exception信息
|
//打印exception信息
|
||||||
_testOutputHelper.WriteLine(ex.ToString());
|
_testOutputHelper.WriteLine(ex.ToString());
|
||||||
Assert.Equal(LexemeErrorType.UnclosedComment, ex.ErrorType);
|
Assert.Equal(LexemeErrorType.UnclosedComment, ex.ErrorType);
|
||||||
|
@ -302,11 +306,108 @@ public class LexicalFileTests
|
||||||
|
|
||||||
program CommentNotClosed;
|
program CommentNotClosed;
|
||||||
""";
|
""";
|
||||||
var lexer = new Lexer(pascalProgram);
|
var ex = Assert.Throws<LexemeException>(() => _lexer.Tokenize(new StringSourceReader(pascalProgram)).ToList());
|
||||||
var ex = Assert.Throws<LexemeException>(() => lexer.Tokenize());
|
_testOutputHelper.WriteLine(ex.ToString());
|
||||||
_testOutputHelper.WriteLine(ex.ToString());
|
|
||||||
Assert.Equal(LexemeErrorType.UnclosedComment, ex.ErrorType);
|
Assert.Equal(LexemeErrorType.UnclosedComment, ex.ErrorType);
|
||||||
Assert.Equal((uint)4, ex.Line);
|
Assert.Equal((uint)4, ex.Line);
|
||||||
Assert.Equal((uint)26, ex.CharPosition);
|
Assert.Equal((uint)26, ex.CharPosition);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
[Fact]
|
||||||
|
public void ClosedCommentFirst()
|
||||||
|
{
|
||||||
|
string pascalProgram = """
|
||||||
|
program exFunction;
|
||||||
|
var
|
||||||
|
a, b, ret : integer;
|
||||||
|
|
||||||
|
begin
|
||||||
|
a := 100;
|
||||||
|
b := 200;
|
||||||
|
(* calling a function to get max value
|
||||||
|
*)
|
||||||
|
ret := a - b;
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
end.
|
||||||
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(pascalProgram));
|
||||||
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
Assert.NotNull(tokens);
|
||||||
|
}
|
||||||
|
|
||||||
|
[Fact]
|
||||||
|
public void ClosedCommentSecond()
|
||||||
|
{
|
||||||
|
string pascalProgram = """
|
||||||
|
program exFunction;
|
||||||
|
var
|
||||||
|
a, b, ret : integer;
|
||||||
|
|
||||||
|
begin
|
||||||
|
a := 100;
|
||||||
|
b := 200;
|
||||||
|
(* calling a function to get max valued *)
|
||||||
|
ret := a - b;
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
end.
|
||||||
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(pascalProgram));
|
||||||
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
Assert.NotNull(tokens);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
[Fact]
|
||||||
|
public void ClosedCommentThird()
|
||||||
|
{
|
||||||
|
string pascalProgram = """
|
||||||
|
{
|
||||||
|
This is a block comment that does closed.
|
||||||
|
}
|
||||||
|
program CommentClosed;
|
||||||
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(pascalProgram));
|
||||||
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
Assert.NotNull(tokens);
|
||||||
|
}
|
||||||
|
|
||||||
|
[Fact]
|
||||||
|
public void ClosedCommentFourth()
|
||||||
|
{
|
||||||
|
string pascalProgram = """
|
||||||
|
{}
|
||||||
|
program CommentClosed;
|
||||||
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(pascalProgram));
|
||||||
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
Assert.NotNull(tokens);
|
||||||
|
}
|
||||||
|
|
||||||
|
[Fact]
|
||||||
|
public void ClosedCommentFifth()
|
||||||
|
{
|
||||||
|
string pascalProgram = """
|
||||||
|
{
|
||||||
|
}
|
||||||
|
program CommentClosed;
|
||||||
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(pascalProgram));
|
||||||
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
Assert.NotNull(tokens);
|
||||||
|
}
|
||||||
|
|
||||||
|
[Fact]
|
||||||
|
public void ClosedCommentSixth()
|
||||||
|
{
|
||||||
|
string pascalProgram = """
|
||||||
|
(**)
|
||||||
|
""";
|
||||||
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(pascalProgram));
|
||||||
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
Assert.NotNull(tokens);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,12 +2,14 @@
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
using Canon.Core.Exceptions;
|
using Canon.Core.Exceptions;
|
||||||
using Xunit.Abstractions;
|
using Xunit.Abstractions;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
namespace Canon.Tests.LexicalParserTests
|
namespace Canon.Tests.LexicalParserTests
|
||||||
{
|
{
|
||||||
|
|
||||||
public class NumberTests
|
public class NumberTests
|
||||||
{
|
{
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
private readonly ITestOutputHelper _testOutputHelper;
|
private readonly ITestOutputHelper _testOutputHelper;
|
||||||
public NumberTests(ITestOutputHelper testOutputHelper)
|
public NumberTests(ITestOutputHelper testOutputHelper)
|
||||||
{
|
{
|
||||||
|
@ -31,8 +33,8 @@ namespace Canon.Tests.LexicalParserTests
|
||||||
[InlineData("$123", "0x123", NumberType.Hex)]
|
[InlineData("$123", "0x123", NumberType.Hex)]
|
||||||
public void TestParseNumber(string input, string expected, NumberType expectedNumberType)
|
public void TestParseNumber(string input, string expected, NumberType expectedNumberType)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(input));
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
SemanticToken token = tokens[0];
|
SemanticToken token = tokens[0];
|
||||||
Assert.Equal(SemanticTokenType.Number, token.TokenType);
|
Assert.Equal(SemanticTokenType.Number, token.TokenType);
|
||||||
NumberSemanticToken numberSemanticToken = (NumberSemanticToken)token;
|
NumberSemanticToken numberSemanticToken = (NumberSemanticToken)token;
|
||||||
|
@ -41,14 +43,13 @@ namespace Canon.Tests.LexicalParserTests
|
||||||
}
|
}
|
||||||
|
|
||||||
[Theory]
|
[Theory]
|
||||||
[InlineData("1E", 1, 3, LexemeErrorType.IllegalNumberFormat)]
|
[InlineData("1E", 1, 2, LexemeErrorType.IllegalNumberFormat)]
|
||||||
[InlineData("123abc", 1, 4, LexemeErrorType.IllegalNumberFormat)]
|
[InlineData("123abc", 1, 4, LexemeErrorType.IllegalNumberFormat)]
|
||||||
[InlineData("123.45.67", 1, 7, LexemeErrorType.IllegalNumberFormat)]
|
[InlineData("123.45.67", 1, 7, LexemeErrorType.IllegalNumberFormat)]
|
||||||
[InlineData("123identifier", 1, 4, LexemeErrorType.IllegalNumberFormat)]
|
[InlineData("123identifier", 1, 4, LexemeErrorType.IllegalNumberFormat)]
|
||||||
public void TestParseNumberError(string input, uint expectedLine, uint expectedCharPosition, LexemeErrorType expectedErrorType)
|
public void TestParseNumberError(string input, uint expectedLine, uint expectedCharPosition, LexemeErrorType expectedErrorType)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
var ex = Assert.Throws<LexemeException>(() => _lexer.Tokenize(new StringSourceReader(input)).ToList());
|
||||||
var ex = Assert.Throws<LexemeException>(() => lexer.Tokenize());
|
|
||||||
_testOutputHelper.WriteLine(ex.ToString());
|
_testOutputHelper.WriteLine(ex.ToString());
|
||||||
Assert.Equal(expectedErrorType, ex.ErrorType);
|
Assert.Equal(expectedErrorType, ex.ErrorType);
|
||||||
Assert.Equal(expectedLine, ex.Line);
|
Assert.Equal(expectedLine, ex.Line);
|
||||||
|
|
|
@ -1,10 +1,13 @@
|
||||||
using Canon.Core.Enums;
|
using Canon.Core.Enums;
|
||||||
using Canon.Core.LexicalParser;
|
using Canon.Core.LexicalParser;
|
||||||
|
using Canon.Tests.Utils;
|
||||||
|
using Canon.Core.Abstractions;
|
||||||
namespace Canon.Tests.LexicalParserTests;
|
namespace Canon.Tests.LexicalParserTests;
|
||||||
|
|
||||||
public class OperatorTypeTests
|
public class OperatorTypeTests
|
||||||
{
|
{
|
||||||
|
private readonly ILexer _lexer = new Lexer();
|
||||||
|
|
||||||
[Theory]
|
[Theory]
|
||||||
[InlineData("+ 123", OperatorType.Plus, true)]
|
[InlineData("+ 123", OperatorType.Plus, true)]
|
||||||
[InlineData("+123", OperatorType.Plus, true)]
|
[InlineData("+123", OperatorType.Plus, true)]
|
||||||
|
@ -22,8 +25,8 @@ public class OperatorTypeTests
|
||||||
[InlineData("m +123", OperatorType.Plus, false)]
|
[InlineData("m +123", OperatorType.Plus, false)]
|
||||||
public void ParseTest(string input, OperatorType result, bool expectedResult)
|
public void ParseTest(string input, OperatorType result, bool expectedResult)
|
||||||
{
|
{
|
||||||
Lexer lexer = new(input);
|
IEnumerable<SemanticToken> tokensEnumerable = _lexer.Tokenize(new StringSourceReader(input));
|
||||||
List<SemanticToken> tokens = lexer.Tokenize();
|
List<SemanticToken> tokens = tokensEnumerable.ToList();
|
||||||
|
|
||||||
SemanticToken token = tokens[0];
|
SemanticToken token = tokens[0];
|
||||||
if (!expectedResult)
|
if (!expectedResult)
|
||||||
|
|
15
Canon.Tests/Utils/EnumerableExtensions.cs
Normal file
15
Canon.Tests/Utils/EnumerableExtensions.cs
Normal file
|
@ -0,0 +1,15 @@
|
||||||
|
namespace Canon.Tests.Utils;
|
||||||
|
|
||||||
|
public static class EnumerableExtensions
|
||||||
|
{
|
||||||
|
/// <summary>
|
||||||
|
/// 含有索引的遍历
|
||||||
|
/// </summary>
|
||||||
|
/// <param name="enumerable">可遍历的接口</param>
|
||||||
|
/// <typeparam name="T"></typeparam>
|
||||||
|
/// <returns></returns>
|
||||||
|
public static IEnumerable<(T, uint)> WithIndex<T>(this IEnumerable<T> enumerable)
|
||||||
|
{
|
||||||
|
return enumerable.Select((value, index) => (value, (uint)index));
|
||||||
|
}
|
||||||
|
}
|
|
@ -6,10 +6,11 @@ namespace Canon.Tests.Utils;
|
||||||
/// <summary>
|
/// <summary>
|
||||||
/// 从字符串中读取源代码
|
/// 从字符串中读取源代码
|
||||||
/// </summary>
|
/// </summary>
|
||||||
public sealed class StringSourceReader(string source) : ISourceReader, IDisposable
|
public sealed class StringSourceReader(string source) : ISourceReader
|
||||||
{
|
{
|
||||||
private readonly IEnumerator<char> _enumerator =
|
private int _pos = -1;
|
||||||
source.GetEnumerator();
|
|
||||||
|
private uint _lastPos;
|
||||||
|
|
||||||
public uint Line { get; private set; } = 1;
|
public uint Line { get; private set; } = 1;
|
||||||
|
|
||||||
|
@ -17,31 +18,70 @@ public sealed class StringSourceReader(string source) : ISourceReader, IDisposab
|
||||||
|
|
||||||
public string FileName => "string";
|
public string FileName => "string";
|
||||||
|
|
||||||
public bool TryReadChar([NotNullWhen(true)] out char? c)
|
public char Current
|
||||||
{
|
{
|
||||||
if (Pos != 0 || Line != 1)
|
get
|
||||||
{
|
{
|
||||||
// 不是第一次读取
|
if (_pos == -1)
|
||||||
if (_enumerator.Current == '\n')
|
|
||||||
{
|
{
|
||||||
Pos = 0;
|
throw new InvalidOperationException("Reader at before the start.");
|
||||||
Line += 1;
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
return source[_pos];
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!_enumerator.MoveNext())
|
public bool Retract()
|
||||||
|
{
|
||||||
|
if (_pos <= 0)
|
||||||
|
{
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
_pos -= 1;
|
||||||
|
if (Current == '\n')
|
||||||
|
{
|
||||||
|
Line -= 1;
|
||||||
|
// TODO: 如果一直回退就完蛋了
|
||||||
|
Pos = _lastPos;
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
Pos -= 1;
|
||||||
|
}
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public bool MoveNext()
|
||||||
|
{
|
||||||
|
if (_pos >= source.Length - 1)
|
||||||
|
{
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (_pos != -1 && Current == '\n')
|
||||||
|
{
|
||||||
|
Line += 1;
|
||||||
|
_lastPos = Pos;
|
||||||
|
Pos = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
_pos += 1;
|
||||||
|
Pos += 1;
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public bool TryPeekChar([NotNullWhen(true)] out char? c)
|
||||||
|
{
|
||||||
|
if (_pos >= source.Length - 1)
|
||||||
{
|
{
|
||||||
c = null;
|
c = null;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
Pos += 1;
|
c = source[_pos + 1];
|
||||||
c = _enumerator.Current;
|
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
public void Dispose()
|
|
||||||
{
|
|
||||||
_enumerator.Dispose();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -8,78 +8,64 @@ public class StringSourceReaderTests
|
||||||
public void LineFeedTest()
|
public void LineFeedTest()
|
||||||
{
|
{
|
||||||
ISourceReader reader = new StringSourceReader("program Main;\nbegin\nend.\n");
|
ISourceReader reader = new StringSourceReader("program Main;\nbegin\nend.\n");
|
||||||
|
reader.MoveNext();
|
||||||
|
|
||||||
Assert.Equal(0u, reader.Pos);
|
CheckLine(reader, "program Main;", 1);
|
||||||
Assert.Equal(1u, reader.Line);
|
reader.MoveNext();
|
||||||
|
CheckLine(reader, "begin", 2);
|
||||||
// program
|
reader.MoveNext();
|
||||||
Assert.True(reader.TryReadChar(out char? c));
|
CheckLine(reader, "end.", 3);
|
||||||
Assert.Equal('p', c);
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out c));
|
|
||||||
Assert.Equal(' ', c);
|
|
||||||
|
|
||||||
// main;
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
|
|
||||||
Assert.True(reader.TryReadChar(out c));
|
|
||||||
Assert.Equal('\n', c);
|
|
||||||
|
|
||||||
// begin
|
|
||||||
for (uint i = 1; i <= 5; i++)
|
|
||||||
{
|
|
||||||
Assert.True(reader.TryReadChar(out char? _));
|
|
||||||
Assert.Equal(i, reader.Pos);
|
|
||||||
Assert.Equal(2u, reader.Line);
|
|
||||||
}
|
|
||||||
|
|
||||||
// \n
|
|
||||||
Assert.True(reader.TryReadChar(out c));
|
|
||||||
Assert.Equal('\n', c);
|
|
||||||
|
|
||||||
// end.
|
|
||||||
foreach (char i in "end.")
|
|
||||||
{
|
|
||||||
Assert.True(reader.TryReadChar(out c));
|
|
||||||
Assert.Equal(i, c);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
[Fact]
|
[Fact]
|
||||||
public void CarriageReturnLineFeedTest()
|
public void CarriageReturnLineFeedTest()
|
||||||
{
|
{
|
||||||
ISourceReader reader = new StringSourceReader("program Main;\r\nbegin\r\nend.\r\n");
|
ISourceReader reader = new StringSourceReader("program Main;\r\nbegin\r\nend.\r\n");
|
||||||
|
reader.MoveNext();
|
||||||
|
|
||||||
// program Main;
|
CheckLine(reader, "program Main;", 1);
|
||||||
foreach ((char value, uint index) in
|
reader.MoveNext();
|
||||||
"program Main;".Select((value, index) => (value, (uint)index)))
|
reader.MoveNext();
|
||||||
{
|
CheckLine(reader, "begin", 2);
|
||||||
Assert.True(reader.TryReadChar(out char? c));
|
reader.MoveNext();
|
||||||
Assert.Equal(value, c);
|
reader.MoveNext();
|
||||||
Assert.Equal(index + 1, reader.Pos);
|
CheckLine(reader, "end.", 3);
|
||||||
Assert.Equal(1u, reader.Line);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
Assert.True(reader.TryReadChar(out _));
|
[Fact]
|
||||||
Assert.True(reader.TryReadChar(out _));
|
public void RetractTest()
|
||||||
|
|
||||||
// begin
|
|
||||||
foreach ((char value, uint index) in
|
|
||||||
"begin".Select((value, index) => (value, (uint)index)))
|
|
||||||
{
|
{
|
||||||
Assert.True(reader.TryReadChar(out char? c));
|
ISourceReader reader = new StringSourceReader("test");
|
||||||
Assert.Equal(value, c);
|
reader.MoveNext();
|
||||||
|
|
||||||
|
Assert.Equal('t', reader.Current);
|
||||||
|
Assert.True(reader.MoveNext());
|
||||||
|
Assert.Equal('e', reader.Current);
|
||||||
|
Assert.True(reader.Retract());
|
||||||
|
Assert.Equal('t', reader.Current);
|
||||||
|
Assert.False(reader.Retract());
|
||||||
|
}
|
||||||
|
|
||||||
|
[Fact]
|
||||||
|
public void PeekTest()
|
||||||
|
{
|
||||||
|
ISourceReader reader = new StringSourceReader("peek");
|
||||||
|
reader.MoveNext();
|
||||||
|
|
||||||
|
Assert.Equal('p', reader.Current);
|
||||||
|
Assert.True(reader.TryPeekChar(out char? c));
|
||||||
|
Assert.Equal('e', c);
|
||||||
|
Assert.Equal('p', reader.Current);
|
||||||
|
}
|
||||||
|
|
||||||
|
private static void CheckLine(ISourceReader reader, string line, uint lineNumber)
|
||||||
|
{
|
||||||
|
foreach ((char value, uint index) in line.WithIndex())
|
||||||
|
{
|
||||||
|
Assert.Equal(value, reader.Current);
|
||||||
|
Assert.Equal(lineNumber, reader.Line);
|
||||||
Assert.Equal(index + 1, reader.Pos);
|
Assert.Equal(index + 1, reader.Pos);
|
||||||
Assert.Equal(2u, reader.Line);
|
reader.MoveNext();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue
Block a user