diff --git a/src/main/java/edu/kit/typicalc/model/Model.java b/src/main/java/edu/kit/typicalc/model/Model.java index 0e78b45..483d14e 100644 --- a/src/main/java/edu/kit/typicalc/model/Model.java +++ b/src/main/java/edu/kit/typicalc/model/Model.java @@ -3,8 +3,6 @@ package edu.kit.typicalc.model; import edu.kit.typicalc.model.parser.ParseError; import edu.kit.typicalc.util.Result; -import java.util.Map; - /** * This interface accepts user input and returns a type inference result. */ @@ -18,5 +16,5 @@ public interface Model { * @return either an error or a TypeInfererInterface on success */ Result getTypeInferer(String lambdaTerm, - Map typeAssumptions); + String typeAssumptions); } diff --git a/src/main/java/edu/kit/typicalc/model/ModelImpl.java b/src/main/java/edu/kit/typicalc/model/ModelImpl.java index d2472a7..acf1e21 100644 --- a/src/main/java/edu/kit/typicalc/model/ModelImpl.java +++ b/src/main/java/edu/kit/typicalc/model/ModelImpl.java @@ -27,7 +27,7 @@ public class ModelImpl implements Model { */ @Override public Result getTypeInferer(String lambdaTerm, - Map typeAssumptions) { + String typeAssumptions) { // Parse Lambda Term LambdaParser parser = new LambdaParser(lambdaTerm); Result result = parser.parse(); diff --git a/src/main/java/edu/kit/typicalc/model/parser/LambdaLexer.java b/src/main/java/edu/kit/typicalc/model/parser/LambdaLexer.java index c9710b1..27f128f 100644 --- a/src/main/java/edu/kit/typicalc/model/parser/LambdaLexer.java +++ b/src/main/java/edu/kit/typicalc/model/parser/LambdaLexer.java @@ -66,7 +66,7 @@ public class LambdaLexer { Token token = tokens.removeFirst(); return new Result<>(token); } else { - return new Result<>(new Token(TokenType.EOF, "", 0)); + return new Result<>(new Token(TokenType.EOF, "", term, term.length())); } } @@ -76,7 +76,7 @@ public class LambdaLexer { } if (pos >= term.length()) { // term ended, return EOF - return new Result<>(new Token(TokenType.EOF, "", pos)); + return new Result<>(new Token(TokenType.EOF, "", term, pos)); } Token t; char c = term.charAt(pos); @@ -84,7 +84,7 @@ public class LambdaLexer { case '-': if (pos + 1 < term.length()) { if (term.charAt(pos + 1) == '>') { - t = new Token(TokenType.ARROW, "->", pos); + t = new Token(TokenType.ARROW, "->", term, pos); advance(); advance(); return new Result<>(t); @@ -97,24 +97,37 @@ public class LambdaLexer { } // bunch of single-character tokens case '.': - t = new Token(TokenType.DOT, ".", pos); + t = new Token(TokenType.DOT, ".", term, pos); + advance(); + return new Result<>(t); + case ',': + t = new Token(TokenType.COMMA, ",", term, pos); + advance(); + return new Result<>(t); + case ':': + t = new Token(TokenType.COLON, ":", term, pos); advance(); return new Result<>(t); case '(': - t = new Token(TokenType.LEFT_PARENTHESIS, "(", pos); + t = new Token(TokenType.LEFT_PARENTHESIS, "(", term, pos); advance(); return new Result<>(t); case ')': - t = new Token(TokenType.RIGHT_PARENTHESIS, ")", pos); + t = new Token(TokenType.RIGHT_PARENTHESIS, ")", term, pos); advance(); return new Result<>(t); case '=': - t = new Token(TokenType.EQUALS, "=", pos); + t = new Token(TokenType.EQUALS, "=", term, pos); advance(); return new Result<>(t); case '\\': case 'λ': - t = new Token(TokenType.LAMBDA, c + "", pos); + t = new Token(TokenType.LAMBDA, Character.toString(c), term, pos); + advance(); + return new Result<>(t); + case '!': + case '∀': + t = new Token(TokenType.UNIVERSAL_QUANTIFIER, Character.toString(c), term, pos); advance(); return new Result<>(t); default: @@ -156,7 +169,7 @@ public class LambdaLexer { type = TokenType.VARIABLE; break; } - return new Result<>(new Token(type, sb.toString(), startPos)); + return new Result<>(new Token(type, sb.toString(), term, startPos)); } else if (Character.isDigit(c)) { int startPos = pos; // number literal @@ -165,7 +178,7 @@ public class LambdaLexer { sb.append(term.charAt(pos)); advance(); } while (pos < term.length() && Character.isDigit(term.charAt(pos))); - return new Result<>(new Token(TokenType.NUMBER, sb.toString(), startPos)); + return new Result<>(new Token(TokenType.NUMBER, sb.toString(), term, startPos)); } else { return new Result<>(null, ParseError.UNEXPECTED_CHARACTER.withCharacter(c, pos, term)); } diff --git a/src/main/java/edu/kit/typicalc/model/parser/LambdaParser.java b/src/main/java/edu/kit/typicalc/model/parser/LambdaParser.java index 8583cf1..5addd87 100644 --- a/src/main/java/edu/kit/typicalc/model/parser/LambdaParser.java +++ b/src/main/java/edu/kit/typicalc/model/parser/LambdaParser.java @@ -30,7 +30,7 @@ public class LambdaParser { * When calling a parseX method, token is the first token of X * (as opposed to the last token of the previous construct). */ - private Token token = new Token(TokenType.EOF, "", -1); + private Token token = new Token(TokenType.EOF, "", "", -1); private static final Set ATOM_START_TOKENS = EnumSet.of(TokenType.VARIABLE, TokenType.NUMBER, TokenType.TRUE, @@ -69,7 +69,7 @@ public class LambdaParser { TokenType current = token.getType(); Optional error = nextToken(); if (current != type) { - return Optional.of(ParseError.UNEXPECTED_TOKEN.withToken(lastToken, lexer.getTerm()).expectedType(type)); + return Optional.of(ParseError.UNEXPECTED_TOKEN.withToken(lastToken).expectedType(type)); } return error; } @@ -91,7 +91,7 @@ public class LambdaParser { } return new Result<>(null, (last.getType() == TokenType.EOF ? ParseError.TOO_FEW_TOKENS : ParseError.UNEXPECTED_TOKEN) - .withToken(last, lexer.getTerm()) + .withToken(last) .expectedTypes(ATOM_START_TOKENS)); } @@ -207,7 +207,7 @@ public class LambdaParser { try { n = Integer.parseInt(number); } catch (NumberFormatException e) { - return new Result<>(null, ParseError.UNEXPECTED_CHARACTER.withToken(token, lexer.getTerm())); + return new Result<>(null, ParseError.UNEXPECTED_CHARACTER.withToken(token)); } error = nextToken(); if (error.isEmpty()) { diff --git a/src/main/java/edu/kit/typicalc/model/parser/ParseError.java b/src/main/java/edu/kit/typicalc/model/parser/ParseError.java index eaeeb91..59a5560 100644 --- a/src/main/java/edu/kit/typicalc/model/parser/ParseError.java +++ b/src/main/java/edu/kit/typicalc/model/parser/ParseError.java @@ -19,7 +19,10 @@ public enum ParseError { /** * some tokens were required, but not provided + * + * DEPRECATED: use UNEXPECTED_TOKEN with TokenType.EOF instead */ + @Deprecated TOO_FEW_TOKENS, /** @@ -39,12 +42,11 @@ public enum ParseError { * Attach a token to this error. * * @param cause the token that caused the error - * @param term the term that is parsed * @return this object */ - public ParseError withToken(Token cause, String term) { + public ParseError withToken(Token cause) { this.cause = Optional.of(cause); - this.term = term; + this.term = cause.getSourceText(); this.position = cause.getPos(); return this; } diff --git a/src/main/java/edu/kit/typicalc/model/parser/Token.java b/src/main/java/edu/kit/typicalc/model/parser/Token.java index 18f3b25..f55d4f4 100644 --- a/src/main/java/edu/kit/typicalc/model/parser/Token.java +++ b/src/main/java/edu/kit/typicalc/model/parser/Token.java @@ -3,7 +3,7 @@ package edu.kit.typicalc.model.parser; import java.util.Objects; /** - * A token of the Prolog language. + * A token of a lambda term or type assumption. */ public class Token { /** @@ -13,6 +13,7 @@ public class Token { * EOF is a special token to indicate that the end of file is reached. */ public enum TokenType { + UNIVERSAL_QUANTIFIER, // ∀ or exclamation mark LAMBDA, // λ or a backslash VARIABLE, // [a-z][a-zA-Z0-9]* except "let" or "in" or constants LET, // let @@ -23,6 +24,8 @@ public class Token { LEFT_PARENTHESIS, // ( RIGHT_PARENTHESIS, // ) DOT, // . + COMMA, // , + COLON, // : EQUALS, // = ARROW, // -> EOF // pseudo token if end of input is reached @@ -36,6 +39,7 @@ public class Token { * the text of this token in the source code */ private final String text; + private final String sourceText; private final int pos; /** @@ -43,11 +47,13 @@ public class Token { * * @param type the token type * @param text text of this token in the source code - * @param pos position this token begins + * @param sourceText context of this token + * @param pos position this token begins in sourceText */ - public Token(TokenType type, String text, int pos) { + public Token(TokenType type, String text, String sourceText, int pos) { this.type = type; this.text = text; + this.sourceText = sourceText; this.pos = pos; } @@ -69,6 +75,13 @@ public class Token { return text; } + /** + * @return the surrounding text of this token + */ + public String getSourceText() { + return sourceText; + } + /** * Returns the position this token is in * @@ -92,11 +105,12 @@ public class Token { return false; } Token token = (Token) o; - return pos == token.pos && type == token.type && Objects.equals(text, token.text); + return pos == token.pos && type == token.type + && Objects.equals(text, token.text) && sourceText.equals(token.sourceText); } @Override public int hashCode() { - return Objects.hash(type, text, pos); + return Objects.hash(type, text, sourceText, pos); } } diff --git a/src/main/java/edu/kit/typicalc/model/parser/TypeAssumptionParser.java b/src/main/java/edu/kit/typicalc/model/parser/TypeAssumptionParser.java index 22bf1c0..bc4a477 100644 --- a/src/main/java/edu/kit/typicalc/model/parser/TypeAssumptionParser.java +++ b/src/main/java/edu/kit/typicalc/model/parser/TypeAssumptionParser.java @@ -4,8 +4,6 @@ import edu.kit.typicalc.model.parser.Token.TokenType; import edu.kit.typicalc.model.term.*; import edu.kit.typicalc.model.type.*; import edu.kit.typicalc.util.Result; -import org.apache.commons.lang3.tuple.ImmutablePair; -import org.apache.commons.lang3.tuple.Pair; import java.util.*; import java.util.regex.Matcher; @@ -16,70 +14,39 @@ import java.util.regex.Pattern; */ public class TypeAssumptionParser { - public static final Pattern TYPE_NAME_PATTERN = Pattern.compile("[a-zA-Z][a-zA-Z0-9]*"); private static final Pattern TYPE_VARIABLE_PATTERN = Pattern.compile("t(\\d+)"); - private static final Set END_TOKENS - = EnumSet.of(TokenType.ARROW, TokenType.RIGHT_PARENTHESIS, TokenType.EOF); - /** * Parse the given type assumptions. * * @param assumptions the type assumptions * @return if successful, a map of the type assumptions, otherwise an error */ - public Result, ParseError> parse(Map assumptions) { - Map typeAssumptions = new LinkedHashMap<>(); - for (Map.Entry entry : assumptions.entrySet()) { - String typeName = entry.getKey(); - if (!TYPE_NAME_PATTERN.matcher(typeName).matches()) { - return new Result<>(null, ParseError.UNEXPECTED_CHARACTER); - } - VarTerm var = new VarTerm(typeName); - Result typeAbs = parseTypeDefinition(entry.getValue()); - if (typeAbs.isError()) { - return new Result<>(typeAbs); - } - typeAssumptions.put(var, typeAbs.unwrap()); - } - return new Result<>(typeAssumptions); - } - - /** - * Parse a type definition that may use the all quantor. - * @param definition type definition - * @return parsed type - */ - public Result parseTypeDefinition(String definition) { - definition = cleanAssumptionText(definition); - Set allQuantified = new HashSet<>(); - if (definition.startsWith("∀")) { - String[] parts = definition.split("\\."); - if (parts.length < 2) { - int colonIndex = definition.indexOf(':'); - if (colonIndex >= 0) { - return new Result<>(null, ParseError.UNEXPECTED_CHARACTER.withCharacter( - ':', colonIndex, definition).expectedCharacter('.') - ); + public Result, ParseError> parse(String assumptions) { + ParserState> state = new InitialState(new LinkedHashMap<>()); + LambdaLexer lexer = new LambdaLexer(cleanAssumptionText(assumptions)); + Optional extraToken = Optional.empty(); + while (true) { + Token token1; + if (extraToken.isPresent()) { + token1 = extraToken.get(); + } else { + Result token = lexer.nextToken(); + if (token.isError()) { + return token.castError(); } - return new Result<>(null, ParseError.TOO_FEW_TOKENS.expectedType(TokenType.DOT)); - } else if (parts.length > 2) { - return new Result<>(null, ParseError.UNEXPECTED_CHARACTER.withCharacter( - '.', parts[0].length() + 1 + parts[1].length(), definition)); + token1 = token.unwrap(); } - for (String quantified : parts[0].substring(1).split(",")) { - quantified = quantified.trim(); - if (!quantified.matches("t\\d+")) { - return new Result<>(null, ParseError.UNEXPECTED_TOKEN.withToken( - new Token(TokenType.VARIABLE, quantified, parts[0].indexOf(quantified)), parts[0] - )); - } - int i = Integer.parseInt(quantified.substring(1)); - allQuantified.add(new TypeVariable(TypeVariableKind.USER_INPUT, i)); + ParserResult> result = state.handle(token1); + if (result.isResult()) { + return new Result<>(result.getResult()); + } else if (result.isError()) { + return new Result<>(null, result.getError()); + } else { + state = result.getState(); + extraToken = result.extraToken(); } - definition = parts[1]; } - return parseType(definition, allQuantified); } private String cleanAssumptionText(String text) { @@ -96,89 +63,403 @@ public class TypeAssumptionParser { .replace('τ', 't'); } - private Result parseType(String type, Set allQuantified) { - // this parser is using the same lexer as the LambdaParser (to avoid types named "let" etc.) - LambdaLexer lexer = new LambdaLexer(type); - Result, ParseError> parsedType = parseType(lexer, 0); - if (parsedType.isError()) { - return new Result<>(null, parsedType.unwrapError()); + private static class ParserResult { + private Optional> newState = Optional.empty(); + private Optional error = Optional.empty(); + private Optional result = Optional.empty(); + private Optional extraToken = Optional.empty(); + + ParserResult(ParseError e) { + this.error = Optional.of(e); + } + + ParserResult(ParserState state) { + this.newState = Optional.of(state); + } + + ParserResult(T result) { + this.result = Optional.of(result); + } + + boolean isTransition() { + return newState.isPresent(); + } + + ParserState getState() { + return newState.get(); + } + + boolean isError() { + return error.isPresent(); + } + + ParseError getError() { + return error.get(); + } + + ParserResult castError() { + return new ParserResult<>(error.get()); + } + + boolean isResult() { + return result.isPresent(); + } + + T getResult() { + return result.get(); + } + + ParserResult attachToken(Token t) { + this.extraToken = Optional.of(t); + return this; + } + + ParserResult copyToken(ParserResult other) { + this.extraToken = other.extraToken; + return this; + } + + Optional extraToken() { + return this.extraToken; } - return new Result<>(new TypeAbstraction(parsedType.unwrap().getLeft(), allQuantified)); } - private Result, ParseError> parseType(LambdaLexer lexer, int parenCount) { - Type type = null; - int removedParens = 0; - while (true) { - Result token = lexer.nextToken(); - if (token.isError()) { - return new Result<>(token); - } - Token t = token.unwrap(); - Result typeResult = null; + private interface ParserState { + ParserResult handle(Token t); + } + + private static class InitialState implements ParserState> { + private Map alreadyParsed = new LinkedHashMap<>(); + + InitialState(Map alreadyParsed) { + this.alreadyParsed = alreadyParsed; + } + + @Override + public ParserResult> handle(Token t) { switch (t.getType()) { - case LEFT_PARENTHESIS: - // recursive call, set number of open parentheses to 1 - Result, ParseError> type2 = parseType(lexer, 1); - typeResult = type2.map(Pair::getLeft); - removedParens += type2.map(Pair::getRight).unwrapOr(1) - 1; - break; case VARIABLE: - // named type or type variable - type = parseLiteral(t.getText()); - break; - case RIGHT_PARENTHESIS: - removedParens += 1; - break; - case ARROW: - if (type == null) { - // there was no type in front of the arrow - return new Result<>(null, ParseError.UNEXPECTED_TOKEN.withToken(t, "")); - } - // recursive call, keep open parentheses count - Result, ParseError> nextType = parseType(lexer, parenCount); - final Type left = type; - typeResult = nextType.map(Pair::getLeft).map(right -> new FunctionType(left, right)); - removedParens += nextType.map(Pair::getRight).unwrapOr(0); - break; + return new ParserResult<>(new ExpectingColon(alreadyParsed, new VarTerm(t.getText()))); case EOF: - break; + return new ParserResult<>(alreadyParsed); default: - return new Result<>(null, ParseError.UNEXPECTED_TOKEN.withToken(t, "")); - } - // update type based on Result - if (typeResult != null && typeResult.isError()) { - return new Result<>(typeResult); - } - type = typeResult != null ? typeResult.unwrap() : type; - // after fully processing one token / a type in parenthesis, - // some type should have been parsed - if (type == null) { - return new Result<>(null, ParseError.TOO_FEW_TOKENS); - } - if (parenCount - removedParens < 0) { - // too many closing parenthesis - return new Result<>(null, ParseError.UNEXPECTED_TOKEN.withToken(t, "")); - } else if (END_TOKENS.contains(t.getType())) { - // potential end of type - if (parenCount - removedParens == 0) { - // opening and closing parentheses match - return new Result<>(new ImmutablePair<>(type, removedParens)); - } else { - // missing closing parenthesis - return new Result<>(null, ParseError.TOO_FEW_TOKENS); - } + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); } } } - private Type parseLiteral(String text) { - Matcher typeVariableMatcher = TYPE_VARIABLE_PATTERN.matcher(text); - if (typeVariableMatcher.matches()) { - int typeVariableIndex = Integer.parseInt(typeVariableMatcher.group(1)); - return new TypeVariable(TypeVariableKind.USER_INPUT, typeVariableIndex); - } else { - return new NamedType(text); + private static class ExpectingColon implements ParserState> { + private Map alreadyParsed; + private final VarTerm var; + ExpectingColon(Map alreadyParsed, VarTerm var) { + this.alreadyParsed = alreadyParsed; + this.var = var; + } + + @Override + public ParserResult> handle(Token t) { + switch (t.getType()) { + case COLON: + return new ParserResult<>(new ExpectingTypeDef(alreadyParsed, var)); + default: + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } + } + } + + private static class ExpectingTypeDef implements ParserState> { + private final Map alreadyParsed; + private final Set typeVariables; + private final VarTerm var; + private Optional> state = Optional.empty(); + ExpectingTypeDef(Map alreadyParsed, VarTerm var) { + this.alreadyParsed = alreadyParsed; + this.typeVariables = new TreeSet<>(); + this.var = var; + } + + ExpectingTypeDef(Map alreadyParsed, Set typeVariables, VarTerm var) { + this.alreadyParsed = alreadyParsed; + this.typeVariables = typeVariables; + this.var = var; + } + + @Override + public ParserResult> handle(Token t) { + switch (t.getType()) { + case UNIVERSAL_QUANTIFIER: + if (typeVariables.isEmpty()) { + return new ParserResult<>(new ExpectingTypeVariables(alreadyParsed, var)); + } else { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } + default: + if (state.isPresent()) { + ParserState status = state.get(); + // already parsing type + ParserResult result = status.handle(t); + if (result.isResult()) { + alreadyParsed.put(var, new TypeAbstraction(result.getResult(), typeVariables)); + return new ParserResult<>(new ExpectingCommaBeforeNextType(alreadyParsed)) + .copyToken(result); + } else if (result.isError()) { + return result.castError(); + } else { + state = Optional.of(result.getState()); + return new ParserResult<>(this); + } + } + // attempt to parse as type + ParserResult result = new ParseTypeState1().handle(t); + if (result.isError()) { + return result.castError(); + } + state = Optional.of(result.getState()); + return new ParserResult<>(this); + } + } + } + + private static class ExpectingCommaBeforeNextType implements ParserState> { + private final Map alreadyParsed; + + ExpectingCommaBeforeNextType(Map alreadyParsed) { + this.alreadyParsed = alreadyParsed; + } + + @Override + public ParserResult> handle(Token t) { + switch (t.getType()) { + case EOF: + return new ParserResult<>(alreadyParsed); + case COMMA: + return new ParserResult<>(new InitialState(alreadyParsed)); + default: + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } + } + } + + /** + * Main type parsing state. + */ + private static class ParseTypeState1 implements ParserState { + private Optional parsedType = Optional.empty(); + private Optional> state = Optional.empty(); + private Optional> stateParenthesis = Optional.empty(); + private int parenthesisInitial = 0; + private int openParens = 0; + + @Override + public ParserResult handle(Token t) { + switch (t.getType()) { + case VARIABLE: + if (state.isPresent()) { + return handleInner(t); + } + if (stateParenthesis.isPresent()) { + return handleInnerParenthesis(t); + } + if (parsedType.isPresent()) { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } + Type type = parseLiteral(t.getText()); + // try parsing function type (see below) + this.parsedType = Optional.of(type); + return new ParserResult<>(this); + case ARROW: + if (state.isPresent()) { + return handleInner(t); + } + if (stateParenthesis.isPresent()) { + return handleInnerParenthesis(t); + } + if (parsedType.isEmpty()) { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } + // parse function type + state = Optional.of(new ParseTypeStateExpectArrow().handle(t).getState()); + return new ParserResult<>(this); + case LEFT_PARENTHESIS: + openParens += 1; + if (state.isPresent()) { + return handleInner(t); + } + if (stateParenthesis.isPresent()) { + return handleInnerParenthesis(t); + } + stateParenthesis = Optional.of(new ParseTypeState1()); + parenthesisInitial = openParens - 1; + return new ParserResult<>(this); + case RIGHT_PARENTHESIS: + openParens -= 1; + if (state.isPresent()) { + return handleInner(t); + } + if (stateParenthesis.isPresent()) { + if (openParens == parenthesisInitial) { + // inner state is done parsing + ParserResult result = handleInnerParenthesis( + new Token(TokenType.EOF, "", "", -1)); + if (result.isError()) { + return result.castError(); + } else { + parsedType = Optional.of(result.getResult()); + stateParenthesis = Optional.empty(); + } + } else { + return handleInnerParenthesis(t); + } + } + if (parsedType.isPresent()) { + return new ParserResult<>(this); // parenthesized part may be start of function + } + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + case COMMA: + case EOF: + if (state.isPresent()) { + return handleInner(t).attachToken(t); + } + if (stateParenthesis.isPresent() && openParens == parenthesisInitial) { + return handleInnerParenthesis(t).attachToken(t); + } + if (parsedType.isPresent() && openParens == parenthesisInitial) { + return new ParserResult<>(parsedType.get()).attachToken(t); + } + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + default: + if (state.isPresent()) { + return handleInner(t); + } + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } + } + + private ParserResult handleInner(Token t) { + ParserState status = state.get(); + // already parsing right type of function + ParserResult result = status.handle(t); + if (result.isResult()) { + return new ParserResult(new FunctionType(parsedType.get(), result.getResult())).copyToken(result); + } else if (result.isError()) { + return result.castError(); + } else { + state = Optional.of(result.getState()); + return new ParserResult<>(this); + } + } + + private ParserResult handleInnerParenthesis(Token t) { + ParserState status = stateParenthesis.get(); + // already parsing right type of function + ParserResult result = status.handle(t); + if (result.isResult()) { + return new ParserResult<>(result.getResult()).copyToken(result); + } else if (result.isError()) { + return result.castError(); + } else { + stateParenthesis = Optional.of(result.getState()); + return new ParserResult<>(this); + } + } + + private static Type parseLiteral(String text) { + Matcher typeVariableMatcher = TYPE_VARIABLE_PATTERN.matcher(text); + if (typeVariableMatcher.matches()) { + int typeVariableIndex = Integer.parseInt(typeVariableMatcher.group(1)); + return new TypeVariable(TypeVariableKind.USER_INPUT, typeVariableIndex); + } else { + return new NamedType(text); + } + } + } + + private static class ParseTypeStateExpectArrow implements ParserState { + private Optional> state = Optional.empty(); + + @Override + public ParserResult handle(Token t) { + switch (t.getType()) { + case ARROW: + if (state.isEmpty()) { + // try parsing remainder as type + state = Optional.of(new ParseTypeState1()); + return new ParserResult<>(this); + } + default: + if (state.isPresent()) { + ParserState status = state.get(); + // already parsing type + ParserResult result = status.handle(t); + if (result.isResult()) { + return result; + } else if (result.isError()) { + return result.castError(); + } else { + state = Optional.of(result.getState()); + return new ParserResult<>(this); + } + } else { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } + } + } + } + + private static class ExpectingTypeVariables implements ParserState> { + private final Map alreadyParsed; + private final VarTerm var; + private final Set variables = new TreeSet<>(); + private boolean expectCommaOrDot = false; + ExpectingTypeVariables(Map alreadyParsed, VarTerm var) { + this.alreadyParsed = alreadyParsed; + this.var = var; + } + + @Override + public ParserResult> handle(Token t) { + switch (t.getType()) { + case VARIABLE: + if (expectCommaOrDot) { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN + .withToken(t) + .expectedTypes(List.of(TokenType.COMMA, Token.TokenType.DOT))); + } + String input = t.getText(); + if (!TYPE_VARIABLE_PATTERN.matcher(input).matches()) { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + // TODO: somehow convey the expected format + } + int i = Integer.parseInt(input.substring(1)); + variables.add(new TypeVariable(TypeVariableKind.USER_INPUT, i)); + expectCommaOrDot = true; + return new ParserResult<>(this); + case COMMA: + if (expectCommaOrDot) { + expectCommaOrDot = false; + return new ParserResult<>(this); + } else { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN + .withToken(t) + .expectedType(TokenType.VARIABLE)); + } + case DOT: + if (expectCommaOrDot) { + // list of type variables is complete + // parse actual type + return new ParserResult<>(new ExpectingTypeDef(alreadyParsed, variables, var)); + } else { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN + .withToken(t) + .expectedType(TokenType.VARIABLE)); + } + default: + if (expectCommaOrDot) { + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t) + .expectedTypes(List.of(TokenType.COMMA, TokenType.DOT))); + } + return new ParserResult<>(ParseError.UNEXPECTED_TOKEN.withToken(t)); + } } } } diff --git a/src/main/java/edu/kit/typicalc/presenter/Presenter.java b/src/main/java/edu/kit/typicalc/presenter/Presenter.java index c65d6d0..45bc066 100644 --- a/src/main/java/edu/kit/typicalc/presenter/Presenter.java +++ b/src/main/java/edu/kit/typicalc/presenter/Presenter.java @@ -1,7 +1,5 @@ package edu.kit.typicalc.presenter; -import java.util.Map; - import com.vaadin.flow.component.UI; import edu.kit.typicalc.model.Model; import edu.kit.typicalc.model.TypeInfererInterface; @@ -32,7 +30,7 @@ public class Presenter implements MainViewListener { } @Override - public void typeInferLambdaString(String lambdaTerm, Map typeAssumptions) { + public void typeInferLambdaString(String lambdaTerm, String typeAssumptions) { Result result = model.getTypeInferer(lambdaTerm, typeAssumptions); if (result.isError()) { view.displayError(result.unwrapError()); diff --git a/src/main/java/edu/kit/typicalc/util/Result.java b/src/main/java/edu/kit/typicalc/util/Result.java index 2bb939a..e8ebdf3 100644 --- a/src/main/java/edu/kit/typicalc/util/Result.java +++ b/src/main/java/edu/kit/typicalc/util/Result.java @@ -116,6 +116,10 @@ public class Result { return new Result<>(value != null ? mapping.apply(value) : null, error); } + public Result castError() { + return new Result<>(null, this.error); + } + @Override public String toString() { if (isOk()) { diff --git a/src/main/java/edu/kit/typicalc/view/main/InputBar.java b/src/main/java/edu/kit/typicalc/view/main/InputBar.java index d5f1531..8b1e9bd 100644 --- a/src/main/java/edu/kit/typicalc/view/main/InputBar.java +++ b/src/main/java/edu/kit/typicalc/view/main/InputBar.java @@ -120,11 +120,8 @@ public class InputBar extends HorizontalLayout implements LocaleChangeObserver { * * @param typeAssumptions the type assumptions as a map */ - protected void setTypeAssumptions(Map typeAssumptions) { - assumptionInputField.setValue( - typeAssumptions.entrySet().stream() - .map(entry -> entry.getKey().trim() + ": " + entry.getValue().trim()) - .collect(Collectors.joining("; "))); + protected void setTypeAssumptions(String typeAssumptions) { + assumptionInputField.setValue(typeAssumptions); } protected void setTermAndClickTypeInfer(String term) { diff --git a/src/main/java/edu/kit/typicalc/view/main/MainView.java b/src/main/java/edu/kit/typicalc/view/main/MainView.java index 9111e15..c6a5e0c 100644 --- a/src/main/java/edu/kit/typicalc/view/main/MainView.java +++ b/src/main/java/edu/kit/typicalc/view/main/MainView.java @@ -3,8 +3,6 @@ package edu.kit.typicalc.view.main; import edu.kit.typicalc.model.TypeInfererInterface; import edu.kit.typicalc.model.parser.ParseError; -import java.util.Map; - /** * Provides an interface for the presenter to interact with the view. The interaction consists of * either passing a TypeInfererInterface or a ParseError to the view. @@ -37,6 +35,6 @@ public interface MainView { * @param lambdaTerm the lambda term to type-infer * @param typeAssumptions type assumptions to use during the calculation */ - void typeInferLambdaString(String lambdaTerm, Map typeAssumptions); + void typeInferLambdaString(String lambdaTerm, String typeAssumptions); } } diff --git a/src/main/java/edu/kit/typicalc/view/main/MainViewImpl.java b/src/main/java/edu/kit/typicalc/view/main/MainViewImpl.java index 1f289ec..00ca04c 100644 --- a/src/main/java/edu/kit/typicalc/view/main/MainViewImpl.java +++ b/src/main/java/edu/kit/typicalc/view/main/MainViewImpl.java @@ -74,14 +74,13 @@ public class MainViewImpl extends AppLayout if (url.getPath().matches(TypeInferenceView.ROUTE + "/.*")) { List segments = url.getSegments(); String term = segments.get(segments.size() - 1); - Map types = url.getQueryParameters().getParameters().entrySet().stream().map(entry -> - Pair.of(entry.getKey(), entry.getValue().get(0)) - ).collect(Collectors.toMap(Pair::getLeft, Pair::getRight, - (existing, replacement) -> existing, LinkedHashMap::new)); + String types = url.getQueryParameters().getParameters() + .entrySet().stream().map(kv -> kv.getKey() + ": " + kv.getValue().get(0)) + .collect(Collectors.joining(", ")); upperBar.inferTerm(decodeURL(term), types); } else if (url.getPath().equals(TypeInferenceView.ROUTE)) { setContent(new StartPageView()); - upperBar.inferTerm(StringUtils.EMPTY, Collections.emptyMap()); + upperBar.inferTerm(StringUtils.EMPTY, ""); } else if (url.getPath().equals(StringUtils.EMPTY)) { setContent(new StartPageView()); } diff --git a/src/main/java/edu/kit/typicalc/view/main/UpperBar.java b/src/main/java/edu/kit/typicalc/view/main/UpperBar.java index d388c40..e4adf67 100644 --- a/src/main/java/edu/kit/typicalc/view/main/UpperBar.java +++ b/src/main/java/edu/kit/typicalc/view/main/UpperBar.java @@ -94,7 +94,7 @@ public class UpperBar extends VerticalLayout implements LocaleChangeObserver { inputConsumer.accept(termAndAssumptions); } - private void startInfer(String term, Map typeAssumptions) { + private void startInfer(String term, String typeAssumptions) { presenter.typeInferLambdaString(term, typeAssumptions); } @@ -104,7 +104,7 @@ public class UpperBar extends VerticalLayout implements LocaleChangeObserver { * @param term the provided string * @param typeAssumptions type assumptions to use */ - protected void inferTerm(String term, Map typeAssumptions) { + protected void inferTerm(String term, String typeAssumptions) { inputBar.setTypeAssumptions(typeAssumptions); inputBar.setTerm(term); startInfer(term, typeAssumptions); diff --git a/src/test/java/edu/kit/typicalc/model/ModelImplTest.java b/src/test/java/edu/kit/typicalc/model/ModelImplTest.java index 32f3181..687034a 100644 --- a/src/test/java/edu/kit/typicalc/model/ModelImplTest.java +++ b/src/test/java/edu/kit/typicalc/model/ModelImplTest.java @@ -23,14 +23,10 @@ class ModelImplTest { @Test void getTypeInferer() { ModelImpl model = new ModelImpl(); - Map map = new LinkedHashMap<>(); - map.put("x", "int"); - Map map2 = new LinkedHashMap<>(); - map2.put("a.x", "3.int"); - assertTrue(model.getTypeInferer("test.x.x.test", map2).isError()); - assertTrue(model.getTypeInferer("x", map2).isError()); + assertTrue(model.getTypeInferer("test.x.x.test", "a.x: 3.int").isError()); + assertTrue(model.getTypeInferer("x", "a.x: 3.int").isError()); - Result result = model.getTypeInferer("λy.x", map); + Result result = model.getTypeInferer("λy.x", "x: int"); VarTerm y = new VarTerm("y"); VarTerm x = new VarTerm("x"); LambdaTerm term = new AbsTerm(y, x); @@ -66,10 +62,9 @@ class ModelImplTest { @Test void quantifiedTypeAssumption() { ModelImpl model = new ModelImpl(); - Map assumptions = new HashMap<>(); - assumptions.put("id", "∀ t1 . t1 -> t1"); - Result result = model.getTypeInferer("(id id) (id true)", assumptions); + Result result = model.getTypeInferer("(id id) (id true)", + "id: ∀ t1 . t1 -> t1"); if (result.isError()) { System.out.println(result.unwrapError()); fail(); @@ -93,7 +88,7 @@ class ModelImplTest { @Test void letTermTypeAssumptions() { Model model = new ModelImpl(); - TypeInfererInterface typer = model.getTypeInferer("(let g = λx.x in g) g", new HashMap<>()).unwrap(); + TypeInfererInterface typer = model.getTypeInferer("(let g = λx.x in g) g", "").unwrap(); AppStep first = (AppStep) typer.getFirstInferenceStep(); LetStep second = (LetStep) first.getPremise1(); VarStep third = (VarStep) second.getPremise(); diff --git a/src/test/java/edu/kit/typicalc/model/parser/LambdaParserFuzzTest.java b/src/test/java/edu/kit/typicalc/model/parser/LambdaParserFuzzTest.java index 20c7a27..5f26e9b 100644 --- a/src/test/java/edu/kit/typicalc/model/parser/LambdaParserFuzzTest.java +++ b/src/test/java/edu/kit/typicalc/model/parser/LambdaParserFuzzTest.java @@ -28,7 +28,7 @@ public class LambdaParserFuzzTest { @Fuzz public void testInference(@From(LambdaTermGenerator.class) String term) { Model model = new ModelImpl(); - Result typer = model.getTypeInferer(term, new HashMap<>()); + Result typer = model.getTypeInferer(term, ""); InferenceStep first = typer.unwrap().getFirstInferenceStep(); } diff --git a/src/test/java/edu/kit/typicalc/model/parser/LambdaParserTest.java b/src/test/java/edu/kit/typicalc/model/parser/LambdaParserTest.java index 7e6bb64..1612cfa 100644 --- a/src/test/java/edu/kit/typicalc/model/parser/LambdaParserTest.java +++ b/src/test/java/edu/kit/typicalc/model/parser/LambdaParserTest.java @@ -98,7 +98,7 @@ class LambdaParserTest { parser = new LambdaParser("x)"); ParseError error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", 1), error.getCause().get()); + assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", "x)", 1), error.getCause().get()); parser = new LambdaParser("??"); assertEquals(ParseError.UNEXPECTED_CHARACTER, parser.parse().unwrapError()); parser = new LambdaParser("aλ"); @@ -110,39 +110,40 @@ class LambdaParserTest { parser = new LambdaParser("x 123333333333333"); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_CHARACTER, error); - assertEquals(new Token(TokenType.NUMBER, "123333333333333", 2), error.getCause().get()); + assertEquals(new Token(TokenType.NUMBER, "123333333333333", "x 123333333333333", 2), + error.getCause().get()); parser = new LambdaParser("λ)"); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", 1), error.getCause().get()); + assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", "λ)", 1), error.getCause().get()); parser = new LambdaParser("λx="); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.EQUALS, "=", 2), error.getCause().get()); + assertEquals(new Token(TokenType.EQUALS, "=", "λx=", 2), error.getCause().get()); parser = new LambdaParser("λx.."); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.DOT, ".", 3), error.getCause().get()); + assertEquals(new Token(TokenType.DOT, ".", "λx..", 3), error.getCause().get()); parser = new LambdaParser("let ) ="); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", 4), error.getCause().get()); + assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", "let ) =", 4), error.getCause().get()); parser = new LambdaParser("let x ."); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.DOT, ".", 6), error.getCause().get()); + assertEquals(new Token(TokenType.DOT, ".", "let x .", 6), error.getCause().get()); parser = new LambdaParser("let x = )"); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", 8), error.getCause().get()); + assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", "let x = )",8), error.getCause().get()); parser = new LambdaParser("let x = y )"); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", 10), error.getCause().get()); + assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", "let x = y )", 10), error.getCause().get()); parser = new LambdaParser("let x = y in )"); error = parser.parse().unwrapError(); assertEquals(ParseError.UNEXPECTED_TOKEN, error); - assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", 13), error.getCause().get()); + assertEquals(new Token(TokenType.RIGHT_PARENTHESIS, ")", "let x = y in )", 13), error.getCause().get()); } @Test diff --git a/src/test/java/edu/kit/typicalc/model/parser/TypeAssumptionParserTest.java b/src/test/java/edu/kit/typicalc/model/parser/TypeAssumptionParserTest.java index 7bcac09..71f7c12 100644 --- a/src/test/java/edu/kit/typicalc/model/parser/TypeAssumptionParserTest.java +++ b/src/test/java/edu/kit/typicalc/model/parser/TypeAssumptionParserTest.java @@ -6,6 +6,7 @@ import edu.kit.typicalc.util.Result; import org.junit.jupiter.api.Test; import java.util.*; +import java.util.stream.Collectors; import static edu.kit.typicalc.model.type.NamedType.BOOLEAN; import static edu.kit.typicalc.model.type.NamedType.INT; @@ -16,9 +17,7 @@ class TypeAssumptionParserTest { @Test void simpleType() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("a", "int"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("a: int"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); @@ -30,9 +29,7 @@ class TypeAssumptionParserTest { @Test void typeVariablesOneDigitIndex() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("x", "t1"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("x: t1"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); @@ -40,23 +37,19 @@ class TypeAssumptionParserTest { assertEquals(new VarTerm("x"), assumption.getKey()); assertEquals(new TypeAbstraction(new TypeVariable(TypeVariableKind.USER_INPUT, 1)), assumption.getValue()); - Map assumptions2 = new LinkedHashMap<>(); - assumptions2.put("x", "t001"); - Result, ParseError> type2 = parser.parse(assumptions2); + Result, ParseError> type2 = parser.parse("x: t009"); assertTrue(type.isOk()); Map types2 = type2.unwrap(); assertEquals(1, types2.size()); Map.Entry assumption2 = types2.entrySet().stream().findFirst().get(); assertEquals(new VarTerm("x"), assumption2.getKey()); - assertEquals(new TypeAbstraction(new TypeVariable(TypeVariableKind.USER_INPUT, 1)), assumption2.getValue()); + assertEquals(new TypeAbstraction(new TypeVariable(TypeVariableKind.USER_INPUT, 9)), assumption2.getValue()); } @Test void typeVariablesMultipleDigitIndex() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("x", "t123456"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("x: t123456"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); @@ -68,9 +61,7 @@ class TypeAssumptionParserTest { @Test void namedTypeStartingWithT() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("x", "tau1"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("x: tau1"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); @@ -82,27 +73,27 @@ class TypeAssumptionParserTest { @Test void functionType() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("id", "int -> int"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("id: int -> int"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); Map.Entry assumption = types.entrySet().stream().findFirst().get(); assertEquals(new VarTerm("id"), assumption.getKey()); assertEquals(new TypeAbstraction(new FunctionType(INT, INT)), assumption.getValue()); + } - Map assumptions2 = new LinkedHashMap<>(); - assumptions2.put("f", "int -> int -> int"); - type = parser.parse(assumptions2); + @Test + void functionType2() { + TypeAssumptionParser parser = new TypeAssumptionParser(); + Result, ParseError> type = parser.parse("f: int -> int -> int"); if (type.isError()) { System.err.println(type.unwrapError()); System.err.println(type.unwrapError().getCause()); } assertTrue(type.isOk()); - types = type.unwrap(); + Map types = type.unwrap(); assertEquals(1, types.size()); - assumption = types.entrySet().stream().findFirst().get(); + Map.Entry assumption = types.entrySet().stream().findFirst().get(); assertEquals(new VarTerm("f"), assumption.getKey()); assertEquals(new TypeAbstraction( new FunctionType( @@ -114,9 +105,7 @@ class TypeAssumptionParserTest { @Test void functionTypeWithVariables() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("fun", "t0 -> t0"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("fun: t0 -> t0"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); @@ -129,9 +118,7 @@ class TypeAssumptionParserTest { @Test void complicatedTypes() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("id", "(int -> int) -> (boolean -> boolean)"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("id: (int -> int) -> (boolean -> boolean)"); if (type.isError()) { System.err.println(type.unwrapError()); System.err.println(type.unwrapError().getCause()); @@ -147,9 +134,7 @@ class TypeAssumptionParserTest { new FunctionType(BOOLEAN, BOOLEAN) )), assumption.getValue()); parser = new TypeAssumptionParser(); - Map assumptions2 = new LinkedHashMap<>(); - assumptions2.put("id", "((int -> int) -> (boolean -> boolean)) -> ((int2 -> boolean2) -> (boolean2 -> int2))"); - type = parser.parse(assumptions2); + type = parser.parse("id: ((int -> int) -> (boolean -> boolean)) -> ((int2 -> boolean2) -> (boolean2 -> int2))"); if (type.isError()) { System.err.println(type.unwrapError()); System.err.println(type.unwrapError().getCause()); @@ -177,9 +162,7 @@ class TypeAssumptionParserTest { @Test void longFunction() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new LinkedHashMap<>(); - assumptions.put("fun", "(a -> b -> c) -> d"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("fun: (a -> b -> c) -> d"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); @@ -195,9 +178,7 @@ class TypeAssumptionParserTest { @Test void allQuantified() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new HashMap<>(); - assumptions.put("id", "∀ t1 . t1 -> t1"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("id: ∀ t1 . t1 -> t1"); assertTrue(type.isOk()); Map types = type.unwrap(); assertEquals(1, types.size()); @@ -212,35 +193,59 @@ class TypeAssumptionParserTest { ), quantified), assumption.getValue()); } + @Test + void allQuantifiedCommaSeparation() { + TypeAssumptionParser parser = new TypeAssumptionParser(); + Result, ParseError> type = parser.parse("transmute: ∀ t1, t2 . t1 -> t2, createAny: ∀ t1 . t1 "); + assertTrue(type.isOk()); + Map types = type.unwrap(); + assertEquals(2, types.size()); + Set quantified = new HashSet<>(); + quantified.add(new TypeVariable(TypeVariableKind.USER_INPUT, 1)); + quantified.add(new TypeVariable(TypeVariableKind.USER_INPUT, 2)); + assertEquals(new TypeAbstraction( + new FunctionType( + new TypeVariable(TypeVariableKind.USER_INPUT, 1), + new TypeVariable(TypeVariableKind.USER_INPUT, 2) + ), quantified), types.get(new VarTerm("transmute"))); + quantified.remove(new TypeVariable(TypeVariableKind.USER_INPUT, 2)); + assertEquals(new TypeAbstraction( + new TypeVariable(TypeVariableKind.USER_INPUT, 1), + quantified), types.get(new VarTerm("createAny"))); + } + @Test void usefulErrors() { TypeAssumptionParser parser = new TypeAssumptionParser(); - Map assumptions = new HashMap<>(); - assumptions.put("id", "∀ t1"); - Result, ParseError> type = parser.parse(assumptions); + Result, ParseError> type = parser.parse("id: ∀ t1"); assertTrue(type.isError()); ParseError error = type.unwrapError(); - assertEquals(ParseError.TOO_FEW_TOKENS, error); + assertEquals(ParseError.UNEXPECTED_TOKEN, error); + assertEquals(Token.TokenType.EOF, error.getCause().get().getType()); Collection expected = error.getExpected().get(); - assertEquals(1, expected.size()); + assertEquals(2, expected.size()); assertTrue(expected.contains(Token.TokenType.DOT)); - // TODO: should also expect a comma, once the parser is fixed! + assertTrue(expected.contains(Token.TokenType.COMMA)); } @Test void errors() { Map tests = new HashMap<>(); - tests.put("", ParseError.TOO_FEW_TOKENS); + tests.put("", + ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.EOF, "", "", 0))); tests.put("ö", ParseError.UNEXPECTED_CHARACTER); - tests.put("(x", ParseError.TOO_FEW_TOKENS); - tests.put("-> x", ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.ARROW, "->", 0), "")); - tests.put("x 11", ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.NUMBER, "11", 2), "")); - tests.put("x )", ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.RIGHT_PARENTHESIS, ")", 2), "")); + tests.put("(x", + ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.EOF, "", "(x", 2))); + tests.put("-> x", + ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.ARROW, "->", "-> x", 0))); + tests.put("x 11", + ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.NUMBER, "11", "x 11", 2))); + tests.put("x )", ParseError.UNEXPECTED_TOKEN.withToken(new Token(Token.TokenType.RIGHT_PARENTHESIS, ")", "x )", 2))); tests.put("x -> (x) )", ParseError.UNEXPECTED_TOKEN - .withToken(new Token(Token.TokenType.RIGHT_PARENTHESIS, ")", 9), "")); + .withToken(new Token(Token.TokenType.RIGHT_PARENTHESIS, ")", "x -> (x) )", 9))); for (Map.Entry entry : tests.entrySet()) { TypeAssumptionParser parser = new TypeAssumptionParser(); - Result, ParseError> type = parser.parse(Map.of("type1", entry.getKey())); + Result, ParseError> type = parser.parse("type1:" + entry.getKey()); assertTrue(type.isError()); assertEquals(entry.getValue(), type.unwrapError()); if (entry.getValue().getCause().isPresent()) { @@ -248,13 +253,15 @@ class TypeAssumptionParserTest { } } TypeAssumptionParser parser = new TypeAssumptionParser(); - Result, ParseError> type = parser.parse(Map.of("föhn", "int")); + Result, ParseError> type = parser.parse("föhn: int"); assertTrue(type.isError()); assertEquals(ParseError.UNEXPECTED_CHARACTER, type.unwrapError()); + assertEquals(1, type.unwrapError().getPosition()); parser = new TypeAssumptionParser(); - type = parser.parse(Map.of("1typ", "int")); + type = parser.parse("1typ:int"); assertTrue(type.isError()); - assertEquals(ParseError.UNEXPECTED_CHARACTER, type.unwrapError()); + assertEquals(ParseError.UNEXPECTED_TOKEN, type.unwrapError()); + assertEquals(0, type.unwrapError().getPosition()); } @Test @@ -281,7 +288,7 @@ class TypeAssumptionParserTest { tests.put("((x) -> ((x)) -> (x)) -> (x -> (x -> (x)))", new TypeAbstraction(xxxxxx)); for (Map.Entry entry : tests.entrySet()) { TypeAssumptionParser parser = new TypeAssumptionParser(); - Result, ParseError> type = parser.parse(Map.of("type1", entry.getKey())); + Result, ParseError> type = parser.parse("type1:" + entry.getKey()); assertTrue(type.isOk()); assertEquals(entry.getValue(), type.unwrap().get(new VarTerm("type1"))); } diff --git a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorConstraintsTest.java b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorConstraintsTest.java index 3fc0fcd..57e6e25 100644 --- a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorConstraintsTest.java +++ b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorConstraintsTest.java @@ -25,7 +25,7 @@ class LatexCreatorConstraintsTest { @Test void singleVarDefaultConstraintTest() { - typeInferer = model.getTypeInferer("x", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("x", "").unwrap(); List actual = new LatexCreatorConstraints(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getEverything(); String constraintSet = AMPERSAND + SPLIT_BEGIN + CONSTRAINT_SET + EQUALS + LATEX_CURLY_LEFT + AMPERSAND @@ -55,7 +55,7 @@ class LatexCreatorConstraintsTest { @Test void singleAbsDefaultConstraintTest() { - typeInferer = model.getTypeInferer("λx.y", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("λx.y", "").unwrap(); List actual = new LatexCreatorConstraints(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getEverything(); String constraintSet1 = AMPERSAND + SPLIT_BEGIN + CONSTRAINT_SET + EQUALS + LATEX_CURLY_LEFT + AMPERSAND @@ -103,12 +103,12 @@ class LatexCreatorConstraintsTest { @Test void lineBreak() { - typeInferer = model.getTypeInferer("a b c d e f", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("a b c d e f", "").unwrap(); List actual = new LatexCreatorConstraints(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getEverything(); assertEquals("\\begin{aligned}&\\begin{split}C=\\{&\\alpha_{2}=\\alpha_{3} \\rightarrow \\alpha_{1},\\alpha_{4}=\\alpha_{5} \\rightarrow \\alpha_{2},\\alpha_{6}=\\alpha_{7} \\rightarrow \\alpha_{4},\\alpha_{8}=\\alpha_{9} \\rightarrow \\alpha_{6},\\alpha_{10}=\\alpha_{11} \\rightarrow \\alpha_{8},\\alpha_{10}=\\beta_{1},\\alpha_{11}=\\beta_{2},\\alpha_{9}=\\beta_{3},\\alpha_{7}=\\beta_{4},\\alpha_{5}=\\beta_{5},\\\\&\\alpha_{3}=\\beta_{6}\\}\\end{split}\\end{aligned}", actual.get(11)); - typeInferer = model.getTypeInferer("let g = a b c d e f in g", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("let g = a b c d e f in g", "").unwrap(); actual = new LatexCreatorConstraints(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getEverything(); assertEquals("\\begin{aligned}&\\begin{split}C=\\{&\\alpha_{1}=\\alpha_{13}\\}\\end{split}\\\\\n" + @@ -117,7 +117,7 @@ class LatexCreatorConstraintsTest { @Test void emptyLetTypeAssumptions() { - typeInferer = model.getTypeInferer("let g = 5 in g", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("let g = 5 in g", "").unwrap(); List actual = new LatexCreatorConstraints(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getEverything(); assertEquals("\\begin{aligned}&\\begin{split}C=\\{&\\alpha_{1}=\\alpha_{3}\\}\\end{split}\\\\\n" + "&\\begin{split}C_{let}=\\{&\\alpha_{2}=\\texttt{int}\\}\\end{split}\\\\&\\begin{split}\\sigma_{let}:=\\textit{mgu}(C_{let})=[&\\alpha_{2}\\mathrel{\\unicode{x21E8}}\\texttt{int}]\\end{split}\\\\&\\sigma_{let}(\\alpha_{2})=\\texttt{int}\\\\&\\Gamma'=\\sigma_{let}(\\emptyset),\\ \\texttt{g}:ta(\\sigma_{let}(\\alpha_{2}),\\sigma_{let}(\\emptyset))=\\texttt{g}:\\texttt{int}\\end{aligned}", actual.get(6)); @@ -126,7 +126,7 @@ class LatexCreatorConstraintsTest { @Test void excessiveMemoryUsageAvoided() { try { - typeInferer = model.getTypeInferer("(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)(λx.x)", "").unwrap(); List actual = new LatexCreatorConstraints(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getEverything(); // should have thrown IllegalStateException by now fail("did not throw exception"); diff --git a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTermTest.java b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTermTest.java index e8b33b4..af57852 100644 --- a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTermTest.java +++ b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTermTest.java @@ -17,7 +17,7 @@ class LatexCreatorTermTest { @Test void absTest() { - typeInferer = model.getTypeInferer("λx.y", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("λx.y", "").unwrap(); assertEquals(LAMBDA + SPACE + MONO_TEXT + "{x}" + DOT_SIGN + LATEX_SPACE + MONO_TEXT + "{y}", new LatexCreatorTerm(typeInferer.getFirstInferenceStep().getConclusion().getLambdaTerm(), @@ -26,7 +26,7 @@ class LatexCreatorTermTest { @Test void appTest() { - typeInferer = model.getTypeInferer("x y", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("x y", "").unwrap(); assertEquals(MONO_TEXT + "{x}" + LATEX_SPACE + MONO_TEXT + "{y}", new LatexCreatorTerm(typeInferer.getFirstInferenceStep().getConclusion().getLambdaTerm(), LatexCreatorMode.NORMAL).getLatex()); @@ -34,7 +34,7 @@ class LatexCreatorTermTest { @Test void varTest() { - typeInferer = model.getTypeInferer("x", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("x", "").unwrap(); assertEquals(MONO_TEXT + "{x}", new LatexCreatorTerm(typeInferer.getFirstInferenceStep().getConclusion().getLambdaTerm(), LatexCreatorMode.NORMAL).getLatex()); @@ -42,7 +42,7 @@ class LatexCreatorTermTest { @Test void integerTest() { - typeInferer = model.getTypeInferer("5", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("5", "").unwrap(); assertEquals(MONO_TEXT + "{5}", new LatexCreatorTerm(typeInferer.getFirstInferenceStep().getConclusion().getLambdaTerm(), LatexCreatorMode.NORMAL).getLatex()); @@ -50,7 +50,7 @@ class LatexCreatorTermTest { @Test void booleanTest() { - typeInferer = model.getTypeInferer("true", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("true", "").unwrap(); assertEquals(MONO_TEXT + "{true}", new LatexCreatorTerm(typeInferer.getFirstInferenceStep().getConclusion().getLambdaTerm(), LatexCreatorMode.NORMAL).getLatex()); @@ -58,7 +58,7 @@ class LatexCreatorTermTest { @Test void letTest() { - typeInferer = model.getTypeInferer("let f = x in f", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("let f = x in f", "").unwrap(); assertEquals(MONO_TEXT + CURLY_LEFT + BOLD_TEXT + "{let}" + CURLY_RIGHT + LATEX_SPACE + MONO_TEXT + "{f}" + EQUALS + MONO_TEXT + "{x}" + LATEX_SPACE + MONO_TEXT + CURLY_LEFT + BOLD_TEXT + "{in}" + CURLY_RIGHT + LATEX_SPACE + MONO_TEXT + "{f}", @@ -68,7 +68,7 @@ class LatexCreatorTermTest { @Test void appWithParenthesesTest() { - typeInferer = model.getTypeInferer("x (y z)", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("x (y z)", "").unwrap(); assertEquals(MONO_TEXT + "{x}" + LATEX_SPACE + PAREN_LEFT + MONO_TEXT + "{y}" + LATEX_SPACE + MONO_TEXT + "{z}" + PAREN_RIGHT, new LatexCreatorTerm(typeInferer.getFirstInferenceStep().getConclusion().getLambdaTerm(), diff --git a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTest.java b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTest.java index c1a7f8b..3701066 100644 --- a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTest.java +++ b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTest.java @@ -14,7 +14,7 @@ class LatexCreatorTest { @Test void testFailedLet() { - TypeInfererInterface typeInferer = model.getTypeInferer("let fun = 5 5 in fun 42", new HashMap<>()).unwrap(); + TypeInfererInterface typeInferer = model.getTypeInferer("let fun = 5 5 in fun 42", "").unwrap(); String latex = new LatexCreator(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getTree(); assertEquals("\\begin{prooftree}\n" + "\\AxiomC{$\\texttt{5} \\in Const$}\n" + @@ -29,7 +29,7 @@ class LatexCreatorTest { "\\LeftLabel{\\textrm L{\\small ET}}\n" + "\\BinaryInfC{$\\vdash\\texttt{\\textbf{let}}\\ \\texttt{fun}=\\texttt{5}\\ \\texttt{5}\\ \\texttt{\\textbf{in}}\\ \\texttt{fun}\\ \\texttt{42}:\\alpha_{1}$}\n" + "\\end{prooftree}", latex); - typeInferer = model.getTypeInferer("(let fun = 5 5 in fun) 42", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("(let fun = 5 5 in fun) 42", "").unwrap(); latex = new LatexCreator(typeInferer, Enum::toString, LatexCreatorMode.NORMAL).getTree(); assertEquals("\\begin{prooftree}\n" + "\\AxiomC{$\\texttt{5} \\in Const$}\n" + diff --git a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTypeTest.java b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTypeTest.java index 70badf1..ddf8520 100644 --- a/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTypeTest.java +++ b/src/test/java/edu/kit/typicalc/view/content/typeinferencecontent/latexcreator/LatexCreatorTypeTest.java @@ -17,30 +17,28 @@ class LatexCreatorTypeTest { @Test void identityTest() { - typeInferer = model.getTypeInferer("λx.x", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("λx.x", "").unwrap(); assertEquals(TREE_VARIABLE + "_{2} " + RIGHT_ARROW + SPACE + TREE_VARIABLE + "_{2}", new LatexCreatorType(typeInferer.getType().get(), LatexCreatorMode.NORMAL).getLatex()); } @Test void generatedTypeTest() { - typeInferer = model.getTypeInferer("x", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("x", "").unwrap(); assertEquals(GENERATED_ASSUMPTION_VARIABLE + "_{1}", new LatexCreatorType(typeInferer.getType().get(), LatexCreatorMode.NORMAL).getLatex()); } @Test void namedTypeTest() { - typeInferer = model.getTypeInferer("5", new HashMap<>()).unwrap(); + typeInferer = model.getTypeInferer("5", "").unwrap(); assertEquals(MONO_TEXT + "{int}", new LatexCreatorType(typeInferer.getType().get(), LatexCreatorMode.NORMAL).getLatex()); } @Test void userVariableTest() { - HashMap map = new HashMap<>(); - map.put("x", "t1"); - typeInferer = model.getTypeInferer("x", map).unwrap(); + typeInferer = model.getTypeInferer("x", "x: t1").unwrap(); assertEquals(USER_VARIABLE + "_{1}", new LatexCreatorType(typeInferer.getType().get(), LatexCreatorMode.NORMAL).getLatex()); }