Merge pull request #40 from Vexu/err

Implement global error set and enum literal completion
This commit is contained in:
Alexandros Naskos 2020-05-28 19:04:06 +03:00 committed by GitHub
commit bc8ecd28f4
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
3 changed files with 108 additions and 9 deletions

View File

@ -413,6 +413,15 @@ pub fn resolveTypeOfNode(analysis_ctx: *AnalysisContext, node: *ast.Node) ?*ast.
else => decl, else => decl,
}; };
}, },
.ErrorSetDecl => {
const set = node.cast(ast.Node.ErrorSetDecl).?;
var i: usize = 0;
while (set.iterate(i)) |decl| : (i += 1) {
// TODO handle errors better?
analysis_ctx.error_completions.add(analysis_ctx.tree(), decl) catch {};
}
return node;
},
.SuffixOp => { .SuffixOp => {
const suffix_op = node.cast(ast.Node.SuffixOp).?; const suffix_op = node.cast(ast.Node.SuffixOp).?;
switch (suffix_op.op) { switch (suffix_op.op) {
@ -505,9 +514,23 @@ pub fn resolveTypeOfNode(analysis_ctx: *AnalysisContext, node: *ast.Node) ?*ast.
}, },
.ContainerDecl => { .ContainerDecl => {
analysis_ctx.onContainer(node.cast(ast.Node.ContainerDecl).?) catch return null; analysis_ctx.onContainer(node.cast(ast.Node.ContainerDecl).?) catch return null;
const container = node.cast(ast.Node.ContainerDecl).?;
const kind = analysis_ctx.tree().token_ids[container.kind_token];
if (kind == .Keyword_struct or (kind == .Keyword_union and container.init_arg_expr == .None)) {
return node;
}
var i: usize = 0;
while (container.iterate(i)) |decl| : (i += 1) {
if (decl.id != .ContainerField) continue;
// TODO handle errors better?
analysis_ctx.enum_completions.add(analysis_ctx.tree(), decl) catch {};
}
return node; return node;
}, },
.MultilineStringLiteral, .StringLiteral, .ErrorSetDecl, .FnProto => return node, .MultilineStringLiteral, .StringLiteral, .FnProto => return node,
else => std.debug.warn("Type resolution case not implemented; {}\n", .{node.id}), else => std.debug.warn("Type resolution case not implemented; {}\n", .{node.id}),
} }
return null; return null;
@ -924,6 +947,7 @@ pub const PositionContext = union(enum) {
string_literal: SourceRange, string_literal: SourceRange,
field_access: SourceRange, field_access: SourceRange,
var_access: SourceRange, var_access: SourceRange,
global_error_set,
enum_literal, enum_literal,
other, other,
empty, empty,
@ -938,6 +962,7 @@ pub const PositionContext = union(enum) {
.enum_literal => null, .enum_literal => null,
.other => null, .other => null,
.empty => null, .empty => null,
.global_error_set => null,
}; };
} }
}; };
@ -1010,6 +1035,7 @@ pub fn documentPositionContext(allocator: *std.mem.Allocator, document: types.Te
.enum_literal => curr_ctx.ctx = .empty, .enum_literal => curr_ctx.ctx = .empty,
.field_access => {}, .field_access => {},
.other => {}, .other => {},
.global_error_set => {},
else => curr_ctx.ctx = .{ else => curr_ctx.ctx = .{
.field_access = tokenRangeAppend(curr_ctx.ctx.range().?, tok), .field_access = tokenRangeAppend(curr_ctx.ctx.range().?, tok),
}, },
@ -1032,6 +1058,7 @@ pub fn documentPositionContext(allocator: *std.mem.Allocator, document: types.Te
(try peek(&stack)).ctx = .empty; (try peek(&stack)).ctx = .empty;
} }
}, },
.Keyword_error => curr_ctx.ctx = .global_error_set,
else => curr_ctx.ctx = .empty, else => curr_ctx.ctx = .empty,
} }
@ -1065,7 +1092,7 @@ fn addOutlineNodes(allocator: *std.mem.Allocator, children: *std.ArrayList(types
try addOutlineNodes(allocator, children, tree, cchild); try addOutlineNodes(allocator, children, tree, cchild);
return; return;
}, },
else => {} else => {},
} }
std.debug.warn("{}\n", .{child.id}); std.debug.warn("{}\n", .{child.id});
_ = try children.append(try getDocumentSymbolsInternal(allocator, tree, child)); _ = try children.append(try getDocumentSymbolsInternal(allocator, tree, child));
@ -1073,7 +1100,6 @@ fn addOutlineNodes(allocator: *std.mem.Allocator, children: *std.ArrayList(types
fn getDocumentSymbolsInternal(allocator: *std.mem.Allocator, tree: *ast.Tree, node: *ast.Node) anyerror!types.DocumentSymbol { fn getDocumentSymbolsInternal(allocator: *std.mem.Allocator, tree: *ast.Tree, node: *ast.Node) anyerror!types.DocumentSymbol {
// const symbols = std.ArrayList(types.DocumentSymbol).init(allocator); // const symbols = std.ArrayList(types.DocumentSymbol).init(allocator);
const start_loc = tree.tokenLocation(0, node.firstToken()); const start_loc = tree.tokenLocation(0, node.firstToken());
const end_loc = tree.tokenLocation(0, node.lastToken()); const end_loc = tree.tokenLocation(0, node.lastToken());
const range = types.Range{ const range = types.Range{
@ -1084,7 +1110,7 @@ fn getDocumentSymbolsInternal(allocator: *std.mem.Allocator, tree: *ast.Tree, no
.end = .{ .end = .{
.line = @intCast(i64, end_loc.line), .line = @intCast(i64, end_loc.line),
.character = @intCast(i64, end_loc.column), .character = @intCast(i64, end_loc.column),
} },
}; };
if (getDeclName(tree, node) == null) { if (getDeclName(tree, node) == null) {
@ -1100,7 +1126,7 @@ fn getDocumentSymbolsInternal(allocator: *std.mem.Allocator, tree: *ast.Tree, no
.FnProto => .Function, .FnProto => .Function,
.VarDecl => .Variable, .VarDecl => .Variable,
.ContainerField => .Field, .ContainerField => .Field,
else => .Variable else => .Variable,
}, },
.range = range, .range = range,
.selectionRange = range, .selectionRange = range,

View File

@ -30,12 +30,57 @@ pub const Handle = struct {
} }
}; };
pub const TagStore = struct {
values: std.StringHashMap(void),
completions: std.ArrayListUnmanaged(types.CompletionItem),
pub fn init(allocator: *std.mem.Allocator) TagStore {
return .{
.values = std.StringHashMap(void).init(allocator),
.completions = .{},
};
}
pub fn deinit(self: *TagStore) void {
const alloc = self.values.allocator;
for (self.completions.items) |item| {
alloc.free(item.label);
if (item.documentation) |some| alloc.free(some.value);
}
self.values.deinit();
self.completions.deinit(self.values.allocator);
}
pub fn add(self: *TagStore, tree: *std.zig.ast.Tree, tag: *std.zig.ast.Node) !void {
const name = analysis.nodeToString(tree, tag).?;
if (self.values.contains(name)) return;
const alloc = self.values.allocator;
const item = types.CompletionItem{
.label = try std.mem.dupe(alloc, u8, name),
.kind = .Constant,
.documentation = if (try analysis.getDocComments(alloc, tree, tag)) |docs|
.{
.kind = .Markdown,
.value = docs,
}
else
null,
};
try self.values.putNoClobber(item.label, {});
try self.completions.append(self.values.allocator, item);
}
};
allocator: *std.mem.Allocator, allocator: *std.mem.Allocator,
handles: std.StringHashMap(*Handle), handles: std.StringHashMap(*Handle),
has_zig: bool, has_zig: bool,
build_files: std.ArrayListUnmanaged(*BuildFile), build_files: std.ArrayListUnmanaged(*BuildFile),
build_runner_path: []const u8, build_runner_path: []const u8,
error_completions: TagStore,
enum_completions: TagStore,
pub fn init( pub fn init(
self: *DocumentStore, self: *DocumentStore,
allocator: *std.mem.Allocator, allocator: *std.mem.Allocator,
@ -47,6 +92,8 @@ pub fn init(
self.has_zig = has_zig; self.has_zig = has_zig;
self.build_files = .{}; self.build_files = .{};
self.build_runner_path = build_runner_path; self.build_runner_path = build_runner_path;
self.error_completions = TagStore.init(allocator);
self.enum_completions = TagStore.init(allocator);
} }
const LoadPackagesContext = struct { const LoadPackagesContext = struct {
@ -478,6 +525,8 @@ pub const AnalysisContext = struct {
scope_nodes: []*std.zig.ast.Node, scope_nodes: []*std.zig.ast.Node,
in_container: *std.zig.ast.Node, in_container: *std.zig.ast.Node,
std_uri: ?[]const u8, std_uri: ?[]const u8,
error_completions: *TagStore,
enum_completions: *TagStore,
pub fn tree(self: AnalysisContext) *std.zig.ast.Tree { pub fn tree(self: AnalysisContext) *std.zig.ast.Tree {
return self.handle.tree; return self.handle.tree;
@ -584,6 +633,8 @@ pub const AnalysisContext = struct {
.scope_nodes = try std.mem.dupe(&self.arena.allocator, *std.zig.ast.Node, self.scope_nodes), .scope_nodes = try std.mem.dupe(&self.arena.allocator, *std.zig.ast.Node, self.scope_nodes),
.in_container = self.in_container, .in_container = self.in_container,
.std_uri = self.std_uri, .std_uri = self.std_uri,
.error_completions = self.error_completions,
.enum_completions = self.enum_completions,
}; };
} }
}; };
@ -623,6 +674,8 @@ pub fn analysisContext(
.scope_nodes = scope_nodes.items, .scope_nodes = scope_nodes.items,
.in_container = in_container, .in_container = in_container,
.std_uri = std_uri, .std_uri = std_uri,
.error_completions = &self.error_completions,
.enum_completions = &self.enum_completions,
}; };
} }
@ -652,4 +705,6 @@ pub fn deinit(self: *DocumentStore) void {
} }
self.build_files.deinit(self.allocator); self.build_files.deinit(self.allocator);
self.error_completions.deinit();
self.enum_completions.deinit();
} }

View File

@ -760,6 +760,24 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config) !v
}), }),
.var_access, .empty => try completeGlobal(id, pos_index, handle, this_config), .var_access, .empty => try completeGlobal(id, pos_index, handle, this_config),
.field_access => |range| try completeFieldAccess(id, handle, pos, range, this_config), .field_access => |range| try completeFieldAccess(id, handle, pos, range, this_config),
.global_error_set => try send(types.Response{
.id = .{ .Integer = id },
.result = .{
.CompletionList = .{
.isIncomplete = false,
.items = document_store.error_completions.completions.items,
},
},
}),
.enum_literal => try send(types.Response{
.id = .{ .Integer = id },
.result = .{
.CompletionList = .{
.isIncomplete = false,
.items = document_store.enum_completions.completions.items,
},
},
}),
else => try respondGeneric(id, no_completions_response), else => try respondGeneric(id, no_completions_response),
} }
} else { } else {