Removed comment preprocessing code from the semantic tokens gap highlighter,
Fixed various comment semantic highlighting issues. Refactored some minor parts of the analysis code.
This commit is contained in:
		
							parent
							
								
									2a57789be8
								
							
						
					
					
						commit
						06518778c1
					
				@ -228,7 +228,6 @@ pub fn getDeclNameToken(tree: ast.Tree, node: ast.Node.Index) ?ast.TokenIndex {
 | 
			
		||||
        .global_var_decl => tree.globalVarDecl(node).ast.mut_token + 1,
 | 
			
		||||
        .simple_var_decl => tree.simpleVarDecl(node).ast.mut_token + 1,
 | 
			
		||||
        .aligned_var_decl => tree.alignedVarDecl(node).ast.mut_token + 1,
 | 
			
		||||
 | 
			
		||||
        // function declaration names
 | 
			
		||||
        .fn_proto,
 | 
			
		||||
        .fn_proto_multi,
 | 
			
		||||
@ -324,8 +323,8 @@ fn resolveVarDeclAliasInternal(
 | 
			
		||||
                .other => |n| n,
 | 
			
		||||
                else => return null,
 | 
			
		||||
            };
 | 
			
		||||
 | 
			
		||||
            if (resolved_node >= node_tags.len or !isContainer(node_tags[resolved_node])) return null;
 | 
			
		||||
            const resolved_tree_tags = resolved.handle.tree.nodes.items(.tag);
 | 
			
		||||
            if (!isContainer(resolved.handle.tree, resolved_node)) return null;
 | 
			
		||||
            break :block NodeWithHandle{ .node = resolved_node, .handle = resolved.handle };
 | 
			
		||||
        } else return null;
 | 
			
		||||
 | 
			
		||||
@ -1117,9 +1116,9 @@ pub const TypeWithHandle = struct {
 | 
			
		||||
        const tree = self.handle.tree;
 | 
			
		||||
        const node = self.type.data.other;
 | 
			
		||||
        const tags = tree.nodes.items(.tag);
 | 
			
		||||
        if (isContainer(tags[node])) {
 | 
			
		||||
        if (isContainer(tree, node)) {
 | 
			
		||||
            var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
            for (declMembers(tree, tags[node], node, &buf)) |child| {
 | 
			
		||||
            for (declMembers(tree, node, &buf)) |child| {
 | 
			
		||||
                if (tags[child].isContainerField()) return false;
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
@ -1478,8 +1477,8 @@ pub fn fnProto(tree: ast.Tree, node: ast.Node.Index, buf: *[1]ast.Node.Index) ?a
 | 
			
		||||
pub fn getImportStr(tree: ast.Tree, node: ast.Node.Index, source_index: usize) ?[]const u8 {
 | 
			
		||||
    const node_tags = tree.nodes.items(.tag);
 | 
			
		||||
    var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
    if (isContainer(node_tags[node])) {
 | 
			
		||||
        const decls = declMembers(tree, node_tags[node], node, &buf);
 | 
			
		||||
    if (isContainer(tree, node)) {
 | 
			
		||||
        const decls = declMembers(tree, node, &buf);
 | 
			
		||||
        for (decls) |decl_idx| {
 | 
			
		||||
            if (getImportStr(tree, decl_idx, source_index)) |name| {
 | 
			
		||||
                return name;
 | 
			
		||||
@ -1806,7 +1805,7 @@ fn addOutlineNodes(allocator: *std.mem.Allocator, tree: ast.Tree, child: ast.Nod
 | 
			
		||||
        .tagged_union_two_trailing,
 | 
			
		||||
        => {
 | 
			
		||||
            var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
            for (declMembers(tree, tree.nodes.items(.tag)[child], child, &buf)) |member|
 | 
			
		||||
            for (declMembers(tree, child, &buf)) |member|
 | 
			
		||||
                try addOutlineNodes(allocator, tree, member, context);
 | 
			
		||||
            return;
 | 
			
		||||
        },
 | 
			
		||||
@ -1894,9 +1893,9 @@ fn getDocumentSymbolsInternal(allocator: *std.mem.Allocator, tree: ast.Tree, nod
 | 
			
		||||
                .encoding = context.encoding,
 | 
			
		||||
            };
 | 
			
		||||
 | 
			
		||||
            if (isContainer(tags[node])) {
 | 
			
		||||
            if (isContainer(tree, node)) {
 | 
			
		||||
                var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
                for (declMembers(tree, tags[node], node, &buf)) |child|
 | 
			
		||||
                for (declMembers(tree, node, &buf)) |child|
 | 
			
		||||
                    try addOutlineNodes(allocator, tree, child, &child_context);
 | 
			
		||||
            }
 | 
			
		||||
 | 
			
		||||
@ -2091,7 +2090,7 @@ fn findContainerScope(container_handle: NodeWithHandle) ?*Scope {
 | 
			
		||||
    const container = container_handle.node;
 | 
			
		||||
    const handle = container_handle.handle;
 | 
			
		||||
 | 
			
		||||
    if (!isContainer(handle.tree.nodes.items(.tag)[container])) return null;
 | 
			
		||||
    if (!isContainer(handle.tree, container)) return null;
 | 
			
		||||
 | 
			
		||||
    // Find the container scope.
 | 
			
		||||
    return for (handle.document_scope.scopes) |*scope| {
 | 
			
		||||
@ -2534,8 +2533,8 @@ fn nodeSourceRange(tree: ast.Tree, node: ast.Node.Index) SourceRange {
 | 
			
		||||
    };
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub fn isContainer(tag: ast.Node.Tag) bool {
 | 
			
		||||
    return switch (tag) {
 | 
			
		||||
pub fn isContainer(tree: ast.Tree, node: ast.Node.Index) bool {
 | 
			
		||||
    return switch (tree.nodes.items(.tag)[node]) {
 | 
			
		||||
        .container_decl,
 | 
			
		||||
        .container_decl_trailing,
 | 
			
		||||
        .container_decl_arg,
 | 
			
		||||
@ -2557,9 +2556,9 @@ pub fn isContainer(tag: ast.Node.Tag) bool {
 | 
			
		||||
 | 
			
		||||
/// Returns the member indices of a given declaration container.
 | 
			
		||||
/// Asserts given `tag` is a container node
 | 
			
		||||
pub fn declMembers(tree: ast.Tree, tag: ast.Node.Tag, node_idx: ast.Node.Index, buffer: *[2]ast.Node.Index) []const ast.Node.Index {
 | 
			
		||||
    std.debug.assert(isContainer(tag));
 | 
			
		||||
    return switch (tag) {
 | 
			
		||||
pub fn declMembers(tree: ast.Tree, node_idx: ast.Node.Index, buffer: *[2]ast.Node.Index) []const ast.Node.Index {
 | 
			
		||||
    std.debug.assert(isContainer(tree, node_idx));
 | 
			
		||||
    return switch (tree.nodes.items(.tag)[node_idx]) {
 | 
			
		||||
        .container_decl, .container_decl_trailing => tree.containerDecl(node_idx).ast.members,
 | 
			
		||||
        .container_decl_arg, .container_decl_arg_trailing => tree.containerDeclArg(node_idx).ast.members,
 | 
			
		||||
        .container_decl_two, .container_decl_two_trailing => tree.containerDeclTwo(buffer, node_idx).ast.members,
 | 
			
		||||
@ -2598,11 +2597,11 @@ fn makeScopeInternal(
 | 
			
		||||
    const token_tags = tree.tokens.items(.tag);
 | 
			
		||||
    const data = tree.nodes.items(.data);
 | 
			
		||||
    const main_tokens = tree.nodes.items(.main_token);
 | 
			
		||||
    const node = tags[node_idx];
 | 
			
		||||
    const node_tag = tags[node_idx];
 | 
			
		||||
 | 
			
		||||
    if (isContainer(node)) {
 | 
			
		||||
    if (isContainer(tree, node_idx)) {
 | 
			
		||||
        var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
        const ast_decls = declMembers(tree, node, node_idx, &buf);
 | 
			
		||||
        const ast_decls = declMembers(tree,node_idx, &buf);
 | 
			
		||||
 | 
			
		||||
        (try scopes.addOne(allocator)).* = .{
 | 
			
		||||
            .range = nodeSourceRange(tree, node_idx),
 | 
			
		||||
@ -2635,7 +2634,7 @@ fn makeScopeInternal(
 | 
			
		||||
                continue;
 | 
			
		||||
            }
 | 
			
		||||
 | 
			
		||||
            if (node == .error_set_decl) {
 | 
			
		||||
            if (node_tag == .error_set_decl) {
 | 
			
		||||
                (try error_completions.addOne(allocator)).* = .{
 | 
			
		||||
                    .label = name,
 | 
			
		||||
                    .kind = .Constant,
 | 
			
		||||
@ -2655,11 +2654,11 @@ fn makeScopeInternal(
 | 
			
		||||
 | 
			
		||||
            if (container_field) |field| {
 | 
			
		||||
                const empty_field = field.ast.type_expr == 0 and field.ast.value_expr == 0;
 | 
			
		||||
                if (empty_field and node == .root) {
 | 
			
		||||
                if (empty_field and node_tag == .root) {
 | 
			
		||||
                    continue;
 | 
			
		||||
                }
 | 
			
		||||
 | 
			
		||||
                const container_decl: ?ast.full.ContainerDecl = switch (node) {
 | 
			
		||||
                const container_decl: ?ast.full.ContainerDecl = switch (node_tag) {
 | 
			
		||||
                    .container_decl, .container_decl_trailing => tree.containerDecl(node_idx),
 | 
			
		||||
                    .container_decl_arg, .container_decl_arg_trailing => tree.containerDeclArg(node_idx),
 | 
			
		||||
                    .container_decl_two, .container_decl_two_trailing => blk: {
 | 
			
		||||
@ -2704,7 +2703,7 @@ fn makeScopeInternal(
 | 
			
		||||
        return;
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    switch (node) {
 | 
			
		||||
    switch (node_tag) {
 | 
			
		||||
        .fn_proto,
 | 
			
		||||
        .fn_proto_one,
 | 
			
		||||
        .fn_proto_simple,
 | 
			
		||||
@ -2782,7 +2781,7 @@ fn makeScopeInternal(
 | 
			
		||||
                uses.deinit();
 | 
			
		||||
            }
 | 
			
		||||
 | 
			
		||||
            const statements: []const ast.Node.Index = switch (node) {
 | 
			
		||||
            const statements: []const ast.Node.Index = switch (node_tag) {
 | 
			
		||||
                .block, .block_semicolon => tree.extra_data[data[node_idx].lhs..data[node_idx].rhs],
 | 
			
		||||
                .block_two, .block_two_semicolon => blk: {
 | 
			
		||||
                    const statements = &[_]ast.Node.Index{ data[node_idx].lhs, data[node_idx].rhs };
 | 
			
		||||
@ -2819,7 +2818,7 @@ fn makeScopeInternal(
 | 
			
		||||
        .@"if",
 | 
			
		||||
        .if_simple,
 | 
			
		||||
        => {
 | 
			
		||||
            const if_node: ast.full.If = if (node == .@"if")
 | 
			
		||||
            const if_node: ast.full.If = if (node_tag == .@"if")
 | 
			
		||||
                tree.ifFull(node_idx)
 | 
			
		||||
            else
 | 
			
		||||
                tree.ifSimple(node_idx);
 | 
			
		||||
@ -2879,8 +2878,8 @@ fn makeScopeInternal(
 | 
			
		||||
        .while_cont,
 | 
			
		||||
        .@"for",
 | 
			
		||||
        .for_simple,
 | 
			
		||||
        => |tag| {
 | 
			
		||||
            const while_node: ast.full.While = switch (node) {
 | 
			
		||||
        => {
 | 
			
		||||
            const while_node: ast.full.While = switch (node_tag) {
 | 
			
		||||
                .@"while" => tree.whileFull(node_idx),
 | 
			
		||||
                .while_simple => tree.whileSimple(node_idx),
 | 
			
		||||
                .while_cont => tree.whileCont(node_idx),
 | 
			
		||||
@ -2889,7 +2888,7 @@ fn makeScopeInternal(
 | 
			
		||||
                else => unreachable,
 | 
			
		||||
            };
 | 
			
		||||
 | 
			
		||||
            const is_for = tag == .@"for" or tag == .for_simple;
 | 
			
		||||
            const is_for = node_tag == .@"for" or node_tag == .for_simple;
 | 
			
		||||
 | 
			
		||||
            if (while_node.label_token) |label| {
 | 
			
		||||
                std.debug.assert(token_tags[label] == .identifier);
 | 
			
		||||
@ -3043,7 +3042,7 @@ fn makeScopeInternal(
 | 
			
		||||
        .async_call_one_comma,
 | 
			
		||||
        => {
 | 
			
		||||
            var buf: [1]ast.Node.Index = undefined;
 | 
			
		||||
            const call: ast.full.Call = switch (node) {
 | 
			
		||||
            const call: ast.full.Call = switch (node_tag) {
 | 
			
		||||
                .async_call,
 | 
			
		||||
                .async_call_comma,
 | 
			
		||||
                .call,
 | 
			
		||||
@ -3071,7 +3070,7 @@ fn makeScopeInternal(
 | 
			
		||||
        .struct_init_one_comma,
 | 
			
		||||
        => {
 | 
			
		||||
            var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
            const struct_init: ast.full.StructInit = switch (node) {
 | 
			
		||||
            const struct_init: ast.full.StructInit = switch (node_tag) {
 | 
			
		||||
                .struct_init, .struct_init_comma => tree.structInit(node_idx),
 | 
			
		||||
                .struct_init_dot, .struct_init_dot_comma => tree.structInitDot(node_idx),
 | 
			
		||||
                .struct_init_dot_two, .struct_init_dot_two_comma => tree.structInitDotTwo(&buf, node_idx),
 | 
			
		||||
@ -3096,7 +3095,7 @@ fn makeScopeInternal(
 | 
			
		||||
        .array_init_one_comma,
 | 
			
		||||
        => {
 | 
			
		||||
            var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
            const array_init: ast.full.ArrayInit = switch (node) {
 | 
			
		||||
            const array_init: ast.full.ArrayInit = switch (node_tag) {
 | 
			
		||||
                .array_init, .array_init_comma => tree.arrayInit(node_idx),
 | 
			
		||||
                .array_init_dot, .array_init_dot_comma => tree.arrayInitDot(node_idx),
 | 
			
		||||
                .array_init_dot_two, .array_init_dot_two_comma => tree.arrayInitDotTwo(&buf, node_idx),
 | 
			
		||||
@ -3129,7 +3128,7 @@ fn makeScopeInternal(
 | 
			
		||||
        .builtin_call_two_comma,
 | 
			
		||||
        => {
 | 
			
		||||
            const b_data = data[node_idx];
 | 
			
		||||
            const params = switch (node) {
 | 
			
		||||
            const params = switch (node_tag) {
 | 
			
		||||
                .builtin_call, .builtin_call_comma => tree.extra_data[b_data.lhs..b_data.rhs],
 | 
			
		||||
                .builtin_call_two, .builtin_call_two_comma => if (b_data.lhs == 0)
 | 
			
		||||
                    &[_]ast.Node.Index{}
 | 
			
		||||
@ -3161,7 +3160,7 @@ fn makeScopeInternal(
 | 
			
		||||
        .slice_open,
 | 
			
		||||
        .slice_sentinel,
 | 
			
		||||
        => {
 | 
			
		||||
            const slice: ast.full.Slice = switch (node) {
 | 
			
		||||
            const slice: ast.full.Slice = switch (node_tag) {
 | 
			
		||||
                .slice => tree.slice(node_idx),
 | 
			
		||||
                .slice_open => tree.sliceOpen(node_idx),
 | 
			
		||||
                .slice_sentinel => tree.sliceSentinel(node_idx),
 | 
			
		||||
 | 
			
		||||
@ -359,7 +359,7 @@ fn nodeToCompletion(
 | 
			
		||||
    else
 | 
			
		||||
        null;
 | 
			
		||||
 | 
			
		||||
    if (analysis.isContainer(node_tags[node])) {
 | 
			
		||||
    if (analysis.isContainer(handle.tree, node)) {
 | 
			
		||||
        const context = DeclToCompletionContext{
 | 
			
		||||
            .completions = list,
 | 
			
		||||
            .config = &config,
 | 
			
		||||
 | 
			
		||||
@ -54,6 +54,30 @@ pub fn documentPosition(doc: types.TextDocument, position: types.Position, encod
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub fn lineSectionLength(tree: ast.Tree, start_index: usize, end_index: usize, encoding: Encoding) !usize {
 | 
			
		||||
    const source = tree.source[start_index..];
 | 
			
		||||
    std.debug.assert(end_index >= start_index and source.len >= end_index - start_index);
 | 
			
		||||
    if (encoding == .utf8) {
 | 
			
		||||
        return end_index - start_index;
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    var result: usize = 0;
 | 
			
		||||
    var i: usize = 0;
 | 
			
		||||
    while (i + start_index < end_index) {
 | 
			
		||||
        std.debug.assert(source[i] != '\n');
 | 
			
		||||
 | 
			
		||||
        const n = try std.unicode.utf8ByteSequenceLength(source[i]);
 | 
			
		||||
        if (i + n >= source.len)
 | 
			
		||||
            return error.CodepointTooLong;
 | 
			
		||||
 | 
			
		||||
        const codepoint = try std.unicode.utf8Decode(source[i .. i + n]);
 | 
			
		||||
 | 
			
		||||
        result += 1 + @as(usize, @boolToInt(codepoint >= 0x10000));
 | 
			
		||||
        i += n;
 | 
			
		||||
    }
 | 
			
		||||
    return result;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub const TokenLocation = struct {
 | 
			
		||||
    line: usize,
 | 
			
		||||
    column: usize,
 | 
			
		||||
@ -71,15 +95,14 @@ pub const TokenLocation = struct {
 | 
			
		||||
    }
 | 
			
		||||
};
 | 
			
		||||
 | 
			
		||||
pub fn tokenRelativeLocation(tree: ast.Tree, start_index: usize, next_token_index: usize, encoding: Encoding) !TokenLocation {
 | 
			
		||||
    const start = next_token_index;
 | 
			
		||||
 | 
			
		||||
pub fn tokenRelativeLocation(tree: ast.Tree, start_index: usize, token_start: usize, encoding: Encoding) !TokenLocation {
 | 
			
		||||
    std.debug.assert(token_start >= start_index);
 | 
			
		||||
    var loc = TokenLocation{
 | 
			
		||||
        .line = 0,
 | 
			
		||||
        .column = 0,
 | 
			
		||||
        .offset = 0,
 | 
			
		||||
    };
 | 
			
		||||
    const token_start = start;
 | 
			
		||||
 | 
			
		||||
    const source = tree.source[start_index..];
 | 
			
		||||
    var i: usize = 0;
 | 
			
		||||
    while (i + start_index < token_start) {
 | 
			
		||||
@ -91,12 +114,11 @@ pub fn tokenRelativeLocation(tree: ast.Tree, start_index: usize, next_token_inde
 | 
			
		||||
        } else {
 | 
			
		||||
            if (encoding == .utf16) {
 | 
			
		||||
                const n = try std.unicode.utf8ByteSequenceLength(c);
 | 
			
		||||
                if (i + n >= source.len)
 | 
			
		||||
                    return error.CodepointTooLong;
 | 
			
		||||
 | 
			
		||||
                const codepoint = try std.unicode.utf8Decode(source[i .. i + n]);
 | 
			
		||||
                if (codepoint < 0x10000) {
 | 
			
		||||
                    loc.column += 1;
 | 
			
		||||
                } else {
 | 
			
		||||
                    loc.column += 2;
 | 
			
		||||
                }
 | 
			
		||||
                loc.column += 1 + @as(usize, @boolToInt(codepoint >= 0x10000));
 | 
			
		||||
                i += n;
 | 
			
		||||
            } else {
 | 
			
		||||
                loc.column += 1;
 | 
			
		||||
 | 
			
		||||
@ -119,7 +119,7 @@ fn symbolReferencesInternal(
 | 
			
		||||
        .error_set_decl,
 | 
			
		||||
        => {
 | 
			
		||||
            var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
            for (analysis.declMembers(tree, node_tags[node], node, &buf)) |member|
 | 
			
		||||
            for (analysis.declMembers(tree, node, &buf)) |member|
 | 
			
		||||
                try symbolReferencesInternal(arena, store, .{ .node = member, .handle = handle }, decl, encoding, context, handler);
 | 
			
		||||
        },
 | 
			
		||||
        .global_var_decl,
 | 
			
		||||
 | 
			
		||||
@ -75,6 +75,36 @@ const Builder = struct {
 | 
			
		||||
        };
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn highlightComment(
 | 
			
		||||
        self: *Builder,
 | 
			
		||||
        prev_end: usize,
 | 
			
		||||
        comment_start: usize,
 | 
			
		||||
        comment_end: usize,
 | 
			
		||||
        token_modifiers: TokenModifiers,
 | 
			
		||||
    ) !void {
 | 
			
		||||
        const comment_delta = offsets.tokenRelativeLocation(
 | 
			
		||||
            self.handle.tree,
 | 
			
		||||
            prev_end,
 | 
			
		||||
            comment_start,
 | 
			
		||||
            self.encoding,
 | 
			
		||||
        ) catch return;
 | 
			
		||||
 | 
			
		||||
        const comment_length = offsets.lineSectionLength(
 | 
			
		||||
            self.handle.tree,
 | 
			
		||||
            comment_start,
 | 
			
		||||
            comment_end,
 | 
			
		||||
            self.encoding,
 | 
			
		||||
        ) catch return;
 | 
			
		||||
 | 
			
		||||
        try self.arr.appendSlice(&.{
 | 
			
		||||
            @truncate(u32, comment_delta.line),
 | 
			
		||||
            @truncate(u32, comment_delta.column),
 | 
			
		||||
            @truncate(u32, comment_length),
 | 
			
		||||
            @enumToInt(TokenType.comment),
 | 
			
		||||
            token_modifiers.toInt(),
 | 
			
		||||
        });
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn add(self: *Builder, token: ast.TokenIndex, token_type: TokenType, token_modifiers: TokenModifiers) !void {
 | 
			
		||||
        const starts = self.handle.tree.tokens.items(.start);
 | 
			
		||||
        var start_idx = if (self.current_token) |current_token|
 | 
			
		||||
@ -85,24 +115,67 @@ const Builder = struct {
 | 
			
		||||
        if (start_idx > starts[token])
 | 
			
		||||
            return;
 | 
			
		||||
 | 
			
		||||
        const delta_loc = while (self.findCommentBetween(start_idx, starts[token])) |comment| {
 | 
			
		||||
            const old_loc = self.handle.tree.tokenLocation(0, self.current_token orelse 0);
 | 
			
		||||
            const comment_delta = offsets.tokenRelativeLocation(self.handle.tree, start_idx, comment.start, self.encoding) catch return;
 | 
			
		||||
        var comments_end: usize = start_idx;
 | 
			
		||||
        var comments_start: usize = start_idx;
 | 
			
		||||
        // Highlight comments in the gap
 | 
			
		||||
        {
 | 
			
		||||
            const source = self.handle.tree.source;
 | 
			
		||||
            var state: enum { none, comment, doc_comment, comment_start } = .none;
 | 
			
		||||
            var prev_byte = source[start_idx];
 | 
			
		||||
            var i: usize = start_idx + 1;
 | 
			
		||||
            while (i < starts[token]) : ({
 | 
			
		||||
                prev_byte = source[i];
 | 
			
		||||
                i += 1;
 | 
			
		||||
            }) {
 | 
			
		||||
                if (prev_byte == '/' and source[i] == '/') {
 | 
			
		||||
                    switch (state) {
 | 
			
		||||
                        .none => {
 | 
			
		||||
                            comments_start = i - 1;
 | 
			
		||||
                            state = .comment_start;
 | 
			
		||||
                        },
 | 
			
		||||
                        .comment_start => state = .doc_comment,
 | 
			
		||||
                        else => {},
 | 
			
		||||
                    }
 | 
			
		||||
                    continue;
 | 
			
		||||
                } else if (prev_byte == '/' and source[i] == '!' and state == .comment_start) {
 | 
			
		||||
                    state = .doc_comment;
 | 
			
		||||
                    continue;
 | 
			
		||||
                }
 | 
			
		||||
 | 
			
		||||
            try self.arr.appendSlice(&[_]u32{
 | 
			
		||||
                @truncate(u32, comment_delta.line),
 | 
			
		||||
                @truncate(u32, comment_delta.column),
 | 
			
		||||
                comment.length,
 | 
			
		||||
                @enumToInt(TokenType.comment),
 | 
			
		||||
                0,
 | 
			
		||||
            });
 | 
			
		||||
                if (source[i] == '\n' and state != .none) {
 | 
			
		||||
                    try self.highlightComment(comments_end, comments_start, i, switch (state) {
 | 
			
		||||
                        .comment, .comment_start => .{},
 | 
			
		||||
                        .doc_comment => .{ .documentation = true },
 | 
			
		||||
                        else => unreachable,
 | 
			
		||||
                    });
 | 
			
		||||
                    comments_end = i;
 | 
			
		||||
                    state = .none;
 | 
			
		||||
                } else if (state == .comment_start) {
 | 
			
		||||
                    state = .comment;
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            if (state != .none) {
 | 
			
		||||
                try self.highlightComment(comments_end, comments_start, i, switch (state) {
 | 
			
		||||
                    .comment, .comment_start => .{},
 | 
			
		||||
                    .doc_comment => .{ .documentation = true },
 | 
			
		||||
                    else => unreachable,
 | 
			
		||||
                });
 | 
			
		||||
                // @@@
 | 
			
		||||
                // comments_end = i;
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
            start_idx = comment.start;
 | 
			
		||||
        } else offsets.tokenRelativeLocation(self.handle.tree, start_idx, starts[token], self.encoding) catch return;
 | 
			
		||||
        std.debug.print("DELTA:\n```\n{s}\n```\n(LEN: {})\n", .{self.handle.tree.source[comments_end..starts[token]], starts[token] - comments_end});
 | 
			
		||||
        const delta = offsets.tokenRelativeLocation(
 | 
			
		||||
            self.handle.tree,
 | 
			
		||||
            comments_start,
 | 
			
		||||
            starts[token],
 | 
			
		||||
            self.encoding,
 | 
			
		||||
        ) catch return;
 | 
			
		||||
 | 
			
		||||
        try self.arr.appendSlice(&[_]u32{
 | 
			
		||||
            @truncate(u32, delta_loc.line),
 | 
			
		||||
            @truncate(u32, delta_loc.column),
 | 
			
		||||
        try self.arr.appendSlice(&.{
 | 
			
		||||
            @truncate(u32, delta.line),
 | 
			
		||||
            @truncate(u32, delta.column),
 | 
			
		||||
            @truncate(u32, offsets.tokenLength(self.handle.tree, token, self.encoding)),
 | 
			
		||||
            @enumToInt(token_type),
 | 
			
		||||
            token_modifiers.toInt(),
 | 
			
		||||
@ -113,15 +186,6 @@ const Builder = struct {
 | 
			
		||||
    fn toOwnedSlice(self: *Builder) []u32 {
 | 
			
		||||
        return self.arr.toOwnedSlice();
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    /// Based on a given start and end index, returns a `Comment` between the positions
 | 
			
		||||
    /// Returns `null` if none was fone
 | 
			
		||||
    fn findCommentBetween(self: Builder, from: u32, to: u32) ?Comment {
 | 
			
		||||
        return for (self.comments.items) |comment| {
 | 
			
		||||
            if (comment.start > from and comment.start < to)
 | 
			
		||||
                break comment;
 | 
			
		||||
        } else null;
 | 
			
		||||
    }
 | 
			
		||||
};
 | 
			
		||||
 | 
			
		||||
fn writeToken(
 | 
			
		||||
@ -710,7 +774,7 @@ fn writeNodeTokens(
 | 
			
		||||
                    .node = struct_init.ast.type_expr,
 | 
			
		||||
                    .handle = handle,
 | 
			
		||||
                })) |struct_type| switch (struct_type.type.data) {
 | 
			
		||||
                    .other => |type_node| if (analysis.isContainer(struct_type.handle.tree.nodes.items(.tag)[type_node]))
 | 
			
		||||
                    .other => |type_node| if (analysis.isContainer(struct_type.handle.tree, type_node))
 | 
			
		||||
                        fieldTokenType(type_node, struct_type.handle)
 | 
			
		||||
                    else
 | 
			
		||||
                        null,
 | 
			
		||||
@ -976,7 +1040,7 @@ fn writeNodeTokens(
 | 
			
		||||
                switch (decl_type.decl.*) {
 | 
			
		||||
                    .ast_node => |decl_node| {
 | 
			
		||||
                        if (decl_type.handle.tree.nodes.items(.tag)[decl_node].isContainerField()) {
 | 
			
		||||
                            const tok_type: ?TokenType = if (analysis.isContainer(lhs_type.handle.tree.nodes.items(.tag)[left_type_node]))
 | 
			
		||||
                            const tok_type: ?TokenType = if (analysis.isContainer(lhs_type.handle.tree, left_type_node))
 | 
			
		||||
                                fieldTokenType(decl_node, lhs_type.handle)
 | 
			
		||||
                            else if (left_type_node == 0)
 | 
			
		||||
                                TokenType.field
 | 
			
		||||
@ -1083,14 +1147,11 @@ fn writeNodeTokens(
 | 
			
		||||
pub fn writeAllSemanticTokens(arena: *std.heap.ArenaAllocator, store: *DocumentStore, handle: *DocumentStore.Handle, encoding: offsets.Encoding) ![]u32 {
 | 
			
		||||
    var builder = Builder.init(arena.child_allocator, handle, encoding);
 | 
			
		||||
 | 
			
		||||
    // as line comments are not nodes, we parse the text then generate the tokens for them
 | 
			
		||||
    try findComments(&builder, handle.tree.source, encoding);
 | 
			
		||||
 | 
			
		||||
    // reverse the ast from the root declarations
 | 
			
		||||
    var gap_highlighter = GapHighlighter.init(&builder, 0);
 | 
			
		||||
 | 
			
		||||
    var buf: [2]ast.Node.Index = undefined;
 | 
			
		||||
    for (analysis.declMembers(handle.tree, .root, 0, &buf)) |child| {
 | 
			
		||||
    for (analysis.declMembers(handle.tree, 0, &buf)) |child| {
 | 
			
		||||
        try gap_highlighter.next(child);
 | 
			
		||||
        try writeNodeTokens(&builder, arena, store, child);
 | 
			
		||||
    }
 | 
			
		||||
@ -1099,54 +1160,3 @@ pub fn writeAllSemanticTokens(arena: *std.heap.ArenaAllocator, store: *DocumentS
 | 
			
		||||
 | 
			
		||||
    return builder.toOwnedSlice();
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/// As the AST does not contain nodes for comments
 | 
			
		||||
/// this will parse through the entire file to search for comments
 | 
			
		||||
/// and generate semantic tokens for them
 | 
			
		||||
fn findComments(builder: *Builder, source: []const u8, encoding: offsets.Encoding) !void {
 | 
			
		||||
    var state: enum { none, comment, doc_comment } = .none;
 | 
			
		||||
 | 
			
		||||
    var prev: u8 = 0;
 | 
			
		||||
    var start: usize = 0;
 | 
			
		||||
    for (source) |c, i| {
 | 
			
		||||
        if (state == .comment and c == '/') {
 | 
			
		||||
            state = .none;
 | 
			
		||||
            continue;
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        if (state == .none and c == '/' and prev == '/') {
 | 
			
		||||
            state = .comment;
 | 
			
		||||
            start = i - 1;
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        if (c == '\n') {
 | 
			
		||||
            if (state == .comment) {
 | 
			
		||||
                state = .none;
 | 
			
		||||
 | 
			
		||||
                const len = if (encoding == .utf8)
 | 
			
		||||
                    i - start
 | 
			
		||||
                else blk: {
 | 
			
		||||
                    var index: usize = start;
 | 
			
		||||
                    var utf16_len: usize = 0;
 | 
			
		||||
                    while (index < i) {
 | 
			
		||||
                        const n = std.unicode.utf8ByteSequenceLength(source[index]) catch unreachable;
 | 
			
		||||
                        const codepoint = std.unicode.utf8Decode(source[index .. index + n]) catch unreachable;
 | 
			
		||||
                        if (codepoint < 0x10000) {
 | 
			
		||||
                            utf16_len += 1;
 | 
			
		||||
                        } else {
 | 
			
		||||
                            utf16_len += 2;
 | 
			
		||||
                        }
 | 
			
		||||
                        index += n;
 | 
			
		||||
                    }
 | 
			
		||||
                    break :blk utf16_len;
 | 
			
		||||
                };
 | 
			
		||||
 | 
			
		||||
                try builder.comments.append(.{
 | 
			
		||||
                    .length = @truncate(u32, len),
 | 
			
		||||
                    .start = @truncate(u32, start),
 | 
			
		||||
                });
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        prev = c;
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
		Loading…
	
		Reference in New Issue
	
	Block a user