2020-07-03 00:33:10 +01:00
|
|
|
const std = @import("std");
|
|
|
|
const types = @import("types.zig");
|
2021-03-05 21:38:42 +00:00
|
|
|
const ast = std.zig.ast;
|
2020-07-03 00:33:10 +01:00
|
|
|
|
|
|
|
pub const Encoding = enum {
|
|
|
|
utf8,
|
|
|
|
utf16,
|
|
|
|
};
|
|
|
|
|
|
|
|
pub const DocumentPosition = struct {
|
|
|
|
line: []const u8,
|
|
|
|
line_index: usize,
|
|
|
|
absolute_index: usize,
|
|
|
|
};
|
|
|
|
|
|
|
|
pub fn documentPosition(doc: types.TextDocument, position: types.Position, encoding: Encoding) !DocumentPosition {
|
|
|
|
var split_iterator = std.mem.split(doc.text, "\n");
|
|
|
|
|
|
|
|
var line_idx: i64 = 0;
|
|
|
|
var line: []const u8 = "";
|
|
|
|
while (line_idx < position.line) : (line_idx += 1) {
|
|
|
|
line = split_iterator.next() orelse return error.InvalidParams;
|
|
|
|
}
|
|
|
|
|
|
|
|
const line_start_idx = split_iterator.index.?;
|
|
|
|
line = split_iterator.next() orelse return error.InvalidParams;
|
|
|
|
|
|
|
|
if (encoding == .utf8) {
|
|
|
|
const index = @intCast(i64, line_start_idx) + position.character;
|
|
|
|
if (index < 0 or index > @intCast(i64, doc.text.len)) {
|
|
|
|
return error.InvalidParams;
|
|
|
|
}
|
|
|
|
return DocumentPosition{ .line = line, .absolute_index = @intCast(usize, index), .line_index = @intCast(usize, position.character) };
|
|
|
|
} else {
|
|
|
|
const utf8 = doc.text[line_start_idx..];
|
|
|
|
var utf8_idx: usize = 0;
|
|
|
|
var utf16_idx: usize = 0;
|
|
|
|
while (utf16_idx < position.character) {
|
|
|
|
if (utf8_idx > utf8.len) {
|
|
|
|
return error.InvalidParams;
|
|
|
|
}
|
|
|
|
|
|
|
|
const n = try std.unicode.utf8ByteSequenceLength(utf8[utf8_idx]);
|
|
|
|
const next_utf8_idx = utf8_idx + n;
|
|
|
|
const codepoint = try std.unicode.utf8Decode(utf8[utf8_idx..next_utf8_idx]);
|
|
|
|
if (codepoint < 0x10000) {
|
|
|
|
utf16_idx += 1;
|
|
|
|
} else {
|
|
|
|
utf16_idx += 2;
|
|
|
|
}
|
|
|
|
utf8_idx = next_utf8_idx;
|
|
|
|
}
|
|
|
|
return DocumentPosition{ .line = line, .absolute_index = line_start_idx + utf8_idx, .line_index = utf8_idx };
|
|
|
|
}
|
|
|
|
}
|
2020-07-03 09:34:42 +01:00
|
|
|
|
|
|
|
pub const TokenLocation = struct {
|
|
|
|
line: usize,
|
|
|
|
column: usize,
|
2020-07-07 09:57:02 +01:00
|
|
|
offset: usize,
|
|
|
|
|
|
|
|
pub fn add(lhs: TokenLocation, rhs: TokenLocation) TokenLocation {
|
|
|
|
return .{
|
|
|
|
.line = lhs.line + rhs.line,
|
|
|
|
.column = if (rhs.line == 0)
|
|
|
|
lhs.column + rhs.column
|
|
|
|
else
|
|
|
|
rhs.column,
|
|
|
|
.offset = rhs.offset,
|
|
|
|
};
|
|
|
|
}
|
2020-07-03 09:34:42 +01:00
|
|
|
};
|
|
|
|
|
2021-03-05 21:38:42 +00:00
|
|
|
pub fn tokenRelativeLocation(tree: ast.Tree, start_index: usize, token: ast.TokenIndex, encoding: Encoding) !TokenLocation {
|
2021-03-01 21:18:38 +00:00
|
|
|
const start = tree.tokens.items(.start)[token];
|
2020-07-03 09:34:42 +01:00
|
|
|
|
|
|
|
var loc = TokenLocation{
|
|
|
|
.line = 0,
|
|
|
|
.column = 0,
|
2020-07-07 09:57:02 +01:00
|
|
|
.offset = 0,
|
2020-07-03 09:34:42 +01:00
|
|
|
};
|
2021-03-01 21:18:38 +00:00
|
|
|
const token_start = start;
|
2020-07-03 09:34:42 +01:00
|
|
|
const source = tree.source[start_index..];
|
|
|
|
var i: usize = 0;
|
2020-09-25 04:59:45 +01:00
|
|
|
while (i + start_index < token_start) {
|
2020-07-03 09:34:42 +01:00
|
|
|
const c = source[i];
|
|
|
|
if (c == '\n') {
|
|
|
|
loc.line += 1;
|
|
|
|
loc.column = 0;
|
|
|
|
i += 1;
|
|
|
|
} else {
|
|
|
|
if (encoding == .utf16) {
|
|
|
|
const n = try std.unicode.utf8ByteSequenceLength(c);
|
2020-07-03 10:24:24 +01:00
|
|
|
const codepoint = try std.unicode.utf8Decode(source[i .. i + n]);
|
2020-07-03 09:34:42 +01:00
|
|
|
if (codepoint < 0x10000) {
|
|
|
|
loc.column += 1;
|
|
|
|
} else {
|
|
|
|
loc.column += 2;
|
|
|
|
}
|
|
|
|
i += n;
|
|
|
|
} else {
|
|
|
|
loc.column += 1;
|
|
|
|
i += 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2020-07-07 09:57:02 +01:00
|
|
|
loc.offset = i + start_index;
|
2020-07-03 09:34:42 +01:00
|
|
|
return loc;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Asserts the token is comprised of valid utf8
|
2021-03-05 21:38:42 +00:00
|
|
|
pub fn tokenLength(tree: ast.Tree, token: ast.TokenIndex, encoding: Encoding) usize {
|
|
|
|
const token_loc = tokenLocation(tree, token);
|
2020-07-03 09:34:42 +01:00
|
|
|
if (encoding == .utf8)
|
2021-03-05 21:38:42 +00:00
|
|
|
return token_loc.end - token_loc.start;
|
2020-07-03 09:34:42 +01:00
|
|
|
|
2021-03-05 21:38:42 +00:00
|
|
|
var i: usize = token_loc.start;
|
2020-07-03 09:34:42 +01:00
|
|
|
var utf16_len: usize = 0;
|
2021-03-05 21:38:42 +00:00
|
|
|
while (i < token_loc.end) {
|
2020-07-03 09:34:42 +01:00
|
|
|
const n = std.unicode.utf8ByteSequenceLength(tree.source[i]) catch unreachable;
|
2020-07-03 10:24:24 +01:00
|
|
|
const codepoint = std.unicode.utf8Decode(tree.source[i .. i + n]) catch unreachable;
|
2020-07-03 09:34:42 +01:00
|
|
|
if (codepoint < 0x10000) {
|
|
|
|
utf16_len += 1;
|
|
|
|
} else {
|
|
|
|
utf16_len += 2;
|
|
|
|
}
|
|
|
|
i += n;
|
|
|
|
}
|
|
|
|
return utf16_len;
|
|
|
|
}
|
2020-07-03 10:39:58 +01:00
|
|
|
|
2021-03-05 21:38:42 +00:00
|
|
|
/// Token location inside source
|
|
|
|
pub const Loc = struct {
|
|
|
|
start: usize,
|
|
|
|
end: usize,
|
|
|
|
};
|
|
|
|
|
|
|
|
pub fn tokenLocation(tree: ast.Tree, token_index: ast.TokenIndex) Loc {
|
|
|
|
const start = tree.tokens.items(.start)[token_index];
|
|
|
|
const tag = tree.tokens.items(.tag)[token_index];
|
|
|
|
|
|
|
|
// For some tokens, re-tokenization is needed to find the end.
|
|
|
|
var tokenizer: std.zig.Tokenizer = .{
|
|
|
|
.buffer = tree.source,
|
|
|
|
.index = start,
|
|
|
|
.pending_invalid_token = null,
|
|
|
|
};
|
|
|
|
|
|
|
|
const token = tokenizer.next();
|
|
|
|
std.debug.assert(token.tag == tag);
|
|
|
|
return .{ .start = token.loc.start, .end = token.loc.end };
|
|
|
|
}
|
|
|
|
|
2020-07-03 10:39:58 +01:00
|
|
|
pub fn documentRange(doc: types.TextDocument, encoding: Encoding) !types.Range {
|
|
|
|
var line_idx: i64 = 0;
|
|
|
|
var curr_line: []const u8 = doc.text;
|
|
|
|
|
|
|
|
var split_iterator = std.mem.split(doc.text, "\n");
|
|
|
|
while (split_iterator.next()) |line| : (line_idx += 1) {
|
|
|
|
curr_line = line;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (encoding == .utf8) {
|
|
|
|
return types.Range{
|
|
|
|
.start = .{
|
|
|
|
.line = 0,
|
|
|
|
.character = 0,
|
|
|
|
},
|
|
|
|
.end = .{
|
|
|
|
.line = line_idx,
|
|
|
|
.character = @intCast(i64, curr_line.len),
|
|
|
|
},
|
|
|
|
};
|
|
|
|
} else {
|
|
|
|
var utf16_len: usize = 0;
|
|
|
|
var line_utf8_idx: usize = 0;
|
|
|
|
while (line_utf8_idx < curr_line.len) {
|
|
|
|
const n = try std.unicode.utf8ByteSequenceLength(curr_line[line_utf8_idx]);
|
|
|
|
const codepoint = try std.unicode.utf8Decode(curr_line[line_utf8_idx .. line_utf8_idx + n]);
|
|
|
|
if (codepoint < 0x10000) {
|
|
|
|
utf16_len += 1;
|
|
|
|
} else {
|
|
|
|
utf16_len += 2;
|
|
|
|
}
|
|
|
|
line_utf8_idx += n;
|
|
|
|
}
|
|
|
|
return types.Range{
|
|
|
|
.start = .{
|
|
|
|
.line = 0,
|
|
|
|
.character = 0,
|
|
|
|
},
|
|
|
|
.end = .{
|
|
|
|
.line = line_idx,
|
|
|
|
.character = @intCast(i64, utf16_len),
|
|
|
|
},
|
|
|
|
};
|
|
|
|
}
|
|
|
|
}
|