Use print instead of warn
This commit is contained in:
parent
385e01f613
commit
b3f6143e08
@ -801,7 +801,7 @@ pub fn resolveTypeOfNodeInternal(
|
||||
|
||||
const import_str = handle.tree.tokenSlice(import_param.cast(ast.Node.StringLiteral).?.token);
|
||||
const new_handle = (store.resolveImport(handle, import_str[1 .. import_str.len - 1]) catch |err| block: {
|
||||
std.debug.warn("Error {} while processing import {}\n", .{ err, import_str });
|
||||
std.debug.print("Error {} while processing import {}\n", .{ err, import_str });
|
||||
return null;
|
||||
}) orelse return null;
|
||||
|
||||
@ -836,7 +836,7 @@ pub fn resolveTypeOfNodeInternal(
|
||||
.type = .{ .data = .{ .other = node }, .is_type_val = false },
|
||||
.handle = handle,
|
||||
},
|
||||
else => {}, //std.debug.warn("Type resolution case not implemented; {}\n", .{node.id}),
|
||||
else => {}, //std.debug.print("Type resolution case not implemented; {}\n", .{node.id}),
|
||||
}
|
||||
return null;
|
||||
}
|
||||
@ -1031,7 +1031,7 @@ pub fn getFieldAccessType(
|
||||
current_type = (try resolveUnwrapOptionalType(store, arena, current_type, &bound_type_params)) orelse return null;
|
||||
},
|
||||
else => {
|
||||
std.debug.warn("Unrecognized token {} after period.\n", .{after_period.id});
|
||||
std.debug.print("Unrecognized token {} after period.\n", .{after_period.id});
|
||||
return null;
|
||||
},
|
||||
}
|
||||
@ -1082,7 +1082,7 @@ pub fn getFieldAccessType(
|
||||
current_type = (try resolveBracketAccessType(store, arena, current_type, if (is_range) .Range else .Single, &bound_type_params)) orelse return null;
|
||||
},
|
||||
else => {
|
||||
std.debug.warn("Unimplemented token: {}\n", .{tok.id});
|
||||
std.debug.print("Unimplemented token: {}\n", .{tok.id});
|
||||
return null;
|
||||
},
|
||||
}
|
||||
@ -1126,7 +1126,7 @@ pub fn nodeToString(tree: *ast.Tree, node: *ast.Node) ?[]const u8 {
|
||||
}
|
||||
},
|
||||
else => {
|
||||
std.debug.warn("INVALID: {}\n", .{node.id});
|
||||
std.debug.print("INVALID: {}\n", .{node.id});
|
||||
},
|
||||
}
|
||||
|
||||
@ -1349,7 +1349,7 @@ fn getDocumentSymbolsInternal(allocator: *std.mem.Allocator, tree: *ast.Tree, no
|
||||
};
|
||||
|
||||
if (getDeclName(tree, node) == null) {
|
||||
std.debug.warn("NULL NAME: {}\n", .{node.id});
|
||||
std.debug.print("NULL NAME: {}\n", .{node.id});
|
||||
}
|
||||
|
||||
const maybe_name = if (getDeclName(tree, node)) |name|
|
||||
@ -1848,7 +1848,7 @@ pub const DocumentScope = struct {
|
||||
|
||||
pub fn debugPrint(self: DocumentScope) void {
|
||||
for (self.scopes) |scope| {
|
||||
std.debug.warn(
|
||||
std.debug.print(
|
||||
\\--------------------------
|
||||
\\Scope {}, range: [{}, {})
|
||||
\\ {} usingnamespaces
|
||||
@ -1863,10 +1863,10 @@ pub const DocumentScope = struct {
|
||||
var decl_it = scope.decls.iterator();
|
||||
var idx: usize = 0;
|
||||
while (decl_it.next()) |name_decl| : (idx += 1) {
|
||||
if (idx != 0) std.debug.warn(", ", .{});
|
||||
std.debug.warn("{}", .{name_decl.key});
|
||||
if (idx != 0) std.debug.print(", ", .{});
|
||||
std.debug.print("{}", .{name_decl.key});
|
||||
}
|
||||
std.debug.warn("\n--------------------------\n", .{});
|
||||
std.debug.print("\n--------------------------\n", .{});
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -108,7 +108,7 @@ fn realloc(allocator: *std.mem.Allocator, old_mem: []u8, old_align: u29, new_siz
|
||||
|
||||
const curr_allocs = self.info.currentlyAllocated();
|
||||
if (self.max_bytes != 0 and curr_allocs >= self.max_bytes) {
|
||||
std.debug.warn("Exceeded maximum bytes {}, exiting.\n", .{self.max_bytes});
|
||||
std.debug.print("Exceeded maximum bytes {}, exiting.\n", .{self.max_bytes});
|
||||
std.process.exit(1);
|
||||
}
|
||||
|
||||
|
@ -152,7 +152,7 @@ fn loadPackages(context: LoadPackagesContext) !void {
|
||||
switch (zig_run_result.term) {
|
||||
.Exited => |exit_code| {
|
||||
if (exit_code == 0) {
|
||||
std.debug.warn("Finished zig run for build file {}\n", .{build_file.uri});
|
||||
std.debug.print("Finished zig run for build file {}\n", .{build_file.uri});
|
||||
|
||||
for (build_file.packages.items) |old_pkg| {
|
||||
allocator.free(old_pkg.name);
|
||||
@ -190,7 +190,7 @@ fn loadPackages(context: LoadPackagesContext) !void {
|
||||
/// This function asserts the document is not open yet and takes ownership
|
||||
/// of the uri and text passed in.
|
||||
fn newDocument(self: *DocumentStore, uri: []const u8, text: []u8) anyerror!*Handle {
|
||||
std.debug.warn("Opened document: {}\n", .{uri});
|
||||
std.debug.print("Opened document: {}\n", .{uri});
|
||||
|
||||
var handle = try self.allocator.create(Handle);
|
||||
errdefer self.allocator.destroy(handle);
|
||||
@ -218,7 +218,7 @@ fn newDocument(self: *DocumentStore, uri: []const u8, text: []u8) anyerror!*Hand
|
||||
// TODO: Better logic for detecting std or subdirectories?
|
||||
const in_std = std.mem.indexOf(u8, uri, "/std/") != null;
|
||||
if (self.zig_exe_path != null and std.mem.endsWith(u8, uri, "/build.zig") and !in_std) {
|
||||
std.debug.warn("Document is a build file, extracting packages...\n", .{});
|
||||
std.debug.print("Document is a build file, extracting packages...\n", .{});
|
||||
// This is a build file.
|
||||
var build_file = try self.allocator.create(BuildFile);
|
||||
errdefer self.allocator.destroy(build_file);
|
||||
@ -240,7 +240,7 @@ fn newDocument(self: *DocumentStore, uri: []const u8, text: []u8) anyerror!*Hand
|
||||
.build_runner_path = self.build_runner_path,
|
||||
.zig_exe_path = self.zig_exe_path.?,
|
||||
}) catch |err| {
|
||||
std.debug.warn("Failed to load packages of build file {} (error: {})\n", .{ build_file.uri, err });
|
||||
std.debug.print("Failed to load packages of build file {} (error: {})\n", .{ build_file.uri, err });
|
||||
};
|
||||
} else if (self.zig_exe_path != null and !in_std) associate_build_file: {
|
||||
// Look into build files to see if we already have one that fits
|
||||
@ -248,7 +248,7 @@ fn newDocument(self: *DocumentStore, uri: []const u8, text: []u8) anyerror!*Hand
|
||||
const build_file_base_uri = build_file.uri[0 .. std.mem.lastIndexOfScalar(u8, build_file.uri, '/').? + 1];
|
||||
|
||||
if (std.mem.startsWith(u8, uri, build_file_base_uri)) {
|
||||
std.debug.warn("Found an associated build file: {}\n", .{build_file.uri});
|
||||
std.debug.print("Found an associated build file: {}\n", .{build_file.uri});
|
||||
build_file.refs += 1;
|
||||
handle.associated_build_file = build_file;
|
||||
break :associate_build_file;
|
||||
@ -292,12 +292,12 @@ fn newDocument(self: *DocumentStore, uri: []const u8, text: []u8) anyerror!*Hand
|
||||
|
||||
pub fn openDocument(self: *DocumentStore, uri: []const u8, text: []const u8) !*Handle {
|
||||
if (self.handles.get(uri)) |entry| {
|
||||
std.debug.warn("Document already open: {}, incrementing count\n", .{uri});
|
||||
std.debug.print("Document already open: {}, incrementing count\n", .{uri});
|
||||
entry.value.count += 1;
|
||||
if (entry.value.is_build_file) |build_file| {
|
||||
build_file.refs += 1;
|
||||
}
|
||||
std.debug.warn("New count: {}\n", .{entry.value.count});
|
||||
std.debug.print("New count: {}\n", .{entry.value.count});
|
||||
return entry.value;
|
||||
}
|
||||
|
||||
@ -312,7 +312,7 @@ pub fn openDocument(self: *DocumentStore, uri: []const u8, text: []const u8) !*H
|
||||
fn decrementBuildFileRefs(self: *DocumentStore, build_file: *BuildFile) void {
|
||||
build_file.refs -= 1;
|
||||
if (build_file.refs == 0) {
|
||||
std.debug.warn("Freeing build file {}\n", .{build_file.uri});
|
||||
std.debug.print("Freeing build file {}\n", .{build_file.uri});
|
||||
for (build_file.packages.items) |pkg| {
|
||||
self.allocator.free(pkg.name);
|
||||
self.allocator.free(pkg.uri);
|
||||
@ -338,7 +338,7 @@ fn decrementCount(self: *DocumentStore, uri: []const u8) void {
|
||||
if (entry.value.count > 0)
|
||||
return;
|
||||
|
||||
std.debug.warn("Freeing document: {}\n", .{uri});
|
||||
std.debug.print("Freeing document: {}\n", .{uri});
|
||||
|
||||
if (entry.value.associated_build_file) |build_file| {
|
||||
self.decrementBuildFileRefs(build_file);
|
||||
@ -378,7 +378,7 @@ pub fn getHandle(self: *DocumentStore, uri: []const u8) ?*Handle {
|
||||
|
||||
// Check if the document text is now sane, move it to sane_text if so.
|
||||
fn refreshDocument(self: *DocumentStore, handle: *Handle, zig_lib_path: ?[]const u8) !void {
|
||||
std.debug.warn("New text for document {}\n", .{handle.uri()});
|
||||
std.debug.print("New text for document {}\n", .{handle.uri()});
|
||||
handle.tree.deinit();
|
||||
handle.tree = try std.zig.parse(self.allocator, handle.document.text);
|
||||
|
||||
@ -424,7 +424,7 @@ fn refreshDocument(self: *DocumentStore, handle: *Handle, zig_lib_path: ?[]const
|
||||
while (idx < still_exist.len) : (idx += 1) {
|
||||
if (still_exist[idx]) continue;
|
||||
|
||||
std.debug.warn("Import removed: {}\n", .{handle.import_uris.items[idx - offset]});
|
||||
std.debug.print("Import removed: {}\n", .{handle.import_uris.items[idx - offset]});
|
||||
const uri = handle.import_uris.orderedRemove(idx - offset);
|
||||
offset += 1;
|
||||
|
||||
@ -441,7 +441,7 @@ pub fn applySave(self: *DocumentStore, handle: *Handle) !void {
|
||||
.build_runner_path = self.build_runner_path,
|
||||
.zig_exe_path = self.zig_exe_path.?,
|
||||
}) catch |err| {
|
||||
std.debug.warn("Failed to load packages of build file {} (error: {})\n", .{ build_file.uri, err });
|
||||
std.debug.print("Failed to load packages of build file {} (error: {})\n", .{ build_file.uri, err });
|
||||
};
|
||||
}
|
||||
}
|
||||
@ -515,7 +515,7 @@ pub fn uriFromImportStr(
|
||||
) !?[]const u8 {
|
||||
if (std.mem.eql(u8, import_str, "std")) {
|
||||
if (self.std_uri) |uri| return try std.mem.dupe(allocator, u8, uri) else {
|
||||
std.debug.warn("Cannot resolve std library import, path is null.\n", .{});
|
||||
std.debug.print("Cannot resolve std library import, path is null.\n", .{});
|
||||
return null;
|
||||
}
|
||||
} else if (std.mem.eql(u8, import_str, "builtin")) {
|
||||
@ -553,7 +553,7 @@ pub fn resolveImport(self: *DocumentStore, handle: *Handle, import_str: []const
|
||||
import_str,
|
||||
)) orelse return null;
|
||||
|
||||
// std.debug.warn("Import final URI: {}\n", .{final_uri});
|
||||
// std.debug.print("Import final URI: {}\n", .{final_uri});
|
||||
var consumed_final_uri = false;
|
||||
defer if (!consumed_final_uri) allocator.free(final_uri);
|
||||
|
||||
@ -582,7 +582,7 @@ pub fn resolveImport(self: *DocumentStore, handle: *Handle, import_str: []const
|
||||
defer allocator.free(file_path);
|
||||
|
||||
var file = std.fs.cwd().openFile(file_path, .{}) catch {
|
||||
std.debug.warn("Cannot open import file {}\n", .{file_path});
|
||||
std.debug.print("Cannot open import file {}\n", .{file_path});
|
||||
return null;
|
||||
};
|
||||
|
||||
@ -594,7 +594,7 @@ pub fn resolveImport(self: *DocumentStore, handle: *Handle, import_str: []const
|
||||
errdefer allocator.free(file_contents);
|
||||
|
||||
file.inStream().readNoEof(file_contents) catch {
|
||||
std.debug.warn("Could not read from file {}\n", .{file_path});
|
||||
std.debug.print("Could not read from file {}\n", .{file_path});
|
||||
return null;
|
||||
};
|
||||
|
||||
@ -615,7 +615,7 @@ fn stdUriFromLibPath(allocator: *std.mem.Allocator, zig_lib_path: ?[]const u8) !
|
||||
const std_path = std.fs.path.resolve(allocator, &[_][]const u8{
|
||||
zpath, "./std/std.zig",
|
||||
}) catch |err| block: {
|
||||
std.debug.warn("Failed to resolve zig std library path, error: {}\n", .{err});
|
||||
std.debug.print("Failed to resolve zig std library path, error: {}\n", .{err});
|
||||
return null;
|
||||
};
|
||||
|
||||
|
60
src/main.zig
60
src/main.zig
@ -845,18 +845,18 @@ fn loadConfig(folder_path: []const u8) ?Config {
|
||||
defer conf_file.close();
|
||||
|
||||
// Max 1MB
|
||||
const file_buf = conf_file.inStream().readAllAlloc(allocator, 0x1000000) catch return null;
|
||||
const file_buf = conf_file.reader().readAllAlloc(allocator, 0x1000000) catch return null;
|
||||
defer allocator.free(file_buf);
|
||||
|
||||
// TODO: Better errors? Doesn't seem like std.json can provide us positions or context.
|
||||
var config = std.json.parse(Config, &std.json.TokenStream.init(file_buf), std.json.ParseOptions{ .allocator = allocator }) catch |err| {
|
||||
std.debug.warn("Error while parsing configuration file: {}\nUsing default config.\n", .{err});
|
||||
std.debug.print("Error while parsing configuration file: {}\nUsing default config.\n", .{err});
|
||||
return null;
|
||||
};
|
||||
|
||||
if (config.zig_lib_path) |zig_lib_path| {
|
||||
if (!std.fs.path.isAbsolute(zig_lib_path)) {
|
||||
std.debug.warn("zig library path is not absolute, defaulting to null.\n", .{});
|
||||
std.debug.print("zig library path is not absolute, defaulting to null.\n", .{});
|
||||
allocator.free(zig_lib_path);
|
||||
config.zig_lib_path = null;
|
||||
}
|
||||
@ -908,7 +908,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const start_time = std.time.milliTimestamp();
|
||||
defer {
|
||||
const end_time = std.time.milliTimestamp();
|
||||
std.debug.warn("Took {}ms to process method {}\n", .{ end_time - start_time, method });
|
||||
std.debug.print("Took {}ms to process method {}\n", .{ end_time - start_time, method });
|
||||
}
|
||||
|
||||
// Core
|
||||
@ -955,11 +955,11 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
if (params.getValue("workspaceFolders")) |workspace_folders| {
|
||||
switch (workspace_folders) {
|
||||
.Array => |folders| {
|
||||
std.debug.warn("Got workspace folders in initialization.\n", .{});
|
||||
std.debug.print("Got workspace folders in initialization.\n", .{});
|
||||
|
||||
for (folders.items) |workspace_folder| {
|
||||
const folder_uri = workspace_folder.Object.getValue("uri").?.String;
|
||||
std.debug.warn("Loaded folder {}\n", .{folder_uri});
|
||||
std.debug.print("Loaded folder {}\n", .{folder_uri});
|
||||
const duped_uri = try std.mem.dupe(allocator, u8, folder_uri);
|
||||
try workspace_folder_configs.putNoClobber(duped_uri, null);
|
||||
}
|
||||
@ -969,7 +969,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
}
|
||||
}
|
||||
|
||||
std.debug.warn("{}\n", .{client_capabilities});
|
||||
std.debug.print("{}\n", .{client_capabilities});
|
||||
try respondGeneric(id, initialize_response);
|
||||
} else if (std.mem.eql(u8, method, "shutdown")) {
|
||||
keep_running.* = false;
|
||||
@ -1025,7 +1025,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const content_changes = params.getValue("contentChanges").?.Array;
|
||||
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to change non existent document {}", .{uri});
|
||||
std.debug.print("Trying to change non existent document {}", .{uri});
|
||||
return;
|
||||
};
|
||||
|
||||
@ -1037,7 +1037,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const text_document = params.getValue("textDocument").?.Object;
|
||||
const uri = text_document.getValue("uri").?.String;
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to save non existent document {}", .{uri});
|
||||
std.debug.print("Trying to save non existent document {}", .{uri});
|
||||
return;
|
||||
};
|
||||
|
||||
@ -1060,7 +1060,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const this_config = configFromUriOr(uri, config);
|
||||
if (this_config.enable_semantic_tokens) {
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to complete in non existent document {}", .{uri});
|
||||
std.debug.print("Trying to complete in non existent document {}", .{uri});
|
||||
return try respondGeneric(id, no_semantic_tokens_response);
|
||||
};
|
||||
|
||||
@ -1083,7 +1083,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const position = params.getValue("position").?.Object;
|
||||
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to complete in non existent document {}", .{uri});
|
||||
std.debug.print("Trying to complete in non existent document {}", .{uri});
|
||||
return try respondGeneric(id, no_completions_response);
|
||||
};
|
||||
|
||||
@ -1149,7 +1149,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const position = params.getValue("position").?.Object;
|
||||
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to got to definition in non existent document {}", .{uri});
|
||||
std.debug.print("Trying to got to definition in non existent document {}", .{uri});
|
||||
return try respondGeneric(id, null_result_response);
|
||||
};
|
||||
|
||||
@ -1179,7 +1179,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const position = params.getValue("position").?.Object;
|
||||
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to got to definition in non existent document {}", .{uri});
|
||||
std.debug.print("Trying to got to definition in non existent document {}", .{uri});
|
||||
return try respondGeneric(id, null_result_response);
|
||||
};
|
||||
|
||||
@ -1206,7 +1206,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const uri = document.getValue("uri").?.String;
|
||||
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to got to definition in non existent document {}", .{uri});
|
||||
std.debug.print("Trying to got to definition in non existent document {}", .{uri});
|
||||
return try respondGeneric(id, null_result_response);
|
||||
};
|
||||
|
||||
@ -1218,7 +1218,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
const uri = document.getValue("uri").?.String;
|
||||
|
||||
const handle = document_store.getHandle(uri) orelse {
|
||||
std.debug.warn("Trying to got to definition in non existent document {}", .{uri});
|
||||
std.debug.print("Trying to got to definition in non existent document {}", .{uri});
|
||||
return try respondGeneric(id, null_result_response);
|
||||
};
|
||||
|
||||
@ -1228,7 +1228,7 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
process.stdout_behavior = .Pipe;
|
||||
|
||||
process.spawn() catch |err| {
|
||||
std.debug.warn("Failed to spawn zig fmt process, error: {}\n", .{err});
|
||||
std.debug.print("Failed to spawn zig fmt process, error: {}\n", .{err});
|
||||
return try respondGeneric(id, null_result_response);
|
||||
};
|
||||
try process.stdin.?.writeAll(handle.document.text);
|
||||
@ -1271,10 +1271,10 @@ fn processJsonRpc(parser: *std.json.Parser, json: []const u8, config: Config, ke
|
||||
// TODO: Unimplemented methods, implement them and add them to server capabilities.
|
||||
try respondGeneric(id, null_result_response);
|
||||
} else if (root.Object.getValue("id")) |_| {
|
||||
std.debug.warn("Method with return value not implemented: {}", .{method});
|
||||
std.debug.print("Method with return value not implemented: {}", .{method});
|
||||
try respondGeneric(id, not_implemented_response);
|
||||
} else {
|
||||
std.debug.warn("Method without return value not implemented: {}", .{method});
|
||||
std.debug.print("Method without return value not implemented: {}", .{method});
|
||||
}
|
||||
}
|
||||
|
||||
@ -1296,12 +1296,12 @@ pub fn main() anyerror!void {
|
||||
}
|
||||
|
||||
defer if (debug_alloc) |dbg| {
|
||||
std.debug.warn("Finished cleanup, last allocation info.\n", .{});
|
||||
std.debug.warn("{}\n", .{dbg.info});
|
||||
std.debug.print("Finished cleanup, last allocation info.\n", .{});
|
||||
std.debug.print("{}\n", .{dbg.info});
|
||||
};
|
||||
|
||||
// Init global vars
|
||||
const in_stream = std.io.getStdIn().inStream();
|
||||
const reader = std.io.getStdIn().reader();
|
||||
stdout = std.io.bufferedOutStream(std.io.getStdOut().outStream());
|
||||
|
||||
// Read the configuration, if any.
|
||||
@ -1339,12 +1339,12 @@ pub fn main() anyerror!void {
|
||||
break :find_zig;
|
||||
}
|
||||
|
||||
std.debug.warn("zig path `{}` is not absolute, will look in path\n", .{exe_path});
|
||||
std.debug.print("zig path `{}` is not absolute, will look in path\n", .{exe_path});
|
||||
}
|
||||
|
||||
const env_path = std.process.getEnvVarOwned(allocator, "PATH") catch |err| switch (err) {
|
||||
error.EnvironmentVariableNotFound => {
|
||||
std.debug.warn("Could not get PATH.\n", .{});
|
||||
std.debug.print("Could not get PATH.\n", .{});
|
||||
break :find_zig;
|
||||
},
|
||||
else => return err,
|
||||
@ -1365,21 +1365,21 @@ pub fn main() anyerror!void {
|
||||
|
||||
var buf: [std.fs.MAX_PATH_BYTES]u8 = undefined;
|
||||
zig_exe_path = std.os.realpath(full_path, &buf) catch continue;
|
||||
std.debug.warn("Found zig in PATH: {}\n", .{zig_exe_path});
|
||||
std.debug.print("Found zig in PATH: {}\n", .{zig_exe_path});
|
||||
break :find_zig;
|
||||
}
|
||||
}
|
||||
|
||||
if (zig_exe_path) |exe_path| {
|
||||
config.zig_exe_path = exe_path;
|
||||
std.debug.warn("Using zig executable {}\n", .{exe_path});
|
||||
std.debug.print("Using zig executable {}\n", .{exe_path});
|
||||
if (config.zig_lib_path == null) {
|
||||
// Set the lib path relative to the executable path.
|
||||
config.zig_lib_path = try std.fs.path.resolve(allocator, &[_][]const u8{
|
||||
std.fs.path.dirname(exe_path).?, "./lib/zig",
|
||||
});
|
||||
|
||||
std.debug.warn("Resolved standard library from executable: {}\n", .{config.zig_lib_path});
|
||||
std.debug.print("Resolved standard library from executable: {}\n", .{config.zig_lib_path});
|
||||
}
|
||||
}
|
||||
|
||||
@ -1404,19 +1404,19 @@ pub fn main() anyerror!void {
|
||||
|
||||
var keep_running = true;
|
||||
while (keep_running) {
|
||||
const headers = readRequestHeader(allocator, in_stream) catch |err| {
|
||||
std.debug.warn("{}; exiting!", .{@errorName(err)});
|
||||
const headers = readRequestHeader(allocator, reader) catch |err| {
|
||||
std.debug.print("{}; exiting!", .{@errorName(err)});
|
||||
return;
|
||||
};
|
||||
defer headers.deinit(allocator);
|
||||
const buf = try allocator.alloc(u8, headers.content_length);
|
||||
defer allocator.free(buf);
|
||||
try in_stream.readNoEof(buf);
|
||||
try reader.readNoEof(buf);
|
||||
try processJsonRpc(&json_parser, buf, config, &keep_running);
|
||||
json_parser.reset();
|
||||
|
||||
if (debug_alloc) |dbg| {
|
||||
std.debug.warn("{}\n", .{dbg.info});
|
||||
std.debug.print("{}\n", .{dbg.info});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -119,6 +119,8 @@ const GapHighlighter = struct {
|
||||
const tok_id = self.builder.handle.tree.token_ids[tok];
|
||||
if (tok_id == .LineComment) {
|
||||
try writeToken(self.builder, tok, .comment);
|
||||
} else if (tok_id == .ContainerDocComment) {
|
||||
try writeTokenMod(self.builder, tok, .comment, TokenModifiers{ .documentation = true });
|
||||
} else if (@enumToInt(tok_id) >= @enumToInt(std.zig.Token.Id.Keyword_align) and
|
||||
@enumToInt(tok_id) <= @enumToInt(std.zig.Token.Id.Keyword_while))
|
||||
{
|
||||
|
Loading…
Reference in New Issue
Block a user