feat: load project file list from git if available
If the loading project has a git repository we now load the file list from git instead of scanning the directory tree. This gives us automatic .gitignore support.
This commit is contained in:
parent
1a0d4ca7b2
commit
ba5dc35913
2 changed files with 105 additions and 50 deletions
120
src/Project.zig
120
src/Project.zig
|
@ -11,23 +11,35 @@ const git = @import("git");
|
||||||
const builtin = @import("builtin");
|
const builtin = @import("builtin");
|
||||||
|
|
||||||
const LSP = @import("LSP.zig");
|
const LSP = @import("LSP.zig");
|
||||||
|
const walk_tree = @import("walk_tree.zig");
|
||||||
|
|
||||||
allocator: std.mem.Allocator,
|
allocator: std.mem.Allocator,
|
||||||
name: []const u8,
|
name: []const u8,
|
||||||
files: std.ArrayList(File),
|
files: std.ArrayListUnmanaged(File) = .empty,
|
||||||
pending: std.ArrayList(File),
|
pending: std.ArrayListUnmanaged(File) = .empty,
|
||||||
longest_file_path: usize = 0,
|
longest_file_path: usize = 0,
|
||||||
open_time: i64,
|
open_time: i64,
|
||||||
language_servers: std.StringHashMap(LSP),
|
language_servers: std.StringHashMap(LSP),
|
||||||
file_language_server: std.StringHashMap(LSP),
|
file_language_server: std.StringHashMap(LSP),
|
||||||
tasks: std.ArrayList(Task),
|
tasks: std.ArrayList(Task),
|
||||||
persistent: bool = false,
|
persistent: bool = false,
|
||||||
|
logger: log.Logger,
|
||||||
logger_lsp: log.Logger,
|
logger_lsp: log.Logger,
|
||||||
logger_git: log.Logger,
|
logger_git: log.Logger,
|
||||||
|
|
||||||
workspace: ?[]const u8 = null,
|
workspace: ?[]const u8 = null,
|
||||||
branch: ?[]const u8 = null,
|
branch: ?[]const u8 = null,
|
||||||
|
|
||||||
|
walker: ?tp.pid = null,
|
||||||
|
|
||||||
|
// async task states
|
||||||
|
state: struct {
|
||||||
|
walk_tree: State = .none,
|
||||||
|
workspace_path: State = .none,
|
||||||
|
current_branch: State = .none,
|
||||||
|
workspace_files: State = .none,
|
||||||
|
} = .{},
|
||||||
|
|
||||||
const Self = @This();
|
const Self = @This();
|
||||||
|
|
||||||
const OutOfMemoryError = error{OutOfMemory};
|
const OutOfMemoryError = error{OutOfMemory};
|
||||||
|
@ -55,22 +67,24 @@ const Task = struct {
|
||||||
mtime: i64,
|
mtime: i64,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
const State = enum { none, running, done, failed };
|
||||||
|
|
||||||
pub fn init(allocator: std.mem.Allocator, name: []const u8) OutOfMemoryError!Self {
|
pub fn init(allocator: std.mem.Allocator, name: []const u8) OutOfMemoryError!Self {
|
||||||
return .{
|
return .{
|
||||||
.allocator = allocator,
|
.allocator = allocator,
|
||||||
.name = try allocator.dupe(u8, name),
|
.name = try allocator.dupe(u8, name),
|
||||||
.files = std.ArrayList(File).init(allocator),
|
|
||||||
.pending = std.ArrayList(File).init(allocator),
|
|
||||||
.open_time = std.time.milliTimestamp(),
|
.open_time = std.time.milliTimestamp(),
|
||||||
.language_servers = std.StringHashMap(LSP).init(allocator),
|
.language_servers = std.StringHashMap(LSP).init(allocator),
|
||||||
.file_language_server = std.StringHashMap(LSP).init(allocator),
|
.file_language_server = std.StringHashMap(LSP).init(allocator),
|
||||||
.tasks = std.ArrayList(Task).init(allocator),
|
.tasks = std.ArrayList(Task).init(allocator),
|
||||||
|
.logger = log.logger("project"),
|
||||||
.logger_lsp = log.logger("lsp"),
|
.logger_lsp = log.logger("lsp"),
|
||||||
.logger_git = log.logger("git"),
|
.logger_git = log.logger("git"),
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn deinit(self: *Self) void {
|
pub fn deinit(self: *Self) void {
|
||||||
|
if (self.walker) |pid| pid.send(.{"stop"}) catch {};
|
||||||
if (self.workspace) |p| self.allocator.free(p);
|
if (self.workspace) |p| self.allocator.free(p);
|
||||||
if (self.branch) |p| self.allocator.free(p);
|
if (self.branch) |p| self.allocator.free(p);
|
||||||
var i_ = self.file_language_server.iterator();
|
var i_ = self.file_language_server.iterator();
|
||||||
|
@ -83,11 +97,13 @@ pub fn deinit(self: *Self) void {
|
||||||
p.value_ptr.*.term();
|
p.value_ptr.*.term();
|
||||||
}
|
}
|
||||||
for (self.files.items) |file| self.allocator.free(file.path);
|
for (self.files.items) |file| self.allocator.free(file.path);
|
||||||
self.files.deinit();
|
self.files.deinit(self.allocator);
|
||||||
|
self.pending.deinit(self.allocator);
|
||||||
for (self.tasks.items) |task| self.allocator.free(task.command);
|
for (self.tasks.items) |task| self.allocator.free(task.command);
|
||||||
self.tasks.deinit();
|
self.tasks.deinit();
|
||||||
self.logger_lsp.deinit();
|
self.logger_lsp.deinit();
|
||||||
self.logger_git.deinit();
|
self.logger_git.deinit();
|
||||||
|
self.logger.deinit();
|
||||||
self.allocator.free(self.name);
|
self.allocator.free(self.name);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -294,11 +310,11 @@ fn make_URI(self: *Self, file_path: ?[]const u8) LspError![]const u8 {
|
||||||
return buf.toOwnedSlice();
|
return buf.toOwnedSlice();
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn sort_files_by_mtime(self: *Self) void {
|
fn sort_files_by_mtime(self: *Self) void {
|
||||||
sort_by_mtime(File, self.files.items);
|
sort_by_mtime(File, self.files.items);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn sort_tasks_by_mtime(self: *Self) void {
|
fn sort_tasks_by_mtime(self: *Self) void {
|
||||||
sort_by_mtime(Task, self.tasks.items);
|
sort_by_mtime(Task, self.tasks.items);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -311,12 +327,14 @@ inline fn sort_by_mtime(T: type, items: []T) void {
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn request_n_most_recent_file(self: *Self, from: tp.pid_ref, n: usize) ClientError!void {
|
pub fn request_n_most_recent_file(self: *Self, from: tp.pid_ref, n: usize) ClientError!void {
|
||||||
|
self.sort_files_by_mtime();
|
||||||
if (n >= self.files.items.len) return error.ClientFailed;
|
if (n >= self.files.items.len) return error.ClientFailed;
|
||||||
const file_path = if (self.files.items.len > 0) self.files.items[n].path else null;
|
const file_path = if (self.files.items.len > 0) self.files.items[n].path else null;
|
||||||
from.send(.{file_path}) catch return error.ClientFailed;
|
from.send(.{file_path}) catch return error.ClientFailed;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn request_recent_files(self: *Self, from: tp.pid_ref, max: usize) ClientError!void {
|
pub fn request_recent_files(self: *Self, from: tp.pid_ref, max: usize) ClientError!void {
|
||||||
|
self.sort_files_by_mtime();
|
||||||
defer from.send(.{ "PRJ", "recent_done", self.longest_file_path, "" }) catch {};
|
defer from.send(.{ "PRJ", "recent_done", self.longest_file_path, "" }) catch {};
|
||||||
for (self.files.items, 0..) |file, i| {
|
for (self.files.items, 0..) |file, i| {
|
||||||
from.send(.{ "PRJ", "recent", self.longest_file_path, file.path }) catch return error.ClientFailed;
|
from.send(.{ "PRJ", "recent", self.longest_file_path, file.path }) catch return error.ClientFailed;
|
||||||
|
@ -386,21 +404,47 @@ pub fn query_recent_files(self: *Self, from: tp.pid_ref, max: usize, query: []co
|
||||||
return @min(max, matches.items.len);
|
return @min(max, matches.items.len);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn add_pending_file(self: *Self, file_path: []const u8, mtime: i128) OutOfMemoryError!void {
|
pub fn walk_tree_entry(self: *Self, file_path: []const u8, mtime: i128) OutOfMemoryError!void {
|
||||||
self.longest_file_path = @max(self.longest_file_path, file_path.len);
|
self.longest_file_path = @max(self.longest_file_path, file_path.len);
|
||||||
(try self.pending.addOne()).* = .{ .path = try self.allocator.dupe(u8, file_path), .mtime = mtime };
|
(try self.pending.addOne(self.allocator)).* = .{ .path = try self.allocator.dupe(u8, file_path), .mtime = mtime };
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn merge_pending_files(self: *Self) OutOfMemoryError!void {
|
pub fn walk_tree_done(self: *Self) OutOfMemoryError!void {
|
||||||
|
self.state.walk_tree = .done;
|
||||||
|
if (self.walker) |pid| pid.deinit();
|
||||||
|
self.walker = null;
|
||||||
|
return self.loaded();
|
||||||
|
}
|
||||||
|
|
||||||
|
fn merge_pending_files(self: *Self) OutOfMemoryError!void {
|
||||||
defer self.sort_files_by_mtime();
|
defer self.sort_files_by_mtime();
|
||||||
const existing = try self.files.toOwnedSlice();
|
const existing = try self.files.toOwnedSlice(self.allocator);
|
||||||
|
defer self.allocator.free(existing);
|
||||||
self.files = self.pending;
|
self.files = self.pending;
|
||||||
self.pending = std.ArrayList(File).init(self.allocator);
|
self.pending = .empty;
|
||||||
|
|
||||||
for (existing) |*file| {
|
for (existing) |*file| {
|
||||||
self.update_mru_internal(file.path, file.mtime, file.pos.row, file.pos.col) catch {};
|
self.update_mru_internal(file.path, file.mtime, file.pos.row, file.pos.col) catch {};
|
||||||
self.allocator.free(file.path);
|
self.allocator.free(file.path);
|
||||||
}
|
}
|
||||||
self.allocator.free(existing);
|
}
|
||||||
|
|
||||||
|
fn loaded(self: *Self) OutOfMemoryError!void {
|
||||||
|
inline for (@typeInfo(@TypeOf(self.state)).@"struct".fields) |f|
|
||||||
|
if (@field(self.state, f.name) == .running) return;
|
||||||
|
|
||||||
|
self.logger.print("project files: {d} restored, {d} {s}", .{
|
||||||
|
self.files.items.len,
|
||||||
|
self.pending.items.len,
|
||||||
|
if (self.state.workspace_files == .done) "tracked" else "walked",
|
||||||
|
});
|
||||||
|
|
||||||
|
try self.merge_pending_files();
|
||||||
|
self.logger.print("opened: {s} with {d} files in {d} ms", .{
|
||||||
|
self.name,
|
||||||
|
self.files.items.len,
|
||||||
|
std.time.milliTimestamp() - self.open_time,
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn update_mru(self: *Self, file_path: []const u8, row: usize, col: usize) OutOfMemoryError!void {
|
pub fn update_mru(self: *Self, file_path: []const u8, row: usize, col: usize) OutOfMemoryError!void {
|
||||||
|
@ -420,14 +464,14 @@ fn update_mru_internal(self: *Self, file_path: []const u8, mtime: i128, row: usi
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
if (row != 0) {
|
if (row != 0) {
|
||||||
(try self.files.addOne()).* = .{
|
(try self.files.addOne(self.allocator)).* = .{
|
||||||
.path = try self.allocator.dupe(u8, file_path),
|
.path = try self.allocator.dupe(u8, file_path),
|
||||||
.mtime = mtime,
|
.mtime = mtime,
|
||||||
.pos = .{ .row = row, .col = col },
|
.pos = .{ .row = row, .col = col },
|
||||||
.visited = true,
|
.visited = true,
|
||||||
};
|
};
|
||||||
} else {
|
} else {
|
||||||
(try self.files.addOne()).* = .{
|
(try self.files.addOne(self.allocator)).* = .{
|
||||||
.path = try self.allocator.dupe(u8, file_path),
|
.path = try self.allocator.dupe(u8, file_path),
|
||||||
.mtime = mtime,
|
.mtime = mtime,
|
||||||
};
|
};
|
||||||
|
@ -1858,23 +1902,55 @@ pub fn get_line(allocator: std.mem.Allocator, buf: []const u8) ![]const u8 {
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn query_git(self: *Self) void {
|
pub fn query_git(self: *Self) void {
|
||||||
git.workspace_path(@intFromPtr(self)) catch {};
|
self.state.workspace_path = .running;
|
||||||
git.current_branch(@intFromPtr(self)) catch {};
|
git.workspace_path(@intFromPtr(self)) catch {
|
||||||
|
self.state.workspace_path = .failed;
|
||||||
|
self.start_walker();
|
||||||
|
};
|
||||||
|
self.state.current_branch = .running;
|
||||||
|
git.current_branch(@intFromPtr(self)) catch {
|
||||||
|
self.state.current_branch = .failed;
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn process_git(self: *Self, m: tp.message) !void {
|
fn start_walker(self: *Self) void {
|
||||||
|
self.state.walk_tree = .running;
|
||||||
|
self.walker = walk_tree.start(self.allocator, self.name) catch blk: {
|
||||||
|
self.state.walk_tree = .failed;
|
||||||
|
break :blk null;
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn process_git(self: *Self, m: tp.message) (OutOfMemoryError || error{Exit})!void {
|
||||||
var value: []const u8 = undefined;
|
var value: []const u8 = undefined;
|
||||||
|
var path: []const u8 = undefined;
|
||||||
if (try m.match(.{ tp.any, tp.any, "workspace_path", tp.null_ })) {
|
if (try m.match(.{ tp.any, tp.any, "workspace_path", tp.null_ })) {
|
||||||
// no git workspace
|
self.state.workspace_path = .done;
|
||||||
|
self.start_walker();
|
||||||
|
try self.loaded();
|
||||||
} else if (try m.match(.{ tp.any, tp.any, "workspace_path", tp.extract(&value) })) {
|
} else if (try m.match(.{ tp.any, tp.any, "workspace_path", tp.extract(&value) })) {
|
||||||
if (self.workspace) |p| self.allocator.free(p);
|
if (self.workspace) |p| self.allocator.free(p);
|
||||||
self.workspace = try self.allocator.dupe(u8, value);
|
self.workspace = try self.allocator.dupe(u8, value);
|
||||||
git.workspace_files(@intFromPtr(self)) catch {};
|
self.state.workspace_path = .done;
|
||||||
|
self.state.workspace_files = .running;
|
||||||
|
git.workspace_files(@intFromPtr(self)) catch {
|
||||||
|
self.state.workspace_files = .failed;
|
||||||
|
};
|
||||||
|
} else if (try m.match(.{ tp.any, tp.any, "current_branch", tp.null_ })) {
|
||||||
|
self.state.current_branch = .done;
|
||||||
|
try self.loaded();
|
||||||
} else if (try m.match(.{ tp.any, tp.any, "current_branch", tp.extract(&value) })) {
|
} else if (try m.match(.{ tp.any, tp.any, "current_branch", tp.extract(&value) })) {
|
||||||
if (self.branch) |p| self.allocator.free(p);
|
if (self.branch) |p| self.allocator.free(p);
|
||||||
self.branch = try self.allocator.dupe(u8, value);
|
self.branch = try self.allocator.dupe(u8, value);
|
||||||
} else if (try m.match(.{ tp.any, tp.any, "workspace_files", tp.extract(&value) })) {
|
self.state.current_branch = .done;
|
||||||
// TODO
|
try self.loaded();
|
||||||
|
} else if (try m.match(.{ tp.any, tp.any, "workspace_files", tp.extract(&path) })) {
|
||||||
|
self.longest_file_path = @max(self.longest_file_path, path.len);
|
||||||
|
const stat = std.fs.cwd().statFile(path) catch return;
|
||||||
|
(try self.pending.addOne(self.allocator)).* = .{ .path = try self.allocator.dupe(u8, path), .mtime = stat.mtime };
|
||||||
|
} else if (try m.match(.{ tp.any, tp.any, "workspace_files", tp.null_ })) {
|
||||||
|
self.state.workspace_files = .done;
|
||||||
|
try self.loaded();
|
||||||
} else {
|
} else {
|
||||||
self.logger_git.err("git", tp.unexpected(m));
|
self.logger_git.err("git", tp.unexpected(m));
|
||||||
}
|
}
|
||||||
|
|
|
@ -9,7 +9,6 @@ const Buffer = @import("Buffer");
|
||||||
const builtin = @import("builtin");
|
const builtin = @import("builtin");
|
||||||
|
|
||||||
const Project = @import("Project.zig");
|
const Project = @import("Project.zig");
|
||||||
const walk_tree = @import("walk_tree.zig");
|
|
||||||
|
|
||||||
pid: tp.pid_ref,
|
pid: tp.pid_ref,
|
||||||
|
|
||||||
|
@ -248,13 +247,12 @@ const Process = struct {
|
||||||
logger: log.Logger,
|
logger: log.Logger,
|
||||||
receiver: Receiver,
|
receiver: Receiver,
|
||||||
projects: ProjectsMap,
|
projects: ProjectsMap,
|
||||||
walker: ?tp.pid = null,
|
|
||||||
|
|
||||||
const InvalidArgumentError = error{InvalidArgument};
|
const InvalidArgumentError = error{InvalidArgument};
|
||||||
const UnsupportedError = error{Unsupported};
|
const UnsupportedError = error{Unsupported};
|
||||||
|
|
||||||
const Receiver = tp.Receiver(*Process);
|
const Receiver = tp.Receiver(*Process);
|
||||||
const ProjectsMap = std.StringHashMap(*Project);
|
const ProjectsMap = std.StringHashMapUnmanaged(*Project);
|
||||||
const RecentProject = struct {
|
const RecentProject = struct {
|
||||||
name: []const u8,
|
name: []const u8,
|
||||||
last_used: i128,
|
last_used: i128,
|
||||||
|
@ -268,7 +266,7 @@ const Process = struct {
|
||||||
.parent = tp.self_pid().clone(),
|
.parent = tp.self_pid().clone(),
|
||||||
.logger = log.logger(module_name),
|
.logger = log.logger(module_name),
|
||||||
.receiver = Receiver.init(Process.receive, self),
|
.receiver = Receiver.init(Process.receive, self),
|
||||||
.projects = ProjectsMap.init(allocator),
|
.projects = .empty,
|
||||||
};
|
};
|
||||||
return tp.spawn_link(self.allocator, self, Process.start, module_name);
|
return tp.spawn_link(self.allocator, self, Process.start, module_name);
|
||||||
}
|
}
|
||||||
|
@ -280,7 +278,7 @@ const Process = struct {
|
||||||
p.value_ptr.*.deinit();
|
p.value_ptr.*.deinit();
|
||||||
self.allocator.destroy(p.value_ptr.*);
|
self.allocator.destroy(p.value_ptr.*);
|
||||||
}
|
}
|
||||||
self.projects.deinit();
|
self.projects.deinit(self.allocator);
|
||||||
self.parent.deinit();
|
self.parent.deinit();
|
||||||
self.logger.deinit();
|
self.logger.deinit();
|
||||||
self.allocator.destroy(self);
|
self.allocator.destroy(self);
|
||||||
|
@ -335,14 +333,10 @@ const Process = struct {
|
||||||
if (try cbor.match(m.buf, .{ "walk_tree_entry", tp.extract(&project_directory), tp.extract(&path), tp.extract(&high), tp.extract(&low) })) {
|
if (try cbor.match(m.buf, .{ "walk_tree_entry", tp.extract(&project_directory), tp.extract(&path), tp.extract(&high), tp.extract(&low) })) {
|
||||||
const mtime = (@as(i128, @intCast(high)) << 64) | @as(i128, @intCast(low));
|
const mtime = (@as(i128, @intCast(high)) << 64) | @as(i128, @intCast(low));
|
||||||
if (self.projects.get(project_directory)) |project|
|
if (self.projects.get(project_directory)) |project|
|
||||||
project.add_pending_file(
|
project.walk_tree_entry(path, mtime) catch |e| self.logger.err("walk_tree_entry", e);
|
||||||
path,
|
|
||||||
mtime,
|
|
||||||
) catch |e| self.logger.err("walk_tree_entry", e);
|
|
||||||
} else if (try cbor.match(m.buf, .{ "walk_tree_done", tp.extract(&project_directory) })) {
|
} else if (try cbor.match(m.buf, .{ "walk_tree_done", tp.extract(&project_directory) })) {
|
||||||
if (self.walker) |pid| pid.deinit();
|
if (self.projects.get(project_directory)) |project|
|
||||||
self.walker = null;
|
project.walk_tree_done() catch |e| return from.forward_error(e, @errorReturnTrace()) catch error.ClientFailed;
|
||||||
self.loaded(project_directory) catch |e| return from.forward_error(e, @errorReturnTrace()) catch error.ClientFailed;
|
|
||||||
} else if (try cbor.match(m.buf, .{ "git", tp.extract(&context), tp.more })) {
|
} else if (try cbor.match(m.buf, .{ "git", tp.extract(&context), tp.more })) {
|
||||||
const project: *Project = @ptrFromInt(context);
|
const project: *Project = @ptrFromInt(context);
|
||||||
project.process_git(m) catch {};
|
project.process_git(m) catch {};
|
||||||
|
@ -404,7 +398,6 @@ const Process = struct {
|
||||||
} else if (try cbor.match(m.buf, .{ "get_mru_position", tp.extract(&project_directory), tp.extract(&path) })) {
|
} else if (try cbor.match(m.buf, .{ "get_mru_position", tp.extract(&project_directory), tp.extract(&path) })) {
|
||||||
self.get_mru_position(from, project_directory, path) catch |e| return from.forward_error(e, @errorReturnTrace()) catch error.ClientFailed;
|
self.get_mru_position(from, project_directory, path) catch |e| return from.forward_error(e, @errorReturnTrace()) catch error.ClientFailed;
|
||||||
} else if (try cbor.match(m.buf, .{"shutdown"})) {
|
} else if (try cbor.match(m.buf, .{"shutdown"})) {
|
||||||
if (self.walker) |pid| pid.send(.{"stop"}) catch {};
|
|
||||||
self.persist_projects();
|
self.persist_projects();
|
||||||
from.send(.{ "project_manager", "shutdown" }) catch return error.ClientFailed;
|
from.send(.{ "project_manager", "shutdown" }) catch return error.ClientFailed;
|
||||||
return error.ExitNormal;
|
return error.ExitNormal;
|
||||||
|
@ -424,10 +417,8 @@ const Process = struct {
|
||||||
self.logger.print("opening: {s}", .{project_directory});
|
self.logger.print("opening: {s}", .{project_directory});
|
||||||
const project = try self.allocator.create(Project);
|
const project = try self.allocator.create(Project);
|
||||||
project.* = try Project.init(self.allocator, project_directory);
|
project.* = try Project.init(self.allocator, project_directory);
|
||||||
try self.projects.put(try self.allocator.dupe(u8, project_directory), project);
|
try self.projects.put(self.allocator, try self.allocator.dupe(u8, project_directory), project);
|
||||||
self.walker = try walk_tree_async(self.allocator, project_directory);
|
|
||||||
self.restore_project(project) catch |e| self.logger.err("restore_project", e);
|
self.restore_project(project) catch |e| self.logger.err("restore_project", e);
|
||||||
project.sort_files_by_mtime();
|
|
||||||
project.query_git();
|
project.query_git();
|
||||||
} else {
|
} else {
|
||||||
self.logger.print("switched to: {s}", .{project_directory});
|
self.logger.print("switched to: {s}", .{project_directory});
|
||||||
|
@ -444,25 +435,13 @@ const Process = struct {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn loaded(self: *Process, project_directory: []const u8) OutOfMemoryError!void {
|
|
||||||
const project = self.projects.get(project_directory) orelse return;
|
|
||||||
try project.merge_pending_files();
|
|
||||||
self.logger.print("opened: {s} with {d} files in {d} ms", .{
|
|
||||||
project_directory,
|
|
||||||
project.files.items.len,
|
|
||||||
std.time.milliTimestamp() - project.open_time,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
fn request_n_most_recent_file(self: *Process, from: tp.pid_ref, project_directory: []const u8, n: usize) (ProjectError || Project.ClientError)!void {
|
fn request_n_most_recent_file(self: *Process, from: tp.pid_ref, project_directory: []const u8, n: usize) (ProjectError || Project.ClientError)!void {
|
||||||
const project = self.projects.get(project_directory) orelse return error.NoProject;
|
const project = self.projects.get(project_directory) orelse return error.NoProject;
|
||||||
project.sort_files_by_mtime();
|
|
||||||
return project.request_n_most_recent_file(from, n);
|
return project.request_n_most_recent_file(from, n);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn request_recent_files(self: *Process, from: tp.pid_ref, project_directory: []const u8, max: usize) (ProjectError || Project.ClientError)!void {
|
fn request_recent_files(self: *Process, from: tp.pid_ref, project_directory: []const u8, max: usize) (ProjectError || Project.ClientError)!void {
|
||||||
const project = self.projects.get(project_directory) orelse return error.NoProject;
|
const project = self.projects.get(project_directory) orelse return error.NoProject;
|
||||||
project.sort_files_by_mtime();
|
|
||||||
return project.request_recent_files(from, max);
|
return project.request_recent_files(from, max);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue