diff options
author | Lorenzo Torres <torres@sideros.org> | 2025-03-29 21:20:16 +0100 |
---|---|---|
committer | Lorenzo Torres <torres@sideros.org> | 2025-03-29 21:20:16 +0100 |
commit | 9ff5e811a8642c59556df077b0c258f1af083480 (patch) | |
tree | e77013a53b692a7c49f82ffdf41befd5ca7b214f /src/mods | |
parent | 16a2a404189cc6b5cd3493df8d19298af7e542ce (diff) | |
parent | 2dc07080dd67ffdd67b6f5fe3ad74b113c6daca9 (diff) |
Merge branch 'master' of ssh://git.sideros.org:/git/sideros
Diffstat (limited to 'src/mods')
-rw-r--r-- | src/mods/Parser.zig | 98 | ||||
-rw-r--r-- | src/mods/ir.zig | 7 | ||||
-rw-r--r-- | src/mods/vm.zig | 417 |
3 files changed, 373 insertions, 149 deletions
diff --git a/src/mods/Parser.zig b/src/mods/Parser.zig index 48a8225..544ef31 100644 --- a/src/mods/Parser.zig +++ b/src/mods/Parser.zig @@ -7,34 +7,10 @@ bytes: []const u8, byte_idx: usize, allocator: Allocator, -// TODO: We don't really need ArrayLists -types: std.ArrayListUnmanaged(Functype) = .{}, -imports: std.ArrayListUnmanaged(Import) = .{}, -exports: std.StringHashMapUnmanaged(u32) = .{}, -functions: std.ArrayListUnmanaged(u32) = .{}, +types: ?[]vm.Functype = null, +functions: ?[]vm.Function = null, memory: ?Memtype = null, -code: std.ArrayListUnmanaged(Func) = .{}, -funcs: std.ArrayListUnmanaged(vm.Func) = .{}, - -pub const FunctionType = struct { - parameters: []u8, - results: []u8, - - pub fn deinit(self: FunctionType, allocator: Allocator) void { - allocator.free(self.parameters); - allocator.free(self.results); - } -}; - -pub const FunctionBody = struct { - locals: []Local, - code: []u8, -}; - -pub const FunctionScope = enum { - external, - internal, -}; +exports: std.StringHashMapUnmanaged(u32) = .{}, const Parser = @This(); @@ -56,6 +32,7 @@ pub const Error = error{ invalid_importdesc, invalid_exportdesc, double_else, + duplicated_funcsec, unresolved_branch, unterminated_wasm, }; @@ -172,11 +149,7 @@ pub fn parseReftype(self: *Parser) !std.wasm.RefType { // NOTE: Parsing of Valtype can be improved but it makes it less close to spec so... // TODO: Do we really need Valtype? -pub const Valtype = union(enum) { - val: std.wasm.Valtype, - ref: std.wasm.RefType, -}; -fn parseValtype(self: *Parser) !Valtype { +fn parseValtype(self: *Parser) !vm.Valtype { const pb = self.peek() orelse return Error.unterminated_wasm; return switch (pb) { 0x7F, 0x7E, 0x7D, 0x7C => .{ .val = try self.parseNumtype() }, @@ -186,24 +159,15 @@ fn parseValtype(self: *Parser) !Valtype { }; } -fn parseResultType(self: *Parser) ![]Valtype { +fn parseResultType(self: *Parser) ![]vm.Valtype { return try self.parseVector(Parser.parseValtype); } -pub const Functype = struct { - parameters: []Valtype, - rt2: []Valtype, - - pub fn deinit(self: Functype, allocator: Allocator) void { - allocator.free(self.parameters); - allocator.free(self.rt2); - } -}; -fn parseFunctype(self: *Parser) !Functype { +fn parseFunctype(self: *Parser) !vm.Functype { if (try self.readByte() != 0x60) return Error.invalid_functype; return .{ .parameters = try self.parseResultType(), - .rt2 = try self.parseResultType(), + .returns = try self.parseResultType(), }; } @@ -245,7 +209,7 @@ fn parseTabletype(self: *Parser) !Tabletype { } const Globaltype = struct { - t: Valtype, + t: vm.Valtype, m: enum { @"const", @"var", @@ -296,11 +260,7 @@ pub fn parseModule(self: *Parser) !vm.Module { .max = self.memory.?.lim.max, }, .exports = self.exports, - .funcs = try self.funcs.toOwnedSlice(self.allocator), - .types = try self.types.toOwnedSlice(self.allocator), - .functions = try self.functions.toOwnedSlice(self.allocator), - .imports = try self.imports.toOwnedSlice(self.allocator), - .code = try self.code.toOwnedSlice(self.allocator), + .functions = self.functions.?, }; } @@ -315,10 +275,9 @@ fn parseTypesec(self: *Parser) !void { const end_idx = self.byte_idx + size; const ft = try self.parseVector(Parser.parseFunctype); - // TODO: Maybe the interface should be better? - try self.types.appendSlice(self.allocator, ft); + self.types = ft; - // TODO: run this check not only on debug + // TODO(ernesto): run this check not only on debug std.debug.assert(self.byte_idx == end_idx); } @@ -349,8 +308,9 @@ fn parseImportsec(self: *Parser) !void { const size = try self.readU32(); const end_idx = self.byte_idx + size; + // TODO(ernesto): this should be used to do name resolution. const imports = try self.parseVector(Parser.parseImport); - try self.imports.appendSlice(self.allocator, imports); + _ = imports; // TODO: run this check not only on debug std.debug.assert(self.byte_idx == end_idx); @@ -361,7 +321,16 @@ fn parseFuncsec(self: *Parser) !void { const end_idx = self.byte_idx + size; const types = try self.parseVector(Parser.readU32); - try self.functions.appendSlice(self.allocator, types); + + if (self.functions != null) return Error.duplicated_funcsec; + self.functions = try self.allocator.alloc(vm.Function, types.len); + + for (types, 0..) |t, i| { + self.functions.?[i].func_type = self.types.?[t]; + } + + // TODO(ernesto): run this check not only in debug + std.debug.assert(types.len == self.functions.?.len); // TODO: run this check not only on debug std.debug.assert(self.byte_idx == end_idx); @@ -446,13 +415,12 @@ fn parseElemsec(self: *Parser) !void { } pub const Func = struct { - locals: []Valtype, - code: []const u8, + locals: []vm.Valtype, ir: IR, }; const Local = struct { n: u32, - t: Valtype, + t: vm.Valtype, }; fn parseLocal(self: *Parser) !Local { return .{ @@ -476,8 +444,7 @@ fn parseCode(self: *Parser) !Func { try ir.print(stdout); const func = Func{ - .locals = try self.allocator.alloc(Valtype, local_count), - .code = &.{}, + .locals = try self.allocator.alloc(vm.Valtype, local_count), .ir = ir, }; @@ -498,10 +465,15 @@ fn parseCodesec(self: *Parser) !void { const end_idx = self.byte_idx + size; const codes = try self.parseVector(Parser.parseCode); - for (codes, 0..) |_, i| { - try self.funcs.append(self.allocator, .{ .internal = @intCast(i) }); + // TODO: run this check not only on debug + std.debug.assert(codes.len == self.functions.?.len); + + for (codes, self.functions.?) |code, *f| { + f.typ = .{ .internal = .{ + .locals = code.locals, + .ir = code.ir, + } }; } - try self.code.appendSlice(self.allocator, codes); // TODO: run this check not only on debug std.debug.assert(self.byte_idx == end_idx); diff --git a/src/mods/ir.zig b/src/mods/ir.zig index 25409e4..e745cb8 100644 --- a/src/mods/ir.zig +++ b/src/mods/ir.zig @@ -1,5 +1,6 @@ const std = @import("std"); const Parser = @import("Parser.zig"); +const vm = @import("vm.zig"); const Allocator = std.mem.Allocator; @@ -44,7 +45,7 @@ opcodes: []Opcode, indices: []Index, // TODO: this could be a byte array and v128.const and i8x16.shuffle could live here too -select_valtypes: []Parser.Valtype, +select_valtypes: []vm.Valtype, pub fn print(self: IR, writer: anytype) !void { for (self.opcodes, 0..) |op, i| { @@ -621,8 +622,8 @@ const IRParserState = struct { fn parseExpression(self: *IRParserState) Parser.Error!void { const b = try self.parser.readByte(); try switch (b) { - 0x00 => {}, // TODO - 0x01 => {}, + 0x00 => self.push(@enumFromInt(b), .{ .u64 = 0 }), + 0x01 => self.push(@enumFromInt(b), .{ .u64 = 0 }), 0x02...0x03 => self.parseBlock(b), 0x04 => self.parseIf(), 0x0C...0x0D => self.parseBranch(b), diff --git a/src/mods/vm.zig b/src/mods/vm.zig index 413069f..0d6cddf 100644 --- a/src/mods/vm.zig +++ b/src/mods/vm.zig @@ -9,29 +9,50 @@ pub const Memory = struct { min: u32, max: ?u32, }; -// TODO: Resolve function calls at parse time -// TODO: Resolve function types at compile time -pub const Func = union(enum) { - internal: u32, - external: u32, + +pub const Valtype = union(enum) { + val: std.wasm.Valtype, + ref: std.wasm.RefType, +}; + +pub const Functype = struct { + parameters: []Valtype, + returns: []Valtype, + + pub fn deinit(self: Functype, allocator: Allocator) void { + allocator.free(self.parameters); + allocator.free(self.returns); + } }; +pub const Function = struct { func_type: Functype, typ: union(enum) { + internal: struct { + locals: []Valtype, + ir: IR, + }, + external: void, +} }; pub const Module = struct { memory: Memory, - funcs: []Func, + functions: []Function, exports: std.StringHashMapUnmanaged(u32), - imports: []Parser.Import, - types: []Parser.Functype, - functions: []u32, - code: []Parser.Func, fn deinit(self: *Module, allocator: Allocator) void { self.exports.deinit(allocator); - allocator.free(self.funcs); - allocator.free(self.imports); - allocator.free(self.types); + for (self.functions) |f| { + allocator.free(f.func_type.parameters); + allocator.free(f.func_type.returns); + switch (f.typ) { + .internal => { + allocator.free(f.typ.internal.ir.opcodes); + allocator.free(f.typ.internal.ir.indices); + allocator.free(f.typ.internal.ir.select_valtypes); + allocator.free(f.typ.internal.locals); + }, + .external => @panic("UNIMPLEMENTED"), + } + } allocator.free(self.functions); - allocator.free(self.code); } }; @@ -41,14 +62,11 @@ pub const CallFrame = struct { locals: []Value, }; -const ValueType = enum { - i32, - i64, -}; - -pub const Value = union(ValueType) { +pub const Value = union(enum) { i32: i32, i64: i64, + f32: f32, + f64: f64, }; pub const Runtime = struct { @@ -61,7 +79,7 @@ pub const Runtime = struct { // if memory max is not set the memory is allowed to grow but it is not supported at the moment const max = module.memory.max orelse 1_000; if (module.memory.max == null) { - std.debug.print("[WARN]: growing memory is not yet supported, usign a default value of 1Kb\n", .{}); + std.log.warn("Growing memory is not yet supported, usign a default value of 1Kb\n", .{}); } const memory = try allocator.alloc(u8, max); return Runtime{ @@ -78,76 +96,328 @@ pub const Runtime = struct { allocator.free(self.memory); } - pub fn executeFrame(self: *Runtime, _: Allocator, frame: *CallFrame) !void { - loop: while (true) { + pub fn executeFrame(self: *Runtime, allocator: Allocator, frame: *CallFrame) !void { + loop: while (frame.program_counter < frame.code.opcodes.len) { const opcode: IR.Opcode = frame.code.opcodes[frame.program_counter]; + const index = frame.code.indices[frame.program_counter]; switch (opcode) { + // TODO(ernesto): How should we handle unreachable? + // Like this + .@"unreachable" => { + std.log.err("Reached unreachable statement at IR counter {any}\n", .{frame.program_counter}); + frame.code.print(std.io.getStdOut().writer()) catch {}; + }, + .nop => @panic("UNIMPLEMENTED"), .br => { - // TODO(luccie-cmd): Branching like this is dangerous, we should do safety things or smth. - frame.program_counter = frame.code.indices[frame.program_counter].u32 - 1; + frame.program_counter = index.u32; + continue; }, .br_if => { if (self.stack.pop().?.i32 != 0) { - // TODO(luccie-cmd): Branching like this is dangerous, we should do safety things or smth. - frame.program_counter = frame.code.indices[frame.program_counter].u32 - 1; + frame.program_counter = index.u32; + continue; } }, - .localget => { - try self.stack.append(frame.locals[frame.code.indices[frame.program_counter].u32]); + .br_table => @panic("UNIMPLEMENTED"), + .@"return" => break :loop, + .call => { + // TODO: figure out how many parameters to push + try self.call(allocator, index.u32, &[_]Value{}); }, - .localset => { - const a = self.stack.pop().?; - frame.locals[frame.code.indices[frame.program_counter].u32] = a; + .call_indirect => @panic("UNIMPLEMENTED"), + + .refnull => @panic("UNIMPLEMENTED"), + .refisnull => @panic("UNIMPLEMENTED"), + .reffunc => @panic("UNIMPLEMENTED"), + + .drop => @panic("UNIMPLEMENTED"), + .select => @panic("UNIMPLEMENTED"), + .select_with_values => @panic("UNIMPLEMENTED"), + + .localget => try self.stack.append(frame.locals[index.u32]), + .localset => frame.locals[index.u32] = self.stack.pop().?, + .localtee => frame.locals[index.u32] = self.stack.items[self.stack.items.len - 1], + .globalget => @panic("UNIMPLEMENTED"), + .globalset => @panic("UNIMPLEMENTED"), + + .tableget => @panic("UNIMPLEMENTED"), + .tableset => @panic("UNIMPLEMENTED"), + .tableinit => @panic("UNIMPLEMENTED"), + .elemdrop => @panic("UNIMPLEMENTED"), + .tablecopy => @panic("UNIMPLEMENTED"), + .tablegrow => @panic("UNIMPLEMENTED"), + .tablesize => @panic("UNIMPLEMENTED"), + .tablefill => @panic("UNIMPLEMENTED"), + + // TODO(ernesto): This code is repeated... + .i32_load => { + const start = index.memarg.alignment + index.memarg.offset; + const end = start + @sizeOf(i32); + try self.stack.append(.{ .i32 = std.mem.littleToNative(i32, std.mem.bytesAsValue(i32, self.memory[start..end]).*) }); }, - .localtee => { - const a = self.stack.pop().?; - try self.stack.append(a); - frame.locals[frame.code.indices[frame.program_counter].u32] = a; + .i64_load => { + const start = index.memarg.alignment + index.memarg.offset; + const end = start + @sizeOf(i64); + try self.stack.append(.{ .i64 = std.mem.littleToNative(i64, std.mem.bytesAsValue(i64, self.memory[start..end]).*) }); }, + .f32_load => { + const start = index.memarg.alignment + index.memarg.offset; + const end = start + @sizeOf(f32); + try self.stack.append(.{ .f32 = std.mem.littleToNative(f32, std.mem.bytesAsValue(f32, self.memory[start..end]).*) }); + }, + .f64_load => { + const start = index.memarg.alignment + index.memarg.offset; + const end = start + @sizeOf(f64); + try self.stack.append(.{ .f64 = std.mem.littleToNative(f64, std.mem.bytesAsValue(f64, self.memory[start..end]).*) }); + }, + .i32_load8_s => @panic("UNIMPLEMENTED"), + .i32_load8_u => @panic("UNIMPLEMENTED"), + .i32_load16_s => @panic("UNIMPLEMENTED"), + .i32_load16_u => @panic("UNIMPLEMENTED"), + .i64_load8_s => @panic("UNIMPLEMENTED"), + .i64_load8_u => @panic("UNIMPLEMENTED"), + .i64_load16_s => @panic("UNIMPLEMENTED"), + .i64_load16_u => @panic("UNIMPLEMENTED"), + .i64_load32_s => @panic("UNIMPLEMENTED"), + .i64_load32_u => @panic("UNIMPLEMENTED"), + .i32_store => { + // TODO(ernesto): I'm pretty sure this is wrong + const start = index.memarg.offset + index.memarg.alignment; + const end = start + @sizeOf(u32); + const val = std.mem.nativeToLittle(i32, self.stack.pop().?.i32); + @memcpy(self.memory[start..end], std.mem.asBytes(&val)); + }, + .i64_store => @panic("UNIMPLEMENTED"), + .f32_store => @panic("UNIMPLEMENTED"), + .f64_store => @panic("UNIMPLEMENTED"), + .i32_store8 => @panic("UNIMPLEMENTED"), + .i32_store16 => @panic("UNIMPLEMENTED"), + .i64_store8 => @panic("UNIMPLEMENTED"), + .i64_store16 => @panic("UNIMPLEMENTED"), + .i64_store32 => @panic("UNIMPLEMENTED"), + + .memorysize => @panic("UNIMPLEMENTED"), + .memorygrow => @panic("UNIMPLEMENTED"), + .memoryinit => @panic("UNIMPLEMENTED"), + .datadrop => @panic("UNIMPLEMENTED"), + .memorycopy => @panic("UNIMPLEMENTED"), + .memoryfill => @panic("UNIMPLEMENTED"), + // 0x36 => { + // const address = leb128Decode(u32, frame.code[frame.program_counter..]); + // frame.program_counter += address.len; + // const offset = leb128Decode(u32, frame.code[frame.program_counter..]); + // frame.program_counter += offset.len; + // const start = (address.val + offset.val); + // const end = start + @sizeOf(u32); + // try self.stack.append(Value{ .i32 = decodeLittleEndian(i32, self.memory[start..end]) }); + // }, + // 0x37 => { + // const address = leb128Decode(u32, frame.code[frame.program_counter..]); + // frame.program_counter += address.len; + // const offset = leb128Decode(u32, frame.code[frame.program_counter..]); + // frame.program_counter += offset.len; + // const start = (address.val + offset.val); + // const end = start + @sizeOf(u32); + // encodeLittleEndian(i32, @constCast(&self.memory[start..end]), self.stack.pop().?.i32); + // }, + // 0x38 => { + // const address = leb128Decode(u32, frame.code[frame.program_counter..]); + // frame.program_counter += address.len; + // const offset = leb128Decode(u32, frame.code[frame.program_counter..]); + // frame.program_counter += offset.len; + // const start = (address.val + offset.val); + // const end = start + @sizeOf(u64); + // encodeLittleEndian(i64, @constCast(&self.memory[start..end]), self.stack.pop().?.i64); + // }, + .i32_const => { try self.stack.append(Value{ .i32 = frame.code.indices[frame.program_counter].i32 }); }, + .i64_const => { + try self.stack.append(Value{ .i64 = frame.code.indices[frame.program_counter].i64 }); + }, + + .f32_const => @panic("UNIMPLEMENTED"), + .f64_const => @panic("UNIMPLEMENTED"), + + .i32_eqz => { + try self.stack.append(Value{ .i32 = @intCast(@as(u1, @bitCast(self.stack.pop().?.i32 == 0))) }); + }, + .i32_eq => @panic("UNIMPLEMENTED"), + .i32_ne => @panic("UNIMPLEMENTED"), + .i32_lt_s => @panic("UNIMPLEMENTED"), .i32_lt_u => { const b = self.stack.pop().?.i32; const a = self.stack.pop().?.i32; try self.stack.append(Value{ .i32 = @intCast(@as(u1, @bitCast(a < b))) }); }, + .i32_gt_s => @panic("UNIMPLEMENTED"), + .i32_gt_u => @panic("UNIMPLEMENTED"), + .i32_le_s => @panic("UNIMPLEMENTED"), + .i32_le_u => @panic("UNIMPLEMENTED"), + .i32_ge_s => @panic("UNIMPLEMENTED"), .i32_ge_u => { const b = self.stack.pop().?.i32; const a = self.stack.pop().?.i32; try self.stack.append(Value{ .i32 = @intCast(@as(u1, @bitCast(a >= b))) }); }, - .i32_eqz => { + + .i64_eqz => { try self.stack.append(Value{ .i32 = @intCast(@as(u1, @bitCast(self.stack.pop().?.i32 == 0))) }); }, + .i64_eq => @panic("UNIMPLEMENTED"), + .i64_ne => @panic("UNIMPLEMENTED"), + .i64_lt_s => @panic("UNIMPLEMENTED"), + .i64_lt_u => { + const b = self.stack.pop().?.i32; + const a = self.stack.pop().?.i32; + try self.stack.append(Value{ .i32 = @intCast(@as(u1, @bitCast(a < b))) }); + }, + .i64_gt_s => @panic("UNIMPLEMENTED"), + .i64_gt_u => @panic("UNIMPLEMENTED"), + .i64_le_s => @panic("UNIMPLEMENTED"), + .i64_le_u => @panic("UNIMPLEMENTED"), + .i64_ge_s => @panic("UNIMPLEMENTED"), + .i64_ge_u => @panic("UNIMPLEMENTED"), + + .f32_eq => @panic("UNIMPLEMENTED"), + .f32_ne => @panic("UNIMPLEMENTED"), + .f32_lt => @panic("UNIMPLEMENTED"), + .f32_gt => @panic("UNIMPLEMENTED"), + .f32_le => @panic("UNIMPLEMENTED"), + .f32_ge => @panic("UNIMPLEMENTED"), + + .f64_eq => @panic("UNIMPLEMENTED"), + .f64_ne => @panic("UNIMPLEMENTED"), + .f64_lt => @panic("UNIMPLEMENTED"), + .f64_gt => @panic("UNIMPLEMENTED"), + .f64_le => @panic("UNIMPLEMENTED"), + .f64_ge => @panic("UNIMPLEMENTED"), + + .i32_clz => @panic("UNIMPLEMENTED"), + .i32_ctz => @panic("UNIMPLEMENTED"), + .i32_popcnt => @panic("UNIMPLEMENTED"), .i32_add => { const a = self.stack.pop().?.i32; const b = self.stack.pop().?.i32; try self.stack.append(Value{ .i32 = a + b }); }, + .i32_sub => @panic("UNIMPLEMENTED"), .i32_and => { const a = self.stack.pop().?.i32; const b = self.stack.pop().?.i32; try self.stack.append(Value{ .i32 = a & b }); }, - .i64_const => { - try self.stack.append(Value{ .i64 = frame.code.indices[frame.program_counter].i64 }); - }, + .i32_mul => @panic("UNIMPLEMENTED"), + .i32_div_s => @panic("UNIMPLEMENTED"), + .i32_div_u => @panic("UNIMPLEMENTED"), + .i32_rem_s => @panic("UNIMPLEMENTED"), + .i32_rem_u => @panic("UNIMPLEMENTED"), + .i32_or => @panic("UNIMPLEMENTED"), + .i32_xor => @panic("UNIMPLEMENTED"), + .i32_shl => @panic("UNIMPLEMENTED"), + .i32_shr_s => @panic("UNIMPLEMENTED"), + .i32_shr_u => @panic("UNIMPLEMENTED"), + .i32_rotl => @panic("UNIMPLEMENTED"), + .i32_rotr => @panic("UNIMPLEMENTED"), + + .i64_clz => @panic("UNIMPLEMENTED"), + .i64_ctz => @panic("UNIMPLEMENTED"), + .i64_popcnt => @panic("UNIMPLEMENTED"), .i64_add => { const a = self.stack.pop().?.i64; const b = self.stack.pop().?.i64; try self.stack.append(Value{ .i64 = a + b }); }, + .i64_sub => @panic("UNIMPLEMENTED"), + .i64_mul => @panic("UNIMPLEMENTED"), + .i64_div_s => @panic("UNIMPLEMENTED"), + .i64_div_u => @panic("UNIMPLEMENTED"), + .i64_rem_s => @panic("UNIMPLEMENTED"), + .i64_rem_u => @panic("UNIMPLEMENTED"), + .i64_and => @panic("UNIMPLEMENTED"), + .i64_or => @panic("UNIMPLEMENTED"), + .i64_xor => @panic("UNIMPLEMENTED"), + .i64_shl => @panic("UNIMPLEMENTED"), + .i64_shr_s => @panic("UNIMPLEMENTED"), + .i64_shr_u => @panic("UNIMPLEMENTED"), + .i64_rotl => @panic("UNIMPLEMENTED"), + .i64_rotr => @panic("UNIMPLEMENTED"), + + .f32_abs => @panic("UNIMPLEMENTED"), + .f32_neg => @panic("UNIMPLEMENTED"), + .f32_ceil => @panic("UNIMPLEMENTED"), + .f32_floor => @panic("UNIMPLEMENTED"), + .f32_trunc => @panic("UNIMPLEMENTED"), + .f32_nearest => @panic("UNIMPLEMENTED"), + .f32_sqrt => @panic("UNIMPLEMENTED"), + .f32_add => @panic("UNIMPLEMENTED"), + .f32_sub => @panic("UNIMPLEMENTED"), + .f32_mul => @panic("UNIMPLEMENTED"), + .f32_div => @panic("UNIMPLEMENTED"), + .f32_min => @panic("UNIMPLEMENTED"), + .f32_max => @panic("UNIMPLEMENTED"), + .f32_copysign => @panic("UNIMPLEMENTED"), + + .f64_abs => @panic("UNIMPLEMENTED"), + .f64_neg => @panic("UNIMPLEMENTED"), + .f64_ceil => @panic("UNIMPLEMENTED"), + .f64_floor => @panic("UNIMPLEMENTED"), + .f64_trunc => @panic("UNIMPLEMENTED"), + .f64_nearest => @panic("UNIMPLEMENTED"), + .f64_sqrt => @panic("UNIMPLEMENTED"), + .f64_add => @panic("UNIMPLEMENTED"), + .f64_sub => @panic("UNIMPLEMENTED"), + .f64_mul => @panic("UNIMPLEMENTED"), + .f64_div => @panic("UNIMPLEMENTED"), + .f64_min => @panic("UNIMPLEMENTED"), + .f64_max => @panic("UNIMPLEMENTED"), + .f64_copysign => @panic("UNIMPLEMENTED"), + + .i32_wrap_i64 => @panic("UNIMPLEMENTED"), + .i32_trunc_f32_s => @panic("UNIMPLEMENTED"), + .i32_trunc_f32_u => @panic("UNIMPLEMENTED"), + .i32_trunc_f64_s => @panic("UNIMPLEMENTED"), + .i32_trunc_f64_u => @panic("UNIMPLEMENTED"), + .i64_extend_i32_s => @panic("UNIMPLEMENTED"), .i64_extend_i32_u => { try self.stack.append(.{ .i64 = self.stack.pop().?.i32 }); }, - .@"return" => { - break :loop; - }, - else => { - std.log.err("instruction {any} not implemented\n", .{opcode}); - std.process.exit(1); - }, + .i64_trunc_f32_s => @panic("UNIMPLEMENTED"), + .i64_trunc_f32_u => @panic("UNIMPLEMENTED"), + .i64_trunc_f64_s => @panic("UNIMPLEMENTED"), + .i64_trunc_f64_u => @panic("UNIMPLEMENTED"), + .f32_convert_i32_s => @panic("UNIMPLEMENTED"), + .f32_convert_i32_u => @panic("UNIMPLEMENTED"), + .f32_convert_i64_s => @panic("UNIMPLEMENTED"), + .f32_convert_i64_u => @panic("UNIMPLEMENTED"), + .f32_demote_f64 => @panic("UNIMPLEMENTED"), + .f64_convert_i32_s => @panic("UNIMPLEMENTED"), + .f64_convert_i32_u => @panic("UNIMPLEMENTED"), + .f64_convert_i64_s => @panic("UNIMPLEMENTED"), + .f64_convert_i64_u => @panic("UNIMPLEMENTED"), + .f64_promote_f32 => @panic("UNIMPLEMENTED"), + .i32_reinterpret_f32 => @panic("UNIMPLEMENTED"), + .i64_reinterpret_f64 => @panic("UNIMPLEMENTED"), + .f32_reinterpret_i32 => @panic("UNIMPLEMENTED"), + .f64_reinterpret_i64 => @panic("UNIMPLEMENTED"), + + .i32_extend8_s => @panic("UNIMPLEMENTED"), + .i32_extend16_s => @panic("UNIMPLEMENTED"), + .i64_extend8_s => @panic("UNIMPLEMENTED"), + .i64_extend16_s => @panic("UNIMPLEMENTED"), + .i64_extend32_s => @panic("UNIMPLEMENTED"), + + .i32_trunc_sat_f32_s => @panic("UNIMPLEMENTED"), + .i32_trunc_sat_f32_u => @panic("UNIMPLEMENTED"), + .i32_trunc_sat_f64_s => @panic("UNIMPLEMENTED"), + .i32_trunc_sat_f64_u => @panic("UNIMPLEMENTED"), + .i64_trunc_sat_f32_s => @panic("UNIMPLEMENTED"), + .i64_trunc_sat_f32_u => @panic("UNIMPLEMENTED"), + .i64_trunc_sat_f64_s => @panic("UNIMPLEMENTED"), + .i64_trunc_sat_f64_u => @panic("UNIMPLEMENTED"), + + .vecinst => @panic("UNIMPLEMENTED"), } // switch (byte) { // 0x02 => { @@ -581,7 +851,7 @@ pub const Runtime = struct { // const integer = leb128Decode(u32, frame.code[frame.program_counter..]); // frame.program_counter += integer.len; - // self.call(allocator, integer.val, &[_]usize{}) catch {}; + // self.call(allocator, integer.val, &[_]usize@panic("UNIMPLEMENTED"),) catch @panic("UNIMPLEMENTED"),; // }, // 0xb => { // _ = self.labels.pop(); @@ -589,17 +859,14 @@ pub const Runtime = struct { // for_loop = false; // } // }, - // else => std.log.err("instruction {} not implemented\n", .{byte}), + // else => std.log.err("instruction @panic("UNIMPLEMENTED"), not implemented\n", .{byte}), // } frame.program_counter += 1; - if (frame.program_counter >= frame.code.opcodes.len) { - break :loop; - } } } - // TODO: Do name resolution - pub fn callExternal(self: *Runtime, allocator: Allocator, name: []const u8, parameters: []usize) !void { + // TODO: Do name resolution at parseTime + pub fn callExternal(self: *Runtime, allocator: Allocator, name: []const u8, parameters: []Value) !void { if (self.module.exports.get(name)) |function| { try self.call(allocator, function, parameters); } else { @@ -607,44 +874,27 @@ pub const Runtime = struct { } } - pub fn call(self: *Runtime, allocator: Allocator, function: usize, parameters: []usize) AllocationError!void { - const f = self.module.funcs[function]; - switch (f) { + pub fn call(self: *Runtime, allocator: Allocator, function: usize, parameters: []Value) AllocationError!void { + const f = self.module.functions[function]; + switch (f.typ) { .internal => { - const ir: IR = self.module.code[f.internal].ir; - const function_type = self.module.types[self.module.functions[f.internal]]; + const ir: IR = f.typ.internal.ir; + const function_type = f.func_type; var frame = CallFrame{ .code = ir, .program_counter = 0x0, - .locals = try allocator.alloc(Value, self.module.code[f.internal].locals.len + function_type.parameters.len), + .locals = try allocator.alloc(Value, f.typ.internal.locals.len + function_type.parameters.len), }; - for (parameters, 0..) |p, i| { - switch (function_type.parameters[i]) { - .val => |v| switch (v) { - .i32 => { - std.debug.print("Local with type i32\n", .{}); - frame.locals[i] = .{ .i32 = @intCast(p) }; - }, - .i64 => { - std.debug.print("Local with type i64\n", .{}); - frame.locals[i] = .{ .i64 = @intCast(p) }; - }, - else => unreachable, - }, - .ref => unreachable, - } - } + @memcpy(frame.locals[0..parameters.len], parameters); - for (self.module.code[f.internal].locals, function_type.parameters.len..) |local, i| { + for (f.typ.internal.locals, function_type.parameters.len..) |local, i| { switch (local) { .val => |v| switch (v) { .i32 => { - std.debug.print("Local with type i32\n", .{}); frame.locals[i] = .{ .i32 = 0 }; }, .i64 => { - std.debug.print("Local with type i64\n", .{}); frame.locals[i] = .{ .i64 = 0 }; }, else => unreachable, @@ -658,10 +908,11 @@ pub const Runtime = struct { allocator.free(frame.locals); }, .external => { - const name = self.module.imports[f.external].name; - if (self.global_runtime.functions.get(name)) |external| { - external(&self.stack); - } + // TODO(ernesto): handle external functions + // const name = self.module.imports[f.external].name; + // if (self.global_runtime.functions.get(name)) |external| { + // external(&self.stack); + // } }, } } |