Update examples to use the corrected logFn API.
This commit is contained in:
parent
754656f2f0
commit
602757e7a9
@ -6,6 +6,7 @@ const flags = @import("tigerbeetle/flags");
|
|||||||
// set log level to debug to print the generated IR
|
// set log level to debug to print the generated IR
|
||||||
pub const std_options = .{
|
pub const std_options = .{
|
||||||
.log_level = .warn,
|
.log_level = .warn,
|
||||||
|
.logFn = asynk.logFn(std.log.defaultLog),
|
||||||
};
|
};
|
||||||
|
|
||||||
pub fn benchmark(a: zml.Tensor, b: zml.Tensor) zml.Tensor {
|
pub fn benchmark(a: zml.Tensor, b: zml.Tensor) zml.Tensor {
|
||||||
|
|||||||
@ -19,6 +19,7 @@ const log = std.log.scoped(.llama);
|
|||||||
|
|
||||||
pub const std_options = .{
|
pub const std_options = .{
|
||||||
.log_level = .info,
|
.log_level = .info,
|
||||||
|
.logFn = asynk.logFn(std.log.defaultLog),
|
||||||
};
|
};
|
||||||
|
|
||||||
pub fn tokenizePromptLlama3(allocator: std.mem.Allocator, tokenizer: zml.tokenizer.Tokenizer, config: LlamaLM.Config, prompt: []const u8) ![]u32 {
|
pub fn tokenizePromptLlama3(allocator: std.mem.Allocator, tokenizer: zml.tokenizer.Tokenizer, config: LlamaLM.Config, prompt: []const u8) ![]u32 {
|
||||||
|
|||||||
@ -1,14 +1,13 @@
|
|||||||
const asynk = @import("async");
|
|
||||||
const std = @import("std");
|
const std = @import("std");
|
||||||
const zml = @import("zml");
|
|
||||||
|
|
||||||
const show_mlir = true;
|
const asynk = @import("async");
|
||||||
|
const zml = @import("zml");
|
||||||
|
|
||||||
const log = std.log.scoped(.mnist);
|
const log = std.log.scoped(.mnist);
|
||||||
|
|
||||||
pub const std_options: std.Options = .{
|
pub const std_options: std.Options = .{
|
||||||
.logFn = asynk.logFn,
|
|
||||||
.log_level = .info,
|
.log_level = .info,
|
||||||
|
.logFn = asynk.logFn(std.log.defaultLog),
|
||||||
};
|
};
|
||||||
|
|
||||||
/// Model definition
|
/// Model definition
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user