Add Llama 3.1 example showcasing optional head‑dimension support.
This commit is contained in:
parent
675cefaf26
commit
7bdce4c078
@ -18,7 +18,7 @@ pub const LlamaLM = struct {
|
||||
int: u32,
|
||||
ints: []u32,
|
||||
}),
|
||||
head_dim: ?u32,
|
||||
head_dim: ?u32 = null,
|
||||
hidden_size: u32,
|
||||
num_hidden_layers: u32,
|
||||
num_attention_heads: u32,
|
||||
|
||||
Loading…
Reference in New Issue
Block a user