Add Llama 3.1 example showcasing optional head‑dimension support.
This commit is contained in:
parent
675cefaf26
commit
7bdce4c078
@ -18,7 +18,7 @@ pub const LlamaLM = struct {
|
|||||||
int: u32,
|
int: u32,
|
||||||
ints: []u32,
|
ints: []u32,
|
||||||
}),
|
}),
|
||||||
head_dim: ?u32,
|
head_dim: ?u32 = null,
|
||||||
hidden_size: u32,
|
hidden_size: u32,
|
||||||
num_hidden_layers: u32,
|
num_hidden_layers: u32,
|
||||||
num_attention_heads: u32,
|
num_attention_heads: u32,
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user