Update example Bazel build files (MODULE.bazel, llama, modernbert) to test the revamped commit workflow.

This commit is contained in:
Foke Singh 2025-01-31 16:28:38 +00:00
parent 0a2ab7c8cb
commit dd52e988b4
4 changed files with 3751 additions and 14 deletions

View File

@ -3,6 +3,7 @@ module(name = "examples")
bazel_dep(name = "aspect_bazel_lib", version = "2.14.0")
bazel_dep(name = "bazel_skylib", version = "1.7.1")
bazel_dep(name = "platforms", version = "0.0.11")
bazel_dep(name = "rules_cc", version = "0.1.1")
bazel_dep(name = "rules_oci", version = "2.2.6")
bazel_dep(name = "rules_rust", version = "0.60.0")
bazel_dep(name = "rules_zig", version = "20250530.0-5084f1f")

File diff suppressed because one or more lines are too long

View File

@ -3,6 +3,7 @@ load("@aspect_bazel_lib//lib:tar.bzl", "mtree_spec", "tar")
load("@aspect_bazel_lib//lib:transitions.bzl", "platform_transition_filegroup")
load("@bazel_skylib//rules:native_binary.bzl", "native_test")
load("@bazel_skylib//rules:write_file.bzl", "write_file")
load("@rules_cc//cc:cc_binary.bzl", "cc_binary")
load("@rules_oci//oci:defs.bzl", "oci_image", "oci_load", "oci_push")
load("@zml//bazel:zig.bzl", "zig_cc_binary")
@ -34,7 +35,7 @@ cc_binary(
"@Meta-Llama-3.1-8B-Instruct//:tokenizer.json",
],
tags = [
"no_ci",
"manual",
],
deps = [":llama_lib"],
)
@ -53,7 +54,7 @@ cc_binary(
"@Meta-Llama-3.1-70B-Instruct//:tokenizer.json",
],
tags = [
"no_ci",
"manual",
],
deps = [":llama_lib"],
)
@ -72,7 +73,7 @@ cc_binary(
"@Meta-Llama-3.2-1B-Instruct//:tokenizer.json",
],
tags = [
"no_ci",
"manual",
],
deps = [":llama_lib"],
)
@ -91,7 +92,7 @@ cc_binary(
"@Meta-Llama-3.2-3B-Instruct//:tokenizer.json",
],
tags = [
"no_ci",
"manual",
],
deps = [":llama_lib"],
)
@ -133,7 +134,7 @@ zig_cc_binary(
],
main = "test.zig",
tags = [
"no_ci",
"manual",
],
deps = [
"@zml//async",
@ -162,7 +163,7 @@ Artificial Intelligence in Healthcare
],
data = ["@Meta-Llama-3.2-1B-Instruct//:tokenizer.json"],
tags = [
"no_ci",
"manual",
],
)
@ -170,7 +171,7 @@ mtree_spec(
name = "mtree",
srcs = [":Llama-3.2-1B-Instruct"],
tags = [
"no_ci",
"manual",
],
)
@ -184,7 +185,7 @@ tar(
compress = "zstd",
mtree = ":mtree",
tags = [
"no_ci",
"manual",
],
)
@ -203,7 +204,7 @@ expand_template(
":tokenizer": "$(rlocationpath @Meta-Llama-3.2-1B-Instruct//:tokenizer.json)",
},
tags = [
"no_ci",
"manual",
],
template = [
"./{}/Llama-3.2-1B-Instruct".format(package_name()),
@ -219,7 +220,7 @@ oci_image(
# entrypoint = ["./{}/Llama-3.2-1B-Instruct".format(package_name())],
entrypoint = ":entrypoint",
tags = [
"no_ci",
"manual",
],
tars = [
"@zml//runtimes:layers",
@ -231,7 +232,7 @@ platform_transition_filegroup(
name = "image",
srcs = [":image_"],
tags = [
"no_ci",
"manual",
],
target_platform = "@zml//platforms:linux_amd64",
)
@ -243,7 +244,7 @@ oci_load(
"distroless/llama-3.2-1b-instruct:latest",
],
tags = [
"no_ci",
"manual",
],
)
@ -253,6 +254,6 @@ oci_push(
remote_tags = ["latest"],
repository = "index.docker.io/steeve/llama-3.2-1b-instruct",
tags = [
"no_ci",
"manual",
],
)

View File

@ -1,3 +1,4 @@
load("@rules_cc//cc:cc_binary.bzl", "cc_binary")
load("@zml//bazel:zig.bzl", "zig_cc_binary")
zig_cc_binary(
@ -24,6 +25,9 @@ cc_binary(
"@ModernBERT-base//:model.safetensors",
"@ModernBERT-base//:tokenizer",
],
tags = [
"manual",
],
deps = [":modernbert_lib"],
)
@ -39,6 +43,9 @@ cc_binary(
"@ModernBERT-large//:model.safetensors",
"@ModernBERT-large//:tokenizer",
],
tags = [
"manual",
],
deps = [":modernbert_lib"],
)
@ -53,7 +60,7 @@ zig_cc_binary(
],
main = "test.zig",
tags = [
"no_ci",
"manual",
],
deps = [
"@com_github_hejsil_clap//:clap",