mirror of
https://github.com/microsoft/BitNet.git
synced 2026-05-06 04:40:52 +00:00
add fc3 support
This commit is contained in:
+4
-1
@@ -19,6 +19,9 @@ SUPPORTED_HF_MODELS = {
|
||||
},
|
||||
"HF1BitLLM/Llama3-8B-1.58-100B-tokens": {
|
||||
"model_name": "Llama3-8B-1.58-100B-tokens",
|
||||
},
|
||||
"tiiuae/falcon3-7b-1.58bit": {
|
||||
"model_name": "falcon3-7b-1.58bit",
|
||||
}
|
||||
}
|
||||
|
||||
@@ -149,7 +152,7 @@ def gen_code():
|
||||
shutil.copyfile(os.path.join(pretuned_kernels, "bitnet-lut-kernels-tl2.h"), "include/bitnet-lut-kernels.h")
|
||||
if get_model_name() == "bitnet_b1_58-large":
|
||||
run_command([sys.executable, "utils/codegen_tl2.py", "--model", "bitnet_b1_58-large", "--BM", "256,128,256", "--BK", "96,192,96", "--bm", "32,32,32"], log_step="codegen")
|
||||
elif get_model_name() == "Llama3-8B-1.58-100B-tokens":
|
||||
elif get_model_name() in ["Llama3-8B-1.58-100B-tokens", "falcon3-7b-1.58bit"]:
|
||||
run_command([sys.executable, "utils/codegen_tl2.py", "--model", "Llama3-8B-1.58-100B-tokens", "--BM", "256,128,256,128", "--BK", "96,96,96,96", "--bm", "32,32,32,32"], log_step="codegen")
|
||||
elif get_model_name() == "bitnet_b1_58-3B":
|
||||
run_command([sys.executable, "utils/codegen_tl2.py", "--model", "bitnet_b1_58-3B", "--BM", "160,320,320", "--BK", "96,96,96", "--bm", "32,32,32"], log_step="codegen")
|
||||
|
||||
@@ -335,6 +335,8 @@ class Model(ABC):
|
||||
if chkhsh == "9c2227e4dd922002fb81bde4fc02b0483ca4f12911410dee2255e4987644e3f8":
|
||||
# ref: https://huggingface.co/CohereForAI/c4ai-command-r-v01
|
||||
res = "command-r"
|
||||
if chkhsh == "9d032fcbd5501f4a38150912590928bfb36091efb5df11b8e2124b0390e3fb1e":
|
||||
res = "falcon3"
|
||||
|
||||
if res is None:
|
||||
logger.warning("\n")
|
||||
|
||||
Reference in New Issue
Block a user