ynankani
|
42928bc14d
|
model : NvFP4 quantized LM head support (#23046)
* NvFP4 quantized LM head support
Signed-off-by: ynankani <ynankani@nvidia.com>
* Address review commnets
Signed-off-by: ynankani <ynankani@nvidia.com>
* Add assert for NvFp4 lm head and tied embeddings
Signed-off-by: ynankani <ynankani@nvidia.com>
* Address review commnets
Signed-off-by: ynankani <ynankani@nvidia.com>
* Create output_s tensor only when LM head NvFp4
Signed-off-by: ynankani <ynankani@nvidia.com>
---------
Signed-off-by: ynankani <ynankani@nvidia.com>
|
2026-05-16 11:09:27 +02:00 |
|
Xuan-Son Nguyen
|
994118a183
|
model: move load_hparams and load_tensors to per-model definition (#22004)
* git-friendly migration
* add build_graph
* nits
* exclude old code from build
* wip
* add llm_arch_model_i
* prepare downstream functions
* nits
* nits
* wip
* wip
* add back create_tensor_qkv
* fix files missing include
* enforce one llm_build per arch
* cmake: use glob
* missing model params
* nits
* wip
* wip (2)
* wip (3)
* test-llama-archs is happy
* improve switch case
* move more stuff into llm_arch_model_i
* fix downstream code
* nits
* nits (2)
* fix order
* llama_model_base
* LLAMA_LOAD_LOCALS
* small fix
* fix build errors
* auto
* rm migration script and ifdef
|
2026-05-04 12:36:59 +02:00 |
|
Sigbjørn Skjæret
|
f772f6e434
|
model : support NVFP4 tensors for Gemma4 (#21971)
* support nvfp4 tensors for Gemma4
* add wo_s to build_attn
* add wo_s to build_attn
* fix glm4
|
2026-04-16 16:51:47 +02:00 |
|
Xuan-Son Nguyen
|
59db9a357d
|
llama: dynamic head_dim and n_rot for SWA (#20301)
* llama: dynamic head_dim and n_rot for SWA
* also add gguf_writer wrappers
* fix build
* build_rope_shift arg reorder
|
2026-03-09 22:22:39 +01:00 |
|
Sigbjørn Skjæret
|
a6fd8ca1fe
|
models : remove unnecessary cont in openelm (#19289)
|
2026-02-03 14:20:57 +01:00 |
|
Sigbjørn Skjæret
|
b164259bba
|
chore : fix models indent after refactor (#16992)
|
2025-11-04 12:29:15 +01:00 |
|
Piotr Wilkin (ilintar)
|
bea04522ff
|
refactor : llama-model.cpp (#16252)
* Sqashed: llama-model.cpp refactoring
* Fix formatting of attn / ffn / ffn_moe calls
* Fix import regression / unify spacing in models.h
* totally DID NOT miss those!
* Add missing qwen3vl(moe) models
* Add missing new .cpp files to build
* Remove extra semicolons
* Editor checker
* Update src/models/models.h
Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>
---------
Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>
|
2025-10-31 23:40:23 +01:00 |
|