Tags: MyselfTry/llama.cpp
Tags
llama : deci : support ffn-free with attention (ggml-org#13296)
common : Add a warning when we can't match samplers from a string or … …char. (ggml-org#13330)
cuda : remove nrows_x in mul_mat_q_process_tile (ggml-org#13325) Signed-off-by: Xiaodong Ye <xiaodong.ye@mthreads.com>
llama : support tie embedding for chatglm models (ggml-org#13328)
CUDA: mix virt/real CUDA archs for GGML_NATIVE=OFF (ggml-org#13135)
clip : refactor graph builder (ggml-org#13321) * mtmd : refactor graph builder * fix qwen2vl * clean up siglip cgraph * pixtral migrated * move minicpmv to a dedicated build function * move max_feature_layer to build_llava * use build_attn for minicpm resampler * fix windows build * add comment for batch_size * also support tinygemma3 test model * qwen2vl does not use RMS norm * fix qwen2vl norm (2)
PreviousNext