mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2026-05-09 19:46:11 +00:00
# Conflicts: # .devops/cann.Dockerfile # .devops/cpu.Dockerfile # .devops/cuda-new.Dockerfile # .devops/cuda.Dockerfile # .devops/intel.Dockerfile # .devops/llama-cli-cann.Dockerfile # .devops/musa.Dockerfile # .devops/nix/package.nix # .devops/rocm.Dockerfile # .devops/s390x.Dockerfile # .devops/vulkan.Dockerfile # .github/workflows/build-cmake-pkg.yml # .github/workflows/build-linux-cross.yml # .github/workflows/build.yml # .github/workflows/copilot-setup-steps.yml # .github/workflows/release.yml # .github/workflows/server-webui.yml # .github/workflows/server.yml # CMakeLists.txt # README.md # build-xcframework.sh # ci/run.sh # cmake/common.cmake # common/CMakeLists.txt # docs/backend/hexagon/CMakeUserPresets.json # docs/backend/hexagon/README.md # docs/build-riscv64-spacemit.md # docs/build.md # examples/debug/debug.cpp # examples/eval-callback/CMakeLists.txt # examples/eval-callback/eval-callback.cpp # examples/llama.android/lib/build.gradle.kts # examples/sycl/build.sh # examples/sycl/win-build-sycl.bat # ggml/src/ggml-hexagon/ggml-hexagon.cpp # ggml/src/ggml-hexagon/htp/CMakeLists.txt # ggml/src/ggml-hexagon/htp/act-ops.c # ggml/src/ggml-hexagon/htp/binary-ops.c # ggml/src/ggml-hexagon/htp/flash-attn-ops.c # ggml/src/ggml-hexagon/htp/get-rows-ops.c # ggml/src/ggml-hexagon/htp/hex-dma.c # ggml/src/ggml-hexagon/htp/hex-dma.h # ggml/src/ggml-hexagon/htp/htp-ctx.h # ggml/src/ggml-hexagon/htp/htp-msg.h # ggml/src/ggml-hexagon/htp/htp-ops.h # ggml/src/ggml-hexagon/htp/hvx-utils.h # ggml/src/ggml-hexagon/htp/main.c # ggml/src/ggml-hexagon/htp/matmul-ops.c # ggml/src/ggml-hexagon/htp/rope-ops.c # ggml/src/ggml-hexagon/htp/set-rows-ops.c # ggml/src/ggml-hexagon/htp/softmax-ops.c # ggml/src/ggml-hexagon/htp/unary-ops.c # ggml/src/ggml-hexagon/htp/worker-pool.c # scripts/debug-test.sh # scripts/serve-static.js # scripts/snapdragon/adb/run-bench.sh # scripts/snapdragon/adb/run-cli.sh # scripts/snapdragon/adb/run-mtmd.sh # scripts/snapdragon/adb/run-tool.sh # scripts/tool_bench.py # tests/CMakeLists.txt # tests/test-backend-ops.cpp # tools/mtmd/clip.cpp
131 lines
5.2 KiB
C
131 lines
5.2 KiB
C
#pragma once
|
|
|
|
#include "ggml.h"
|
|
#include "mtmd.h"
|
|
|
|
#include <stddef.h>
|
|
#include <stdint.h>
|
|
|
|
// !!! Internal header, to be used by mtmd only !!!
|
|
|
|
#define MTMD_INTERNAL_HEADER
|
|
|
|
struct clip_ctx;
|
|
|
|
struct clip_image_size {
|
|
int width;
|
|
int height;
|
|
};
|
|
|
|
struct clip_image_f32;
|
|
struct clip_image_u8_batch;
|
|
struct clip_image_f32_batch;
|
|
|
|
enum clip_modality {
|
|
CLIP_MODALITY_VISION,
|
|
CLIP_MODALITY_AUDIO,
|
|
};
|
|
|
|
enum clip_flash_attn_type {
|
|
CLIP_FLASH_ATTN_TYPE_AUTO = -1,
|
|
CLIP_FLASH_ATTN_TYPE_DISABLED = 0,
|
|
CLIP_FLASH_ATTN_TYPE_ENABLED = 1,
|
|
};
|
|
|
|
struct clip_context_params {
|
|
bool use_gpu;
|
|
enum clip_flash_attn_type flash_attn_type;
|
|
int image_min_tokens;
|
|
int image_max_tokens;
|
|
bool warmup;
|
|
ggml_backend_sched_eval_callback cb_eval;
|
|
void * cb_eval_user_data;
|
|
};
|
|
|
|
struct clip_init_result {
|
|
struct clip_ctx * ctx_v; // vision context
|
|
struct clip_ctx * ctx_a; // audio context
|
|
};
|
|
|
|
struct clip_init_result clip_init(const char * fname, struct clip_context_params ctx_params);
|
|
|
|
void clip_free(struct clip_ctx * ctx);
|
|
|
|
size_t clip_embd_nbytes(const struct clip_ctx * ctx);
|
|
size_t clip_embd_nbytes_by_img(const struct clip_ctx * ctx, int img_w, int img_h);
|
|
|
|
int32_t clip_get_image_size (const struct clip_ctx * ctx);
|
|
int32_t clip_get_patch_size (const struct clip_ctx * ctx);
|
|
int32_t clip_get_hidden_size(const struct clip_ctx * ctx);
|
|
|
|
// TODO: should be enum, not string
|
|
const char * clip_patch_merge_type(const struct clip_ctx * ctx);
|
|
|
|
int clip_n_output_tokens(const struct clip_ctx * ctx, struct clip_image_f32 * img);
|
|
|
|
// for M-RoPE, this will be the number of token positions in X and Y directions
|
|
// for other models, X will be the total number of tokens and Y will be 1
|
|
int clip_n_output_tokens_x(const struct clip_ctx * ctx, struct clip_image_f32 * img);
|
|
int clip_n_output_tokens_y(const struct clip_ctx * ctx, struct clip_image_f32 * img);
|
|
|
|
// this should be equal to the embedding dimension of the text model
|
|
int clip_n_mmproj_embd(const struct clip_ctx * ctx);
|
|
|
|
struct clip_image_size * clip_image_size_init(void);
|
|
struct clip_image_u8 * clip_image_u8_init (void);
|
|
struct clip_image_f32 * clip_image_f32_init(void);
|
|
struct clip_image_f32_batch * clip_image_f32_batch_init(void); // only used by libllava
|
|
|
|
// nx, ny are the output image dimensions
|
|
unsigned char * clip_image_u8_get_data(struct clip_image_u8 * img, uint32_t * nx, uint32_t * ny);
|
|
|
|
void clip_image_size_free (struct clip_image_size * img_size);
|
|
void clip_image_u8_free (struct clip_image_u8 * img);
|
|
void clip_image_f32_free(struct clip_image_f32 * img);
|
|
void clip_image_u8_batch_free (struct clip_image_u8_batch * batch);
|
|
void clip_image_f32_batch_free(struct clip_image_f32_batch * batch);
|
|
|
|
// use for accessing underlay data of clip_image_f32_batch
|
|
size_t clip_image_f32_batch_n_images(const struct clip_image_f32_batch * batch); // equivalent to batch->size()
|
|
size_t clip_image_f32_batch_nx(const struct clip_image_f32_batch * batch, int idx); // equivalent to batch[idx]->nx
|
|
size_t clip_image_f32_batch_ny(const struct clip_image_f32_batch * batch, int idx); // equivalent to batch[idx]->ny
|
|
struct clip_image_f32 * clip_image_f32_get_img(const struct clip_image_f32_batch * batch, int idx); // equivalent to batch[idx]->data
|
|
|
|
/**
|
|
* Build image from pixels decoded by other libraries instead of stb_image.h for better performance.
|
|
* The memory layout is RGBRGBRGB..., input buffer length must be 3*nx*ny bytes
|
|
*/
|
|
void clip_build_img_from_pixels(const unsigned char * rgb_pixels, int nx, int ny, struct clip_image_u8 * img);
|
|
|
|
bool clip_image_load_from_file(const char * fname, struct clip_image_u8 * img);
|
|
|
|
/** interpret bytes as an image file with length bytes_length, and use the result to populate img */
|
|
bool clip_image_load_from_bytes(const unsigned char * bytes, size_t bytes_length, struct clip_image_u8 * img, const int maxdims);
|
|
|
|
/** preprocess img and store the result in res_imgs, pad_to_square may be overridden to false depending on model configuration */
|
|
bool clip_image_preprocess(struct clip_ctx * ctx, const struct clip_image_u8 * img, struct clip_image_f32_batch * res_imgs );
|
|
|
|
struct ggml_tensor * clip_get_newline_tensor(const struct clip_ctx * ctx);
|
|
|
|
bool clip_image_encode (struct clip_ctx * ctx, int n_threads, struct clip_image_f32 * img, float * vec);
|
|
bool clip_image_batch_encode(struct clip_ctx * ctx, int n_threads, const struct clip_image_f32_batch * imgs, float * vec);
|
|
|
|
int clip_is_minicpmv(const struct clip_ctx * ctx);
|
|
bool clip_is_glm(const struct clip_ctx * ctx);
|
|
bool clip_is_mrope(const struct clip_ctx * ctx);
|
|
bool clip_is_llava(const struct clip_ctx * ctx);
|
|
bool clip_is_gemma3(const struct clip_ctx * ctx);
|
|
bool clip_is_pixtral(const struct clip_ctx * ctx);
|
|
void set_clip_uses_gpu(bool usegpu);
|
|
int clip_get_projector_type_ext(clip_ctx * ctx);
|
|
|
|
bool clip_encode_float_image (struct clip_ctx * ctx, int n_threads, float * img, int h, int w, float * vec);
|
|
|
|
// use by audio input
|
|
void clip_image_f32_batch_add_mel(struct clip_image_f32_batch * batch, int n_mel, int n_frames, float * mel);
|
|
|
|
bool clip_has_vision_encoder(const struct clip_ctx * ctx);
|
|
bool clip_has_audio_encoder(const struct clip_ctx * ctx);
|
|
bool clip_has_whisper_encoder(const struct clip_ctx * ctx);
|
|
|
|
bool clip_model_quantize(const char * fname_inp, const char * fname_out, const int itype) ;
|