diff --git a/expose.h b/expose.h
index 5c67570e2..c9d59d25f 100644
--- a/expose.h
+++ b/expose.h
@@ -169,6 +169,7 @@ struct sd_load_model_inputs
const char * vae_filename = nullptr;
const char * lora_filename = nullptr;
const float lora_multiplier = 1.0f;
+ const char * photomaker_filename = nullptr;
const int img_hard_limit = 0;
const int img_soft_limit = 0;
const bool quiet = false;
@@ -180,6 +181,7 @@ struct sd_generation_inputs
const char * negative_prompt = nullptr;
const char * init_images = "";
const char * mask = "";
+ const char * photomaker_image = "";
const bool flip_mask = false;
const float denoising_strength = 0.0f;
const float cfg_scale = 0.0f;
diff --git a/kcpp_sdui.embd b/kcpp_sdui.embd
index e38790274..8ea800f9d 100644
--- a/kcpp_sdui.embd
+++ b/kcpp_sdui.embd
@@ -5,18 +5,18 @@
Stable UI for KoboldCpp
-
-
diff --git a/koboldcpp.py b/koboldcpp.py
index 79369699b..d8fc1c6d6 100644
--- a/koboldcpp.py
+++ b/koboldcpp.py
@@ -279,6 +279,7 @@ class sd_load_model_inputs(ctypes.Structure):
("vae_filename", ctypes.c_char_p),
("lora_filename", ctypes.c_char_p),
("lora_multiplier", ctypes.c_float),
+ ("photomaker_filename", ctypes.c_char_p),
("img_hard_limit", ctypes.c_int),
("img_soft_limit", ctypes.c_int),
("quiet", ctypes.c_bool),
@@ -289,6 +290,7 @@ class sd_generation_inputs(ctypes.Structure):
("negative_prompt", ctypes.c_char_p),
("init_images", ctypes.c_char_p),
("mask", ctypes.c_char_p),
+ ("photomaker_image", ctypes.c_char_p),
("flip_mask", ctypes.c_bool),
("denoising_strength", ctypes.c_float),
("cfg_scale", ctypes.c_float),
@@ -657,6 +659,13 @@ def is_incomplete_utf8_sequence(byte_seq): #note, this will only flag INCOMPLETE
return True #incomplete sequence
return False #invalid sequence, but not incomplete
+def strip_base64_prefix(encoded_data):
+ if not encoded_data:
+ return ""
+ if encoded_data.startswith("data:image"):
+ encoded_data = encoded_data.split(',', 1)[-1]
+ return encoded_data
+
def unpack_to_dir(destpath = ""):
srcpath = os.path.abspath(os.path.dirname(__file__))
cliunpack = False if destpath == "" else True
@@ -1523,7 +1532,7 @@ def generate(genparams, stream_flag=False):
return {"text":outstr,"status":ret.status,"stopreason":ret.stopreason,"prompt_tokens":ret.prompt_tokens, "completion_tokens": ret.completion_tokens}
-def sd_load_model(model_filename,vae_filename,lora_filename,t5xxl_filename,clipl_filename,clipg_filename):
+def sd_load_model(model_filename,vae_filename,lora_filename,t5xxl_filename,clipl_filename,clipg_filename,photomaker_filename):
global args
inputs = sd_load_model_inputs()
inputs.model_filename = model_filename.encode("UTF-8")
@@ -1547,6 +1556,7 @@ def sd_load_model(model_filename,vae_filename,lora_filename,t5xxl_filename,clipl
inputs.t5xxl_filename = t5xxl_filename.encode("UTF-8")
inputs.clipl_filename = clipl_filename.encode("UTF-8")
inputs.clipg_filename = clipg_filename.encode("UTF-8")
+ inputs.photomaker_filename = photomaker_filename.encode("UTF-8")
inputs.img_hard_limit = args.sdclamped
inputs.img_soft_limit = args.sdclampedsoft
inputs = set_backend_props(inputs)
@@ -1617,7 +1627,8 @@ def sd_generate(genparams):
prompt = forced_posprompt
init_images_arr = genparams.get("init_images", [])
init_images = ("" if (not init_images_arr or len(init_images_arr)==0 or not init_images_arr[0]) else init_images_arr[0])
- mask = genparams.get("mask", "")
+ init_images = strip_base64_prefix(init_images)
+ mask = strip_base64_prefix(genparams.get("mask", ""))
flip_mask = genparams.get("inpainting_mask_invert", 0)
denoising_strength = tryparsefloat(genparams.get("denoising_strength", 0.6),0.6)
cfg_scale = tryparsefloat(genparams.get("cfg_scale", 5),5)
@@ -1629,6 +1640,7 @@ def sd_generate(genparams):
seed = random.randint(100000, 999999)
sample_method = genparams.get("sampler_name", "k_euler_a")
clip_skip = tryparseint(genparams.get("clip_skip", -1),-1)
+ photomaker_image = strip_base64_prefix(genparams.get("photomaker_image", ""))
#clean vars
cfg_scale = (1 if cfg_scale < 1 else (25 if cfg_scale > 25 else cfg_scale))
@@ -1642,6 +1654,7 @@ def sd_generate(genparams):
inputs.negative_prompt = negative_prompt.encode("UTF-8")
inputs.init_images = init_images.encode("UTF-8")
inputs.mask = "".encode("UTF-8") if not mask else mask.encode("UTF-8")
+ inputs.photomaker_image = "".encode("UTF-8") if not photomaker_image else photomaker_image.encode("UTF-8")
inputs.flip_mask = flip_mask
inputs.cfg_scale = cfg_scale
inputs.denoising_strength = denoising_strength
@@ -4288,6 +4301,7 @@ def show_gui():
sd_t5xxl_var = ctk.StringVar()
sd_clipl_var = ctk.StringVar()
sd_clipg_var = ctk.StringVar()
+ sd_photomaker_var = ctk.StringVar()
sd_vaeauto_var = ctk.IntVar(value=0)
sd_notile_var = ctk.IntVar(value=0)
sd_clamped_var = ctk.StringVar(value="0")
@@ -5002,13 +5016,12 @@ def show_gui():
makefileentry(images_tab, "Image LoRA (safetensors/gguf):", "Select SD lora file",sd_lora_var, 20, width=280, singlecol=True, filetypes=[("*.safetensors *.gguf", "*.safetensors *.gguf")],tooltiptxt="Select a .safetensors or .gguf SD LoRA model file to be loaded. Should be unquantized!")
makelabelentry(images_tab, "Image LoRA Multiplier:" , sd_loramult_var, 22, 50,padx=290,singleline=True,tooltip="What mutiplier value to apply the SD LoRA with.")
-
-
makefileentry(images_tab, "T5-XXL File:", "Select Optional T5-XXL model file (SD3 or flux)",sd_t5xxl_var, 24, width=280, singlerow=True, filetypes=[("*.safetensors *.gguf","*.safetensors *.gguf")],tooltiptxt="Select a .safetensors t5xxl file to be loaded.")
makefileentry(images_tab, "Clip-L File:", "Select Optional Clip-L model file (SD3 or flux)",sd_clipl_var, 26, width=280, singlerow=True, filetypes=[("*.safetensors *.gguf","*.safetensors *.gguf")],tooltiptxt="Select a .safetensors t5xxl file to be loaded.")
makefileentry(images_tab, "Clip-G File:", "Select Optional Clip-G model file (SD3)",sd_clipg_var, 28, width=280, singlerow=True, filetypes=[("*.safetensors *.gguf","*.safetensors *.gguf")],tooltiptxt="Select a .safetensors t5xxl file to be loaded.")
+ makefileentry(images_tab, "PhotoMaker:", "Select Optional PhotoMaker model file (SDXL)",sd_photomaker_var, 30, width=280, singlerow=True, filetypes=[("*.safetensors *.gguf","*.safetensors *.gguf")],tooltiptxt="PhotoMaker is a model that allows face cloning.\nSelect a .safetensors PhotoMaker file to be loaded (SDXL only).")
- sdvaeitem1,sdvaeitem2,sdvaeitem3 = makefileentry(images_tab, "Image VAE:", "Select Optional SD VAE file",sd_vae_var, 30, width=280, singlerow=True, filetypes=[("*.safetensors *.gguf", "*.safetensors *.gguf")],tooltiptxt="Select a .safetensors or .gguf SD VAE file to be loaded.")
+ sdvaeitem1,sdvaeitem2,sdvaeitem3 = makefileentry(images_tab, "Image VAE:", "Select Optional SD VAE file",sd_vae_var, 40, width=280, singlerow=True, filetypes=[("*.safetensors *.gguf", "*.safetensors *.gguf")],tooltiptxt="Select a .safetensors or .gguf SD VAE file to be loaded.")
def toggletaesd(a,b,c):
if sd_vaeauto_var.get()==1:
sdvaeitem1.grid_remove()
@@ -5019,8 +5032,8 @@ def show_gui():
sdvaeitem1.grid()
sdvaeitem2.grid()
sdvaeitem3.grid()
- makecheckbox(images_tab, "Use TAE SD (AutoFix Broken VAE)", sd_vaeauto_var, 32,command=toggletaesd,tooltiptxt="Replace VAE with TAESD. May fix bad VAE.")
- makecheckbox(images_tab, "No VAE Tiling", sd_notile_var, 34,tooltiptxt="Disables VAE tiling, may not work for large images.")
+ makecheckbox(images_tab, "Use TAE SD (AutoFix Broken VAE)", sd_vaeauto_var, 42,command=toggletaesd,tooltiptxt="Replace VAE with TAESD. May fix bad VAE.")
+ makecheckbox(images_tab, "No VAE Tiling", sd_notile_var, 44,tooltiptxt="Disables VAE tiling, may not work for large images.")
# audio tab
audio_tab = tabcontent["Audio"]
@@ -5268,6 +5281,8 @@ def show_gui():
args.sdclipl = sd_clipl_var.get()
if sd_clipg_var.get() != "":
args.sdclipg = sd_clipg_var.get()
+ if sd_photomaker_var.get() != "":
+ args.sdphotomaker = sd_photomaker_var.get()
if sd_quant_var.get()==1:
args.sdquant = True
if sd_lora_var.get() != "":
@@ -5471,6 +5486,7 @@ def show_gui():
sd_t5xxl_var.set(dict["sdt5xxl"] if ("sdt5xxl" in dict and dict["sdt5xxl"]) else "")
sd_clipl_var.set(dict["sdclipl"] if ("sdclipl" in dict and dict["sdclipl"]) else "")
sd_clipg_var.set(dict["sdclipg"] if ("sdclipg" in dict and dict["sdclipg"]) else "")
+ sd_photomaker_var.set(dict["sdphotomaker"] if ("sdphotomaker" in dict and dict["sdphotomaker"]) else "")
sd_vaeauto_var.set(1 if ("sdvaeauto" in dict and dict["sdvaeauto"]) else 0)
sd_notile_var.set(1 if ("sdnotile" in dict and dict["sdnotile"]) else 0)
sd_lora_var.set(dict["sdlora"] if ("sdlora" in dict and dict["sdlora"]) else "")
@@ -6509,6 +6525,10 @@ def kcpp_main_process(launch_args, g_memory=None, gui_launcher=False):
dlfile = download_model_from_url(args.sdclipg,[".gguf",".safetensors"],min_file_size=500000)
if dlfile:
args.sdclipg = dlfile
+ if args.sdphotomaker and args.sdphotomaker!="":
+ dlfile = download_model_from_url(args.sdphotomaker,[".gguf",".safetensors"],min_file_size=500000)
+ if dlfile:
+ args.sdphotomaker = dlfile
if args.sdvae and args.sdvae!="":
dlfile = download_model_from_url(args.sdvae,[".gguf",".safetensors"],min_file_size=500000)
if dlfile:
@@ -6785,6 +6805,7 @@ def kcpp_main_process(launch_args, g_memory=None, gui_launcher=False):
imgt5xxl = ""
imgclipl = ""
imgclipg = ""
+ imgphotomaker = ""
if args.sdlora:
if os.path.exists(args.sdlora):
imglora = os.path.abspath(args.sdlora)
@@ -6810,13 +6831,18 @@ def kcpp_main_process(launch_args, g_memory=None, gui_launcher=False):
imgclipg = os.path.abspath(args.sdclipg)
else:
print("Missing SD Clip-G model file...")
+ if args.sdphotomaker:
+ if os.path.exists(args.sdphotomaker):
+ imgphotomaker = os.path.abspath(args.sdphotomaker)
+ else:
+ print("Missing SD Photomaker model file...")
imgmodel = os.path.abspath(imgmodel)
fullsdmodelpath = imgmodel
friendlysdmodelname = os.path.basename(imgmodel)
friendlysdmodelname = os.path.splitext(friendlysdmodelname)[0]
friendlysdmodelname = sanitize_string(friendlysdmodelname)
- loadok = sd_load_model(imgmodel,imgvae,imglora,imgt5xxl,imgclipl,imgclipg)
+ loadok = sd_load_model(imgmodel,imgvae,imglora,imgt5xxl,imgclipl,imgclipg,imgphotomaker)
print("Load Image Model OK: " + str(loadok))
if not loadok:
exitcounter = 999
@@ -7235,6 +7261,7 @@ if __name__ == '__main__':
sdparsergroup.add_argument("--sdt5xxl", metavar=('[filename]'), help="Specify a T5-XXL safetensors model for use in SD3 or Flux. Leave blank if prebaked or unused.", default="")
sdparsergroup.add_argument("--sdclipl", metavar=('[filename]'), help="Specify a Clip-L safetensors model for use in SD3 or Flux. Leave blank if prebaked or unused.", default="")
sdparsergroup.add_argument("--sdclipg", metavar=('[filename]'), help="Specify a Clip-G safetensors model for use in SD3. Leave blank if prebaked or unused.", default="")
+ sdparsergroup.add_argument("--sdphotomaker", metavar=('[filename]'), help="PhotoMaker is a model that allows face cloning. Specify a PhotoMaker safetensors model which will be applied replacing img2img. SDXL models only. Leave blank if unused.", default="")
sdparsergroupvae = sdparsergroup.add_mutually_exclusive_group()
sdparsergroupvae.add_argument("--sdvae", metavar=('[filename]'), help="Specify an image generation safetensors VAE which replaces the one in the model.", default="")
sdparsergroupvae.add_argument("--sdvaeauto", help="Uses a built-in VAE via TAE SD, which is very fast, and fixed bad VAEs.", action='store_true')
diff --git a/otherarch/sdcpp/conditioner.hpp b/otherarch/sdcpp/conditioner.hpp
index a8769649e..59538f683 100644
--- a/otherarch/sdcpp/conditioner.hpp
+++ b/otherarch/sdcpp/conditioner.hpp
@@ -597,6 +597,7 @@ struct FrozenCLIPEmbedderWithCustomWords : public Conditioner {
GGML_ASSERT(it != tokens.end()); // prompt must have trigger word
tokens.erase(it);
return decode(tokens);
+ //return prompt; //kcpp we don't care about photomaker trigger words
}
SDCondition get_learned_condition(ggml_context* work_ctx,
diff --git a/otherarch/sdcpp/sdtype_adapter.cpp b/otherarch/sdcpp/sdtype_adapter.cpp
index 2bf44babf..cfec7efc8 100644
--- a/otherarch/sdcpp/sdtype_adapter.cpp
+++ b/otherarch/sdcpp/sdtype_adapter.cpp
@@ -57,7 +57,7 @@ struct SDParams {
std::string controlnet_path;
std::string embeddings_path;
std::string stacked_id_embeddings_path;
- std::string input_id_images_path;
+ std::string input_id_images_path = "";
sd_type_t wtype = SD_TYPE_COUNT;
std::string lora_model_dir;
std::string output_path = "output.png";
@@ -116,6 +116,7 @@ static int sddebugmode = 0;
static std::string recent_data = "";
static uint8_t * input_image_buffer = NULL;
static uint8_t * input_mask_buffer = NULL;
+static uint8_t * input_photomaker_buffer = NULL;
static std::string sdplatformenv, sddeviceenv, sdvulkandeviceenv;
static bool notiling = false;
@@ -134,6 +135,7 @@ bool sdtype_load_model(const sd_load_model_inputs inputs) {
std::string t5xxl_filename = inputs.t5xxl_filename;
std::string clipl_filename = inputs.clipl_filename;
std::string clipg_filename = inputs.clipg_filename;
+ std::string photomaker_filename = inputs.photomaker_filename;
notiling = inputs.notile;
cfg_side_limit = inputs.img_hard_limit;
cfg_square_limit = inputs.img_soft_limit;
@@ -164,6 +166,10 @@ bool sdtype_load_model(const sd_load_model_inputs inputs) {
{
printf("With Custom Clip-G Model: %s\n",clipg_filename.c_str());
}
+ if(photomaker_filename!="")
+ {
+ printf("With PhotoMaker Model: %s\n",photomaker_filename.c_str());
+ }
if(inputs.quant)
{
printf("Note: Loading a pre-quantized model is always faster than using compress weights!\n");
@@ -205,6 +211,7 @@ bool sdtype_load_model(const sd_load_model_inputs inputs) {
sd_params->t5xxl_path = t5xxl_filename;
sd_params->clip_l_path = clipl_filename;
sd_params->clip_g_path = clipg_filename;
+ sd_params->stacked_id_embeddings_path = photomaker_filename;
//if t5 is set, and model is a gguf, load it as a diffusion model path
bool endswithgguf = (sd_params->model_path.rfind(".gguf") == sd_params->model_path.size() - 5);
if(sd_params->t5xxl_path!="" && endswithgguf)
@@ -423,6 +430,7 @@ sd_generation_outputs sdtype_generate(const sd_generation_inputs inputs)
std::string cleannegprompt = clean_input_prompt(inputs.negative_prompt);
std::string img2img_data = std::string(inputs.init_images);
std::string img2img_mask = std::string(inputs.mask);
+ std::string photomaker_image_data = std::string(inputs.photomaker_image);
std::string sampler = inputs.sample_method;
sd_params->prompt = cleanprompt;
@@ -490,15 +498,17 @@ sd_generation_outputs sdtype_generate(const sd_generation_inputs inputs)
//for img2img
sd_image_t input_image = {0,0,0,nullptr};
+ sd_image_t photomaker_reference = {0,0,0,nullptr};
std::vector image_buffer;
std::vector image_mask_buffer;
+ std::vector photomaker_buffer;
int nx, ny, nc;
- int nx2, ny2, nc2;
int img2imgW = sd_params->width; //for img2img input
int img2imgH = sd_params->height;
int img2imgC = 3; // Assuming RGB image
std::vector resized_image_buf(img2imgW * img2imgH * img2imgC);
std::vector resized_mask_buf(img2imgW * img2imgH * img2imgC);
+ std::vector resized_photomaker_buf(img2imgW * img2imgH * img2imgC);
std::string ts = get_timestamp_str();
if(!sd_is_quiet)
@@ -543,6 +553,35 @@ sd_generation_outputs sdtype_generate(const sd_generation_inputs inputs)
sd_params->sample_method = sample_method_t::EULER_A;
}
+ if(photomaker_image_data!="")
+ {
+ if(input_photomaker_buffer!=nullptr) //just in time free old buffer
+ {
+ stbi_image_free(input_photomaker_buffer);
+ input_photomaker_buffer = nullptr;
+ }
+ int nx2, ny2, nc2;
+ photomaker_buffer = kcpp_base64_decode(photomaker_image_data);
+ input_photomaker_buffer = stbi_load_from_memory(photomaker_buffer.data(), photomaker_buffer.size(), &nx2, &ny2, &nc2, 1);
+ // Resize the image
+ int resok = stbir_resize_uint8(input_photomaker_buffer, nx2, ny2, 0, resized_photomaker_buf.data(), img2imgW, img2imgH, 0, 1);
+ if (!resok) {
+ printf("\nKCPP SD: resize photomaker image failed!\n");
+ output.data = "";
+ output.status = 0;
+ return output;
+ }
+ photomaker_reference.width = img2imgW;
+ photomaker_reference.height = img2imgH;
+ photomaker_reference.channel = img2imgC;
+ photomaker_reference.data = resized_photomaker_buf.data();
+
+ //ensure prompt has img keyword, otherwise append it
+ if (sd_params->prompt.find("img") == std::string::npos) {
+ sd_params->prompt += " img";
+ }
+ }
+
if (sd_params->mode == TXT2IMG) {
if(!sd_is_quiet && sddebugmode==1)
@@ -585,7 +624,8 @@ sd_generation_outputs sdtype_generate(const sd_generation_inputs inputs)
sd_params->skip_layers.size(),
sd_params->slg_scale,
sd_params->skip_layer_start,
- sd_params->skip_layer_end);
+ sd_params->skip_layer_end,
+ (photomaker_image_data!=""?&photomaker_reference:nullptr));
} else {
if (sd_params->width <= 0 || sd_params->width % 64 != 0 || sd_params->height <= 0 || sd_params->height % 64 != 0) {
@@ -596,18 +636,11 @@ sd_generation_outputs sdtype_generate(const sd_generation_inputs inputs)
}
image_buffer = kcpp_base64_decode(img2img_data);
-
if(input_image_buffer!=nullptr) //just in time free old buffer
{
stbi_image_free(input_image_buffer);
input_image_buffer = nullptr;
}
- if(input_mask_buffer!=nullptr) //just in time free old buffer
- {
- stbi_image_free(input_mask_buffer);
- input_mask_buffer = nullptr;
- }
-
input_image_buffer = stbi_load_from_memory(image_buffer.data(), image_buffer.size(), &nx, &ny, &nc, 3);
if (nx < 64 || ny < 64 || nx > 1024 || ny > 1024 || nc!= 3) {
@@ -634,6 +667,12 @@ sd_generation_outputs sdtype_generate(const sd_generation_inputs inputs)
if(img2img_mask!="")
{
+ int nx2, ny2, nc2;
+ if(input_mask_buffer!=nullptr) //just in time free old buffer
+ {
+ stbi_image_free(input_mask_buffer);
+ input_mask_buffer = nullptr;
+ }
image_mask_buffer = kcpp_base64_decode(img2img_mask);
input_mask_buffer = stbi_load_from_memory(image_mask_buffer.data(), image_mask_buffer.size(), &nx2, &ny2, &nc2, 1);
// Resize the image
@@ -709,7 +748,8 @@ sd_generation_outputs sdtype_generate(const sd_generation_inputs inputs)
sd_params->skip_layers.size(),
sd_params->slg_scale,
sd_params->skip_layer_start,
- sd_params->skip_layer_end);
+ sd_params->skip_layer_end,
+ (photomaker_image_data!=""?&photomaker_reference:nullptr));
}
if (results == NULL) {
diff --git a/otherarch/sdcpp/stable-diffusion.cpp b/otherarch/sdcpp/stable-diffusion.cpp
index 6475b926c..45de5834c 100644
--- a/otherarch/sdcpp/stable-diffusion.cpp
+++ b/otherarch/sdcpp/stable-diffusion.cpp
@@ -328,7 +328,7 @@ public:
LOG_WARN(
"!!!It looks like you are using SDXL model. "
"If you find that the generated images are completely black, "
- "try specifying SDXL VAE FP16 Fix with the --vae parameter. "
+ "try specifying a different VAE. "
"You can find it here: https://huggingface.co/madebyollin/sdxl-vae-fp16-fix/blob/main/sdxl_vae.safetensors");
}
} else if (sd_version_is_sd3(version)) {
@@ -1408,7 +1408,8 @@ sd_image_t* generate_image(sd_ctx_t* sd_ctx,
float slg_scale = 0,
float skip_layer_start = 0.01,
float skip_layer_end = 0.2,
- ggml_tensor* masked_image = NULL) {
+ ggml_tensor* masked_image = NULL,
+ const sd_image_t* photomaker_reference = nullptr) {
if (seed < 0) {
// Generally, when using the provided command line, the seed is always >0.
// However, to prevent potential issues if 'stable-diffusion.cpp' is invoked as a library
@@ -1451,6 +1452,10 @@ sd_image_t* generate_image(sd_ctx_t* sd_ctx,
ggml_tensor* init_img = NULL;
SDCondition id_cond;
std::vector class_tokens_mask;
+ if (sd_ctx->sd->pmid_model && photomaker_reference!=nullptr)
+ {
+ sd_ctx->sd->stacked_id = true; //turn on photomaker if needed
+ }
if (sd_ctx->sd->stacked_id) {
if (!sd_ctx->sd->pmid_lora->applied) {
t0 = ggml_time_ms();
@@ -1493,6 +1498,30 @@ sd_image_t* generate_image(sd_ctx_t* sd_ctx,
input_id_images.push_back(input_image);
}
}
+
+ // handle single photomaker image passed in by kcpp
+ if (sd_ctx->sd->pmid_model && photomaker_reference!=nullptr)
+ {
+ int c = 0;
+ int width, height;
+ width = photomaker_reference->width;
+ height = photomaker_reference->height;
+ c = photomaker_reference->channel;
+ uint8_t* input_image_buffer = photomaker_reference->data;
+ sd_image_t* input_image = NULL;
+ input_image = new sd_image_t{(uint32_t)width,
+ (uint32_t)height,
+ 3,
+ input_image_buffer};
+ input_image = preprocess_id_image(input_image);
+ if (input_image == NULL) {
+ LOG_ERROR("\npreprocess input id image from kcpp photomaker failed\n");
+ } else {
+ LOG_INFO("\nPhotoMaker loaded image from kcpp\n");
+ input_id_images.push_back(input_image);
+ }
+ }
+
if (input_id_images.size() > 0) {
sd_ctx->sd->pmid_model->style_strength = style_ratio;
int32_t w = input_id_images[0]->width;
@@ -1744,7 +1773,8 @@ sd_image_t* txt2img(sd_ctx_t* sd_ctx,
size_t skip_layers_count = 0,
float slg_scale = 0,
float skip_layer_start = 0.01,
- float skip_layer_end = 0.2) {
+ float skip_layer_end = 0.2,
+ const sd_image_t* photomaker_reference = nullptr) {
std::vector skip_layers_vec(skip_layers, skip_layers + skip_layers_count);
LOG_DEBUG("txt2img %dx%d", width, height);
if (sd_ctx == NULL) {
@@ -1822,7 +1852,9 @@ sd_image_t* txt2img(sd_ctx_t* sd_ctx,
skip_layers_vec,
slg_scale,
skip_layer_start,
- skip_layer_end);
+ skip_layer_end,
+ nullptr,
+ photomaker_reference);
size_t t1 = ggml_time_ms();
@@ -1856,7 +1888,8 @@ sd_image_t* img2img(sd_ctx_t* sd_ctx,
size_t skip_layers_count = 0,
float slg_scale = 0,
float skip_layer_start = 0.01,
- float skip_layer_end = 0.2) {
+ float skip_layer_end = 0.2,
+ const sd_image_t* photomaker_reference = nullptr) {
std::vector skip_layers_vec(skip_layers, skip_layers + skip_layers_count);
LOG_DEBUG("img2img %dx%d", width, height);
if (sd_ctx == NULL) {
@@ -2002,7 +2035,8 @@ sd_image_t* img2img(sd_ctx_t* sd_ctx,
slg_scale,
skip_layer_start,
skip_layer_end,
- masked_image);
+ masked_image,
+ photomaker_reference);
size_t t2 = ggml_time_ms();