This commit is contained in:
Ali Ghalel 2025-09-03 22:19:33 -04:00 committed by GitHub
commit 736c315920
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
4 changed files with 679 additions and 713 deletions

View File

@ -1 +1 @@
version = '2.5.5'
version = '2.5.6'

View File

@ -1,9 +1,13 @@
import threading
prompts_processed = False
processed_tasks_dictionary = []
enhanced_tasks_dictionary = []
enhanced_clip_encode_dict = {}
clip_encodes = {}
import re
from extras.inpaint_mask import generate_mask_from_image, SAMOptions
from modules.patch import PatchSettings, patch_settings, patch_all
import modules.config
patch_all()
@ -33,9 +37,14 @@ class AsyncTask:
self.performance_selection = Performance(args.pop())
self.steps = self.performance_selection.steps()
self.original_steps = self.steps
self.aspect_ratios_selection = args.pop()
self.image_number = args.pop()
self.output_format = args.pop()
self.seed = int(args.pop())
@ -50,13 +59,17 @@ class AsyncTask:
self.input_image_checkbox = args.pop()
self.current_tab = args.pop()
self.uov_method = args.pop()
self.uov_input_image = args.pop()
self.outpaint_selections = args.pop()
self.inpaint_input_image = args.pop()
self.inpaint_additional_prompt = args.pop()
self.inpaint_mask_image_upload = args.pop()
self.disable_preview = args.pop()
self.disable_intermediate_results = args.pop()
self.disable_seed_increment = args.pop()
self.black_out_nsfw = args.pop()
@ -71,7 +84,11 @@ class AsyncTask:
self.overwrite_step = args.pop()
self.overwrite_switch = args.pop()
self.overwrite_width = args.pop()
self.overwrite_height = args.pop()
self.overwrite_vary_strength = args.pop()
self.overwrite_upscale_strength = args.pop()
self.mixing_image_prompt_and_vary_upscale = args.pop()
@ -90,11 +107,17 @@ class AsyncTask:
self.debugging_inpaint_preprocessor = args.pop()
self.inpaint_disable_initial_latent = args.pop()
self.inpaint_engine = args.pop()
self.inpaint_strength = args.pop()
self.inpaint_respective_field = args.pop()
self.inpaint_advanced_masking_checkbox = args.pop()
self.invert_mask_checkbox = args.pop()
self.inpaint_erode_or_dilate = args.pop()
self.save_final_enhanced_image_only = args.pop() if not args_manager.args.disable_image_log else False
self.save_metadata_to_images = args.pop() if not args_manager.args.disable_metadata else False
self.metadata_scheme = MetadataScheme(
@ -113,28 +136,52 @@ class AsyncTask:
self.dino_erode_or_dilate = args.pop()
self.debugging_enhance_masks_checkbox = args.pop()
self.enhance_input_image = args.pop()
self.enhance_checkbox = args.pop()
self.enhance_uov_method = args.pop()
self.enhance_uov_processing_order = args.pop()
self.enhance_uov_prompt_type = args.pop()
self.enhance_ctrls = []
for _ in range(modules.config.default_enhance_tabs):
enhance_enabled = args.pop()
enhance_mask_dino_prompt_text = args.pop()
enhance_prompt = args.pop()
enhance_negative_prompt = args.pop()
enhance_mask_model = args.pop()
enhance_mask_cloth_category = args.pop()
enhance_mask_sam_model = args.pop()
enhance_mask_text_threshold = args.pop()
enhance_mask_box_threshold = args.pop()
enhance_mask_sam_max_detections = args.pop()
enhance_inpaint_disable_initial_latent = args.pop()
enhance_inpaint_engine = args.pop()
enhance_inpaint_strength = args.pop()
enhance_inpaint_respective_field = args.pop()
enhance_inpaint_erode_or_dilate = args.pop()
enhance_mask_invert = args.pop()
if enhance_enabled:
self.enhance_ctrls.append([
@ -161,6 +208,7 @@ class AsyncTask:
async_tasks = []
class EarlyReturnException(BaseException):
pass
@ -207,12 +255,20 @@ def worker():
try:
async_gradio_app = shared.gradio_root
flag = f'''App started successful. Use the app with {str(async_gradio_app.local_url)} or {str(async_gradio_app.server_name)}:{str(async_gradio_app.server_port)}'''
public_url = async_gradio_app.share_url
if public_url:
with open("/content/public_url.txt", "w") as f:
f.write(public_url)
print(f"Public URL saved: {public_url}")
if async_gradio_app.share:
flag += f''' or {async_gradio_app.share_url}'''
print(flag)
except Exception as e:
print(e)
def debug_print(variable_name, value):
return print(f"[DEBUG] {variable_name}: {value}")
def progressbar(async_task, number, text):
print(f'[Fooocus] {text}')
async_task.yields.append(['preview', (number, text, None)])
@ -228,6 +284,7 @@ def worker():
async_task.results = async_task.results + imgs
if do_not_show_finished_images:
return
async_task.yields.append(['results', async_task.results])
@ -235,6 +292,11 @@ def worker():
def build_image_wall(async_task):
results = []
print(f"async_task.results: {str(async_task.results)}")
for img in async_task.results:
print(f"img: {str(img)}")
print(f"len(async_task.results) : {len(async_task.results)}")
if len(async_task.results) < 2:
return
@ -277,10 +339,27 @@ def worker():
async_task.results = async_task.results + [wall]
return
def process_task(all_steps, async_task, callback, controlnet_canny_path, controlnet_cpds_path, current_task_id,
denoising_strength, final_scheduler_name, goals, initial_latent, steps, switch, positive_cond,
negative_cond, task, loras, tiled, use_expansion, width, height, base_progress, preparation_steps,
total_count, show_intermediate_results, persist_image=True):
global processed_tasks_dictionary
task = processed_tasks_dictionary[current_task_id]
image_seed = task['task_seed'] ##
print(f"[Seed] Using Seed: {image_seed} for task: {current_task_id +1}") ##
##set new height and width:
width = task['overwrite_width'] ##
#width = 1152
height = task['overwrite_height'] ##
#height = 896
print(f"[ASPECT RATIO] using width, height: {width}, {height}") ##
if async_task.last_stop is not False:
ldm_patched.modules.model_management.interrupt_current_processing()
if 'cn' in goals:
@ -299,7 +378,7 @@ def worker():
switch=switch,
width=width,
height=height,
image_seed=task['task_seed'],
image_seed=image_seed ,##
callback=callback,
sampler_name=async_task.sampler_name,
scheduler_name=final_scheduler_name,
@ -319,6 +398,10 @@ def worker():
imgs = default_censor(imgs)
progressbar(async_task, current_progress, f'Saving image {current_task_id + 1}/{total_count} to system ...')
img_paths = save_and_log(async_task, height, imgs, task, use_expansion, width, loras, persist_image)
#debug_print("debug process_task Fun img_paths", img_paths)##
yield_result(async_task, img_paths, current_progress, async_task.black_out_nsfw, False,
do_not_show_finished_images=not show_intermediate_results or async_task.disable_intermediate_results)
@ -393,6 +476,7 @@ def worker():
return img_paths
def apply_control_nets(async_task, height, ip_adapter_face_path, ip_adapter_path, width, current_progress):
for task in async_task.cn_tasks[flags.cn_canny]:
cn_img, cn_stop, cn_weight = task
@ -638,51 +722,89 @@ def worker():
height = async_task.overwrite_height
return steps, switch, width, height
def process_prompt(async_task, prompt, negative_prompt, base_model_additional_loras, image_number, disable_seed_increment, use_expansion, use_style,
use_synthetic_refiner, current_progress, advance_progress=False):
def process_prompts(async_task, prompt, negative_prompt, base_model_additional_loras, image_number, disable_seed_increment, use_expansion, use_style, use_synthetic_refiner, current_progress, advance_progress=False):
global prompts_processed, processed_tasks_dictionary, enhanced_tasks_dictionary
if prompts_processed:
print("[PROMPTS] already been processed!")
loras = async_task.loras + async_task.performance_loras
return processed_tasks_dictionary, enhanced_tasks_dictionary, use_expansion, loras, current_progress, enhanced_tasks_dictionary
# Process main prompts
prompts = remove_empty_str([safe_str(p) for p in prompt.splitlines()], default='')
negative_prompts = remove_empty_str([safe_str(p) for p in negative_prompt.splitlines()], default='')
prompt = prompts[0]
negative_prompt = negative_prompts[0]
if prompt == '':
# disable expansion when empty since it is not meaningful and influences image prompt
use_expansion = False
print("[Fooocus Expansion] Disabled")
extra_positive_prompts = prompts[1:] if len(prompts) > 1 else []
extra_negative_prompts = negative_prompts[1:] if len(negative_prompts) > 1 else []
if advance_progress:
current_progress += 1
progressbar(async_task, current_progress, 'Loading models ...')
lora_filenames = modules.util.remove_performance_lora(modules.config.lora_filenames,
async_task.performance_selection)
loras, prompt = parse_lora_references_from_prompt(prompt, async_task.loras,
modules.config.default_max_lora_number,
lora_filenames=lora_filenames)
loras += async_task.performance_loras
pipeline.refresh_everything(refiner_model_name=async_task.refiner_model_name,
base_model_name=async_task.base_model_name,
loras=loras, base_model_additional_loras=base_model_additional_loras,
use_synthetic_refiner=use_synthetic_refiner, vae_name=async_task.vae_name)
loras = async_task.loras + async_task.performance_loras
pipeline.set_clip_skip(async_task.clip_skip)
if advance_progress:
current_progress += 1
progressbar(async_task, current_progress, 'Processing prompts ...')
tasks = []
wildcard_start_row = {}
if '__' in prompt:
matches = re.findall(r'__([\w\.\(\)ぁ-んァ-ヶ一-龯ー々・-]+?)__(\d+)_', prompt)
debug_print("matches", matches)
for wildcard_name, start_row in matches:
debug_print("wildcard_name", wildcard_name)
debug_print("start_row", start_row)
wildcard_start_row[wildcard_name] = int(start_row)
if advance_progress:
current_progress += 1
progressbar(async_task, current_progress, 'Loading models ...')
lora_filenames = modules.util.remove_performance_lora(modules.config.lora_filenames, async_task.performance_selection)
processed_tasks_dictionary = []
enhanced_tasks_dictionary = []
for i in range(image_number):
if disable_seed_increment:
task_seed = async_task.seed % (constants.MAX_SEED + 1)
else:
task_seed = (async_task.seed + i) % (constants.MAX_SEED + 1) # randint is inclusive, % is not
task_seed = (async_task.seed + i) % (constants.MAX_SEED + 1)
task_rng = random.Random(task_seed) # may bind to inpaint noise in the future
task_prompt = apply_wildcards(prompt, task_rng, i, async_task.read_wildcards_in_order)
task_rng = random.Random(task_seed)
task_prompt = apply_wildcards(prompt, task_rng, i, async_task.read_wildcards_in_order, wildcard_start_row=wildcard_start_row)
task_prompt = apply_arrays(task_prompt, i)
task_negative_prompt = apply_wildcards(negative_prompt, task_rng, i, async_task.read_wildcards_in_order)
task_extra_positive_prompts = [apply_wildcards(pmt, task_rng, i, async_task.read_wildcards_in_order) for pmt
in
extra_positive_prompts]
task_extra_negative_prompts = [apply_wildcards(pmt, task_rng, i, async_task.read_wildcards_in_order) for pmt
in
extra_negative_prompts]
print(f"[PROMPTS] task_prompt: {task_prompt}")
loras, task_prompt = parse_lora_references_from_prompt(task_prompt, async_task.loras, modules.config.default_max_lora_number, lora_filenames=lora_filenames)
loras += async_task.performance_loras
aspect_ratio_match = re.search(r'(\d+)\s*[×xX*]\s*(\d+)', task_prompt)
if aspect_ratio_match:
overwrite_width = int(aspect_ratio_match.group(1))
overwrite_height = int(aspect_ratio_match.group(2))
print(f"[Aspect Ratio] Using Custom Aspect Ratio: {overwrite_width}×{overwrite_height}")
task_prompt = re.sub(r'(\d+)\s*[×xX*]\s*(\d+)', '', task_prompt).strip().rstrip(',')
else:
match = re.match(r'(\d+)\s*[×xX*]\s*(\d+)', async_task.aspect_ratios_selection)
if match:
print("[Aspect Ratio] Using UI's Selected Resolution")
overwrite_width = int(match.group(1))
overwrite_height = int(match.group(2))
else:
print("[Aspect Ratio] Using Default Aspect Ratio of 896×1152")
overwrite_width = 896
overwrite_height = 1152
task_negative_prompt = apply_wildcards(negative_prompt, task_rng, i, async_task.read_wildcards_in_order, wildcard_start_row=wildcard_start_row)
task_extra_positive_prompts = [apply_wildcards(pmt, task_rng, i, async_task.read_wildcards_in_order, wildcard_start_row=wildcard_start_row) for pmt in extra_positive_prompts]
task_extra_negative_prompts = [apply_wildcards(pmt, task_rng, i, async_task.read_wildcards_in_order, wildcard_start_row=wildcard_start_row) for pmt in extra_negative_prompts]
positive_basic_workloads = []
negative_basic_workloads = []
@ -714,45 +836,200 @@ def worker():
positive_basic_workloads = remove_empty_str(positive_basic_workloads, default=task_prompt)
negative_basic_workloads = remove_empty_str(negative_basic_workloads, default=task_negative_prompt)
tasks.append(dict(
processed_tasks_dictionary.append(dict(
task_seed=task_seed,
task_prompt=task_prompt,
task_negative_prompt=task_negative_prompt,
positive=positive_basic_workloads,
negative=negative_basic_workloads,
expansion='',
c=None,
uc=None,
positive_top_k=len(positive_basic_workloads),
negative_top_k=len(negative_basic_workloads),
log_positive_prompt='\n'.join([task_prompt] + task_extra_positive_prompts),
log_negative_prompt='\n'.join([task_negative_prompt] + task_extra_negative_prompts),
styles=task_styles
styles=async_task.style_selections.copy(),
overwrite_width=overwrite_width,
overwrite_height=overwrite_height,
loras=loras
))
# Process enhance_ctrls prompts
for enhance_ctrl in async_task.enhance_ctrls:
mask_enhance_prompt = enhance_ctrl[1]
mask_enhance_negative_prompt = enhance_ctrl[2]
# Apply wildcards and arrays to enhanced prompts
enhanced_prompt = apply_wildcards(mask_enhance_prompt, task_rng, i, async_task.read_wildcards_in_order, wildcard_start_row=wildcard_start_row)
enhanced_prompt = apply_arrays(enhanced_prompt, i)
enhanced_negative_prompt = apply_wildcards(mask_enhance_negative_prompt, task_rng, i, async_task.read_wildcards_in_order, wildcard_start_row=wildcard_start_row)
# Apply styles to enhanced prompts (if use_style is True)
enhanced_positive_basic_workloads = []
enhanced_negative_basic_workloads = []
if use_style:
placeholder_replaced = False
for j, s in enumerate(task_styles):
if s == random_style_name:
s = get_random_style(task_rng)
task_styles[j] = s
p, n, style_has_placeholder = apply_style(s, positive=enhanced_prompt)
if style_has_placeholder:
placeholder_replaced = True
enhanced_positive_basic_workloads = enhanced_positive_basic_workloads + p
enhanced_negative_basic_workloads = enhanced_negative_basic_workloads + n
if not placeholder_replaced:
enhanced_positive_basic_workloads = [enhanced_prompt] + enhanced_positive_basic_workloads
else:
enhanced_positive_basic_workloads.append(enhanced_prompt)
enhanced_negative_basic_workloads.append(enhanced_negative_prompt)
# Save enhanced prompts to the dictionary with the same structure as processed_tasks_dictionary
enhanced_tasks_dictionary.append(dict(
task_seed=task_seed,
task_prompt=enhanced_prompt,
task_negative_prompt=enhanced_negative_prompt,
positive=enhanced_positive_basic_workloads,
negative=enhanced_negative_basic_workloads,
expansion='',
positive_top_k=len(enhanced_positive_basic_workloads),
negative_top_k=len(enhanced_negative_basic_workloads),
log_positive_prompt=enhanced_prompt,
log_negative_prompt=enhanced_negative_prompt,
styles=async_task.style_selections.copy(),
overwrite_width=overwrite_width,
overwrite_height=overwrite_height,
loras=loras
))
if use_expansion:
if advance_progress:
current_progress += 1
for i, t in enumerate(tasks):
for i, t in enumerate(processed_tasks_dictionary):
progressbar(async_task, current_progress, f'Preparing Fooocus text #{i + 1} ...')
expansion = pipeline.final_expansion(t['task_prompt'], t['task_seed'])
print(f'[Prompt Expansion] {expansion}')
t['expansion'] = expansion
t['positive'] = copy.deepcopy(t['positive']) + [expansion] # Deep copy.
for i, t in enumerate(enhanced_tasks_dictionary):
progressbar(async_task, current_progress, f'Preparing Fooocus enhance text #{i + 1} ...')
expansion = pipeline.final_expansion(t['task_prompt'], t['task_seed'])
print(f'[Prompt Expansion] {expansion}')
t['expansion'] = expansion
t['positive'] = copy.deepcopy(t['positive']) + [expansion] # Deep copy.
if advance_progress:
current_progress += 1
for i, t in enumerate(tasks):
progressbar(async_task, current_progress, f'Encoding positive #{i + 1} ...')
t['c'] = pipeline.clip_encode(texts=t['positive'], pool_top_k=t['positive_top_k'])
if advance_progress:
current_progress += 1
for i, t in enumerate(tasks):
if not prompts_processed:
prompts_processed = True
print("[PROMPTS] Finished Processing the Prompts!")
return processed_tasks_dictionary, use_expansion, loras, current_progress, enhanced_tasks_dictionary
def prepare_task(async_task, prompt, negative_prompt, base_model_additional_loras, image_number, disable_seed_increment, use_expansion, use_style, use_synthetic_refiner, current_progress, advance_progress=False, current_task_id=0):
"""
inputs: tasks dictionary, current_task_id, base_model_additional_loras
calls process_prompts function if prompts_processed bool is false
deletes all clip_encode['c'], clip_encode['uc']
refreshes the pipeline with correct task Loras
then encode the current task prompt and neg prompt
"""
global processed_tasks_dictionary, prompts_processed, clip_encode, enhanced_tasks_dictionary, enhanced_clip_encode_dict
if not prompts_processed:
processed_tasks_dictionary, enhanced_tasks_dictionary, use_expansion, loras, current_progress, enhanced_tasks_dictionary = process_prompts(
async_task, async_task.prompt, async_task.negative_prompt,
base_model_additional_loras, async_task.image_number,
async_task.disable_seed_increment, use_expansion, use_style,
use_synthetic_refiner, current_progress, advance_progress=True
)
tasks = processed_tasks_dictionary # Bind to tasks
task = tasks[current_task_id]
clip_encode = {}
enhanced_clip_encode_dict = {} # Dictionary to store enhanced CLIP encodings
# Debug current_task_id
print(f"[PREPARE TASK] Current Task id: {current_task_id}")
# Update the Loras for the current task
loras = task['loras']
# Refresh the pipeline:
try:
# Attempt to refresh with the loras
pipeline.refresh_everything(
refiner_model_name=async_task.refiner_model_name,
base_model_name=async_task.base_model_name,
loras=loras, # Use only valid LoRAs
base_model_additional_loras=base_model_additional_loras, # Added base_model_additional_loras instead of []
use_synthetic_refiner=False,
vae_name=async_task.vae_name
)
print(f"[LORAS] Using Unique LoRAs for Task {current_task_id + 1}")
except Exception as e:
print(f"[ERROR] {e}")
# Default LoRAs as fallback
default_loras = []
# Refresh with default LoRAs
pipeline.refresh_everything(
refiner_model_name=async_task.refiner_model_name,
base_model_name=async_task.base_model_name,
loras=default_loras, # Use only valid LoRAs
base_model_additional_loras=base_model_additional_loras, # Added base_model_additional_loras instead of []
use_synthetic_refiner=False,
vae_name=async_task.vae_name
)
print(f"[LORAS] There was an error loading a lora file, disabled all loras ")
print(f"[LORAS] Using Loras: {str(loras)}")
# Encode the current task prompt
progressbar(async_task, current_progress, f'Encoding positive #{current_task_id + 1} ...')
print(f"[CLIP] task['positive']: {task['positive']}")
print(f"[CLIP] task['negative']: {task['negative']}")
if 'positive' not in task or 'positive_top_k' not in task:
print("Error: 'positive' or 'positive_top_k' key missing in task")
# clip encode the current task
clip_encode['c'] = pipeline.clip_encode(texts=task['positive'], pool_top_k=task['positive_top_k'])
# print(f"[CLIP] clip_encode['c']: {clip_encode['c']}")
# Encode negative prompt for the current task
if abs(float(async_task.cfg_scale) - 1.0) < 1e-4:
t['uc'] = pipeline.clone_cond(t['c'])
clip_encode['uc'] = pipeline.clone_cond(clip_encode['c'])
else:
progressbar(async_task, current_progress, f'Encoding negative #{i + 1} ...')
t['uc'] = pipeline.clip_encode(texts=t['negative'], pool_top_k=t['negative_top_k'])
return tasks, use_expansion, loras, current_progress
progressbar(async_task, current_progress, f'Encoding negative #{current_task_id + 1} ...')
clip_encode['uc'] = pipeline.clip_encode(texts=task['negative'], pool_top_k=task['negative_top_k'])
# Encode enhanced prompts
for index, enhance_ctrl in enumerate(async_task.enhance_ctrls):
enhance_tasks = enhanced_tasks_dictionary # Bind to enhance_tasks
enhance_task = enhance_tasks[index]
enhance_mask_dino_prompt_text = enhance_ctrl[0]
enhance_positive = enhance_task['positive'] if enhance_ctrl[1] != "" else task['positive']
enhance_negative = enhance_task['negative'] if enhance_ctrl[2] != "" else task['negative']
# Use enhance_mask_dino_prompt_text as the unique key for enhanced CLIP encodings
enhanced_clip_encode = {}
# Encode enhanced positive prompt
progressbar(async_task, current_progress, f'Encoding enhanced positive for "{enhance_mask_dino_prompt_text}" ...')
enhanced_clip_encode['c'] = pipeline.clip_encode(texts=enhance_positive, pool_top_k=enhance_task['positive_top_k'])
# Encode enhanced negative prompt
if abs(float(async_task.cfg_scale) - 1.0) < 1e-4:
enhanced_clip_encode['uc'] = pipeline.clone_cond(enhanced_clip_encode['c'])
else:
progressbar(async_task, current_progress, f'Encoding enhanced negative for "{enhance_mask_dino_prompt_text}" ...')
enhanced_clip_encode['uc'] = pipeline.clip_encode(texts=enhance_negative, pool_top_k=enhance_task['negative_top_k'])
# Save enhanced CLIP encodings in the dictionary
enhanced_clip_encode_dict[enhance_mask_dino_prompt_text] = enhanced_clip_encode
return tasks, use_expansion, loras, current_progress, clip_encode, enhanced_clip_encode_dict
def apply_freeu(async_task):
print(f'FreeU is enabled!')
@ -942,8 +1219,8 @@ def worker():
if 'fast' in uov_method:
skip_prompt_processing = True
steps = 0
else:
steps = performance.steps_uov()
if advance_progress:
current_progress += 1
@ -951,12 +1228,15 @@ def worker():
modules.config.downloading_upscale_model()
return uov_input_image, skip_prompt_processing, steps
def prepare_enhance_prompt(prompt: str, fallback_prompt: str):
def prepare_enhance_prompt(prompt, fallback_prompt):
if isinstance(prompt, list):
# Convert the list back into a single string if needed
prompt = '\n'.join(prompt)
if safe_str(prompt) == '' or len(remove_empty_str([safe_str(p) for p in prompt.splitlines()], default='')) == 0:
prompt = fallback_prompt
return prompt
def stop_processing(async_task, processing_start_time):
async_task.processing = False
processing_time = time.perf_counter() - processing_start_time
@ -965,17 +1245,21 @@ def worker():
def process_enhance(all_steps, async_task, callback, controlnet_canny_path, controlnet_cpds_path,
current_progress, current_task_id, denoising_strength, inpaint_disable_initial_latent,
inpaint_engine, inpaint_respective_field, inpaint_strength,
prompt, negative_prompt, final_scheduler_name, goals, height, img, mask,
enhance_clip_positive, enhance_clip_negative, final_scheduler_name, goals, height, img, mask,
preparation_steps, steps, switch, tiled, total_count, use_expansion, use_style,
use_synthetic_refiner, width, show_intermediate_results=True, persist_image=True):
global processed_tasks_dictionary
tasks = processed_tasks_dictionary
task_enhance = tasks[current_task_id]
loras = task_enhance['loras']
prompt = task_enhance['task_prompt']
negative_prompt = task_enhance['task_negative_prompt']
base_model_additional_loras = []
inpaint_head_model_path = None
inpaint_parameterized = inpaint_engine != 'None' # inpaint_engine = None, improve detail
initial_latent = None
prompt = prepare_enhance_prompt(prompt, async_task.prompt)
negative_prompt = prepare_enhance_prompt(negative_prompt, async_task.negative_prompt)
if 'vary' in goals:
img, denoising_strength, initial_latent, width, height, current_progress = apply_vary(
async_task, async_task.enhance_uov_method, denoising_strength, img, switch, current_progress)
@ -1000,11 +1284,9 @@ def worker():
if inpaint_patch_model_path not in base_model_additional_loras:
base_model_additional_loras += [(inpaint_patch_model_path, 1.0)]
progressbar(async_task, current_progress, 'Preparing enhance prompts ...')
# positive and negative conditioning aren't available here anymore, process prompt again
tasks_enhance, use_expansion, loras, current_progress = process_prompt(
async_task, prompt, negative_prompt, base_model_additional_loras, 1, True,
use_expansion, use_style, use_synthetic_refiner, current_progress)
task_enhance = tasks_enhance[0]
# TODO could support vary, upscale and CN in the future
# if 'cn' in goals:
# apply_control_nets(async_task, height, ip_adapter_face_path, ip_adapter_path, width)
@ -1020,17 +1302,17 @@ def worker():
imgs, img_paths, current_progress = process_task(all_steps, async_task, callback, controlnet_canny_path,
controlnet_cpds_path, current_task_id, denoising_strength,
final_scheduler_name, goals, initial_latent, steps, switch,
task_enhance['c'], task_enhance['uc'], task_enhance, loras,
enhance_clip_positive, enhance_clip_negative, task_enhance, loras,
tiled, use_expansion, width, height, current_progress,
preparation_steps, total_count, show_intermediate_results,
persist_image)
del task_enhance['c'], task_enhance['uc'] # Save memory
#del task_enhance['c'], task_enhance['uc'] # Save memory
return current_progress, imgs[0], prompt, negative_prompt
def enhance_upscale(all_steps, async_task, base_progress, callback, controlnet_canny_path, controlnet_cpds_path,
current_task_id, denoising_strength, done_steps_inpainting, done_steps_upscaling, enhance_steps,
prompt, negative_prompt, final_scheduler_name, height, img, preparation_steps, switch, tiled,
enhance_clip_positive, enhance_clip_negative, final_scheduler_name, height, img, preparation_steps, switch, tiled,
total_count, use_expansion, use_style, use_synthetic_refiner, width, persist_image=True):
# reset inpaint worker to prevent tensor size issues and not mix upscale and inpainting
inpaint_worker.current_task = None
@ -1047,18 +1329,20 @@ def worker():
current_progress, img, prompt, negative_prompt = process_enhance(
all_steps, async_task, callback, controlnet_canny_path,
controlnet_cpds_path, current_progress, current_task_id, denoising_strength, False,
'None', 0.0, 0.0, prompt, negative_prompt, final_scheduler_name,
'None', 0.0, 0.0, enhance_clip_positive, enhance_clip_negative, final_scheduler_name,
goals_enhance, height, img, None, preparation_steps, steps, switch, tiled, total_count,
use_expansion, use_style, use_synthetic_refiner, width, persist_image=persist_image)
except ldm_patched.modules.model_management.InterruptProcessingException:
if async_task.last_stop == 'skip':
print('User skipped')
current_task_id += 1 ##
async_task.last_stop = False
# also skip all enhance steps for this image, but add the steps to the progress bar
if async_task.enhance_uov_processing_order == flags.enhancement_uov_before:
done_steps_inpainting += len(async_task.enhance_ctrls) * enhance_steps
exception_result = 'continue'
print("debug enhance_upscale FUN exception_result continue ( didn't upscale)")
else:
print('User stopped')
exception_result = 'break'
@ -1069,6 +1353,9 @@ def worker():
@torch.no_grad()
@torch.inference_mode()
def handler(async_task: AsyncTask):
global processed_tasks_dictionary,enhanced_tasks_dictionary, prompts_processed, clip_encode, enhanced_clip_encode_dict
skipped_images_ids = []
preparation_start_time = time.perf_counter()
async_task.processing = True
@ -1156,12 +1443,14 @@ def worker():
progressbar(async_task, current_progress, 'Initializing ...')
loras = async_task.loras
## start
if not skip_prompt_processing:
tasks, use_expansion, loras, current_progress = process_prompt(async_task, async_task.prompt, async_task.negative_prompt,
tasks, use_expansion, loras, current_progress, enhanced_tasks_dictionary = process_prompts(async_task, async_task.prompt, async_task.negative_prompt,
base_model_additional_loras, async_task.image_number,
async_task.disable_seed_increment, use_expansion, use_style,
use_synthetic_refiner, current_progress, advance_progress=True)
if len(goals) > 0:
current_progress += 1
progressbar(async_task, current_progress, 'Image processing ...')
@ -1277,17 +1566,30 @@ def worker():
show_intermediate_results = len(tasks) > 1 or async_task.should_enhance
persist_image = not async_task.should_enhance or not async_task.save_final_enhanced_image_only
stop_processing_all = False
#Start Generating / process Tasks
for current_task_id, task in enumerate(tasks):
task = tasks[current_task_id]
prompt = task['task_prompt']
negative_prompt = task['task_negative_prompt']
progressbar(async_task, current_progress, f'Preparing task {current_task_id + 1}/{async_task.image_number} ...')
execution_start_time = time.perf_counter()
#print(f"debug the first loop in handler Ran, and current_task_id = {current_task_id}")
## Prepare Task
tasks, use_expansion, loras, current_progress, clip_encode, enhanced_clip_encode_dict = prepare_task(async_task, async_task.prompt, async_task.negative_prompt,
base_model_additional_loras, async_task.image_number,
async_task.disable_seed_increment, use_expansion, use_style,
use_synthetic_refiner, current_progress, advance_progress=True, current_task_id=current_task_id)
try:
imgs, img_paths, current_progress = process_task(all_steps, async_task, callback, controlnet_canny_path,
controlnet_cpds_path, current_task_id,
denoising_strength, final_scheduler_name, goals,
initial_latent, async_task.steps, switch, task['c'],
task['uc'], task, loras, tiled, use_expansion, width,
initial_latent, async_task.steps, switch, clip_encode['c'],
clip_encode['uc'], task, loras, tiled, use_expansion, width,
height, current_progress, preparation_steps,
async_task.image_number, show_intermediate_results,
persist_image)
@ -1298,17 +1600,19 @@ def worker():
except ldm_patched.modules.model_management.InterruptProcessingException:
if async_task.last_stop == 'skip':
print('User skipped')
skipped_images_ids.append(current_task_id)
async_task.last_stop = False
#current_task_id += 1 ##
continue
else:
print('User stopped')
break
del task['c'], task['uc'] # Save memory
#####del task['c'], task['uc'] # Save memory
execution_time = time.perf_counter() - execution_start_time
print(f'Generating and saving time: {execution_time:.2f} seconds')
if not async_task.should_enhance:
if async_task.enhance_checkbox and not async_task.should_enhance:
print(f'[Enhance] Skipping, preconditions aren\'t met')
stop_processing(async_task, processing_start_time)
return
@ -1323,29 +1627,62 @@ def worker():
active_enhance_tabs += 1
enhance_uov_before = async_task.enhance_uov_processing_order == flags.enhancement_uov_before
enhance_uov_after = async_task.enhance_uov_processing_order == flags.enhancement_uov_after
total_count = len(images_to_enhance) * active_enhance_tabs
total_count = len(tasks)
async_task.images_to_enhance_count = len(images_to_enhance)
base_progress = current_progress
current_task_id = -1
current_task_id = 0
done_steps_upscaling = 0
done_steps_inpainting = 0
enhance_steps, _, _, _ = apply_overrides(async_task, async_task.original_steps, height, width)
exception_result = None
## start enhance
if async_task.enhance_checkbox:
if async_task.input_image_checkbox:
total_count = 1
tasks, use_expansion, loras, current_progress, clip_encode, enhanced_clip_encode_dict = prepare_task(async_task, async_task.prompt, async_task.negative_prompt,
base_model_additional_loras, async_task.image_number,
async_task.disable_seed_increment, use_expansion, use_style,
use_synthetic_refiner, current_progress, advance_progress=True)
for index, img in enumerate(images_to_enhance):
while current_task_id in skipped_images_ids:
print(f"Skipped Enhancing Task with id: {current_task_id}")
current_task_id += 1
task = tasks[current_task_id]
prompt = task['task_prompt']
negative_prompt = task['task_negative_prompt']
enhance_clip_positive = clip_encode['c']
enhance_clip_negative = clip_encode['uc']
if stop_processing_all:
break
skip_to_next_image = False
#print(f"debug The Enhance Loop ran, index = {index} and current_task_id: {current_task_id}")
async_task.enhance_stats[index] = 0
enhancement_image_start_time = time.perf_counter()
last_enhance_prompt = async_task.prompt
last_enhance_negative_prompt = async_task.negative_prompt
if enhance_uov_before:
current_task_id += 1
if current_task_id >= len(processed_tasks_dictionary):
print("Error: current_task_id exceeds the number of images to enhance.")
break
if enhance_uov_before and not skip_to_next_image and not stop_processing_all:
persist_image = not async_task.save_final_enhanced_image_only or active_enhance_tabs == 0
if not async_task.enhance_checkbox:
persist_image = True
current_task_id, done_steps_inpainting, done_steps_upscaling, img, exception_result = enhance_upscale(
all_steps, async_task, base_progress, callback, controlnet_canny_path, controlnet_cpds_path,
current_task_id, denoising_strength, done_steps_inpainting, done_steps_upscaling, enhance_steps,
async_task.prompt, async_task.negative_prompt, final_scheduler_name, height, img, preparation_steps,
enhance_clip_positive, enhance_clip_negative, final_scheduler_name, height, img, preparation_steps,
switch, tiled, total_count, use_expansion, use_style, use_synthetic_refiner, width, persist_image)
async_task.enhance_stats[index] += 1
@ -1354,13 +1691,47 @@ def worker():
elif exception_result == 'break':
break
# inpaint for all other tabs
for enhance_mask_dino_prompt_text, enhance_prompt, enhance_negative_prompt, enhance_mask_model, enhance_mask_cloth_category, enhance_mask_sam_model, enhance_mask_text_threshold, enhance_mask_box_threshold, enhance_mask_sam_max_detections, enhance_inpaint_disable_initial_latent, enhance_inpaint_engine, enhance_inpaint_strength, enhance_inpaint_respective_field, enhance_inpaint_erode_or_dilate, enhance_mask_invert in async_task.enhance_ctrls:
current_task_id += 1
# inpaint for all other tabs (enhance #1,#2,#3,....)
for enhance_mask_dino_prompt_text, mask_enhance_prompt, mask_enhance_negative_prompt, enhance_mask_model, enhance_mask_cloth_category, enhance_mask_sam_model, enhance_mask_text_threshold, enhance_mask_box_threshold, enhance_mask_sam_max_detections, enhance_inpaint_disable_initial_latent, enhance_inpaint_engine, enhance_inpaint_strength, enhance_inpaint_respective_field, enhance_inpaint_erode_or_dilate, enhance_mask_invert in async_task.enhance_ctrls:
# prepare the task loras and clip_encode the positive and Negative prompts Unique for current task
tasks_enhance, use_expansion, loras, current_progress, clip_encode, enhanced_clip_encode_dict = prepare_task(
async_task, prompt, negative_prompt, base_model_additional_loras, 1, True,
use_expansion, use_style, use_synthetic_refiner, current_progress, current_task_id=current_task_id)
task_enhance = tasks_enhance[current_task_id] ## was tasks_enhance[0]
enhance_prompt = mask_enhance_prompt if mask_enhance_prompt != "" else prompt
enhance_negative_prompt = mask_enhance_negative_prompt if mask_enhance_negative_prompt != "" else negative_prompt
if mask_enhance_prompt:
print(f"[CLIP] using Positive: {enhanced_tasks_dictionary[current_task_id]['positive']}")
enhance_clip_positive = enhanced_clip_encode_dict[enhance_mask_dino_prompt_text]['c']
else:
enhance_clip_positive = clip_encode['c']
if mask_enhance_negative_prompt:
enhance_clip_negative = enhanced_clip_encode_dict[enhance_mask_dino_prompt_text]['uc']
else:
enhance_clip_negative = clip_encode['uc']
print(f"\n[ENHANCE]: {enhance_mask_dino_prompt_text} for Task: {current_task_id+1}")
## Handle skipping and stop
if skip_to_next_image or stop_processing_all:
print(f"debug Skipped All Enhance for current_task_id: {current_task_id}")
break
## Handles Input image enhance
if async_task.input_image_checkbox:
tasks, use_expansion, loras, current_progress, clip_encode, enhanced_clip_encode_dict = prepare_task(async_task, enhance_prompt, enhance_negative_prompt,
base_model_additional_loras, async_task.image_number,
async_task.disable_seed_increment, use_expansion, use_style,
use_synthetic_refiner, current_progress, advance_progress=True)
current_progress = int(base_progress + (100 - preparation_steps) / float(all_steps) * (done_steps_upscaling + done_steps_inpainting))
progressbar(async_task, current_progress, f'Preparing enhancement {current_task_id + 1}/{total_count} ...')
progressbar(async_task, current_progress, f'Preparing enhancement {enhance_mask_dino_prompt_text} for task: {current_task_id + 1}/{total_count} ...')
enhancement_task_start_time = time.perf_counter()
is_last_enhance_for_image = (current_task_id + 1) % active_enhance_tabs == 0 and not enhance_uov_after
is_last_enhance_for_image = (current_task_id) % active_enhance_tabs == 0 and not enhance_uov_after ##
persist_image = not async_task.save_final_enhanced_image_only or is_last_enhance_for_image
extras = {}
@ -1369,6 +1740,8 @@ def worker():
elif enhance_mask_model == 'u2net_cloth_seg':
extras['cloth_category'] = enhance_mask_cloth_category
mask, dino_detection_count, sam_detection_count, sam_detection_on_mask_count = generate_mask_from_image(
img, mask_model=enhance_mask_model, extras=extras, sam_options=SAMOptions(
dino_prompt=enhance_mask_dino_prompt_text,
@ -1405,15 +1778,17 @@ def worker():
goals_enhance = ['inpaint']
try:
current_progress, img, enhance_prompt_processed, enhance_negative_prompt_processed = process_enhance(
all_steps, async_task, callback, controlnet_canny_path, controlnet_cpds_path,
current_progress, current_task_id, denoising_strength, enhance_inpaint_disable_initial_latent,
enhance_inpaint_engine, enhance_inpaint_respective_field, enhance_inpaint_strength,
enhance_prompt, enhance_negative_prompt, final_scheduler_name, goals_enhance, height, img, mask,
enhance_clip_positive, enhance_clip_negative, final_scheduler_name, goals_enhance, height, img, mask,
preparation_steps, enhance_steps, switch, tiled, total_count, use_expansion, use_style,
use_synthetic_refiner, width, persist_image=persist_image)
async_task.enhance_stats[index] += 1
if (should_process_enhance_uov and async_task.enhance_uov_processing_order == flags.enhancement_uov_after
and async_task.enhance_uov_prompt_type == flags.enhancement_uov_prompt_type_last_filled):
if enhance_prompt_processed != '':
@ -1425,9 +1800,11 @@ def worker():
if async_task.last_stop == 'skip':
print('User skipped')
async_task.last_stop = False
continue
skip_to_next_image = True
break ###continue
else:
print('User stopped')
stop_processing_all = True
exception_result = 'break'
break
finally:
@ -1436,19 +1813,25 @@ def worker():
enhancement_task_time = time.perf_counter() - enhancement_task_start_time
print(f'Enhancement time: {enhancement_task_time:.2f} seconds')
if exception_result == 'break':
break
if enhance_uov_after:
current_task_id += 1
# last step in enhance, always save
if enhance_uov_after and not skip_to_next_image and not stop_processing_all:
tasks, use_expansion, loras, current_progress, clip_encode, enhanced_clip_encode_dict = prepare_task(async_task, enhance_prompt, enhance_negative_prompt,
base_model_additional_loras, async_task.image_number,
async_task.disable_seed_increment, use_expansion, use_style,
use_synthetic_refiner, current_progress, advance_progress=True)
task = tasks[current_task_id]
prompt = task['positive']
negative_prompt = task['negative']
#last step in enhance, always save
persist_image = True
#print("debug enhance_uov_after Section")
current_task_id, done_steps_inpainting, done_steps_upscaling, img, exception_result = enhance_upscale(
all_steps, async_task, base_progress, callback, controlnet_canny_path, controlnet_cpds_path,
current_task_id, denoising_strength, done_steps_inpainting, done_steps_upscaling, enhance_steps,
last_enhance_prompt, last_enhance_negative_prompt, final_scheduler_name, height, img,
preparation_steps, switch, tiled, total_count, use_expansion, use_style, use_synthetic_refiner,
width, persist_image)
enhance_clip_positive, enhance_clip_negative, final_scheduler_name, height, img, preparation_steps,
switch, tiled, total_count, use_expansion, use_style, use_synthetic_refiner, width, persist_image)
async_task.enhance_stats[index] += 1
if exception_result == 'continue':
@ -1456,10 +1839,18 @@ def worker():
elif exception_result == 'break':
break
if exception_result == 'break':
break
current_task_id += 1
enhancement_image_time = time.perf_counter() - enhancement_image_start_time
print(f'Enhancement image time: {enhancement_image_time:.2f} seconds')
stop_processing(async_task, processing_start_time)
stop_processing_all = False
return
while True:
@ -1475,8 +1866,11 @@ def worker():
pipeline.prepare_text_encoder(async_call=True)
except:
traceback.print_exc()
task.yields.append(['finish', task.results])
####task.yields.append(['finish', task.results])
finally:
global processed_tasks_dictionary, prompts_processed
processed_tasks_dictionary = []
prompts_processed = False
if pid in modules.patch.patch_settings:
del modules.patch.patch_settings[pid]
pass

View File

@ -465,33 +465,65 @@ def cleanup_prompt(prompt):
return cleaned_prompt[:-2]
def apply_wildcards(wildcard_text, rng, i, read_wildcards_in_order) -> str:
def apply_wildcards(wildcard_text, rng, i, read_wildcards_in_order, wildcard_start_row=None) -> str:
if wildcard_start_row is None:
wildcard_start_row = {}
for _ in range(modules.config.wildcards_max_bfs_depth):
placeholders = re.findall(r'__([\w-]+)__', wildcard_text)
if len(placeholders) == 0:
# Match both ordered wildcards (_o__) and regular wildcards (__)
# Ensure full filenames are captured including dots, numbers, special characters
placeholders = re.findall(r'(_o__|__)?([\w\.\(\)ぁ-んァ-ヶ一-龯ー々・-]+?)__(\d+_)?', wildcard_text)
if not placeholders:
return wildcard_text
print(f'[Wildcards] processing: {wildcard_text}')
for placeholder in placeholders:
for prefix, placeholder, suffix in placeholders:
try:
# Ensure the full wildcard file name is matched correctly
matches = [x for x in modules.config.wildcard_filenames if os.path.splitext(os.path.basename(x))[0] == placeholder]
if not matches:
raise FileNotFoundError(f"Wildcard file for '{placeholder}' not found.")
words = open(os.path.join(modules.config.path_wildcards, matches[0]), encoding='utf-8').read().splitlines()
words = [x for x in words if x != '']
assert len(words) > 0
if read_wildcards_in_order:
wildcard_text = wildcard_text.replace(f'__{placeholder}__', words[i % len(words)], 1)
words = [x.strip() for x in words if x.strip()]
if not words:
raise ValueError(f"Wildcard file '{matches[0]}' is empty.")
# Handling ordered and unordered wildcards
if prefix == '_o__':
start_row = wildcard_start_row.get(placeholder, 0) - 1
if not suffix:
start_row = 0 # Ensure no negative index errors
selected_word = words[(start_row + i) % len(words)]
else:
wildcard_text = wildcard_text.replace(f'__{placeholder}__', rng.choice(words), 1)
except:
selected_word = words[i % len(words)] if read_wildcards_in_order else rng.choice(words)
print(f"[Wildcards] selected_word: {selected_word}")
# Recursively process nested wildcards
selected_word = apply_wildcards(selected_word, rng, i, read_wildcards_in_order, wildcard_start_row)
# Replace the entire wildcard pattern (including suffix) with the selected word
wildcard_pattern = f'{prefix}{placeholder}__{suffix}' if suffix else f'{prefix}{placeholder}__'
wildcard_text = wildcard_text.replace(wildcard_pattern, selected_word, 1)
except Exception as e:
print(f'[Wildcards] Warning: {placeholder}.txt missing or empty. '
f'Using "{placeholder}" as a normal word.')
wildcard_text = wildcard_text.replace(f'__{placeholder}__', placeholder)
f'Using "{placeholder}" as a normal word. Error: {e}')
wildcard_pattern = f'{prefix}{placeholder}__{suffix}' if suffix else f'{prefix}{placeholder}__'
wildcard_text = wildcard_text.replace(wildcard_pattern, placeholder, 1)
print(f'[Wildcards] {wildcard_text}')
print(f'[Wildcards] BFS stack overflow. Current text: {wildcard_text}')
return wildcard_text
def get_image_size_info(image: np.ndarray, aspect_ratios: list) -> str:
try:
image = Image.fromarray(np.uint8(image))

512
readme.md
View File

@ -1,486 +1,26 @@
<div align=center>
<img src="https://github.com/lllyasviel/Fooocus/assets/19834515/483fb86d-c9a2-4c20-997c-46dafc124f25">
</div>
# Fooocus
[>>> Click Here to Install Fooocus <<<](#download)
Fooocus is an image generating software (based on [Gradio](https://www.gradio.app/) <a href='https://github.com/gradio-app/gradio'><img src='https://img.shields.io/github/stars/gradio-app/gradio'></a>).
Fooocus presents a rethinking of image generator designs. The software is offline, open source, and free, while at the same time, similar to many online image generators like Midjourney, the manual tweaking is not needed, and users only need to focus on the prompts and images. Fooocus has also simplified the installation: between pressing "download" and generating the first image, the number of needed mouse clicks is strictly limited to less than 3. Minimal GPU memory requirement is 4GB (Nvidia).
**Recently many fake websites exist on Google when you search “fooocus”. Do not trust those here is the only official source of Fooocus.**
# Project Status: Limited Long-Term Support (LTS) with Bug Fixes Only
The Fooocus project, built entirely on the **Stable Diffusion XL** architecture, is now in a state of limited long-term support (LTS) with bug fixes only. As the existing functionalities are considered as nearly free of programmartic issues (Thanks to [mashb1t](https://github.com/mashb1t)'s huge efforts), future updates will focus exclusively on addressing any bugs that may arise.
**There are no current plans to migrate to or incorporate newer model architectures.** However, this may change during time with the development of open-source community. For example, if the community converge to one single dominant method for image generation (which may really happen in half or one years given the current status), Fooocus may also migrate to that exact method.
For those interested in utilizing newer models such as **Flux**, we recommend exploring alternative platforms such as [WebUI Forge](https://github.com/lllyasviel/stable-diffusion-webui-forge) (also from us), [ComfyUI/SwarmUI](https://github.com/comfyanonymous/ComfyUI). Additionally, several [excellent forks of Fooocus](https://github.com/lllyasviel/Fooocus?tab=readme-ov-file#forks) are available for experimentation.
Again, recently many fake websites exist on Google when you search “fooocus”. Do **NOT** get Fooocus from those websites this page is the only official source of Fooocus. We never have any website like such as “fooocus.com”, “fooocus.net”, “fooocus.co”, “fooocus.ai”, “fooocus.org”, “fooocus.pro”, “fooocus.one”. Those websites are ALL FAKE. **They have ABSOLUTLY no relationship to us. Fooocus is a 100% non-commercial offline open-source software.**
# Features
Below is a quick list using Midjourney's examples:
| Midjourney | Fooocus |
| - | - |
| High-quality text-to-image without needing much prompt engineering or parameter tuning. <br> (Unknown method) | High-quality text-to-image without needing much prompt engineering or parameter tuning. <br> (Fooocus has an offline GPT-2 based prompt processing engine and lots of sampling improvements so that results are always beautiful, no matter if your prompt is as short as “house in garden” or as long as 1000 words) |
| V1 V2 V3 V4 | Input Image -> Upscale or Variation -> Vary (Subtle) / Vary (Strong)|
| U1 U2 U3 U4 | Input Image -> Upscale or Variation -> Upscale (1.5x) / Upscale (2x) |
| Inpaint / Up / Down / Left / Right (Pan) | Input Image -> Inpaint or Outpaint -> Inpaint / Up / Down / Left / Right <br> (Fooocus uses its own inpaint algorithm and inpaint models so that results are more satisfying than all other software that uses standard SDXL inpaint method/model) |
| Image Prompt | Input Image -> Image Prompt <br> (Fooocus uses its own image prompt algorithm so that result quality and prompt understanding are more satisfying than all other software that uses standard SDXL methods like standard IP-Adapters or Revisions) |
| --style | Advanced -> Style |
| --stylize | Advanced -> Advanced -> Guidance |
| --niji | [Multiple launchers: "run.bat", "run_anime.bat", and "run_realistic.bat".](https://github.com/lllyasviel/Fooocus/discussions/679) <br> Fooocus support SDXL models on Civitai <br> (You can google search “Civitai” if you do not know about it) |
| --quality | Advanced -> Quality |
| --repeat | Advanced -> Image Number |
| Multi Prompts (::) | Just use multiple lines of prompts |
| Prompt Weights | You can use " I am (happy:1.5)". <br> Fooocus uses A1111's reweighting algorithm so that results are better than ComfyUI if users directly copy prompts from Civitai. (Because if prompts are written in ComfyUI's reweighting, users are less likely to copy prompt texts as they prefer dragging files) <br> To use embedding, you can use "(embedding:file_name:1.1)" |
| --no | Advanced -> Negative Prompt |
| --ar | Advanced -> Aspect Ratios |
| InsightFace | Input Image -> Image Prompt -> Advanced -> FaceSwap |
| Describe | Input Image -> Describe |
Below is a quick list using LeonardoAI's examples:
| LeonardoAI | Fooocus |
| - | - |
| Prompt Magic | Advanced -> Style -> Fooocus V2 |
| Advanced Sampler Parameters (like Contrast/Sharpness/etc) | Advanced -> Advanced -> Sampling Sharpness / etc |
| User-friendly ControlNets | Input Image -> Image Prompt -> Advanced |
Also, [click here to browse the advanced features.](https://github.com/lllyasviel/Fooocus/discussions/117)
# Download
### Windows
You can directly download Fooocus with:
**[>>> Click here to download <<<](https://github.com/lllyasviel/Fooocus/releases/download/v2.5.0/Fooocus_win64_2-5-0.7z)**
After you download the file, please uncompress it and then run the "run.bat".
![image](https://github.com/lllyasviel/Fooocus/assets/19834515/c49269c4-c274-4893-b368-047c401cc58c)
The first time you launch the software, it will automatically download models:
1. It will download [default models](#models) to the folder "Fooocus\models\checkpoints" given different presets. You can download them in advance if you do not want automatic download.
2. Note that if you use inpaint, at the first time you inpaint an image, it will download [Fooocus's own inpaint control model from here](https://huggingface.co/lllyasviel/fooocus_inpaint/resolve/main/inpaint_v26.fooocus.patch) as the file "Fooocus\models\inpaint\inpaint_v26.fooocus.patch" (the size of this file is 1.28GB).
After Fooocus 2.1.60, you will also have `run_anime.bat` and `run_realistic.bat`. They are different model presets (and require different models, but they will be automatically downloaded). [Check here for more details](https://github.com/lllyasviel/Fooocus/discussions/679).
After Fooocus 2.3.0 you can also switch presets directly in the browser. Keep in mind to add these arguments if you want to change the default behavior:
* Use `--disable-preset-selection` to disable preset selection in the browser.
* Use `--always-download-new-model` to download missing models on preset switch. Default is fallback to `previous_default_models` defined in the corresponding preset, also see terminal output.
![image](https://github.com/lllyasviel/Fooocus/assets/19834515/d386f817-4bd7-490c-ad89-c1e228c23447)
If you already have these files, you can copy them to the above locations to speed up installation.
Note that if you see **"MetadataIncompleteBuffer" or "PytorchStreamReader"**, then your model files are corrupted. Please download models again.
Below is a test on a relatively low-end laptop with **16GB System RAM** and **6GB VRAM** (Nvidia 3060 laptop). The speed on this machine is about 1.35 seconds per iteration. Pretty impressive nowadays laptops with 3060 are usually at very acceptable price.
![image](https://github.com/lllyasviel/Fooocus/assets/19834515/938737a5-b105-4f19-b051-81356cb7c495)
Besides, recently many other software report that Nvidia driver above 532 is sometimes 10x slower than Nvidia driver 531. If your generation time is very long, consider download [Nvidia Driver 531 Laptop](https://www.nvidia.com/download/driverResults.aspx/199991/en-us/) or [Nvidia Driver 531 Desktop](https://www.nvidia.com/download/driverResults.aspx/199990/en-us/).
Note that the minimal requirement is **4GB Nvidia GPU memory (4GB VRAM)** and **8GB system memory (8GB RAM)**. This requires using Microsofts Virtual Swap technique, which is automatically enabled by your Windows installation in most cases, so you often do not need to do anything about it. However, if you are not sure, or if you manually turned it off (would anyone really do that?), or **if you see any "RuntimeError: CPUAllocator"**, you can enable it here:
<details>
<summary>Click here to see the image instructions. </summary>
![image](https://github.com/lllyasviel/Fooocus/assets/19834515/2a06b130-fe9b-4504-94f1-2763be4476e9)
**And make sure that you have at least 40GB free space on each drive if you still see "RuntimeError: CPUAllocator" !**
</details>
Please open an issue if you use similar devices but still cannot achieve acceptable performances.
Note that the [minimal requirement](#minimal-requirement) for different platforms is different.
See also the common problems and troubleshoots [here](troubleshoot.md).
### Colab
(Last tested - 2024 Aug 12 by [mashb1t](https://github.com/mashb1t))
| Colab | Info
| --- | --- |
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lllyasviel/Fooocus/blob/main/fooocus_colab.ipynb) | Fooocus Official
In Colab, you can modify the last line to `!python entry_with_update.py --share --always-high-vram` or `!python entry_with_update.py --share --always-high-vram --preset anime` or `!python entry_with_update.py --share --always-high-vram --preset realistic` for Fooocus Default/Anime/Realistic Edition.
You can also change the preset in the UI. Please be aware that this may lead to timeouts after 60 seconds. If this is the case, please wait until the download has finished, change the preset to initial and back to the one you've selected or reload the page.
Note that this Colab will disable refiner by default because Colab free's resources are relatively limited (and some "big" features like image prompt may cause free-tier Colab to disconnect). We make sure that basic text-to-image is always working on free-tier Colab.
Using `--always-high-vram` shifts resource allocation from RAM to VRAM and achieves the overall best balance between performance, flexibility and stability on the default T4 instance. Please find more information [here](https://github.com/lllyasviel/Fooocus/pull/1710#issuecomment-1989185346).
Thanks to [camenduru](https://github.com/camenduru) for the template!
### Linux (Using Anaconda)
If you want to use Anaconda/Miniconda, you can
git clone https://github.com/lllyasviel/Fooocus.git
cd Fooocus
conda env create -f environment.yaml
conda activate fooocus
pip install -r requirements_versions.txt
Then download the models: download [default models](#models) to the folder "Fooocus\models\checkpoints". **Or let Fooocus automatically download the models** using the launcher:
conda activate fooocus
python entry_with_update.py
Or, if you want to open a remote port, use
conda activate fooocus
python entry_with_update.py --listen
Use `python entry_with_update.py --preset anime` or `python entry_with_update.py --preset realistic` for Fooocus Anime/Realistic Edition.
### Linux (Using Python Venv)
Your Linux needs to have **Python 3.10** installed, and let's say your Python can be called with the command **python3** with your venv system working; you can
git clone https://github.com/lllyasviel/Fooocus.git
cd Fooocus
python3 -m venv fooocus_env
source fooocus_env/bin/activate
pip install -r requirements_versions.txt
See the above sections for model downloads. You can launch the software with:
source fooocus_env/bin/activate
python entry_with_update.py
Or, if you want to open a remote port, use
source fooocus_env/bin/activate
python entry_with_update.py --listen
Use `python entry_with_update.py --preset anime` or `python entry_with_update.py --preset realistic` for Fooocus Anime/Realistic Edition.
### Linux (Using native system Python)
If you know what you are doing, and your Linux already has **Python 3.10** installed, and your Python can be called with the command **python3** (and Pip with **pip3**), you can
git clone https://github.com/lllyasviel/Fooocus.git
cd Fooocus
pip3 install -r requirements_versions.txt
See the above sections for model downloads. You can launch the software with:
python3 entry_with_update.py
Or, if you want to open a remote port, use
python3 entry_with_update.py --listen
Use `python entry_with_update.py --preset anime` or `python entry_with_update.py --preset realistic` for Fooocus Anime/Realistic Edition.
### Linux (AMD GPUs)
Note that the [minimal requirement](#minimal-requirement) for different platforms is different.
Same with the above instructions. You need to change torch to the AMD version
pip uninstall torch torchvision torchaudio torchtext functorch xformers
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.6
AMD is not intensively tested, however. The AMD support is in beta.
Use `python entry_with_update.py --preset anime` or `python entry_with_update.py --preset realistic` for Fooocus Anime/Realistic Edition.
### Windows (AMD GPUs)
Note that the [minimal requirement](#minimal-requirement) for different platforms is different.
Same with Windows. Download the software and edit the content of `run.bat` as:
.\python_embeded\python.exe -m pip uninstall torch torchvision torchaudio torchtext functorch xformers -y
.\python_embeded\python.exe -m pip install torch-directml
.\python_embeded\python.exe -s Fooocus\entry_with_update.py --directml
pause
Then run the `run.bat`.
AMD is not intensively tested, however. The AMD support is in beta.
For AMD, use `.\python_embeded\python.exe Fooocus\entry_with_update.py --directml --preset anime` or `.\python_embeded\python.exe Fooocus\entry_with_update.py --directml --preset realistic` for Fooocus Anime/Realistic Edition.
### Mac
Note that the [minimal requirement](#minimal-requirement) for different platforms is different.
Mac is not intensively tested. Below is an unofficial guideline for using Mac. You can discuss problems [here](https://github.com/lllyasviel/Fooocus/pull/129).
You can install Fooocus on Apple Mac silicon (M1 or M2) with macOS 'Catalina' or a newer version. Fooocus runs on Apple silicon computers via [PyTorch](https://pytorch.org/get-started/locally/) MPS device acceleration. Mac Silicon computers don't come with a dedicated graphics card, resulting in significantly longer image processing times compared to computers with dedicated graphics cards.
1. Install the conda package manager and pytorch nightly. Read the [Accelerated PyTorch training on Mac](https://developer.apple.com/metal/pytorch/) Apple Developer guide for instructions. Make sure pytorch recognizes your MPS device.
1. Open the macOS Terminal app and clone this repository with `git clone https://github.com/lllyasviel/Fooocus.git`.
1. Change to the new Fooocus directory, `cd Fooocus`.
1. Create a new conda environment, `conda env create -f environment.yaml`.
1. Activate your new conda environment, `conda activate fooocus`.
1. Install the packages required by Fooocus, `pip install -r requirements_versions.txt`.
1. Launch Fooocus by running `python entry_with_update.py`. (Some Mac M2 users may need `python entry_with_update.py --disable-offload-from-vram` to speed up model loading/unloading.) The first time you run Fooocus, it will automatically download the Stable Diffusion SDXL models and will take a significant amount of time, depending on your internet connection.
Use `python entry_with_update.py --preset anime` or `python entry_with_update.py --preset realistic` for Fooocus Anime/Realistic Edition.
### Docker
See [docker.md](docker.md)
### Download Previous Version
See the guidelines [here](https://github.com/lllyasviel/Fooocus/discussions/1405).
## Minimal Requirement
Below is the minimal requirement for running Fooocus locally. If your device capability is lower than this spec, you may not be able to use Fooocus locally. (Please let us know, in any case, if your device capability is lower but Fooocus still works.)
| Operating System | GPU | Minimal GPU Memory | Minimal System Memory | [System Swap](troubleshoot.md) | Note |
|-------------------|------------------------------|------------------------------|---------------------------|--------------------------------|----------------------------------------------------------------------------|
| Windows/Linux | Nvidia RTX 4XXX | 4GB | 8GB | Required | fastest |
| Windows/Linux | Nvidia RTX 3XXX | 4GB | 8GB | Required | usually faster than RTX 2XXX |
| Windows/Linux | Nvidia RTX 2XXX | 4GB | 8GB | Required | usually faster than GTX 1XXX |
| Windows/Linux | Nvidia GTX 1XXX | 8GB (&ast; 6GB uncertain) | 8GB | Required | only marginally faster than CPU |
| Windows/Linux | Nvidia GTX 9XX | 8GB | 8GB | Required | faster or slower than CPU |
| Windows/Linux | Nvidia GTX < 9XX | Not supported | / | / | / |
| Windows | AMD GPU | 8GB (updated 2023 Dec 30) | 8GB | Required | via DirectML (&ast; ROCm is on hold), about 3x slower than Nvidia RTX 3XXX |
| Linux | AMD GPU | 8GB | 8GB | Required | via ROCm, about 1.5x slower than Nvidia RTX 3XXX |
| Mac | M1/M2 MPS | Shared | Shared | Shared | about 9x slower than Nvidia RTX 3XXX |
| Windows/Linux/Mac | only use CPU | 0GB | 32GB | Required | about 17x slower than Nvidia RTX 3XXX |
&ast; AMD GPU ROCm (on hold): The AMD is still working on supporting ROCm on Windows.
&ast; Nvidia GTX 1XXX 6GB uncertain: Some people report 6GB success on GTX 10XX, but some other people report failure cases.
*Note that Fooocus is only for extremely high quality image generating. We will not support smaller models to reduce the requirement and sacrifice result quality.*
## Troubleshoot
See the common problems [here](troubleshoot.md).
## Default Models
<a name="models"></a>
Given different goals, the default models and configs of Fooocus are different:
| Task | Windows | Linux args | Main Model | Refiner | Config |
|-----------| --- | --- |-----------------------------| --- |--------------------------------------------------------------------------------|
| General | run.bat | | juggernautXL_v8Rundiffusion | not used | [here](https://github.com/lllyasviel/Fooocus/blob/main/presets/default.json) |
| Realistic | run_realistic.bat | --preset realistic | realisticStockPhoto_v20 | not used | [here](https://github.com/lllyasviel/Fooocus/blob/main/presets/realistic.json) |
| Anime | run_anime.bat | --preset anime | animaPencilXL_v500 | not used | [here](https://github.com/lllyasviel/Fooocus/blob/main/presets/anime.json) |
Note that the download is **automatic** - you do not need to do anything if the internet connection is okay. However, you can download them manually if you (or move them from somewhere else) have your own preparation.
## UI Access and Authentication
In addition to running on localhost, Fooocus can also expose its UI in two ways:
* Local UI listener: use `--listen` (specify port e.g. with `--port 8888`).
* API access: use `--share` (registers an endpoint at `.gradio.live`).
In both ways the access is unauthenticated by default. You can add basic authentication by creating a file called `auth.json` in the main directory, which contains a list of JSON objects with the keys `user` and `pass` (see example in [auth-example.json](./auth-example.json)).
## List of "Hidden" Tricks
<a name="tech_list"></a>
<details>
<summary>Click to see a list of tricks. Those are based on SDXL and are not very up-to-date with latest models.</summary>
1. GPT2-based [prompt expansion as a dynamic style "Fooocus V2".](https://github.com/lllyasviel/Fooocus/discussions/117#raw) (similar to Midjourney's hidden pre-processing and "raw" mode, or the LeonardoAI's Prompt Magic).
2. Native refiner swap inside one single k-sampler. The advantage is that the refiner model can now reuse the base model's momentum (or ODE's history parameters) collected from k-sampling to achieve more coherent sampling. In Automatic1111's high-res fix and ComfyUI's node system, the base model and refiner use two independent k-samplers, which means the momentum is largely wasted, and the sampling continuity is broken. Fooocus uses its own advanced k-diffusion sampling that ensures seamless, native, and continuous swap in a refiner setup. (Update Aug 13: Actually, I discussed this with Automatic1111 several days ago, and it seems that the “native refiner swap inside one single k-sampler” is [merged]( https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/12371) into the dev branch of webui. Great!)
3. Negative ADM guidance. Because the highest resolution level of XL Base does not have cross attentions, the positive and negative signals for XL's highest resolution level cannot receive enough contrasts during the CFG sampling, causing the results to look a bit plastic or overly smooth in certain cases. Fortunately, since the XL's highest resolution level is still conditioned on image aspect ratios (ADM), we can modify the adm on the positive/negative side to compensate for the lack of CFG contrast in the highest resolution level. (Update Aug 16, the IOS App [Draw Things](https://apps.apple.com/us/app/draw-things-ai-generation/id6444050820) will support Negative ADM Guidance. Great!)
4. We implemented a carefully tuned variation of Section 5.1 of ["Improving Sample Quality of Diffusion Models Using Self-Attention Guidance"](https://arxiv.org/pdf/2210.00939.pdf). The weight is set to very low, but this is Fooocus's final guarantee to make sure that the XL will never yield an overly smooth or plastic appearance (examples [here](https://github.com/lllyasviel/Fooocus/discussions/117#sharpness)). This can almost eliminate all cases for which XL still occasionally produces overly smooth results, even with negative ADM guidance. (Update 2023 Aug 18, the Gaussian kernel of SAG is changed to an anisotropic kernel for better structure preservation and fewer artifacts.)
5. We modified the style templates a bit and added the "cinematic-default".
6. We tested the "sd_xl_offset_example-lora_1.0.safetensors" and it seems that when the lora weight is below 0.5, the results are always better than XL without lora.
7. The parameters of samplers are carefully tuned.
8. Because XL uses positional encoding for generation resolution, images generated by several fixed resolutions look a bit better than those from arbitrary resolutions (because the positional encoding is not very good at handling int numbers that are unseen during training). This suggests that the resolutions in UI may be hard coded for best results.
9. Separated prompts for two different text encoders seem unnecessary. Separated prompts for the base model and refiner may work, but the effects are random, and we refrain from implementing this.
10. The DPM family seems well-suited for XL since XL sometimes generates overly smooth texture, but the DPM family sometimes generates overly dense detail in texture. Their joint effect looks neutral and appealing to human perception.
11. A carefully designed system for balancing multiple styles as well as prompt expansion.
12. Using automatic1111's method to normalize prompt emphasizing. This significantly improves results when users directly copy prompts from civitai.
13. The joint swap system of the refiner now also supports img2img and upscale in a seamless way.
14. CFG Scale and TSNR correction (tuned for SDXL) when CFG is bigger than 10.
</details>
## Customization
After the first time you run Fooocus, a config file will be generated at `Fooocus\config.txt`. This file can be edited to change the model path or default parameters.
For example, an edited `Fooocus\config.txt` (this file will be generated after the first launch) may look like this:
```json
{
"path_checkpoints": "D:\\Fooocus\\models\\checkpoints",
"path_loras": "D:\\Fooocus\\models\\loras",
"path_embeddings": "D:\\Fooocus\\models\\embeddings",
"path_vae_approx": "D:\\Fooocus\\models\\vae_approx",
"path_upscale_models": "D:\\Fooocus\\models\\upscale_models",
"path_inpaint": "D:\\Fooocus\\models\\inpaint",
"path_controlnet": "D:\\Fooocus\\models\\controlnet",
"path_clip_vision": "D:\\Fooocus\\models\\clip_vision",
"path_fooocus_expansion": "D:\\Fooocus\\models\\prompt_expansion\\fooocus_expansion",
"path_outputs": "D:\\Fooocus\\outputs",
"default_model": "realisticStockPhoto_v10.safetensors",
"default_refiner": "",
"default_loras": [["lora_filename_1.safetensors", 0.5], ["lora_filename_2.safetensors", 0.5]],
"default_cfg_scale": 3.0,
"default_sampler": "dpmpp_2m",
"default_scheduler": "karras",
"default_negative_prompt": "low quality",
"default_positive_prompt": "",
"default_styles": [
"Fooocus V2",
"Fooocus Photograph",
"Fooocus Negative"
]
}
```
Many other keys, formats, and examples are in `Fooocus\config_modification_tutorial.txt` (this file will be generated after the first launch).
Consider twice before you really change the config. If you find yourself breaking things, just delete `Fooocus\config.txt`. Fooocus will go back to default.
A safer way is just to try "run_anime.bat" or "run_realistic.bat" - they should already be good enough for different tasks.
~Note that `user_path_config.txt` is deprecated and will be removed soon.~ (Edit: it is already removed.)
### All CMD Flags
```
entry_with_update.py [-h] [--listen [IP]] [--port PORT]
[--disable-header-check [ORIGIN]]
[--web-upload-size WEB_UPLOAD_SIZE]
[--hf-mirror HF_MIRROR]
[--external-working-path PATH [PATH ...]]
[--output-path OUTPUT_PATH]
[--temp-path TEMP_PATH] [--cache-path CACHE_PATH]
[--in-browser] [--disable-in-browser]
[--gpu-device-id DEVICE_ID]
[--async-cuda-allocation | --disable-async-cuda-allocation]
[--disable-attention-upcast]
[--all-in-fp32 | --all-in-fp16]
[--unet-in-bf16 | --unet-in-fp16 | --unet-in-fp8-e4m3fn | --unet-in-fp8-e5m2]
[--vae-in-fp16 | --vae-in-fp32 | --vae-in-bf16]
[--vae-in-cpu]
[--clip-in-fp8-e4m3fn | --clip-in-fp8-e5m2 | --clip-in-fp16 | --clip-in-fp32]
[--directml [DIRECTML_DEVICE]]
[--disable-ipex-hijack]
[--preview-option [none,auto,fast,taesd]]
[--attention-split | --attention-quad | --attention-pytorch]
[--disable-xformers]
[--always-gpu | --always-high-vram | --always-normal-vram | --always-low-vram | --always-no-vram | --always-cpu [CPU_NUM_THREADS]]
[--always-offload-from-vram]
[--pytorch-deterministic] [--disable-server-log]
[--debug-mode] [--is-windows-embedded-python]
[--disable-server-info] [--multi-user] [--share]
[--preset PRESET] [--disable-preset-selection]
[--language LANGUAGE]
[--disable-offload-from-vram] [--theme THEME]
[--disable-image-log] [--disable-analytics]
[--disable-metadata] [--disable-preset-download]
[--disable-enhance-output-sorting]
[--enable-auto-describe-image]
[--always-download-new-model]
[--rebuild-hash-cache [CPU_NUM_THREADS]]
```
## Inline Prompt Features
### Wildcards
Example prompt: `__color__ flower`
Processed for positive and negative prompt.
Selects a random wildcard from a predefined list of options, in this case the `wildcards/color.txt` file.
The wildcard will be replaced with a random color (randomness based on seed).
You can also disable randomness and process a wildcard file from top to bottom by enabling the checkbox `Read wildcards in order` in Developer Debug Mode.
Wildcards can be nested and combined, and multiple wildcards can be used in the same prompt (example see `wildcards/color_flower.txt`).
### Array Processing
Example prompt: `[[red, green, blue]] flower`
Processed only for positive prompt.
Processes the array from left to right, generating a separate image for each element in the array. In this case 3 images would be generated, one for each color.
Increase the image number to 3 to generate all 3 variants.
Arrays can not be nested, but multiple arrays can be used in the same prompt.
Does support inline LoRAs as array elements!
### Inline LoRAs
Example prompt: `flower <lora:sunflowers:1.2>`
Processed only for positive prompt.
Applies a LoRA to the prompt. The LoRA file must be located in the `models/loras` directory.
## Advanced Features
[Click here to browse the advanced features.](https://github.com/lllyasviel/Fooocus/discussions/117)
## Forks
Below are some Forks to Fooocus:
| Fooocus' forks |
| - |
| [fenneishi/Fooocus-Control](https://github.com/fenneishi/Fooocus-Control) </br>[runew0lf/RuinedFooocus](https://github.com/runew0lf/RuinedFooocus) </br> [MoonRide303/Fooocus-MRE](https://github.com/MoonRide303/Fooocus-MRE) </br> [mashb1t/Fooocus](https://github.com/mashb1t/Fooocus) </br> and so on ... |
## Thanks
Many thanks to [twri](https://github.com/twri) and [3Diva](https://github.com/3Diva) and [Marc K3nt3L](https://github.com/K3nt3L) for creating additional SDXL styles available in Fooocus.
The project starts from a mixture of [Stable Diffusion WebUI](https://github.com/AUTOMATIC1111/stable-diffusion-webui) and [ComfyUI](https://github.com/comfyanonymous/ComfyUI) codebases.
Also, thanks [daswer123](https://github.com/daswer123) for contributing the Canvas Zoom!
## Update Log
The log is [here](update_log.md).
## Localization/Translation/I18N
You can put json files in the `language` folder to translate the user interface.
For example, below is the content of `Fooocus/language/example.json`:
```json
{
"Generate": "生成",
"Input Image": "入力画像",
"Advanced": "고급",
"SAI 3D Model": "SAI 3D Modèle"
}
```
If you add `--language example` arg, Fooocus will read `Fooocus/language/example.json` to translate the UI.
For example, you can edit the ending line of Windows `run.bat` as
.\python_embeded\python.exe -s Fooocus\entry_with_update.py --language example
Or `run_anime.bat` as
.\python_embeded\python.exe -s Fooocus\entry_with_update.py --language example --preset anime
Or `run_realistic.bat` as
.\python_embeded\python.exe -s Fooocus\entry_with_update.py --language example --preset realistic
For practical translation, you may create your own file like `Fooocus/language/jp.json` or `Fooocus/language/cn.json` and then use flag `--language jp` or `--language cn`. Apparently, these files do not exist now. **We need your help to create these files!**
Note that if no `--language` is given and at the same time `Fooocus/language/default.json` exists, Fooocus will always load `Fooocus/language/default.json` for translation. By default, the file `Fooocus/language/default.json` does not exist.
# Fooocus - Enhanced Version
### Key Enhancements:
- Fixed issues with incorrect prompts and seeds during enhance upscale and other enhancement processes.
- Added functionality for inline LoRA usage, now fully compatible with wildcards. You can use it like this: `<lora:lora_name:1>`.
- Also supports LoRA names that include special characters like `(` and Japanese characters.
- Introduced support for inline aspect ratios, which also work seamlessly with wildcards.
- You can specify aspect ratios in the prompt like `1280×720`, `1280x720`, or `1280*720`.
- You can also copy-paste them from the available aspect ratios in the Fooocus UI.
- Improved wildcard processing:
- You can now choose to process some wildcards in order while keeping others random, all within the same prompt.
- To process a wildcard in order, prefix it with `_o__` instead of `__`.
- Specify the starting row for wildcard processing using a suffix, like this: `_o__color__6_` (starts processing from the 6th row).
- **Important:** Keep the "process wildcards in order" flag **disabled** in the advanced section.
- Optimized the skip functionality to avoid using incorrect prompts when skipping generations.
- Various bug fixes and optimizations for improved performance and stability.
- Probably more enhancements/bug fixes that I can't remember 😀.
This version retains Fooocus' simplicity while enhancing its flexibility and accuracy in prompt handling.
---
### Support My Work
Ive been developing and improving Fooocus despite not having a GPU powerful enough to run it myself. If you find this project useful, consider supporting me so I can continue working on it and eventually get better hardware for testing and development.
☕ [Support me on Ko-fi](https://ko-fi.com/edward_no_genso)