Vaes[]Comfy{'prompt': {'4': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, '17': {'inputs': {'cfg': 10, 'seed': 715571678028107, 'model': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'steps': 28, 'denoise': 1, 'negative': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 14, 'height': 4096, 'text_g': 'embedding:negativeXL_D.safetensors, military headwear', 'text_l': 'embedding:negativeXL_D.safetensors, military headwear', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, 'positive': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 0, 'height': 4044, 'text_g': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'text_l': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, 'scheduler': 'normal', 'latent_image': {'inputs': {'width': 856, 'height': 1160, 'batch_size': 2}, 'class_type': 'EmptyLatentImage'}, 'sampler_name': 'dpmpp_2s_ancestral'}, 'class_type': 'KSampler'}, '19': {'inputs': {'images': {'inputs': {'vae': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'samples': {'inputs': {'cfg': 10, 'seed': 715571678028107, 'model': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'steps': 28, 'denoise': 1, 'negative': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 14, 'height': 4096, 'text_g': 'embedding:negativeXL_D.safetensors, military headwear', 'text_l': 'embedding:negativeXL_D.safetensors, military headwear', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, 'positive': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 0, 'height': 4044, 'text_g': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'text_l': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, 'scheduler': 'normal', 'latent_image': {'inputs': {'width': 856, 'height': 1160, 'batch_size': 2}, 'class_type': 'EmptyLatentImage'}, 'sampler_name': 'dpmpp_2s_ancestral'}, 'class_type': 'KSampler'}}, 'class_type': 'VAEDecode'}, 'filename_prefix': 'xl_output'}, 'class_type': 'SaveImage'}, '26': {'inputs': {'vae': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'samples': {'inputs': {'cfg': 10, 'seed': 715571678028107, 'model': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'steps': 28, 'denoise': 1, 'negative': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 14, 'height': 4096, 'text_g': 'embedding:negativeXL_D.safetensors, military headwear', 'text_l': 'embedding:negativeXL_D.safetensors, military headwear', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, 'positive': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 0, 'height': 4044, 'text_g': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'text_l': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, 'scheduler': 'normal', 'latent_image': {'inputs': {'width': 856, 'height': 1160, 'batch_size': 2}, 'class_type': 'EmptyLatentImage'}, 'sampler_name': 'dpmpp_2s_ancestral'}, 'class_type': 'KSampler'}}, 'class_type': 'VAEDecode'}, '30': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 0, 'height': 4044, 'text_g': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'text_l': 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, '33': {'inputs': {'clip': {'inputs': {'ckpt_name': 'CounterfeitXL_β.safetensors'}, 'class_type': 'CheckpointLoaderSimple'}, 'width': 4096, 'crop_h': 0, 'crop_w': 14, 'height': 4096, 'text_g': 'embedding:negativeXL_D.safetensors, military headwear', 'text_l': 'embedding:negativeXL_D.safetensors, military headwear', 'target_width': 1024, 'target_height': 1024}, 'class_type': 'CLIPTextEncodeSDXL'}, '43': {'inputs': {'width': 1536, 'height': 640, 'batch_size': 2}, 'class_type': 'EmptyLatentImage'}, '59': {'inputs': {'width': 856, 'height': 1160, 'batch_size': 2}, 'class_type': 'EmptyLatentImage'}}, 'workflow': {'extra': {}, 'links': [[45, 17, 0, 26, 0, 'LATENT'], [46, 4, 2, 26, 1, 'VAE'], [48, 26, 0, 19, 0, 'IMAGE'], [49, 4, 1, 30, 0, 'CLIP'], [50, 30, 0, 17, 1, 'CONDITIONING'], [51, 32, 0, 30, 1, 'STRING'], [52, 32, 0, 30, 2, 'STRING'], [55, 16, 0, 33, 1, 'STRING'], [56, 33, 0, 17, 2, 'CONDITIONING'], [57, 16, 0, 33, 2, 'STRING'], [242, 4, 1, 33, 0, 'CLIP'], [261, 4, 0, 17, 0, 'MODEL'], [262, 59, 0, 17, 3, 'LATENT']], 'nodes': [{'id': 26, 'pos': [1088, -227], 'mode': 0, 'size': {'0': 140, '1': 46}, 'type': 'VAEDecode', 'flags': {}, 'order': 8, 'inputs': [{'link': 45, 'name': 'samples', 'type': 'LATENT', 'slot_index': 0}, {'link': 46, 'name': 'vae', 'type': 'VAE', 'slot_index': 1}], 'outputs': [{'name': 'IMAGE', 'type': 'IMAGE', 'links': [48], 'shape': 3, 'slot_index': 0}], 'properties': {'Node name for S&R': 'VAEDecode'}}, {'id': 33, 'pos': [389, 95], 'mode': 0, 'size': {'0': 400, '1': 270.0000305175781}, 'type': 'CLIPTextEncodeSDXL', 'flags': {}, 'order': 5, 'inputs': [{'link': 242, 'name': 'clip', 'type': 'CLIP', 'slot_index': 0}, {'link': 55, 'name': 'text_g', 'type': 'STRING', 'widget': {'name': 'text_g', 'config': ['STRING', {'default': 'CLIP_G', 'multiline': True}]}}, {'link': 57, 'name': 'text_l', 'type': 'STRING', 'widget': {'name': 'text_l', 'config': ['STRING', {'default': 'CLIP_L', 'multiline': True}]}, 'slot_index': 2}], 'outputs': [{'name': 'CONDITIONING', 'type': 'CONDITIONING', 'links': [56], 'shape': 3, 'slot_index': 0}], 'properties': {'Node name for S&R': 'CLIPTextEncodeSDXL'}, 'widgets_values': [4096, 4096, 14, 0, 1024, 1024, 'embedding:negativeXL_D.safetensors, military headwear', 'embedding:negativeXL_D.safetensors, military headwear']}, {'id': 30, 'pos': [382, -235], 'mode': 0, 'size': {'0': 400, '1': 270.0000305175781}, 'type': 'CLIPTextEncodeSDXL', 'flags': {}, 'order': 6, 'inputs': [{'link': 49, 'name': 'clip', 'type': 'CLIP', 'slot_index': 0}, {'link': 51, 'name': 'text_g', 'type': 'STRING', 'widget': {'name': 'text_g', 'config': ['STRING', {'default': 'CLIP_G', 'multiline': True}]}}, {'link': 52, 'name': 'text_l', 'type': 'STRING', 'widget': {'name': 'text_l', 'config': ['STRING', {'default': 'CLIP_L', 'multiline': True}]}}], 'outputs': [{'name': 'CONDITIONING', 'type': 'CONDITIONING', 'links': [50], 'shape': 3, 'slot_index': 0}], 'properties': {'Node name for S&R': 'CLIPTextEncodeSDXL'}, 'widgets_values': [4096, 4044, 0, 0, 1024, 1024, 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant', 'close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant']}, {'id': 16, 'pos': [-101, -32], 'mode': 0, 'size': {'0': 398, '1': 140}, 'type': 'PrimitiveNode', 'flags': {}, 'order': 0, 'title': 'Negative Prompt', 'outputs': [{'name': 'STRING', 'type': 'STRING', 'links': [55, 57], 'widget': {'name': 'text_g', 'config': ['STRING', {'default': 'CLIP_G', 'multiline': True}]}, 'slot_index': 0}], 'properties': {}, 'widgets_values': ['embedding:negativeXL_D.safetensors, military headwear']}, {'id': 17, 'pos': [833, -121], 'mode': 0, 'size': {'0': 315, '1': 474}, 'type': 'KSampler', 'flags': {}, 'order': 7, 'inputs': [{'link': 261, 'name': 'model', 'type': 'MODEL'}, {'link': 50, 'name': 'positive', 'type': 'CONDITIONING'}, {'link': 56, 'name': 'negative', 'type': 'CONDITIONING'}, {'link': 262, 'name': 'latent_image', 'type': 'LATENT'}], 'outputs': [{'name': 'LATENT', 'type': 'LATENT', 'links': [45], 'shape': 3, 'slot_index': 0}], 'properties': {'Node name for S&R': 'KSampler'}, 'widgets_values': [715571678028107, 'randomize', 28, 10, 'dpmpp_2s_ancestral', 'normal', 1]}, {'id': 43, 'pos': [389, 430], 'mode': 0, 'size': {'0': 315, '1': 106}, 'type': 'EmptyLatentImage', 'flags': {}, 'order': 1, 'outputs': [{'name': 'LATENT', 'type': 'LATENT', 'links': [], 'shape': 3, 'slot_index': 0}], 'properties': {'Node name for S&R': 'EmptyLatentImage'}, 'widgets_values': [1536, 640, 2]}, {'id': 59, 'pos': [747, 435], 'mode': 0, 'size': {'0': 315, '1': 106}, 'type': 'EmptyLatentImage', 'flags': {}, 'order': 2, 'outputs': [{'name': 'LATENT', 'type': 'LATENT', 'links': [262], 'shape': 3, 'slot_index': 0}], 'properties': {'Node name for S&R': 'EmptyLatentImage'}, 'widgets_values': [856, 1160, 2]}, {'id': 4, 'pos': [-158, 197], 'mode': 0, 'size': {'0': 397, '1': 98}, 'type': 'CheckpointLoaderSimple', 'flags': {}, 'order': 3, 'outputs': [{'name': 'MODEL', 'type': 'MODEL', 'links': [261], 'slot_index': 0}, {'name': 'CLIP', 'type': 'CLIP', 'links': [49, 242], 'slot_index': 1}, {'name': 'VAE', 'type': 'VAE', 'links': [46], 'slot_index': 2}], 'properties': {'Node name for S&R': 'CheckpointLoaderSimple'}, 'widgets_values': ['CounterfeitXL_β.safetensors']}, {'id': 32, 'pos': [-102, -254], 'mode': 0, 'size': {'0': 398, '1': 140}, 'type': 'PrimitiveNode', 'flags': {}, 'order': 4, 'title': 'Positive Prompt', 'outputs': [{'name': 'STRING', 'type': 'STRING', 'links': [51, 52], 'widget': {'name': 'text_g', 'config': ['STRING', {'default': 'CLIP_G', 'multiline': True}]}, 'slot_index': 0}], 'properties': {}, 'widgets_values': ['close up, masterpiece, best quality, bokeh, cute, 1girl, solo, beret, hair clip, beige hair, round eyewear, brown jacket, turtleneck sweater, upper body, looking at viewer, yellow eyes, closed mouth, leaf, white background, plant']}, {'id': 19, 'pos': [1179, 65], 'mode': 0, 'size': {'0': 509.235107421875, '1': 320.00152587890625}, 'type': 'SaveImage', 'flags': {}, 'order': 9, 'inputs': [{'link': 48, 'name': 'images', 'type': 'IMAGE'}], 'properties': {}, 'widgets_values': ['xl_output']}], 'config': {}, 'groups': [], 'version': 0.4, 'last_link_id': 262, 'last_node_id': 75}}Width856Height1160Models['CounterfeitXL_β.safetensors']Denoise1Modelids[]SchedulernormalUpscalers[]Versionids[]Controlnets[]Additionalresources[]