mitek's picture
Upload 1159 files
7eb3676 verified
from .logger import logger
def image(src):
return f'<img src={src} style="width: 0px; min-width: 100%">'
def video(src):
return f'<video src={src} autoplay muted loop controls controlslist="nodownload noremoteplayback noplaybackrate" style="width: 0px; min-width: 100%" class="VHS_loopedvideo">'
def short_desc(desc):
return f'<div id=VHS_shortdesc>{desc}</div>'
def format_each(desc, **kwargs):
if isinstance(desc, dict):
res = {}
for k,v in desc.items():
res[format_each(k, **kwargs)] = format_each(v, **kwargs)
return res
if isinstance(desc, list):
res = []
for v in desc:
res.append(format_each(v, **kwargs))
return res
return desc.format(**kwargs)
def format_type(desc, lower, lowers=None, upper=None, uppers=None, cap=None):
"""Utility function for nodes with image/latent/mask variants"""
if lowers is None:
lowers = lower + 's'
if cap is None:
cap = lower.capitalize()
if upper is None:
upper = lower.upper()
if uppers is None:
uppers = lowers.upper()
return format_each(desc, lower=lower, lowers=lowers, upper=upper, uppers=uppers, cap=cap)
common_descriptions = {
'merge_strategy': [
'Determines what the output resolution will be if input resolutions don\'t match',
{'match A': 'Always use the resolution for A',
'match B': 'Always use the resolution for B',
'match smaller': 'Pick the smaller resolution by area',
'match larger': 'Pick the larger resolution by area',
}],
'scale_method': [
'Determines what method to use if scaling is required',
],
'crop_method': 'When sizes don\'t match, should the resized image have it\'s aspect ratio changed, or be cropped to maintain aspect ratio',
'VHS_PATH': [
'This is a VHS_PATH input. When edited, it provides a list of possible valid files or directories',
video('https://github.com/Kosinkadink/ComfyUI-VideoHelperSuite/assets/4284322/729b7185-1fca-41d8-bc8d-a770bb2a5ce6'),
'The current top-most completion may be selected with Tab',
'You can navigate up a directory by pressing Ctrl+B (or Ctrl+W if supported by browser)',
'The filter on suggested file types can be disabled by pressing Ctrl+G.',
'If converted to an input, this functions as a string',
],
"GetCount": ['Get {cap} Count πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Return the number of {lowers} in an input as an INT'),
{'Inputs': {
'{lowers}': 'The input {lower}',
},
'Outputs': {
'count': 'The number of {lowers} in the input',
},
}],
"SelectEveryNth": ['Select Every Nth {cap} πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Keep only 1 {lower} for every interval'),
{'Inputs': {
'{lowers}': 'The input {lower}',
},
'Outputs': {
'{upper}': 'The output {lowers}',
'count': 'The number of {lowers} in the input',
},
'Widgets':{
'select_every_nth': 'The interval from which one frame is kept. 1 means no frames are skipped.',
'skip_first_{lowers}': 'A number of frames which that is skipped from the start. This applies before select_every_nth. As a result, multiple copies of the node can each have a different skip_first_frames to divide the {lower} into groups'
},
}],
}
descriptions = {
'VHS_VideoCombine': ['Video Combine πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Combine an image sequence into a video'), {
'Inputs': {
'images': 'The images to be turned into a video',
'audio':'(optional) audio to add to the video',
'meta_batch': '(optional) Connect to a Meta Batch manager to divide extremely long image sequences into sub batches. See the documentation for Meta Batch Manager',
'vae':['(optional) If provided, the node will take latents as input instead of images. This drastically reduces the required RAM (not VRAM) when working with long (100+ frames) sequences',
"Unlike on Load Video, this isn't always a strict upgrade over using a standalone VAE Decode.",
"If you have multiple Video Combine outputs, then the VAE decode will be performed for each output node increasing execution time",
"If you make any change to output settings on the Video Combine (such as changing the output format), the VAE decode will be performed again as the decoded result is (by design) not cached",
]
},
'Widgets':{
'frame_rate': 'The frame rate which will be used for the output video. Consider converting this to an input and connecting this to a Load Video with Video Info(Loaded)->fps. When including audio, failure to properly set this will result in audio desync',
'loop_count': 'The number of additional times the video should repeat. Can cause performance issues when used with long (100+ frames) sequences',
'filename_prefix': 'A prefix to add to the name of the output filename. This can include subfolders or format strings.',
'format': 'The output format to use. Formats starting with, \'image\' are saved with PIL, but formats starting with \'video\' utilize the video_formats system. \'video\' options require ffmpeg and selecting one frequently adds additional options to the node.',
'pingpong': 'Play the video normally, then repeat the video in reverse so that it \'pingpongs\' back and forth. This is frequently used to minimize the appearance of skips on very short animations.',
'save_output': 'Specifies if output files should be saved to the output folder, or the temporary output folder',
'videopreview': 'Displays a preview for the processed result. If advanced previews is enabled, the output is always converted to a format viewable from the browser. If the video has audio, it will also be previewed when moused over. Additional preview options can be accessed with right click.',
},
'Common Format Widgets': {
'crf': 'Determines how much to prioritize quality over filesize. Numbers vary between formats, but on each format that includes it, the default value provides visually loss less output',
'pix_fmt': ['The pixel format to use for output. Alternative options will often have higher quality at the cost of increased file size and reduced compatibility with external software.', {
'yuv420p': 'The most common and default format',
'yuv420p10le': 'Use 10 bit color depth. This can improve color quality when combined with 16bit input color depth',
'yuva420p': 'Include transparency in the output video'
}],
'input_color_depth': 'VHS supports outputting 16bit images. While this produces higher quality output, the difference usually isn\'t visible without postprocessing and it significantly increases file size and processing time.',
'save_metadata': 'Determines if metadata for the workflow should be included in the output video file',
}
}],
'VHS_LoadVideo': ['Load Video πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Loads a video from the input folder'),
{'Inputs': {
'meta_batch': '(optional) Connect to a Meta Batch manager to divide extremely long sequences into sub batches. See the documentation for Meta Batch Manager',
'vae': ['(optional) If provided the node will output latents instead of images. This drastically reduces the required RAM (not VRAM) when working with long (100+ frames) sequences',
'Using this is strongly encouraged unless connecting to a node that requires a blue image connection such as Apply Controllnet',
],
},
'Outputs': {
'IMAGE': 'The loaded images',
'frame_count': 'The length of images just returned',
'audio': 'The audio from the loaded video',
'video_info': 'Exposes additional info about the video such as the source frame rate, or the total length',
'LATENT': 'The loaded images pre-converted to latents. Only available when a vae is connected',
},
'Widgets': {
'video': 'The video file to be loaded. Lists all files with a video extension in the ComfyUI/Input folder',
'force_rate': 'Drops or duplicates frames so that the produced output has the target frame rate. Many motion models are trained on videos of a specific frame rate and will give better results if input matches that frame rate. If set to 0, all frames are returned. May give unusual results with inputs that have a variable frame rate like animated gifs. Reducing this value can also greatly reduce the execution time and memory requirements.',
'force_size': ['Allows for conveniently scaling the input without requiring an additional node. Provides options to maintain aspect ratio or conveniently target common training formats for Animate Diff', {'custom_width': 'Allows for an arbitrary width to be entered, cropping to maintain aspect ratio if both are set',
'custom_height': 'Allows for an arbitrary height to be entered, cropping to maintain aspect ratio if both are set'}],
'frame_load_cap': 'The maximum number of frames to load. If 0, all frames are loaded.',
'skip_first_frames': 'A number of frames which are discarded before producing output.',
'select_every_nth': 'Similar to frame rate. Keeps only the first of every n frames and discard the rest. Has better compatibility with variable frame rate inputs such as gifs. When combined with force_rate, select_every_nth_applies after force_rate so the resulting output has a frame rate equivalent to force_rate/select_every_nth. select_every_nth does not apply to skip_first_frames',
'choose video to upload': 'An upload button is provided to upload local files to the input folder',
'videopreview': 'Displays a preview for the selected video input. If advanced previews is enabled, this preview will reflect the frame_load_cap, force_rate, skip_first_frames, and select_every_nth values chosen. If the video has audio, it will also be previewed when moused over. Additional preview options can be accessed with right click.',
}
}],
'VHS_LoadVideoPath': ['Load Video (Path) πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Loads a video from an arbitrary path'),
{'Inputs': {
'meta_batch': '(optional) Connect to a Meta Batch manager to divide extremely long sequences into sub batches. See the documentation for Meta Batch Manager',
'vae': ['(optional) If provided the node will output latents instead of images. This drastically reduces the required RAM (not VRAM) when working with long (100+ frames) sequences',
'Using this is strongly encouraged unless connecting to a node that requires a blue image connection such as Apply Controllnet',
],
},
'Outputs': {
'IMAGE': 'The loaded images',
'frame_count': 'The length of images just returned',
'audio': 'The audio from the loaded video',
'video_info': 'Exposes additional info about the video such as the source frame rate, or the total length',
'LATENT': 'The loaded images pre-converted to latents. Only available when a vae is connected',
},
'Widgets': {
'video': ['The video file to be loaded.', 'You can also select an image to load it as a single frame'] + common_descriptions['VHS_PATH'],
'force_rate': 'Drops or duplicates frames so that the produced output has the target frame rate. Many motion models are trained on videos of a specific frame rate and will give better results if input matches that frame rate. If set to 0, all frames are returned. May give unusual results with inputs that have a variable frame rate like animated gifs. Reducing this value can also greatly reduce the execution time and memory requirements.',
'force_size': ['Allows for conveniently scaling the input without requiring an additional node. Provides options to maintain aspect ratio or conveniently target common training formats for Animate Diff', {'custom_width': 'Allows for an arbitrary width to be entered, cropping to maintain aspect ratio if both are set',
'custom_height': 'Allows for an arbitrary height to be entered, cropping to maintain aspect ratio if both are set'}],
'frame_load_cap': 'The maximum number of frames to load. If 0, all frames are loaded.',
'skip_first_frames': 'A number of frames which are discarded before producing output.',
'select_every_nth': 'Similar to frame rate. Keeps only the first of every n frames and discard the rest. Has better compatibility with variable frame rate inputs such as gifs. When combined with force_rate, select_every_nth_applies after force_rate so the resulting output has a frame rate equivalent to force_rate/select_every_nth. select_every_nth does not apply to skip_first_frames',
'videopreview': 'Displays a preview for the selected video input. Will only be shown if Advanced Previews is enabled. This preview will reflect the frame_load_cap, force_rate, skip_first_frames, and select_every_nth values chosen. If the video has audio, it will also be previewed when moused over. Additional preview options can be accessed with right click.',
}
}],
'VHS_LoadImages': ['Load Images πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Loads a sequence of images from a subdirectory of the input folder'),
{'Inputs': {
'meta_batch': '(optional) Connect to a Meta Batch manager to divide extremely long sequences into sub batches. See the documentation for Meta Batch Manager',
},
'Outputs': {
'IMAGE': 'The loaded images',
'MASK': 'The alpha channel of the loaded images.',
'frame_count': 'The length of images just returned',
},
'Widgets': {
'directory': 'The directory images will be loaded from. Filtered to process jpg, png, ppm, bmp, tif, and webp files',
'image_load_cap': 'The maximum number of images to load. If 0, all images are loaded.',
'skip_first_images': 'A number of images which are discarded before producing output.',
'select_every_nth': 'Keeps only the first of every n frames and discard the rest.',
'choose folder to upload': 'An upload button is provided to upload a local folder containing images to the input folder',
'videopreview': 'Displays a preview for the selected video input. Will only be shown if Advanced Previews is enabled. This preview will reflect the image_load_cap, skip_first_images, and select_every_nth values chosen. Additional preview options can be accessed with right click.',
}
}],
'VHS_LoadImagesPath': ['Load Images (Path) πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Loads a sequence of images from a subdirectory of the input folder'),
{'Inputs': {
'meta_batch': '(optional) Connect to a Meta Batch manager to divide extremely long sequences into sub batches. See the documentation for Meta Batch Manager',
},
'Outputs': {
'IMAGE': 'The loaded images',
'MASK': 'The alpha channel of the loaded images.',
'frame_count': 'The length of images just returned',
},
'Widgets': {
'directory': ['The directory images will be loaded from. Filtered to process jpg, png, ppm, bmp, tif, and webp files'] + common_descriptions['VHS_PATH'],
'image_load_cap': 'The maximum number of images to load. If 0, all images are loaded.',
'skip_first_images': 'A number of images which are discarded before producing output.',
'select_every_nth': 'Keeps only the first of every n frames and discard the rest.',
'videopreview': 'Displays a preview for the selected video input. Will only be shown if Advanced Previews is enabled. This preview will reflect the image_load_cap, skip_first_images, and select_every_nth values chosen. Additional preview options can be accessed with right click.',
}
}],
"VHS_LoadAudio": ['Load Audio (Path) πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Loads an audio file from an arbitrary path'),
{'Outputs': {
'audio': 'The loaded audio',
},
'Widgets': {
'audio_file': ['The audio file to be loaded.'] + common_descriptions['VHS_PATH'],
'seek_seconds': 'An offset from the start of the sound file that the audio should start from',
}
}],
"VHS_LoadAudioUpload": ['Load Audio (Upload) πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Loads an audio file from the input directory'),
"Very similar in functionality to the built-in LoadAudio. It was originally added before VHS swapped to use Comfy's internal AUDIO format, but provides the additional options for start time and duration",
{'Outputs': {
'audio': 'The loaded audio',
},
'Widgets': {
'audio': 'The audio file to be loaded.',
'start_time': 'An offset from the start of the sound file that the audio should start from',
'duration': 'A maximum limit for the audio. Disabled if 0',
'choose audio to upload': 'An upload button is provided to upload an audio file to the input folder',
}
}],
"VHS_AudioToVHSAudio": ['Audio to legacy VHS_AUDIO πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('utility function for compatibility with external nodes'),
"VHS used to use an internal VHS_AUDIO format for routing audio between inputs and outputs. This format was intended to only be used internally and was designed with a focus on performance over ease of use. Since ComfyUI now has an internal AUDIO format, VHS now uses this format. However, some custom node packs were made that are external to both ComfyUI and VHS that use VHS_AUDIO. This node was added so that those external nodes can still function",
{'Inputs': {
'audio': 'An input in the standardized AUDIO format',
},
'Outputs': {
'vhs_audio': 'An output in the legacy VHS_AUDIO format for use with external nodes',
},
}],
"VHS_VHSAudioToAudio": ['Legacy VHS_AUDIO to Audio πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('utility function for compatibility with external nodes'),
"VHS used to use an internal VHS_AUDIO format for routing audio between inputs and outputs. This format was intended to only be used internally and was designed with a focus on performance over ease of use. Since ComfyUI now has an internal AUDIO format, VHS now uses this format. However, some custom node packs were made that are external to both ComfyUI and VHS that use VHS_AUDIO. This node was added so that those external nodes can still function",
{'Inputs': {
'vhs_audio': 'An input in the legacy VHS_AUDIO format produced by an external node',
},
'Outputs': {
'vhs_audio': 'An output in the standardized AUDIO format',
},
}],
"VHS_PruneOutputs": ['Prune Outputs πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Automates deletion of undesired outputs from a Video Combine node.'),
'Video Combine produces a number of file outputs in addition to the final output. Some of these, such as a video file without audio included, are implementation limitations and are not feasible to solve. As an alternative, the Prune Outputs node is added to automate the deletion of these file outputs if they are not desired',
{'Inputs': {
'filenames': 'A connection from a Video Combine node to indicate which outputs should be pruned',
},
'Widgets': {
'options': ['Which files should be deleted',
{'Intermediate': 'Delete any files that were required for intermediate processing but are not the final output, like the no-audio output file when audio is included',
'Intermediate and Utility': 'Delete all produced files that aren\'t the final output, including the first frame png',
}]}
}],
"VHS_BatchManager": ['Meta Batch Manager πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Split the processing of a very long video into sets of smaller Meta Batches'),
"The Meta Batch Manager allows for extremely long input videos to be processed when all other methods for fitting the content in RAM fail. It does not effect VRAM usage.",
"It must be connected to at least one Input (a Load Video or Load Images) AND at least one Video Combine",
image("https://github.com/Kosinkadink/ComfyUI-VideoHelperSuite/assets/4284322/7cb3fb7e-59d8-4cb2-a09f-9c6698de8b1f"),
"It functions by holding both the inputs and ouputs open between executions, and automatically requeue's the workflow until one of the inputs is unable to provide additional images.",
"Because each sub execution only contains a subset of the total frames, each sub execution creates a hard window which temporal smoothing can not be applied across. This results in jumps in the output.",
{'Outputs': {
'meta_batch': 'Add all connected nodes to this Meta Batch',
},
'Widgets': {
'frames_per_batch': 'How many frames to process for each sub execution. If loading as image, each frame will use about 50MB of RAM (not VRAM), and this can safely be set in the 100-1000 range, depending on available memory. When loading and combining from latent space (no blue image noodles exist), this value can be much higher, around the 2,000 to 20,000 range',
}
}],
"VHS_VideoInfo": ['Video Info πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Splits information on a video into a numerous outputs'),
{'Inputs': {
'video_info': 'A connection to a Load Video node',
},
'Outputs': {
'source_fps🟨': 'The frame rate of the video',
'source_frame_count🟨': 'How many total frames the video contains before accounting for frame rate or select_every_nth',
'source_duration🟨': 'The length of images just returned in seconds',
'source_width🟨': 'The width',
'source_height🟨': 'The height',
'loaded_fps🟦': 'The frame rate after accounting for force_rate and select_every_nth. This output is of particular use as it can be connected to the converted frame_rate input of a Video Combine node to ensure audio remains synchronized.',
'loaded_frame_count🟦': 'The number of frames returned by the current execution. Identical to the frame_count returned by the node itself',
'loaded_duration🟦': 'The duration in seconds of returned images after accounting for frame_load_cap',
'loaded_width🟦': 'The width of the video after scaling. These coordinates are in image space even if loading to latent space',
'loaded_height🟦': 'The height of the video after scaling. These coordinates are in image space even if loading to latent space',
},
}],
"VHS_VideoInfoSource": ['Video Info Source πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Splits information on a video into a numerous outputs describing the file itself without accounting for load options'),
{'Inputs': {
'video_info': 'A connection to a Load Video node',
},
'Outputs': {
'source_fps🟨': 'The frame rate of the video',
'source_frame_count🟨': 'How many total frames the video contains before accounting for frame rate or select_every_nth',
'source_duration🟨': 'The length of images just returned in seconds',
'source_width🟨': 'The original width',
'source_height🟨': 'The original height',
}
}],
"VHS_VideoInfoLoaded": ['Video Info Loaded πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Splits information on a video into a numerous outputs describing the file itself after accounting for load options'),
{'Inputs': {
'video_info': 'A connection to a Load Video node',
},
'Outputs': {
'loaded_fps🟦': 'The frame rate after accounting for force_rate and select_every_nth. This output is of particular use as it can be connected to the converted frame_rate input of a Video Combine node to ensure audio remains synchronized.',
'loaded_frame_count🟦': 'The number of frames returned by the current execution. Identical to the frame_count returned by the node itself',
'loaded_duration🟦': 'The duration in seconds of returned images after accounting for frame_load_cap',
'loaded_width🟦': 'The width of the video after scaling. This is the dimension of the corresponding image even if loading as a latent directly',
'loaded_height🟦': 'The height of the video after scaling. This is the dimension of the corresponding image even if loading as a latent directly',
}
}],
"VHS_SelectFilename": ['VAE Select Filename πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Select a single filename from the VHS_FILENAMES output by a Video Combine and return it as a string'),
'Take care when combining this node with Prune Outputs. The VHS_FILENAMES object is immutable and will always contain the full list of output files, but execution order is undefined behavior (currently, Prune Outputs will generally execute first) and SelectFilename may return a path to a file that no longer exists.',
{'Inputs': {
'filenames': 'A VHS_FILENAMES from a Video Combine node',
},
'Outputs': {
'filename': 'A string representation of the full output path for the chosen file',
},
'Widgets': {
'index': 'The index of which file should be selected. The default, -1, chooses the most complete output',
},
}],
# Batched Nodes
"VHS_VAEEncodeBatched": ['VAE Encode Batched πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Encode images as latents with a manually specified batch size.'),
"Some people have ran into VRAM issues when encoding or decoding large batches of images. As a workaround, this node lets you manually set a batch size when encoding images.",
"Unless these issues have been encountered, it is simpler to use the native VAE Encode or to encode directly from a Load Video",
{'Inputs': {
'pixels': 'The images to be encoded.',
'vae': 'The VAE to use when encoding.',
},
'Outputs': {
'LATENT': 'The encoded latents.',
},
'Widgets': {
'per_batch': 'The maximum number of images to encode in each batch.',
},
}],
"VHS_VAEDecodeBatched": ['VAE Decode Batched πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Decode latents to images with a manually specified batch size'),
"Some people have ran into VRAM issues when encoding or decoding large batches of images. As a workaround, this node lets you manually set a batch size when decoding latents.",
"Unless these issues have been encountered, it is simpler to use the native VAE Decode or to decode from a Video Combine directly",
{'Inputs': {
'samples': 'The latents to be decoded.',
'vae': 'The VAE to use when decoding.',
},
'Outputs': {
'IMAGE': 'The decoded images.',
},
'Widgets': {
'per_batch': 'The maximum number of images to decode in each batch.',
},
}],
# Latent and Image nodes
"VHS_SplitLatents": ['Split Latents πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Split a set of latents into two groups'),
{'Inputs': {
'latents': 'The latents to be split.',
},
'Outputs': {
'LATENT_A': 'The first group of latents',
'A_count': 'The number of latents in group A. This will be equal to split_index unless the latents input has length less than split_index',
'LATENT_B': 'The second group of latents',
'B_count': 'The number of latents in group B'
},
'Widgets': {
'split_index': 'The index of the first latent that will be in the second output groups.',
},
}],
"VHS_SplitImages": ['Split Images πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Split a set of images into two groups'),
{'Inputs': {
'images': 'The images to be split.',
},
'Outputs': {
'IMAGE_A': 'The first group of images',
'A_count': 'The number of images in group A. This will be equal to split_index unless the images input has length less than split_index',
'IMAGE_B': 'The second group of images',
'B_count': 'The number of images in group B'
},
'Widgets': {
'split_index': 'The index of the first latent that will be in the second output groups.',
},
}],
"VHS_SplitMasks": ['Split Masks πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Split a set of masks into two groups'),
{'Inputs': {
'mask': 'The masks to be split.',
},
'Outputs': {
'MASK_A': 'The first group of masks',
'A_count': 'The number of masks in group A. This will be equal to split_index unless the mask input has length less than split_index',
'MASK_B': 'The second group of masks',
'B_count': 'The number of masks in group B'
},
'Widgets': {
'split_index': 'The index of the first latent that will be in the second output groups.',
},
}],
"VHS_MergeLatents": ['Merge Latents πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Combine two groups of latents into a single group of latents'),
{'Inputs': {
'latents_A': 'The first group of latents',
'latents_B': 'The first group of latents',
},
'Outputs': {
'LATENT': 'The combined group of latents',
'count': 'The length of the combined group',
},
'Widgets': {
'merge_strategy': common_descriptions['merge_strategy'],
'scale_method': common_descriptions['scale_method'],
'crop': common_descriptions['crop_method'],
},
}],
"VHS_MergeImages": ['Merge Images πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Combine two groups of images into a single group of images'),
{'Inputs': {
'images_A': 'The first group of images',
'images_B': 'The first group of images',
},
'Outputs': {
'IMAGE': 'The combined group of images',
'count': 'The length of the combined group',
},
'Widgets': {
'merge_strategy': common_descriptions['merge_strategy'],
'scale_method': common_descriptions['scale_method'],
'crop': common_descriptions['crop_method'],
},
}],
"VHS_MergeMasks": ['Merge Masks πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Combine two groups of masks into a single group of masks'),
{'Inputs': {
'mask_A': 'The first group of masks',
'mask_B': 'The first group of masks',
},
'Outputs': {
'MASK': 'The combined group of masks',
'count': 'The length of the combined group',
},
'Widgets': {
'merge_strategy': common_descriptions['merge_strategy'],
'scale_method': common_descriptions['scale_method'],
'crop': common_descriptions['crop_method'],
},
}],
"VHS_GetLatentCount": format_type(common_descriptions['GetCount'], 'latent'),
"VHS_GetImageCount": format_type(common_descriptions['GetCount'], 'image'),
"VHS_GetMaskCount": format_type(common_descriptions['GetCount'], 'mask'),
"VHS_DuplicateLatents": ['Repeat Latents πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Append copies of a latent to itself so it repeats'),
{'Inputs': {
'latents': 'The latents to be repeated',
},
'Outputs': {
'LATENT': 'The latent with repeats',
'count': 'The number of latents in the output. Equal to the length of the input latent * multiply_by',
},
'Widgets': {
'multiply_by': 'Controls the number of times the latent should repeat. 1, the default, means no change.',
},
}],
"VHS_DuplicateImages": ['Repeat Images πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Append copies of a image to itself so it repeats'),
{'Inputs': {
'IMAGES': 'The image to be repeated',
},
'Outputs': {
'IMAGE': 'The image with repeats',
'count': 'The number of image in the output. Equal to the length of the input image * multiply_by',
},
'Widgets': {
'multiply_by': 'Controls the number of times the mask should repeat. 1, the default, means no change.',
},
}],
"VHS_DuplicateMasks": ['Repeat Masks πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Append copies of a mask to itself so it repeats'),
{'Inputs': {
'masks': 'The masks to be repeated',
},
'Outputs': {
'LATENT': 'The mask with repeats',
'count': 'The number of mask in the output. Equal to the length of the input mask * multiply_by',
},
'Widgets': {
'multiply_by': 'Controls the number of times the mask should repeat. 1, the default, means no change.',
},
}],
"VHS_SelectEveryNthLatent": format_type(common_descriptions['SelectEveryNth'], 'latent'),
"VHS_SelectEveryNthImage": format_type(common_descriptions['SelectEveryNth'], 'image'),
#TODO: fix discrepency of input being mask instead of masks?
"VHS_SelectEveryNthMask": format_type(common_descriptions['SelectEveryNth'], 'mask', lowers='mask'),
#TODO: port documentation for select nodes to new system
#"VHS_SelectLatents": None,
#"VHS_SelectImages": None,
#"VHS_SelectMasks": None,
"VHS_Unbatch": ['Unbatch πŸŽ₯πŸ…₯πŸ…—πŸ…’', short_desc('Experimental node to unbatch a list of items into a single concatenated item'),
"Useful for when you want a single video output from a complex workflow",
"Has no relation to the Meta Batch system of VHS",
{'Inputs': {
'batched': 'Any input which may or may not be batched',
},
'Outputs': {
'unbatched': 'A single output element. Torch tensors are concatenated across dim 0, all other types are added which functions as concatenation for strings and arrays, but may give undesired results for other types',
},
}],
}
def as_html(entry, depth=0):
if isinstance(entry, dict):
size = 0.8 if depth < 2 else 1
html = ''
for k in entry:
if k == "collapsed":
continue
collapse_single = k.endswith("_collapsed")
if collapse_single:
name = k[:-len("_collapsed")]
else:
name = k
collapse_flag = ' VHS_precollapse' if entry.get("collapsed", False) or collapse_single else ''
html += f'<div vhs_title=\"{name}\" style=\"display: flex; font-size: {size}em\" class=\"VHS_collapse{collapse_flag}\"><div style=\"color: #AAA; height: 1.5em;\">[<span style=\"font-family: monospace\">-</span>]</div><div style=\"width: 100%\">{name}: {as_html(entry[k], depth=depth+1)}</div></div>'
return html
if isinstance(entry, list):
if depth == 0:
depth += 1
size = .8
else:
size = 1
html = ''
html += entry[0]
for i in entry[1:]:
html += f'<div style=\"font-size: {size}em\">{as_html(i, depth=depth)}</div>'
return html
return str(entry)
def format_descriptions(nodes):
for k in descriptions:
if k.endswith("_collapsed"):
k = k[:-len("_collapsed")]
nodes[k].DESCRIPTION = as_html(descriptions[k])
undocumented_nodes = []
for k in nodes:
if not hasattr(nodes[k], "DESCRIPTION"):
undocumented_nodes.append(k)
if len(undocumented_nodes) > 0:
logger.warning('Some nodes have not been documented %s', undocumented_nodes)