Added new models to Groq LLM. Added a new node: Tiktoken Tokenizer Info.
This repository hosts a collection of nodes developed for ComfyUI. It aims to share useful components that enhance the functionality of ComfyUI projects. Some nodes are forks or versions of nodes from other packs, some are bespoke and useful, and some are experimental and are quite useless, so they have been marked with a Caution
label in this document.
๐ Get File Path - Returns the file path in different formats to a file in your /input-folder.
๐พ Save Text File With Path Node - Save text file, and return the saved file's path.
๐ผ๏ธ Download Image from URL Node - Download an image from the web.
๐ Tiktoken Tokenizer Info - Returns token information about input text and lets you split it.
๐งน String Cleaning - Cleans up text strings.
โจ๐ฌ Groq LLM API Node - Query Groq large language model.
โจ๐ท Groq VLM API Node - Query Groq vision language model.
โจ๐ Groq ALM API Node - Query Groq Audio Model.
โ Generate Negative Prompt Node - Generate negative prompts automatically.
You may need to manually install the requirements. They should be listed in requirements.txt
You may need to install the following libraries using pip install XXX
:
configparser
groq
transformers
torch
This node returns the file path of a given file in the \input-folder.
It is meant to have a browse-button so you can browse to any file, but it doesn't yet.
If you know how to add this, please let me know or do a pull request.
This node is adapted and enhanced from the Save Text File node found in the YMC GitHub ymc-node-suite-comfyui pack.
The node can now give you a full file path output if you need it, as well as output the file-name as a separate output, in case you need it for something else.
[!IMPORTANT]
2024-06-05 - Version 1.1.1
The node was severely updated so existing workflows are going to break. I won't do another overhaul like this.
The new node is more consistent in functionality and more intentional with the inputs and outputs.
It now handles more edge cases and supports both a prefix, suffix, a dynamic counting with customizable separator before/after the counter in the right circumstances.
Sorry for any troubles caused.
This node downloads an image from an URL and lets you use it.
It also outputs the Width/Height of the image.
save_path
line to prevent saving the image (it will still be saved in the TEMP-folder).save_file_name_override
section, the file will be saved with this name.
[!IMPORTANT]
2024-09-14 - Version 1.2.0
This node was renamed in the code to match the functionality. This may break existing nodes.
This node takes text as input, and returns a bunch of data from the tiktoken tokenizer.
It returns the following values:
token_count
: Total number of tokenscharacter_count
: Total number of charactersword_count
: Total number of wordssplit_string
: Tokenized list of stringssplit_string_list
: Tokenized list of strings (output as list)split_token_ids
: List of token IDssplit_token_ids_list
: List of token IDs (output as list)text_hash
: Text hashspecial_tokens_used
: Special tokens usedspecial_tokens_used_list
: Special tokens used (output as list)token_chunk_by_size
: Returns the input text, split into different strings in a list by the token_chunk_size
value.token_chunk_by_size_to_word
: Same as above but respects "words" by stripping backwards to the nearest space and splitting the chunk there.token_chunk_by_size_to_section
: Same as above, but strips backwards to the nearest newline, period or comma.This node helps you quickly clean up and format strings by letting you remove leading or trailing spaces, periods, commas, or custom text, as well as removing linebreaks, or replacing them with a period.
input_string
: Your input string. Use ComfyUI-Easy-Use for looping through a list of strings.collapse_sequential_spaces
: Collapses sequential spaces (" ") in a string into one.strip_leading_spaces
: Removes any leading spaces from each line of the input string.strip_trailing_spaces
: Removes any trailing spaces from each line of the input string.strip_leading_symbols
: Removes leading punctuation symbols (, . ! ? : ;) from each line of the input string.strip_trailing_symbols
: Removes leading punctuation symbols (, . ! ? : ;) from each line of the input string.strip_inside_tags
: Removes any tags and the characters inside. <> would strip out anything like <html>
or </div>
, including the <
and >
strip_newlines
: Removes any linebreaks in the input string.replace_newlines_with_period_space
: Replaces any linebreaks in the input string with a ". ". If multiple linebreaks are found in a row, they will be replaced with a single ". ".strip_leading_custom
: Removes any leading characters, words or symbols from each line of the input string. One entry per line. Space (" ") is supported. Will be processed in order, so you can combine multiple lines. Does not support linebreak removal.strip_trailing_custom
: Removes any trailing characters, words or symbols from each line of the input string. One entry per line. Space (" ") is supported. Will be processed in order, so you can combine multiple lines. Does not support linebreak removal.strip_all_custom
: Removes any characters, words or symbols found anywhere in the text. One entry per line. Space (" ") is supported. Will be processed in order, so you can combine multiple lines. Does not support linebreak removal.multiline_find
: Find and replace for multiple entries. Will be processed in order.multiline_replace
: Find and replace for multiple entries. Will be processed in order.[!IMPORTANT]
2024-09-14 - Version 1.2.0
This node was renamed to match the new VLM and ALM nodes added.
This node makes an API call to groq, and returns the response in text format.
You need to manually enter your groq API key into the GroqConfig.ini
file.
Currently, the Groq API can be used for free, with very friendly and generous rate limits.
model: Choose from a drop-down one of the available models. The list need to be manually updated when they add additional models.
preset: This is a dropdown with a few preset prompts, the user's own presets, or the option to use a fully custom prompt. See examples and presets below.
system_message: The system message to send to the API. This is only used with the Use [system_message] and [user_input]
option in the preset list. The other presets provide their own system message.
user_input: This is used with the Use [system_message] and [user_input]
, but can also be used with presets. In the system message, just mention the USER to refer to this input field. See the presets for examples.
temperature: Controls the randomness of the response. A higher temperature leads to more varied responses.
max_tokens: The maximum number of tokens that the model can process in a single response. Limits can be found here.
top_p: The threshold for the most probable next token to use. Higher values result in more predictable results.
seed: Random seed. Change the control_after_generate
option below if you want to re-use the seed or get a new generation each time.
control_after_generate: Standard comfy seed controls. Set it to fixed
or randomize
based on your needs.
stop: Enter a word or stopping sequence which will terminate the AI's output. The string itself will not be returned.
stop
is not compatible with json_mode
.json_mode: If enabled, the model will output the result in JSON format.
Note: You must include a description of the desired JSON format in the system message. See the examples below.
Note: json_mode
is not compatible with stop
.
The following presets can be found in the \nodes\groq\DefaultPrompts.json
file. They can be edited, but it's better to copy the presets to the UserPrompts.json
-file.
This preset, (default), means that the next two fields are fully utilized. Manually enter the instruction to the AI in the system_message
field, and if you have any specific requests in the user_input
field. Combined they make up the complete instruction to the LLM. Sometimes a system message is enough, and inside the system message you could even refer to the contents of the user input.
This is a tailored instruction that will return a randomized Stable Diffusion-like prompt. If you enter some text in the user_input
area, you should get a prompt about this subject. You can also leave it empty and it will create its own examples based on the underlying prompt.
You should get better result from providing it with a short sentence to start it off.
This will return a negative prompt which intends to be used together with the user_input
string to complement it and enhance a resulting image.
This will return a list format of 10 subjects for an image, described in a simple and short style. These work good as user_input
for the Generate a prompt about [user_input]
preset.
You should also manually turn on json_mode
when using this prompt. You should get a stable json formatted output from it in a similar style to the Generate a prompt about [user_input]
above.
Note: You can actually use the entire result (JSON and all), as your prompt. Stable Diffusion seem to handle it quite fine.
Edit the \nodes\groq\UserPrompts.json
file to create your own presets.
Follow the existing structure and look at the DefaultPrompts.json
for examples.
[!IMPORTANT]
2024-09-27 - Version 1.2.1
Added new Llama 3.2 vision model to the list, but this model is not yet officially available. Once it is, this should automatically work.
This node makes an API call to groq with an attached image and then uses Vision Language Models to return a description of the image, or answer to a question about the image in text format.
You need to manually enter your groq API key into the GroqConfig.ini
file.
Currently, the Groq API can be used for free, with very friendly and generous rate limits.
Image Size Limit: The maximum allowed size for a request containing an image URL as input is 20MB. Requests larger than this limit will return a 400 error.
Request Size Limit (Base64 Enconded Images): The maximum allowed size for a request containing a base64 encoded image is 4MB. Requests larger than this limit will return a 413 error.
This node makes an API call to groq with an attached audio file and then uses Audio Language Models to transcribe the audio and return the text in different output formats.
The model distil-whisper-large-v3-en
only supports the language en
.
The model whisper-large-v3
supports the languages listed below. It can also be left empty, but this provides worse results than running the model locally.
[!NOTE] The presets / prompt do very little. They are meant to help you guide the output, but I don't get any relevant results.
You can convert the file_path
to input to use the Get File Path node to find your files.
https://www.wikiwand.com/en/articles/List_of_ISO_639_language_codes
is tg uz zh ru tr hi la tk haw fr vi cs hu kk he cy bs sw ht mn gl si mg sa es ja pt lt mr fa sl kn uk ms ta hr bg pa yi fo th lv ln ca br sq jv sn gu ba te bn et sd tl ha de hy so oc nn az km yo ko pl da mi ml ka am tt su yue nl no ne mt my ur ps ar id fi el ro as en it sk be lo lb bo sv sr mk eu
You need to manually enter your groq API key into the GroqConfig.ini
file.
Currently, the Groq API can be used for free, with very friendly and generous rate limits.
You can use this to generate files to use in a Karaoke app.
[!CAUTION] This node is highly experimental, and does not produce any useful result right now. It also requires you to download a specially trained model for it. It's just not worth the effort. It's mostly here to share a work in progress project.
This node utilizes a GPT-2 text inference model to generate a negative prompt that is supposed to enhance the aspects of the positive prompt.
[!IMPORTANT] Installation Step: Download the weights.pt file from the project's Hugging Face repository.
Place the
weights.pt
file in the following directory of your ComfyUI setup without renaming it:\ComfyUI\custom_nodes\ComfyUI-mnemic-nodes\nodes\negativeprompt
The directory should resemble the following structure:
For additional information, please visit the project's GitHub page.