More API Nodes (#7956)
* Add Ideogram generate node.
* Add staging api.
* Add API_NODE and common error for missing auth token (#5)
* Add Minimax Video Generation + Async Task queue polling example (#6)
* [Minimax] Show video preview and embed workflow in ouput (#7)
* Remove uv.lock
* Remove polling operations.
* Revert "Remove polling operations."
* Update stubs.
* Added Ideogram and Minimax back in.
* Added initial BFL Flux 1.1 [pro] Ultra node (#11)
* Add --comfy-api-base launch arg (#13)
* Add instructions for staging development. (#14)
* remove validation to make it easier to run against LAN copies of the API
* Manually add BFL polling status response schema (#15)
* Add function for uploading files. (#18)
* Add Luma nodes (#16)
* Refactor util functions (#20)
* Add VIDEO type (#21)
* Add rest of Luma node functionality (#19)
* Fix image_luma_ref not working (#28)
* [Bug] Remove duplicated option T2V-01 in MinimaxTextToVideoNode (#31)
* Add utils to map from pydantic model fields to comfy node inputs (#30)
* add veo2, bump av req (#32)
* Add Recraft nodes (#29)
* Add Kling Nodes (#12)
* Add Camera Concepts (luma_concepts) to Luma Video nodes (#33)
* Add Runway nodes (#17)
* Convert Minimax node to use VIDEO output type (#34)
* Standard `CATEGORY` system for api nodes (#35)
* Set `Content-Type` header when uploading files (#36)
* add better error propagation to veo2 (#37)
* Add Realistic Image and Logo Raster styles for Recraft v3 (#38)
* Fix runway image upload and progress polling (#39)
* Fix image upload for Luma: only include `Content-Type` header field if it's set explicitly (#40)
* Moved Luma nodes to nodes_luma.py (#47)
* Moved Recraft nodes to nodes_recraft.py (#48)
* Add Pixverse nodes (#46)
* Move and fix BFL nodes to node_bfl.py (#49)
* Move and edit Minimax node to nodes_minimax.py (#50)
* Add Minimax Image to Video node + Cleanup (#51)
* Add Recraft Text to Vector node, add Save SVG node to handle its output (#53)
* Added pixverse_template support to Pixverse Text to Video node (#54)
* Added Recraft Controls + Recraft Color RGB nodes (#57)
* split remaining nodes out of nodes_api, make utility lib, refactor ideogram (#61)
* Add types and doctstrings to utils file (#64)
* Fix: `PollingOperation` progress bar update progress by absolute value (#65)
* Use common download function in kling nodes module (#67)
* Fix: Luma video nodes in `api nodes/image` category (#68)
* Set request type explicitly (#66)
* Add `control_after_generate` to all seed inputs (#69)
* Fix bug: deleting `Content-Type` when property does not exist (#73)
* Add preview to Save SVG node (#74)
* change default poll interval (#76), rework veo2
* Add Pixverse and updated Kling types (#75)
* Added Pixverse Image to VIdeo node (#77)
* Add Pixverse Transition Video node (#79)
* Proper ray-1-6 support as fix has been applied in backend (#80)
* Added Recraft Style - Infinite Style Library node (#82)
* add ideogram v3 (#83)
* [Kling] Split Camera Control config to its own node (#81)
* Add Pika i2v and t2v nodes (#52)
* Temporary Fix for Runway (#87)
* Added Stability Stable Image Ultra node (#86)
* Remove Runway nodes (#88)
* Fix: Prompt text can't be validated in Kling nodes when using primitive nodes (#90)
* Fix: typo in node name "Stabiliy" => "Stability" (#91)
* Add String (Multiline) node (#93)
* Update Pika Duration and Resolution options (#94)
* Change base branch to master. Not main. (#95)
* Fix UploadRequest file_name param (#98)
* Removed Infinite Style Library until later (#99)
* fix ideogram style types (#100)
* fix multi image return (#101)
* add metadata saving to SVG (#102)
* Bump templates version to include API node template workflows (#104)
* Fix: `download_url_to_video_output` return type (#103)
* fix 4o generation bug (#106)
* Serve SVG files directly (#107)
* Add a bunch of nodes, 3 ready to use, the rest waiting for endpoint support (#108)
* Revert "Serve SVG files directly" (#111)
* Expose 4 remaining Recraft nodes (#112)
* [Kling] Add `Duration` and `Video ID` outputs (#105)
* Fix: datamodel-codegen sets string#binary type to non-existent `bytes_aliased` variable (#114)
* Fix: Dall-e 2 not setting request content-type dynamically (#113)
* Default request timeout: one hour. (#116)
* Add Kling nodes: camera control, start-end frame, lip-sync, video extend (#115)
* Add 8 nodes - 4 BFL, 4 Stability (#117)
* Fix error for Recraft ImageToImage error for nonexistent random_seed param (#118)
* Add remaining Pika nodes (#119)
* Make controls input work for Recraft Image to Image node (#120)
* Use upstream PR: Support saving Comfy VIDEO type to buffer (#123)
* Use Upstream PR: "Fix: Error creating video when sliced audio tensor chunks are non-c-contiguous" (#127)
* Improve audio upload utils (#128)
* Fix: Nested `AnyUrl` in request model cannot be serialized (Kling, Runway) (#129)
* Show errors and API output URLs to the user (change log levels) (#131)
* Fix: Luma I2I fails when weight is <=0.01 (#132)
* Change category of `LumaConcepts` node from image to video (#133)
* Fix: `image.shape` accessed before `image` is null-checked (#134)
* Apply small fixes and most prompt validation (if needed to avoid API error) (#135)
* Node name/category modifications (#140)
* Add back Recraft Style - Infinite Style Library node (#141)
* Fixed Kling: Check attributes of pydantic types. (#144)
* Bump `comfyui-workflow-templates` version (#142)
* [Kling] Print response data when error validating response (#146)
* Fix: error validating Kling image response, trying to use `"key" in` on Pydantic class instance (#147)
* [Kling] Fix: Correct/verify supported subset of input combos in Kling nodes (#149)
* [Kling] Fix typo in node description (#150)
* [Kling] Fix: CFG min/max not being enforced (#151)
* Rebase launch-rebase (private) on prep-branch (public copy of master) (#153)
* Bump templates version (#154)
* Fix: Kling image gen nodes don't return entire batch when `n` > 1 (#152)
* Remove pixverse_template from PixVerse Transition Video node (#155)
* Invert image_weight value on Luma Image to Image node (#156)
* Invert and resize mask for Ideogram V3 node to match masking conventions (#158)
* [Kling] Fix: image generation nodes not returning Tuple (#159)
* [Bug] [Kling] Fix Kling camera control (#161)
* Kling Image Gen v2 + improve node descriptions for Flux/OpenAI (#160)
* [Kling] Don't return video_id from dual effect video (#162)
* Bump frontend to 1.18.8 (#163)
* Use 3.9 compat syntax (#164)
* Use Python 3.10
* add example env var
* Update templates to 0.1.11
* Bump frontend to 1.18.9
---------
Co-authored-by: Robin Huang <robin.j.huang@gmail.com>
Co-authored-by: Christian Byrne <cbyrne@comfy.org>
Co-authored-by: thot experiment <94414189+thot-experiment@users.noreply.github.com>
2025-05-06 03:23:00 -05:00
from __future__ import annotations
from enum import Enum
from typing import Any , Dict , Optional
from pydantic import BaseModel , Field , confloat , conint
class BFLOutputFormat ( str , Enum ) :
png = ' png '
jpeg = ' jpeg '
class BFLFluxExpandImageRequest ( BaseModel ) :
prompt : str = Field ( . . . , description = ' The description of the changes you want to make. This text guides the expansion process, allowing you to specify features, styles, or modifications for the expanded areas. ' )
prompt_upsampling : Optional [ bool ] = Field (
None , description = ' Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation. '
)
seed : Optional [ int ] = Field ( None , description = ' The seed value for reproducibility. ' )
top : conint ( ge = 0 , le = 2048 ) = Field ( . . . , description = ' Number of pixels to expand at the top of the image ' )
bottom : conint ( ge = 0 , le = 2048 ) = Field ( . . . , description = ' Number of pixels to expand at the bottom of the image ' )
left : conint ( ge = 0 , le = 2048 ) = Field ( . . . , description = ' Number of pixels to expand at the left side of the image ' )
right : conint ( ge = 0 , le = 2048 ) = Field ( . . . , description = ' Number of pixels to expand at the right side of the image ' )
steps : conint ( ge = 15 , le = 50 ) = Field ( . . . , description = ' Number of steps for the image generation process ' )
guidance : confloat ( ge = 1.5 , le = 100 ) = Field ( . . . , description = ' Guidance strength for the image generation process ' )
safety_tolerance : Optional [ conint ( ge = 0 , le = 6 ) ] = Field (
6 , description = ' Tolerance level for input and output moderation. Between 0 and 6, 0 being most strict, 6 being least strict. Defaults to 2. '
)
output_format : Optional [ BFLOutputFormat ] = Field (
BFLOutputFormat . png , description = " Output format for the generated image. Can be ' jpeg ' or ' png ' . " , examples = [ ' png ' ]
)
image : str = Field ( None , description = ' A Base64-encoded string representing the image you wish to expand ' )
class BFLFluxFillImageRequest ( BaseModel ) :
prompt : str = Field ( . . . , description = ' The description of the changes you want to make. This text guides the expansion process, allowing you to specify features, styles, or modifications for the expanded areas. ' )
prompt_upsampling : Optional [ bool ] = Field (
None , description = ' Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation. '
)
seed : Optional [ int ] = Field ( None , description = ' The seed value for reproducibility. ' )
steps : conint ( ge = 15 , le = 50 ) = Field ( . . . , description = ' Number of steps for the image generation process ' )
guidance : confloat ( ge = 1.5 , le = 100 ) = Field ( . . . , description = ' Guidance strength for the image generation process ' )
safety_tolerance : Optional [ conint ( ge = 0 , le = 6 ) ] = Field (
6 , description = ' Tolerance level for input and output moderation. Between 0 and 6, 0 being most strict, 6 being least strict. Defaults to 2. '
)
output_format : Optional [ BFLOutputFormat ] = Field (
BFLOutputFormat . png , description = " Output format for the generated image. Can be ' jpeg ' or ' png ' . " , examples = [ ' png ' ]
)
image : str = Field ( None , description = ' A Base64-encoded string representing the image you wish to modify. Can contain alpha mask if desired. ' )
mask : str = Field ( None , description = ' A Base64-encoded string representing the mask of the areas you with to modify. ' )
class BFLFluxCannyImageRequest ( BaseModel ) :
prompt : str = Field ( . . . , description = ' Text prompt for image generation ' )
prompt_upsampling : Optional [ bool ] = Field (
None , description = ' Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation. '
)
canny_low_threshold : Optional [ int ] = Field ( None , description = ' Low threshold for Canny edge detection ' )
canny_high_threshold : Optional [ int ] = Field ( None , description = ' High threshold for Canny edge detection ' )
seed : Optional [ int ] = Field ( None , description = ' The seed value for reproducibility. ' )
steps : conint ( ge = 15 , le = 50 ) = Field ( . . . , description = ' Number of steps for the image generation process ' )
guidance : confloat ( ge = 1 , le = 100 ) = Field ( . . . , description = ' Guidance strength for the image generation process ' )
safety_tolerance : Optional [ conint ( ge = 0 , le = 6 ) ] = Field (
6 , description = ' Tolerance level for input and output moderation. Between 0 and 6, 0 being most strict, 6 being least strict. Defaults to 2. '
)
output_format : Optional [ BFLOutputFormat ] = Field (
BFLOutputFormat . png , description = " Output format for the generated image. Can be ' jpeg ' or ' png ' . " , examples = [ ' png ' ]
)
control_image : Optional [ str ] = Field ( None , description = ' Base64 encoded image to use as control input if no preprocessed image is provided ' )
preprocessed_image : Optional [ str ] = Field ( None , description = ' Optional pre-processed image that will bypass the control preprocessing step ' )
class BFLFluxDepthImageRequest ( BaseModel ) :
prompt : str = Field ( . . . , description = ' Text prompt for image generation ' )
prompt_upsampling : Optional [ bool ] = Field (
None , description = ' Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation. '
)
seed : Optional [ int ] = Field ( None , description = ' The seed value for reproducibility. ' )
steps : conint ( ge = 15 , le = 50 ) = Field ( . . . , description = ' Number of steps for the image generation process ' )
guidance : confloat ( ge = 1 , le = 100 ) = Field ( . . . , description = ' Guidance strength for the image generation process ' )
safety_tolerance : Optional [ conint ( ge = 0 , le = 6 ) ] = Field (
6 , description = ' Tolerance level for input and output moderation. Between 0 and 6, 0 being most strict, 6 being least strict. Defaults to 2. '
)
output_format : Optional [ BFLOutputFormat ] = Field (
BFLOutputFormat . png , description = " Output format for the generated image. Can be ' jpeg ' or ' png ' . " , examples = [ ' png ' ]
)
control_image : Optional [ str ] = Field ( None , description = ' Base64 encoded image to use as control input if no preprocessed image is provided ' )
preprocessed_image : Optional [ str ] = Field ( None , description = ' Optional pre-processed image that will bypass the control preprocessing step ' )
class BFLFluxProGenerateRequest ( BaseModel ) :
prompt : str = Field ( . . . , description = ' The text prompt for image generation. ' )
prompt_upsampling : Optional [ bool ] = Field (
None , description = ' Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation. '
)
seed : Optional [ int ] = Field ( None , description = ' The seed value for reproducibility. ' )
width : conint ( ge = 256 , le = 1440 ) = Field ( 1024 , description = ' Width of the generated image in pixels. Must be a multiple of 32. ' )
height : conint ( ge = 256 , le = 1440 ) = Field ( 768 , description = ' Height of the generated image in pixels. Must be a multiple of 32. ' )
safety_tolerance : Optional [ conint ( ge = 0 , le = 6 ) ] = Field (
6 , description = ' Tolerance level for input and output moderation. Between 0 and 6, 0 being most strict, 6 being least strict. Defaults to 2. '
)
output_format : Optional [ BFLOutputFormat ] = Field (
BFLOutputFormat . png , description = " Output format for the generated image. Can be ' jpeg ' or ' png ' . " , examples = [ ' png ' ]
)
image_prompt : Optional [ str ] = Field ( None , description = ' Optional image to remix in base64 format ' )
# image_prompt_strength: Optional[confloat(ge=0.0, le=1.0)] = Field(
# None, description='Blend between the prompt and the image prompt.'
# )
2025-05-29 10:27:40 -07:00
class BFLFluxKontextProGenerateRequest ( BaseModel ) :
prompt : str = Field ( . . . , description = ' The text prompt for what you wannt to edit. ' )
input_image : Optional [ str ] = Field ( None , description = ' Image to edit in base64 format ' )
seed : Optional [ int ] = Field ( None , description = ' The seed value for reproducibility. ' )
guidance : confloat ( ge = 0.1 , le = 99.0 ) = Field ( . . . , description = ' Guidance strength for the image generation process ' )
steps : conint ( ge = 1 , le = 150 ) = Field ( . . . , description = ' Number of steps for the image generation process ' )
safety_tolerance : Optional [ conint ( ge = 0 , le = 2 ) ] = Field (
2 , description = ' Tolerance level for input and output moderation. Between 0 and 2, 0 being most strict, 6 being least strict. Defaults to 2. '
)
output_format : Optional [ BFLOutputFormat ] = Field (
BFLOutputFormat . png , description = " Output format for the generated image. Can be ' jpeg ' or ' png ' . " , examples = [ ' png ' ]
)
aspect_ratio : Optional [ str ] = Field ( None , description = ' Aspect ratio of the image between 21:9 and 9:21. ' )
prompt_upsampling : Optional [ bool ] = Field (
None , description = ' Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation. '
)
More API Nodes (#7956)
* Add Ideogram generate node.
* Add staging api.
* Add API_NODE and common error for missing auth token (#5)
* Add Minimax Video Generation + Async Task queue polling example (#6)
* [Minimax] Show video preview and embed workflow in ouput (#7)
* Remove uv.lock
* Remove polling operations.
* Revert "Remove polling operations."
* Update stubs.
* Added Ideogram and Minimax back in.
* Added initial BFL Flux 1.1 [pro] Ultra node (#11)
* Add --comfy-api-base launch arg (#13)
* Add instructions for staging development. (#14)
* remove validation to make it easier to run against LAN copies of the API
* Manually add BFL polling status response schema (#15)
* Add function for uploading files. (#18)
* Add Luma nodes (#16)
* Refactor util functions (#20)
* Add VIDEO type (#21)
* Add rest of Luma node functionality (#19)
* Fix image_luma_ref not working (#28)
* [Bug] Remove duplicated option T2V-01 in MinimaxTextToVideoNode (#31)
* Add utils to map from pydantic model fields to comfy node inputs (#30)
* add veo2, bump av req (#32)
* Add Recraft nodes (#29)
* Add Kling Nodes (#12)
* Add Camera Concepts (luma_concepts) to Luma Video nodes (#33)
* Add Runway nodes (#17)
* Convert Minimax node to use VIDEO output type (#34)
* Standard `CATEGORY` system for api nodes (#35)
* Set `Content-Type` header when uploading files (#36)
* add better error propagation to veo2 (#37)
* Add Realistic Image and Logo Raster styles for Recraft v3 (#38)
* Fix runway image upload and progress polling (#39)
* Fix image upload for Luma: only include `Content-Type` header field if it's set explicitly (#40)
* Moved Luma nodes to nodes_luma.py (#47)
* Moved Recraft nodes to nodes_recraft.py (#48)
* Add Pixverse nodes (#46)
* Move and fix BFL nodes to node_bfl.py (#49)
* Move and edit Minimax node to nodes_minimax.py (#50)
* Add Minimax Image to Video node + Cleanup (#51)
* Add Recraft Text to Vector node, add Save SVG node to handle its output (#53)
* Added pixverse_template support to Pixverse Text to Video node (#54)
* Added Recraft Controls + Recraft Color RGB nodes (#57)
* split remaining nodes out of nodes_api, make utility lib, refactor ideogram (#61)
* Add types and doctstrings to utils file (#64)
* Fix: `PollingOperation` progress bar update progress by absolute value (#65)
* Use common download function in kling nodes module (#67)
* Fix: Luma video nodes in `api nodes/image` category (#68)
* Set request type explicitly (#66)
* Add `control_after_generate` to all seed inputs (#69)
* Fix bug: deleting `Content-Type` when property does not exist (#73)
* Add preview to Save SVG node (#74)
* change default poll interval (#76), rework veo2
* Add Pixverse and updated Kling types (#75)
* Added Pixverse Image to VIdeo node (#77)
* Add Pixverse Transition Video node (#79)
* Proper ray-1-6 support as fix has been applied in backend (#80)
* Added Recraft Style - Infinite Style Library node (#82)
* add ideogram v3 (#83)
* [Kling] Split Camera Control config to its own node (#81)
* Add Pika i2v and t2v nodes (#52)
* Temporary Fix for Runway (#87)
* Added Stability Stable Image Ultra node (#86)
* Remove Runway nodes (#88)
* Fix: Prompt text can't be validated in Kling nodes when using primitive nodes (#90)
* Fix: typo in node name "Stabiliy" => "Stability" (#91)
* Add String (Multiline) node (#93)
* Update Pika Duration and Resolution options (#94)
* Change base branch to master. Not main. (#95)
* Fix UploadRequest file_name param (#98)
* Removed Infinite Style Library until later (#99)
* fix ideogram style types (#100)
* fix multi image return (#101)
* add metadata saving to SVG (#102)
* Bump templates version to include API node template workflows (#104)
* Fix: `download_url_to_video_output` return type (#103)
* fix 4o generation bug (#106)
* Serve SVG files directly (#107)
* Add a bunch of nodes, 3 ready to use, the rest waiting for endpoint support (#108)
* Revert "Serve SVG files directly" (#111)
* Expose 4 remaining Recraft nodes (#112)
* [Kling] Add `Duration` and `Video ID` outputs (#105)
* Fix: datamodel-codegen sets string#binary type to non-existent `bytes_aliased` variable (#114)
* Fix: Dall-e 2 not setting request content-type dynamically (#113)
* Default request timeout: one hour. (#116)
* Add Kling nodes: camera control, start-end frame, lip-sync, video extend (#115)
* Add 8 nodes - 4 BFL, 4 Stability (#117)
* Fix error for Recraft ImageToImage error for nonexistent random_seed param (#118)
* Add remaining Pika nodes (#119)
* Make controls input work for Recraft Image to Image node (#120)
* Use upstream PR: Support saving Comfy VIDEO type to buffer (#123)
* Use Upstream PR: "Fix: Error creating video when sliced audio tensor chunks are non-c-contiguous" (#127)
* Improve audio upload utils (#128)
* Fix: Nested `AnyUrl` in request model cannot be serialized (Kling, Runway) (#129)
* Show errors and API output URLs to the user (change log levels) (#131)
* Fix: Luma I2I fails when weight is <=0.01 (#132)
* Change category of `LumaConcepts` node from image to video (#133)
* Fix: `image.shape` accessed before `image` is null-checked (#134)
* Apply small fixes and most prompt validation (if needed to avoid API error) (#135)
* Node name/category modifications (#140)
* Add back Recraft Style - Infinite Style Library node (#141)
* Fixed Kling: Check attributes of pydantic types. (#144)
* Bump `comfyui-workflow-templates` version (#142)
* [Kling] Print response data when error validating response (#146)
* Fix: error validating Kling image response, trying to use `"key" in` on Pydantic class instance (#147)
* [Kling] Fix: Correct/verify supported subset of input combos in Kling nodes (#149)
* [Kling] Fix typo in node description (#150)
* [Kling] Fix: CFG min/max not being enforced (#151)
* Rebase launch-rebase (private) on prep-branch (public copy of master) (#153)
* Bump templates version (#154)
* Fix: Kling image gen nodes don't return entire batch when `n` > 1 (#152)
* Remove pixverse_template from PixVerse Transition Video node (#155)
* Invert image_weight value on Luma Image to Image node (#156)
* Invert and resize mask for Ideogram V3 node to match masking conventions (#158)
* [Kling] Fix: image generation nodes not returning Tuple (#159)
* [Bug] [Kling] Fix Kling camera control (#161)
* Kling Image Gen v2 + improve node descriptions for Flux/OpenAI (#160)
* [Kling] Don't return video_id from dual effect video (#162)
* Bump frontend to 1.18.8 (#163)
* Use 3.9 compat syntax (#164)
* Use Python 3.10
* add example env var
* Update templates to 0.1.11
* Bump frontend to 1.18.9
---------
Co-authored-by: Robin Huang <robin.j.huang@gmail.com>
Co-authored-by: Christian Byrne <cbyrne@comfy.org>
Co-authored-by: thot experiment <94414189+thot-experiment@users.noreply.github.com>
2025-05-06 03:23:00 -05:00
class BFLFluxProUltraGenerateRequest ( BaseModel ) :
prompt : str = Field ( . . . , description = ' The text prompt for image generation. ' )
prompt_upsampling : Optional [ bool ] = Field (
None , description = ' Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation. '
)
seed : Optional [ int ] = Field ( None , description = ' The seed value for reproducibility. ' )
aspect_ratio : Optional [ str ] = Field ( None , description = ' Aspect ratio of the image between 21:9 and 9:21. ' )
safety_tolerance : Optional [ conint ( ge = 0 , le = 6 ) ] = Field (
6 , description = ' Tolerance level for input and output moderation. Between 0 and 6, 0 being most strict, 6 being least strict. Defaults to 2. '
)
output_format : Optional [ BFLOutputFormat ] = Field (
BFLOutputFormat . png , description = " Output format for the generated image. Can be ' jpeg ' or ' png ' . " , examples = [ ' png ' ]
)
raw : Optional [ bool ] = Field ( None , description = ' Generate less processed, more natural-looking images. ' )
image_prompt : Optional [ str ] = Field ( None , description = ' Optional image to remix in base64 format ' )
image_prompt_strength : Optional [ confloat ( ge = 0.0 , le = 1.0 ) ] = Field (
None , description = ' Blend between the prompt and the image prompt. '
)
class BFLFluxProGenerateResponse ( BaseModel ) :
id : str = Field ( . . . , description = ' The unique identifier for the generation task. ' )
polling_url : str = Field ( . . . , description = ' URL to poll for the generation result. ' )
class BFLStatus ( str , Enum ) :
task_not_found = " Task not found "
pending = " Pending "
request_moderated = " Request Moderated "
content_moderated = " Content Moderated "
ready = " Ready "
error = " Error "
class BFLFluxProStatusResponse ( BaseModel ) :
id : str = Field ( . . . , description = " The unique identifier for the generation task. " )
status : BFLStatus = Field ( . . . , description = " The status of the task. " )
result : Optional [ Dict [ str , Any ] ] = Field (
None , description = " The result of the task (null if not completed). "
)
progress : confloat ( ge = 0.0 , le = 1.0 ) = Field (
. . . , description = " The progress of the task (0.0 to 1.0). "
)
details : Optional [ Dict [ str , Any ] ] = Field (
None , description = " Additional details about the task (null if not available). "
)