From 0ef6b440b2b49566d2eadbb6b3af79474725bba9 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 27 Nov 2023 17:13:04 +0000 Subject: [PATCH] auto-organize --- .domino/compiled_metadata.json | 972 ++++++++++++++++++--------------- .domino/dependencies_map.json | 23 +- 2 files changed, 555 insertions(+), 440 deletions(-) diff --git a/.domino/compiled_metadata.json b/.domino/compiled_metadata.json index 014800e..b68b29c 100644 --- a/.domino/compiled_metadata.json +++ b/.domino/compiled_metadata.json @@ -1,16 +1,16 @@ { - "TextTaggingPiece": { - "name": "TextTaggingPiece", + "TextGeneratorPiece": { + "name": "TextGeneratorPiece", "dependency": { "dockerfile": "Dockerfile_01", "requirements_file": null }, "tags": [ "openai", - "text classification" + "text generation" ], "style": { - "node_label": "Text Tagging", + "node_label": "Text Generator", "node_type": "default", "node_style": { "backgroundColor": "#ebebeb" @@ -21,7 +21,203 @@ "cursor": "pointer" } }, - "description": "classification of text with user defined tags", + "description": "This Piece allows you to create a custom prompt with your own arguments and send it to OpenAI for text generation.", + "input_schema": { + "$defs": { + "InnerArgModel": { + "description": "Inner argument model to use in the prompt args", + "properties": { + "arg_name": { + "description": "Name of the prompt argument.", + "from_upstream": "never", + "title": "Arg Name", + "type": "string" + }, + "arg_value": { + "anyOf": [ + { + "type": "string" + }, + { + "items": {}, + "type": "array" + }, + { + "type": "integer" + }, + { + "type": "number" + }, + { + "type": "boolean" + }, + { + "type": "object" + }, + { + "format": "date", + "type": "string" + }, + { + "format": "time", + "type": "string" + }, + { + "format": "date-time", + "type": "string" + } + ], + "description": "Value of the prompt argument.", + "from_upstream": "always", + "title": "Arg Value" + } + }, + "required": [ + "arg_name", + "arg_value" + ], + "title": "InnerArgModel", + "type": "object" + }, + "LLMModelType": { + "description": "OpenAI model type", + "enum": [ + "gpt-3.5-turbo", + "gpt-4" + ], + "title": "LLMModelType", + "type": "string" + }, + "OutputTypeType": { + "description": "Output type for the generated text", + "enum": [ + "file", + "string", + "file_and_string" + ], + "title": "OutputTypeType", + "type": "string" + } + }, + "description": "TextGeneratorPiece Input model", + "properties": { + "template": { + "default": "What is the capital city of {country}?", + "description": "Compose a prompt template using the {arg_name} notation to insert arguments.", + "title": "Template", + "type": "string", + "widget": "textarea" + }, + "prompt_args": { + "default": [ + { + "arg_name": "country", + "arg_value": "Brazil" + } + ], + "description": "List of arguments to insert into the prompt.", + "items": { + "$ref": "#/$defs/InnerArgModel" + }, + "title": "Prompt Args", + "type": "array" + }, + "output_type": { + "allOf": [ + { + "$ref": "#/$defs/OutputTypeType" + } + ], + "default": "string", + "description": "The type of output to return." + }, + "openai_model": { + "allOf": [ + { + "$ref": "#/$defs/LLMModelType" + } + ], + "default": "gpt-3.5-turbo", + "description": "OpenAI model name." + }, + "completion_max_tokens": { + "default": 500, + "description": "The maximum number of tokens in the generated text.", + "title": "Completion Max Tokens", + "type": "integer" + }, + "temperature": { + "default": 0.3, + "description": "Temperature of the model, between 0 (more precise) and 1 (more creative).", + "exclusiveMaximum": 1.0, + "exclusiveMinimum": 0.0, + "title": "Temperature", + "type": "number" + } + }, + "title": "InputModel", + "type": "object" + }, + "output_schema": { + "description": "TextGeneratorPiece Output model", + "properties": { + "string_generated_text": { + "default": null, + "description": "The generated text as a string", + "title": "String Generated Text", + "type": "string" + }, + "file_path_generated_text": { + "default": null, + "description": "The path to text file containing generated text", + "title": "File Path Generated Text", + "type": "string" + } + }, + "title": "OutputModel", + "type": "object" + }, + "secrets_schema": { + "description": "TextGeneratorPiece Secrets model", + "properties": { + "OPENAI_API_KEY": { + "description": "Your OpenAI API key", + "title": "Openai Api Key", + "type": "string" + } + }, + "required": [ + "OPENAI_API_KEY" + ], + "title": "SecretsModel", + "type": "object" + }, + "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/TextGeneratorPiece" + }, + "InformationExtractionPiece": { + "name": "InformationExtractionPiece", + "dependency": { + "dockerfile": "Dockerfile_01", + "requirements_file": null + }, + "tags": [ + "text", + "information extraction", + "openai" + ], + "style": { + "node_label": "Information Extraction", + "node_type": "default", + "node_style": { + "backgroundColor": "#ebebeb" + }, + "useIcon": true, + "icon_class_name": "fa-solid:align-right", + "iconStyle": { + "cursor": "pointer" + } + }, + "description": "Extracts user-defined information from the input text.", "input_schema": { "$defs": { "LLMModelType": { @@ -46,7 +242,6 @@ "type": "string" }, "OutputModifierModel": { - "description": "OutputModifierModel with extra fields", "properties": { "name": { "description": "Name of the output argument.", @@ -70,12 +265,6 @@ "from_upstream": "never", "title": "Description", "type": "string" - }, - "enum": { - "default": "", - "description": "Comma separated list of possible values for the output modifier. Example: 'negative,neutral,positive'. If not provided, this will be ignored. ", - "title": "Enum", - "type": "string" } }, "required": [ @@ -85,10 +274,11 @@ "type": "object" } }, - "description": "TextTaggingPiece Input model", + "description": "InformationExtractionPiece Input model", "properties": { "input_text": { - "description": "Source text to be tagged.", + "description": "Source text from where information should be extracted.", + "from_upstream": "always", "title": "Input Text", "type": "string" }, @@ -99,32 +289,28 @@ } ], "default": "gpt-3.5-turbo-1106", - "description": "OpenAI model name to use for tagging." + "description": "OpenAI model name to use for information extraction." }, - "tags": { + "extract_items": { "default": [ { - "description": "Sentiment of the text. Should be a number between -1 and 1.", - "enum": "", - "name": "sentiment", - "type": "float" + "description": "Name of the person.", + "name": "name", + "type": "string" + }, + { + "description": "Age of the person.", + "name": "age", + "type": "integer" } ], - "description": "Tags to classify the source text.", + "description": "Information items to be extracted from source text.", "from_upstream": "never", "items": { "$ref": "#/$defs/OutputModifierModel" }, - "title": "Tags", + "title": "Extract Items", "type": "array" - }, - "temperature": { - "default": 0.0, - "description": "Temperature of the model, between 0 (more precise) and 1 (more creative).", - "exclusiveMaximum": 1.0, - "exclusiveMinimum": 0.0, - "title": "Temperature", - "type": "number" } }, "required": [ @@ -135,13 +321,13 @@ }, "output_schema": { "additionalProperties": true, - "description": "TextTaggingPiece Output Model", + "description": "InformationExtractionPiece Output Model", "properties": {}, "title": "OutputModel", "type": "object" }, "secrets_schema": { - "description": "TextTaggingPiece Secrets model", + "description": "InformationExtractionPiece Secrets model", "properties": { "OPENAI_API_KEY": { "description": "Your OpenAI API key.", @@ -155,10 +341,119 @@ "title": "SecretsModel", "type": "object" }, - "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/TextTaggingPiece" + "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/InformationExtractionPiece" }, - "PromptCreatorForImageGeneratorPiece": { - "name": "PromptCreatorForImageGeneratorPiece", + "AudioTranscriptionPiece": { + "name": "AudioTranscriptionPiece", + "dependency": { + "dockerfile": "Dockerfile_01", + "requirements_file": null + }, + "tags": [ + "OpenAI" + ], + "style": { + "node_label": "OpenAI Audio Transcript", + "node_type": "default", + "node_style": { + "backgroundColor": "#ebebeb" + }, + "useIcon": true, + "icon_class_name": "fa-solid:comment-dots", + "iconStyle": { + "cursor": "pointer" + } + }, + "description": "This Piece uses the OpenAI API to extract text transcripts from audio.", + "container_resources": { + "requests": { + "cpu": "100m", + "memory": "128Mi" + }, + "limits": { + "cpu": "500m", + "memory": "512Mi" + } + }, + "input_schema": { + "$defs": { + "OutputTypeType": { + "enum": [ + "file", + "string", + "both" + ], + "title": "OutputTypeType", + "type": "string" + } + }, + "properties": { + "audio_file_path": { + "description": "The path to the audio file to process.", + "from_upstream": "always", + "title": "Audio File Path", + "type": "string" + }, + "output_type": { + "allOf": [ + { + "$ref": "#/$defs/OutputTypeType" + } + ], + "default": "string", + "description": "The type of output for the result text. Options are `string`, `file` or `both`. Default is `string`." + }, + "temperature": { + "default": 0.1, + "description": "What sampling temperature to use, between 0 and 1", + "exclusiveMinimum": 0.0, + "maximum": 1.0, + "title": "Temperature", + "type": "number" + } + }, + "required": [ + "audio_file_path" + ], + "title": "InputModel", + "type": "object" + }, + "output_schema": { + "properties": { + "transcription_result": { + "default": "", + "description": "The result transcription text as a string.", + "title": "Transcription Result", + "type": "string" + }, + "file_path_transcription_result": { + "default": "", + "description": "The path to the text file with the transcription result.", + "title": "File Path Transcription Result", + "type": "string" + } + }, + "title": "OutputModel", + "type": "object" + }, + "secrets_schema": { + "properties": { + "OPENAI_API_KEY": { + "description": "OpenAI API key", + "title": "Openai Api Key", + "type": "string" + } + }, + "required": [ + "OPENAI_API_KEY" + ], + "title": "SecretsModel", + "type": "object" + }, + "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/AudioTranscriptionPiece" + }, + "PromptCreatorForImageGeneratorPiece": { + "name": "PromptCreatorForImageGeneratorPiece", "dependency": { "dockerfile": "Dockerfile_01", "requirements_file": null @@ -287,18 +582,18 @@ }, "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/PromptCreatorForImageGeneratorPiece" }, - "TextGeneratorPiece": { - "name": "TextGeneratorPiece", + "TextTaggingPiece": { + "name": "TextTaggingPiece", "dependency": { "dockerfile": "Dockerfile_01", "requirements_file": null }, "tags": [ "openai", - "text generation" + "text classification" ], "style": { - "node_label": "Text Generator", + "node_label": "Text Tagging", "node_type": "default", "node_style": { "backgroundColor": "#ebebeb" @@ -309,115 +604,76 @@ "cursor": "pointer" } }, - "description": "This Piece allows you to create a custom prompt with your own arguments and send it to OpenAI for text generation.", + "description": "classification of text with user defined tags", "input_schema": { "$defs": { - "InnerArgModel": { - "description": "Inner argument model to use in the prompt args", - "properties": { - "arg_name": { - "description": "Name of the prompt argument.", - "from_upstream": "never", - "title": "Arg Name", - "type": "string" - }, - "arg_value": { - "anyOf": [ - { - "type": "string" - }, - { - "items": {}, - "type": "array" - }, - { - "type": "integer" - }, - { - "type": "number" - }, - { - "type": "boolean" - }, - { - "type": "object" - }, - { - "format": "date", - "type": "string" - }, - { - "format": "time", - "type": "string" - }, - { - "format": "date-time", - "type": "string" - } - ], - "description": "Value of the prompt argument.", - "from_upstream": "always", - "title": "Arg Value" - } - }, - "required": [ - "arg_name", - "arg_value" - ], - "title": "InnerArgModel", - "type": "object" - }, "LLMModelType": { "description": "OpenAI model type", "enum": [ - "gpt-3.5-turbo", + "gpt-3.5-turbo-1106", "gpt-4" ], "title": "LLMModelType", "type": "string" }, - "OutputTypeType": { - "description": "Output type for the generated text", + "OutputModifierItemType": { + "description": "OutputArgsType Enum", "enum": [ - "file", "string", - "file_and_string" + "integer", + "float", + "boolean", + "array" ], - "title": "OutputTypeType", + "title": "OutputModifierItemType", "type": "string" - } - }, - "description": "TextGeneratorPiece Input model", - "properties": { - "template": { - "default": "What is the capital city of {country}?", - "description": "Compose a prompt template using the {arg_name} notation to insert arguments.", - "title": "Template", - "type": "string", - "widget": "textarea" }, - "prompt_args": { - "default": [ - { - "arg_name": "country", - "arg_value": "Brazil" + "OutputModifierModel": { + "description": "OutputModifierModel with extra fields", + "properties": { + "name": { + "description": "Name of the output argument.", + "from_upstream": "never", + "title": "Name", + "type": "string" + }, + "type": { + "allOf": [ + { + "$ref": "#/$defs/OutputModifierItemType" + } + ], + "default": "string", + "description": "Type of the output argument.", + "from_upstream": "never" + }, + "description": { + "default": "", + "description": "Description of the output argument.", + "from_upstream": "never", + "title": "Description", + "type": "string" + }, + "enum": { + "default": "", + "description": "Comma separated list of possible values for the output modifier. Example: 'negative,neutral,positive'. If not provided, this will be ignored. ", + "title": "Enum", + "type": "string" } - ], - "description": "List of arguments to insert into the prompt.", - "items": { - "$ref": "#/$defs/InnerArgModel" }, - "title": "Prompt Args", - "type": "array" - }, - "output_type": { - "allOf": [ - { - "$ref": "#/$defs/OutputTypeType" - } + "required": [ + "name" ], - "default": "string", - "description": "The type of output to return." + "title": "OutputModifierModel", + "type": "object" + } + }, + "description": "TextTaggingPiece Input model", + "properties": { + "input_text": { + "description": "Source text to be tagged.", + "title": "Input Text", + "type": "string" }, "openai_model": { "allOf": [ @@ -425,17 +681,28 @@ "$ref": "#/$defs/LLMModelType" } ], - "default": "gpt-3.5-turbo", - "description": "OpenAI model name." + "default": "gpt-3.5-turbo-1106", + "description": "OpenAI model name to use for tagging." }, - "completion_max_tokens": { - "default": 500, - "description": "The maximum number of tokens in the generated text.", - "title": "Completion Max Tokens", - "type": "integer" + "tags": { + "default": [ + { + "description": "Sentiment of the text. Should be a number between -1 and 1.", + "enum": "", + "name": "sentiment", + "type": "float" + } + ], + "description": "Tags to classify the source text.", + "from_upstream": "never", + "items": { + "$ref": "#/$defs/OutputModifierModel" + }, + "title": "Tags", + "type": "array" }, "temperature": { - "default": 0.3, + "default": 0.0, "description": "Temperature of the model, between 0 (more precise) and 1 (more creative).", "exclusiveMaximum": 1.0, "exclusiveMinimum": 0.0, @@ -443,33 +710,24 @@ "type": "number" } }, + "required": [ + "input_text" + ], "title": "InputModel", "type": "object" }, "output_schema": { - "description": "TextGeneratorPiece Output model", - "properties": { - "string_generated_text": { - "default": null, - "description": "The generated text as a string", - "title": "String Generated Text", - "type": "string" - }, - "file_path_generated_text": { - "default": null, - "description": "The path to text file containing generated text", - "title": "File Path Generated Text", - "type": "string" - } - }, + "additionalProperties": true, + "description": "TextTaggingPiece Output Model", + "properties": {}, "title": "OutputModel", "type": "object" }, "secrets_schema": { - "description": "TextGeneratorPiece Secrets model", + "description": "TextTaggingPiece Secrets model", "properties": { "OPENAI_API_KEY": { - "description": "Your OpenAI API key", + "description": "Your OpenAI API key.", "title": "Openai Api Key", "type": "string" } @@ -480,275 +738,241 @@ "title": "SecretsModel", "type": "object" }, - "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/TextGeneratorPiece" + "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/TextTaggingPiece" }, - "ImageGeneratorPiece": { - "name": "ImageGeneratorPiece", + "AudioTranscriptionLocalPiece": { + "name": "AudioTranscriptionLocalPiece", "dependency": { - "dockerfile": "Dockerfile_01", + "dockerfile": "Dockerfile_whisper", "requirements_file": null }, "tags": [], "style": { - "node_label": "DALL-E Image Generator", + "node_label": "Audio Transcription Local", "node_type": "default", "node_style": { "backgroundColor": "#ebebeb" }, "useIcon": true, - "icon_class_name": "fa-solid:image", + "icon_class_name": "fa-solid:comment-dots", "iconStyle": { "cursor": "pointer" } }, - "description": "This Piece utilizes DALL-E, an OpenAI model that generates images based on a prompt.", + "description": "Runs transcription locally using Whisper, a general-purpose speech recognition model. Ref: https://github.com/openai/whisper", + "container_resources": { + "use_gpu": true, + "requests": { + "cpu": "1000m", + "memory": "3Gi" + }, + "limits": { + "cpu": "5000m", + "memory": "15Gi" + } + }, "input_schema": { "$defs": { - "ImageFormat": { - "description": "Image format to return", + "ModelSizeType": { "enum": [ - "url", - "image_png", - "base64_string" + "tiny", + "base", + "small", + "medium", + "large" ], - "title": "ImageFormat", - "type": "string" - }, - "ImageSize": { - "description": "Image size to generate", - "enum": [ - "1024x1024", - "512x512", - "256x256" - ], - "title": "ImageSize", + "title": "ModelSizeType", "type": "string" }, "OutputTypeType": { - "description": "Output type for the result text", "enum": [ - "file", "string", - "file_and_string" + "file", + "both" ], "title": "OutputTypeType", "type": "string" } }, - "description": "ImageGeneratorPiece input model", "properties": { - "prompt": { - "description": "A text description of the desired image", - "title": "Prompt", + "audio_file_path": { + "description": "The path to the audio file to process.", + "from_upstream": "always", + "title": "Audio File Path", "type": "string" }, - "size": { - "allOf": [ - { - "$ref": "#/$defs/ImageSize" - } - ], - "default": "1024x1024", - "description": "The size of the generated images" - }, - "image_format": { + "output_type": { "allOf": [ { - "$ref": "#/$defs/ImageFormat" + "$ref": "#/$defs/OutputTypeType" } ], - "default": "url", - "description": "The format in which the generated image is returned" + "default": "string", + "description": "The type of output for the result text. Options are `string`, `file` or `both`. Default is `string`." }, - "output_type": { + "model_size": { "allOf": [ { - "$ref": "#/$defs/OutputTypeType" + "$ref": "#/$defs/ModelSizeType" } ], - "default": "string", - "description": "The type of the output. Attention: if Response Format equals to image_png, then Output Type must be file type." + "default": "tiny", + "description": "The size of the model to use. Default is tiny." } }, "required": [ - "prompt" + "audio_file_path" ], "title": "InputModel", "type": "object" }, "output_schema": { - "description": "ImageGeneratorPiece output model", "properties": { - "output_string": { - "default": null, - "description": "The generated image as string", - "title": "Output String", + "transcription_result": { + "default": "", + "description": "The result transcription text as a string.", + "title": "Transcription Result", "type": "string" }, - "output_file_path": { - "default": null, - "description": "Path to the generated image", - "title": "Output File Path", - "type": "string" + "file_path_transcription_result": { + "anyOf": [ + { + "format": "file-path", + "type": "string" + }, + { + "type": "string" + } + ], + "default": "", + "description": "The path to the text file with the transcription result.", + "title": "File Path Transcription Result" } }, "title": "OutputModel", "type": "object" }, - "secrets_schema": { - "description": "ImageGeneratorPiece secrets model", - "properties": { - "OPENAI_API_KEY": { - "description": "Your OpenAI API key", - "title": "Openai Api Key", - "type": "string" - } - }, - "required": [ - "OPENAI_API_KEY" - ], - "title": "SecretsModel", - "type": "object" - }, - "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/ImageGeneratorPiece" + "secrets_schema": null, + "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/AudioTranscriptionLocalPiece" }, - "InformationExtractionPiece": { - "name": "InformationExtractionPiece", + "ImageGeneratorPiece": { + "name": "ImageGeneratorPiece", "dependency": { "dockerfile": "Dockerfile_01", "requirements_file": null }, - "tags": [ - "text", - "information extraction", - "openai" - ], + "tags": [], "style": { - "node_label": "Information Extraction", + "node_label": "DALL-E Image Generator", "node_type": "default", "node_style": { "backgroundColor": "#ebebeb" }, "useIcon": true, - "icon_class_name": "fa-solid:align-right", + "icon_class_name": "fa-solid:image", "iconStyle": { "cursor": "pointer" } }, - "description": "Extracts user-defined information from the input text.", + "description": "This Piece utilizes DALL-E, an OpenAI model that generates images based on a prompt.", "input_schema": { "$defs": { - "LLMModelType": { - "description": "OpenAI model type", + "ImageFormat": { + "description": "Image format to return", "enum": [ - "gpt-3.5-turbo-1106", - "gpt-4" + "url", + "image_png", + "base64_string" ], - "title": "LLMModelType", + "title": "ImageFormat", "type": "string" }, - "OutputModifierItemType": { - "description": "OutputArgsType Enum", + "ImageSize": { + "description": "Image size to generate", "enum": [ - "string", - "integer", - "float", - "boolean", - "array" + "1024x1024", + "512x512", + "256x256" ], - "title": "OutputModifierItemType", + "title": "ImageSize", "type": "string" }, - "OutputModifierModel": { - "properties": { - "name": { - "description": "Name of the output argument.", - "from_upstream": "never", - "title": "Name", - "type": "string" - }, - "type": { - "allOf": [ - { - "$ref": "#/$defs/OutputModifierItemType" - } - ], - "default": "string", - "description": "Type of the output argument.", - "from_upstream": "never" - }, - "description": { - "default": "", - "description": "Description of the output argument.", - "from_upstream": "never", - "title": "Description", - "type": "string" - } - }, - "required": [ - "name" + "OutputTypeType": { + "description": "Output type for the result text", + "enum": [ + "file", + "string", + "file_and_string" ], - "title": "OutputModifierModel", - "type": "object" + "title": "OutputTypeType", + "type": "string" } }, - "description": "InformationExtractionPiece Input model", + "description": "ImageGeneratorPiece input model", "properties": { - "input_text": { - "description": "Source text from where information should be extracted.", - "from_upstream": "always", - "title": "Input Text", + "prompt": { + "description": "A text description of the desired image", + "title": "Prompt", "type": "string" }, - "openai_model": { + "size": { "allOf": [ { - "$ref": "#/$defs/LLMModelType" + "$ref": "#/$defs/ImageSize" } ], - "default": "gpt-3.5-turbo-1106", - "description": "OpenAI model name to use for information extraction." + "default": "1024x1024", + "description": "The size of the generated images" }, - "extract_items": { - "default": [ + "image_format": { + "allOf": [ { - "description": "Name of the person.", - "name": "name", - "type": "string" - }, + "$ref": "#/$defs/ImageFormat" + } + ], + "default": "url", + "description": "The format in which the generated image is returned" + }, + "output_type": { + "allOf": [ { - "description": "Age of the person.", - "name": "age", - "type": "integer" + "$ref": "#/$defs/OutputTypeType" } ], - "description": "Information items to be extracted from source text.", - "from_upstream": "never", - "items": { - "$ref": "#/$defs/OutputModifierModel" - }, - "title": "Extract Items", - "type": "array" + "default": "string", + "description": "The type of the output. Attention: if Response Format equals to image_png, then Output Type must be file type." } }, "required": [ - "input_text" + "prompt" ], "title": "InputModel", "type": "object" }, "output_schema": { - "additionalProperties": true, - "description": "InformationExtractionPiece Output Model", - "properties": {}, + "description": "ImageGeneratorPiece output model", + "properties": { + "output_string": { + "default": null, + "description": "The generated image as string", + "title": "Output String", + "type": "string" + }, + "output_file_path": { + "default": null, + "description": "Path to the generated image", + "title": "Output File Path", + "type": "string" + } + }, "title": "OutputModel", "type": "object" }, "secrets_schema": { - "description": "InformationExtractionPiece Secrets model", + "description": "ImageGeneratorPiece secrets model", "properties": { "OPENAI_API_KEY": { - "description": "Your OpenAI API key.", + "description": "Your OpenAI API key", "title": "Openai Api Key", "type": "string" } @@ -759,7 +983,7 @@ "title": "SecretsModel", "type": "object" }, - "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/InformationExtractionPiece" + "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/ImageGeneratorPiece" }, "TextSummarizerPiece": { "name": "TextSummarizerPiece", @@ -918,125 +1142,5 @@ "type": "object" }, "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/TextSummarizerPiece" - }, - "AudioTranscriptionPiece": { - "name": "AudioTranscriptionPiece", - "dependency": { - "dockerfile": "Dockerfile_01", - "requirements_file": null - }, - "tags": [ - "OpenAI" - ], - "style": { - "node_label": "OpenAI Audio Transcript", - "node_type": "default", - "node_style": { - "backgroundColor": "#ebebeb" - }, - "useIcon": true, - "icon_class_name": "fa-solid:headset", - "iconStyle": { - "cursor": "pointer" - } - }, - "description": "This Piece uses the OpenAI API to extract text transcripts from audio.", - "container_resources": { - "requests": { - "cpu": "100m", - "memory": "128Mi" - }, - "limits": { - "cpu": "500m", - "memory": "512Mi" - } - }, - "input_schema": { - "$defs": { - "OutputTypeType": { - "description": "Output type for the result text", - "enum": [ - "file", - "string", - "file_and_string" - ], - "title": "OutputTypeType", - "type": "string" - } - }, - "description": "AudioTranscriptPiece input model", - "properties": { - "audio_file_path": { - "description": "The path to the audio file to process.", - "title": "Audio File Path", - "type": "string" - }, - "output_type": { - "allOf": [ - { - "$ref": "#/$defs/OutputTypeType" - } - ], - "default": "string", - "description": "The type of output for the result text" - }, - "temperature": { - "default": 0.1, - "description": "What sampling temperature to use, between 0 and 1", - "exclusiveMinimum": 0.0, - "maximum": 1.0, - "title": "Temperature", - "type": "number" - } - }, - "required": [ - "audio_file_path" - ], - "title": "InputModel", - "type": "object" - }, - "output_schema": { - "description": "AudioTranscriptPiece output model", - "properties": { - "message": { - "description": "Output message to log", - "title": "Message", - "type": "string" - }, - "string_transcription_result": { - "default": null, - "description": "The result transcription text as a string.", - "title": "String Transcription Result", - "type": "string" - }, - "file_path_transcription_result": { - "default": null, - "description": "The result transcription text as a file path.", - "title": "File Path Transcription Result", - "type": "string" - } - }, - "required": [ - "message" - ], - "title": "OutputModel", - "type": "object" - }, - "secrets_schema": { - "description": "AudioTranscriptPiece secret model", - "properties": { - "OPENAI_API_KEY": { - "description": "OpenAI API key", - "title": "Openai Api Key", - "type": "string" - } - }, - "required": [ - "OPENAI_API_KEY" - ], - "title": "SecretsModel", - "type": "object" - }, - "source_url": "https://github.com/Tauffer-Consulting/openai_domino_pieces/tree/main/pieces/AudioTranscriptionPiece" } } \ No newline at end of file diff --git a/.domino/dependencies_map.json b/.domino/dependencies_map.json index 8bae9d3..f2613da 100644 --- a/.domino/dependencies_map.json +++ b/.domino/dependencies_map.json @@ -5,17 +5,28 @@ "requirements_file": null }, "pieces": [ - "TextTaggingPiece", - "PromptCreatorForImageGeneratorPiece", "TextGeneratorPiece", - "ImageGeneratorPiece", "InformationExtractionPiece", - "TextSummarizerPiece", - "AudioTranscriptionPiece" + "AudioTranscriptionPiece", + "PromptCreatorForImageGeneratorPiece", + "TextTaggingPiece", + "ImageGeneratorPiece", + "TextSummarizerPiece" ], "secrets": [ "OPENAI_API_KEY" ], - "source_image": "ghcr.io/tauffer-consulting/openai_domino_pieces:dev-group0" + "source_image": "ghcr.io/tauffer-consulting/openai_domino_pieces:0.7.0-group0" + }, + "group1": { + "dependency": { + "dockerfile": "Dockerfile_whisper", + "requirements_file": null + }, + "pieces": [ + "AudioTranscriptionLocalPiece" + ], + "secrets": [], + "source_image": "ghcr.io/tauffer-consulting/openai_domino_pieces:0.7.0-group1" } } \ No newline at end of file