mirror of
https://github.com/Skyvern-AI/skyvern.git
synced 2025-09-02 18:50:24 +00:00
878 lines
38 KiB
Python
878 lines
38 KiB
Python
import abc
|
|
import json
|
|
import os
|
|
import smtplib
|
|
import uuid
|
|
from dataclasses import dataclass
|
|
from email.message import EmailMessage
|
|
from enum import StrEnum
|
|
from pathlib import Path
|
|
from tempfile import NamedTemporaryFile
|
|
from typing import Annotated, Any, Literal, Union
|
|
from urllib.parse import urlparse
|
|
|
|
import filetype
|
|
import structlog
|
|
from pydantic import BaseModel, Field
|
|
|
|
from skyvern.exceptions import (
|
|
ContextParameterValueNotFound,
|
|
MissingBrowserStatePage,
|
|
TaskNotFound,
|
|
UnexpectedTaskStatus,
|
|
)
|
|
from skyvern.forge import app
|
|
from skyvern.forge.prompts import prompt_engine
|
|
from skyvern.forge.sdk.api.aws import AsyncAWSClient
|
|
from skyvern.forge.sdk.api.files import download_file, get_path_for_workflow_download_directory
|
|
from skyvern.forge.sdk.api.llm.api_handler_factory import LLMAPIHandlerFactory
|
|
from skyvern.forge.sdk.schemas.tasks import TaskStatus
|
|
from skyvern.forge.sdk.settings_manager import SettingsManager
|
|
from skyvern.forge.sdk.workflow.context_manager import WorkflowRunContext
|
|
from skyvern.forge.sdk.workflow.exceptions import InvalidEmailClientConfiguration
|
|
from skyvern.forge.sdk.workflow.models.parameter import (
|
|
PARAMETER_TYPE,
|
|
AWSSecretParameter,
|
|
ContextParameter,
|
|
OutputParameter,
|
|
WorkflowParameter,
|
|
)
|
|
|
|
LOG = structlog.get_logger()
|
|
|
|
|
|
class BlockType(StrEnum):
|
|
TASK = "task"
|
|
FOR_LOOP = "for_loop"
|
|
CODE = "code"
|
|
TEXT_PROMPT = "text_prompt"
|
|
DOWNLOAD_TO_S3 = "download_to_s3"
|
|
UPLOAD_TO_S3 = "upload_to_s3"
|
|
SEND_EMAIL = "send_email"
|
|
|
|
|
|
@dataclass(frozen=True)
|
|
class BlockResult:
|
|
success: bool
|
|
output_parameter: OutputParameter | None = None
|
|
output_parameter_value: dict[str, Any] | list | str | None = None
|
|
|
|
|
|
class Block(BaseModel, abc.ABC):
|
|
# Must be unique within workflow definition
|
|
label: str
|
|
block_type: BlockType
|
|
output_parameter: OutputParameter | None = None
|
|
|
|
@classmethod
|
|
def get_subclasses(cls) -> tuple[type["Block"], ...]:
|
|
return tuple(cls.__subclasses__())
|
|
|
|
@staticmethod
|
|
def get_workflow_run_context(workflow_run_id: str) -> WorkflowRunContext:
|
|
return app.WORKFLOW_CONTEXT_MANAGER.get_workflow_run_context(workflow_run_id)
|
|
|
|
@staticmethod
|
|
def get_async_aws_client() -> AsyncAWSClient:
|
|
return app.WORKFLOW_CONTEXT_MANAGER.aws_client
|
|
|
|
@abc.abstractmethod
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
pass
|
|
|
|
async def execute_safe(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
try:
|
|
return await self.execute(workflow_run_id, **kwargs)
|
|
except Exception:
|
|
LOG.exception(
|
|
"Block execution failed",
|
|
exc_info=True,
|
|
workflow_run_id=workflow_run_id,
|
|
block_label=self.label,
|
|
block_type=self.block_type,
|
|
)
|
|
return BlockResult(success=False)
|
|
|
|
@abc.abstractmethod
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
pass
|
|
|
|
|
|
class TaskBlock(Block):
|
|
block_type: Literal[BlockType.TASK] = BlockType.TASK
|
|
|
|
url: str | None = None
|
|
title: str = "Untitled Task"
|
|
navigation_goal: str | None = None
|
|
data_extraction_goal: str | None = None
|
|
data_schema: dict[str, Any] | None = None
|
|
# error code to error description for the LLM
|
|
error_code_mapping: dict[str, str] | None = None
|
|
max_retries: int = 0
|
|
parameters: list[PARAMETER_TYPE] = []
|
|
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
parameters = self.parameters
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
|
|
if self.url and workflow_run_context.has_parameter(self.url):
|
|
parameters.append(workflow_run_context.get_parameter(self.url))
|
|
|
|
return parameters
|
|
|
|
@staticmethod
|
|
async def get_task_order(workflow_run_id: str, current_retry: int) -> tuple[int, int]:
|
|
"""
|
|
Returns the order and retry for the next task in the workflow run as a tuple.
|
|
"""
|
|
last_task_for_workflow_run = await app.DATABASE.get_last_task_for_workflow_run(workflow_run_id=workflow_run_id)
|
|
# If there is no previous task, the order will be 0 and the retry will be 0.
|
|
if last_task_for_workflow_run is None:
|
|
return 0, 0
|
|
# If there is a previous task but the current retry is 0, the order will be the order of the last task + 1
|
|
# and the retry will be 0.
|
|
order = last_task_for_workflow_run.order or 0
|
|
if current_retry == 0:
|
|
return order + 1, 0
|
|
# If there is a previous task and the current retry is not 0, the order will be the order of the last task
|
|
# and the retry will be the retry of the last task + 1. (There is a validation that makes sure the retry
|
|
# of the last task is equal to current_retry - 1) if it is not, we use last task retry + 1.
|
|
retry = last_task_for_workflow_run.retry or 0
|
|
if retry + 1 != current_retry:
|
|
LOG.error(
|
|
f"Last task for workflow run is retry number {last_task_for_workflow_run.retry}, "
|
|
f"but current retry is {current_retry}. Could be race condition. Using last task retry + 1",
|
|
workflow_run_id=workflow_run_id,
|
|
last_task_id=last_task_for_workflow_run.task_id,
|
|
last_task_retry=last_task_for_workflow_run.retry,
|
|
current_retry=current_retry,
|
|
)
|
|
|
|
return order, retry + 1
|
|
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
current_retry = 0
|
|
# initial value for will_retry is True, so that the loop runs at least once
|
|
will_retry = True
|
|
workflow_run = await app.WORKFLOW_SERVICE.get_workflow_run(workflow_run_id=workflow_run_id)
|
|
workflow = await app.WORKFLOW_SERVICE.get_workflow(workflow_id=workflow_run.workflow_id)
|
|
# if the task url is parameterized, we need to get the value from the workflow run context
|
|
if self.url and workflow_run_context.has_parameter(self.url) and workflow_run_context.has_value(self.url):
|
|
task_url_parameter_value = workflow_run_context.get_value(self.url)
|
|
if task_url_parameter_value:
|
|
LOG.info(
|
|
"Task URL is parameterized, using parameter value",
|
|
task_url_parameter_value=task_url_parameter_value,
|
|
task_url_parameter_key=self.url,
|
|
)
|
|
self.url = task_url_parameter_value
|
|
|
|
# TODO (kerem) we should always retry on terminated. We should make a distinction between retriable and
|
|
# non-retryable terminations
|
|
while will_retry:
|
|
task_order, task_retry = await self.get_task_order(workflow_run_id, current_retry)
|
|
task, step = await app.agent.create_task_and_step_from_block(
|
|
task_block=self,
|
|
workflow=workflow,
|
|
workflow_run=workflow_run,
|
|
workflow_run_context=workflow_run_context,
|
|
task_order=task_order,
|
|
task_retry=task_retry,
|
|
)
|
|
organization = await app.DATABASE.get_organization(organization_id=workflow.organization_id)
|
|
if not organization:
|
|
raise Exception(f"Organization is missing organization_id={workflow.organization_id}")
|
|
browser_state = await app.BROWSER_MANAGER.get_or_create_for_workflow_run(
|
|
workflow_run=workflow_run, url=self.url
|
|
)
|
|
if not browser_state.page:
|
|
LOG.error("BrowserState has no page", workflow_run_id=workflow_run.workflow_run_id)
|
|
raise MissingBrowserStatePage(workflow_run_id=workflow_run.workflow_run_id)
|
|
|
|
LOG.info(
|
|
f"Navigating to page",
|
|
url=self.url,
|
|
workflow_run_id=workflow_run_id,
|
|
task_id=task.task_id,
|
|
workflow_id=workflow.workflow_id,
|
|
organization_id=workflow.organization_id,
|
|
step_id=step.step_id,
|
|
)
|
|
|
|
if self.url:
|
|
await browser_state.page.goto(self.url)
|
|
|
|
try:
|
|
await app.agent.execute_step(organization=organization, task=task, step=step, workflow_run=workflow_run)
|
|
except Exception as e:
|
|
# Make sure the task is marked as failed in the database before raising the exception
|
|
await app.DATABASE.update_task(
|
|
task.task_id,
|
|
status=TaskStatus.failed,
|
|
organization_id=workflow.organization_id,
|
|
failure_reason=str(e),
|
|
)
|
|
raise e
|
|
|
|
# Check task status
|
|
updated_task = await app.DATABASE.get_task(task_id=task.task_id, organization_id=workflow.organization_id)
|
|
if not updated_task:
|
|
raise TaskNotFound(task.task_id)
|
|
if not updated_task.status.is_final():
|
|
raise UnexpectedTaskStatus(task_id=updated_task.task_id, status=updated_task.status)
|
|
if updated_task.status == TaskStatus.completed:
|
|
LOG.info(
|
|
f"Task completed",
|
|
task_id=updated_task.task_id,
|
|
workflow_run_id=workflow_run_id,
|
|
workflow_id=workflow.workflow_id,
|
|
organization_id=workflow.organization_id,
|
|
)
|
|
if self.output_parameter:
|
|
await workflow_run_context.register_output_parameter_value_post_execution(
|
|
parameter=self.output_parameter,
|
|
value=updated_task.extracted_information,
|
|
)
|
|
await app.DATABASE.create_workflow_run_output_parameter(
|
|
workflow_run_id=workflow_run_id,
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=updated_task.extracted_information,
|
|
)
|
|
LOG.info(
|
|
f"Registered output parameter value",
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=updated_task.extracted_information,
|
|
workflow_run_id=workflow_run_id,
|
|
workflow_id=workflow.workflow_id,
|
|
task_id=updated_task.task_id,
|
|
)
|
|
return BlockResult(
|
|
success=True,
|
|
output_parameter=self.output_parameter,
|
|
output_parameter_value=updated_task.extracted_information,
|
|
)
|
|
return BlockResult(success=True)
|
|
else:
|
|
current_retry += 1
|
|
will_retry = current_retry <= self.max_retries
|
|
retry_message = f", retrying task {current_retry}/{self.max_retries}" if will_retry else ""
|
|
LOG.warning(
|
|
f"Task failed with status {updated_task.status}{retry_message}",
|
|
task_id=updated_task.task_id,
|
|
status=updated_task.status,
|
|
workflow_run_id=workflow_run_id,
|
|
workflow_id=workflow.workflow_id,
|
|
organization_id=workflow.organization_id,
|
|
current_retry=current_retry,
|
|
max_retries=self.max_retries,
|
|
)
|
|
|
|
return BlockResult(success=False)
|
|
|
|
|
|
class ForLoopBlock(Block):
|
|
block_type: Literal[BlockType.FOR_LOOP] = BlockType.FOR_LOOP
|
|
|
|
# TODO (kerem): Add support for ContextParameter
|
|
loop_over: PARAMETER_TYPE
|
|
loop_blocks: list["BlockTypeVar"]
|
|
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
parameters = {self.loop_over}
|
|
|
|
for loop_block in self.loop_blocks:
|
|
for parameter in loop_block.get_all_parameters(workflow_run_id):
|
|
parameters.add(parameter)
|
|
return list(parameters)
|
|
|
|
def get_loop_block_context_parameters(self, workflow_run_id: str, loop_data: Any) -> list[ContextParameter]:
|
|
if not isinstance(loop_data, dict):
|
|
# TODO (kerem): Should we add support for other types?
|
|
raise ValueError("loop_data should be a dict")
|
|
|
|
loop_block_parameters = self.get_all_parameters(workflow_run_id)
|
|
context_parameters = [
|
|
parameter for parameter in loop_block_parameters if isinstance(parameter, ContextParameter)
|
|
]
|
|
for context_parameter in context_parameters:
|
|
if context_parameter.key not in loop_data:
|
|
raise ContextParameterValueNotFound(
|
|
parameter_key=context_parameter.key,
|
|
existing_keys=list(loop_data.keys()),
|
|
workflow_run_id=workflow_run_id,
|
|
)
|
|
context_parameter.value = loop_data[context_parameter.key]
|
|
|
|
return context_parameters
|
|
|
|
def get_loop_over_parameter_values(self, workflow_run_context: WorkflowRunContext) -> list[Any]:
|
|
if isinstance(self.loop_over, WorkflowParameter) or isinstance(self.loop_over, OutputParameter):
|
|
parameter_value = workflow_run_context.get_value(self.loop_over.key)
|
|
if isinstance(parameter_value, list):
|
|
return parameter_value
|
|
else:
|
|
# TODO (kerem): Should we raise an error here?
|
|
return [parameter_value]
|
|
else:
|
|
# TODO (kerem): Implement this for context parameters
|
|
raise NotImplementedError
|
|
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
loop_over_values = self.get_loop_over_parameter_values(workflow_run_context)
|
|
LOG.info(
|
|
f"Number of loop_over values: {len(loop_over_values)}",
|
|
block_type=self.block_type,
|
|
workflow_run_id=workflow_run_id,
|
|
num_loop_over_values=len(loop_over_values),
|
|
)
|
|
outputs_with_loop_values = []
|
|
for loop_over_value in loop_over_values:
|
|
context_parameters_with_value = self.get_loop_block_context_parameters(workflow_run_id, loop_over_value)
|
|
for context_parameter in context_parameters_with_value:
|
|
workflow_run_context.set_value(context_parameter.key, context_parameter.value)
|
|
try:
|
|
block_outputs = [
|
|
await loop_block.execute(workflow_run_id=workflow_run_id) for loop_block in self.loop_blocks
|
|
]
|
|
except Exception as e:
|
|
LOG.error("ForLoopBlock: Failed to execute loop block", exc_info=True)
|
|
raise e
|
|
outputs_with_loop_values.append(
|
|
[
|
|
{
|
|
"loop_value": loop_over_value,
|
|
"output_parameter": block_output.output_parameter,
|
|
"output_value": workflow_run_context.get_value(block_output.output_parameter.key),
|
|
}
|
|
for block_output in block_outputs
|
|
if block_output.output_parameter
|
|
]
|
|
)
|
|
|
|
# If all block outputs are successful, the loop is successful
|
|
success = all([block_output.success for block_output in block_outputs])
|
|
if not success:
|
|
LOG.info(
|
|
"ForLoopBlock: Encountered an failure processing block, terminating early",
|
|
block_outputs=block_outputs,
|
|
)
|
|
break
|
|
|
|
if self.output_parameter:
|
|
await workflow_run_context.register_output_parameter_value_post_execution(
|
|
parameter=self.output_parameter,
|
|
value=outputs_with_loop_values,
|
|
)
|
|
await app.DATABASE.create_workflow_run_output_parameter(
|
|
workflow_run_id=workflow_run_id,
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=outputs_with_loop_values,
|
|
)
|
|
return BlockResult(
|
|
success=success, output_parameter=self.output_parameter, output_parameter_value=outputs_with_loop_values
|
|
)
|
|
|
|
return BlockResult(success=success)
|
|
|
|
|
|
class CodeBlock(Block):
|
|
block_type: Literal[BlockType.CODE] = BlockType.CODE
|
|
|
|
code: str
|
|
parameters: list[PARAMETER_TYPE] = []
|
|
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
return self.parameters
|
|
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
# get workflow run context
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
# get all parameters into a dictionary
|
|
parameter_values = {}
|
|
for parameter in self.parameters:
|
|
value = workflow_run_context.get_value(parameter.key)
|
|
secret_value = workflow_run_context.get_original_secret_value_or_none(value)
|
|
if secret_value is not None:
|
|
parameter_values[parameter.key] = secret_value
|
|
else:
|
|
parameter_values[parameter.key] = value
|
|
|
|
local_variables: dict[str, Any] = {}
|
|
exec(self.code, parameter_values, local_variables)
|
|
result = {"result": local_variables.get("result")}
|
|
if self.output_parameter:
|
|
await workflow_run_context.register_output_parameter_value_post_execution(
|
|
parameter=self.output_parameter,
|
|
value=result,
|
|
)
|
|
await app.DATABASE.create_workflow_run_output_parameter(
|
|
workflow_run_id=workflow_run_id,
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=result,
|
|
)
|
|
return BlockResult(success=True, output_parameter=self.output_parameter, output_parameter_value=result)
|
|
|
|
return BlockResult(success=True)
|
|
|
|
|
|
class TextPromptBlock(Block):
|
|
block_type: Literal[BlockType.TEXT_PROMPT] = BlockType.TEXT_PROMPT
|
|
|
|
llm_key: str
|
|
prompt: str
|
|
parameters: list[PARAMETER_TYPE] = []
|
|
json_schema: dict[str, Any] | None = None
|
|
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
return self.parameters
|
|
|
|
async def send_prompt(self, prompt: str, parameter_values: dict[str, Any]) -> dict[str, Any]:
|
|
llm_api_handler = LLMAPIHandlerFactory.get_llm_api_handler(self.llm_key)
|
|
if not self.json_schema:
|
|
self.json_schema = {
|
|
"type": "object",
|
|
"properties": {
|
|
"llm_response": {
|
|
"type": "string",
|
|
"description": "Your response to the prompt",
|
|
}
|
|
},
|
|
}
|
|
|
|
prompt = prompt_engine.load_prompt_from_string(prompt, **parameter_values)
|
|
prompt += (
|
|
"\n\n"
|
|
+ "Please respond to the prompt above using the following JSON definition:\n\n"
|
|
+ "```json\n"
|
|
+ json.dumps(self.json_schema, indent=2)
|
|
+ "\n```\n\n"
|
|
)
|
|
LOG.info("TextPromptBlock: Sending prompt to LLM", prompt=prompt, llm_key=self.llm_key)
|
|
response = await llm_api_handler(prompt=prompt)
|
|
LOG.info("TextPromptBlock: Received response from LLM", response=response)
|
|
return response
|
|
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
# get workflow run context
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
# get all parameters into a dictionary
|
|
parameter_values = {}
|
|
for parameter in self.parameters:
|
|
value = workflow_run_context.get_value(parameter.key)
|
|
secret_value = workflow_run_context.get_original_secret_value_or_none(value)
|
|
if secret_value is not None:
|
|
parameter_values[parameter.key] = secret_value
|
|
else:
|
|
parameter_values[parameter.key] = value
|
|
|
|
response = await self.send_prompt(self.prompt, parameter_values)
|
|
if self.output_parameter:
|
|
await workflow_run_context.register_output_parameter_value_post_execution(
|
|
parameter=self.output_parameter,
|
|
value=response,
|
|
)
|
|
await app.DATABASE.create_workflow_run_output_parameter(
|
|
workflow_run_id=workflow_run_id,
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=response,
|
|
)
|
|
return BlockResult(success=True, output_parameter=self.output_parameter, output_parameter_value=response)
|
|
|
|
return BlockResult(success=True)
|
|
|
|
|
|
class DownloadToS3Block(Block):
|
|
block_type: Literal[BlockType.DOWNLOAD_TO_S3] = BlockType.DOWNLOAD_TO_S3
|
|
|
|
url: str
|
|
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
|
|
if self.url and workflow_run_context.has_parameter(self.url):
|
|
return [workflow_run_context.get_parameter(self.url)]
|
|
|
|
return []
|
|
|
|
async def _upload_file_to_s3(self, uri: str, file_path: str) -> None:
|
|
try:
|
|
client = self.get_async_aws_client()
|
|
await client.upload_file_from_path(uri=uri, file_path=file_path)
|
|
finally:
|
|
# Clean up the temporary file since it's created with delete=False
|
|
os.unlink(file_path)
|
|
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
# get workflow run context
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
# get all parameters into a dictionary
|
|
if self.url and workflow_run_context.has_parameter(self.url) and workflow_run_context.has_value(self.url):
|
|
task_url_parameter_value = workflow_run_context.get_value(self.url)
|
|
if task_url_parameter_value:
|
|
LOG.info(
|
|
"DownloadToS3Block: Task URL is parameterized, using parameter value",
|
|
task_url_parameter_value=task_url_parameter_value,
|
|
task_url_parameter_key=self.url,
|
|
)
|
|
self.url = task_url_parameter_value
|
|
|
|
try:
|
|
file_path = await download_file(self.url, max_size_mb=10)
|
|
except Exception as e:
|
|
LOG.error("DownloadToS3Block: Failed to download file", url=self.url, error=str(e))
|
|
raise e
|
|
|
|
uri = None
|
|
try:
|
|
uri = f"s3://{SettingsManager.get_settings().AWS_S3_BUCKET_DOWNLOADS}/{SettingsManager.get_settings().ENV}/{workflow_run_id}/{uuid.uuid4()}"
|
|
await self._upload_file_to_s3(uri, file_path)
|
|
except Exception as e:
|
|
LOG.error("DownloadToS3Block: Failed to upload file to S3", uri=uri, error=str(e))
|
|
raise e
|
|
|
|
LOG.info("DownloadToS3Block: File downloaded and uploaded to S3", uri=uri)
|
|
if self.output_parameter:
|
|
LOG.info("DownloadToS3Block: Output parameter defined, registering output parameter value")
|
|
await workflow_run_context.register_output_parameter_value_post_execution(
|
|
parameter=self.output_parameter,
|
|
value=uri,
|
|
)
|
|
await app.DATABASE.create_workflow_run_output_parameter(
|
|
workflow_run_id=workflow_run_id,
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=uri,
|
|
)
|
|
return BlockResult(success=True, output_parameter=self.output_parameter, output_parameter_value=uri)
|
|
|
|
LOG.info("DownloadToS3Block: No output parameter defined, returning None")
|
|
return BlockResult(success=True)
|
|
|
|
|
|
class UploadToS3Block(Block):
|
|
block_type: Literal[BlockType.UPLOAD_TO_S3] = BlockType.UPLOAD_TO_S3
|
|
|
|
# TODO (kerem): A directory upload is supported but we should also support a list of files
|
|
path: str | None = None
|
|
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
|
|
if self.path and workflow_run_context.has_parameter(self.path):
|
|
return [workflow_run_context.get_parameter(self.path)]
|
|
|
|
return []
|
|
|
|
@staticmethod
|
|
def _get_s3_uri(workflow_run_id: str, path: str) -> str:
|
|
s3_bucket = SettingsManager.get_settings().AWS_S3_BUCKET_UPLOADS
|
|
s3_key = f"{SettingsManager.get_settings().ENV}/{workflow_run_id}/{uuid.uuid4()}_{Path(path).name}"
|
|
return f"s3://{s3_bucket}/{s3_key}"
|
|
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
# get workflow run context
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
# get all parameters into a dictionary
|
|
if self.path and workflow_run_context.has_parameter(self.path) and workflow_run_context.has_value(self.path):
|
|
file_path_parameter_value = workflow_run_context.get_value(self.path)
|
|
if file_path_parameter_value:
|
|
LOG.info(
|
|
"UploadToS3Block: File path is parameterized, using parameter value",
|
|
file_path_parameter_value=file_path_parameter_value,
|
|
file_path_parameter_key=self.path,
|
|
)
|
|
self.path = file_path_parameter_value
|
|
# if the path is WORKFLOW_DOWNLOAD_DIRECTORY_PARAMETER_KEY, use the download directory for the workflow run
|
|
elif self.path == SettingsManager.get_settings().WORKFLOW_DOWNLOAD_DIRECTORY_PARAMETER_KEY:
|
|
self.path = str(get_path_for_workflow_download_directory(workflow_run_id).absolute())
|
|
|
|
if not self.path or not os.path.exists(self.path):
|
|
raise FileNotFoundError(f"UploadToS3Block: File not found at path: {self.path}")
|
|
|
|
try:
|
|
client = self.get_async_aws_client()
|
|
# is the file path a file or a directory?
|
|
if os.path.isdir(self.path):
|
|
# get all files in the directory, if there are more than 10 files, we will not upload them
|
|
files = os.listdir(self.path)
|
|
if len(files) > 10:
|
|
raise ValueError("Too many files in the directory, not uploading")
|
|
for file in files:
|
|
# if the file is a directory, we will not upload it
|
|
if os.path.isdir(os.path.join(self.path, file)):
|
|
LOG.warning("UploadToS3Block: Skipping directory", file=file)
|
|
continue
|
|
file_path = os.path.join(self.path, file)
|
|
await client.upload_file_from_path(
|
|
uri=self._get_s3_uri(workflow_run_id, file_path), file_path=file_path
|
|
)
|
|
else:
|
|
await client.upload_file_from_path(
|
|
uri=self._get_s3_uri(workflow_run_id, self.path), file_path=self.path
|
|
)
|
|
except Exception as e:
|
|
LOG.exception("UploadToS3Block: Failed to upload file to S3", file_path=self.path, exc_info=True)
|
|
raise e
|
|
|
|
LOG.info("UploadToS3Block: File(s) uploaded to S3", file_path=self.path)
|
|
return BlockResult(success=True)
|
|
|
|
|
|
class SendEmailBlock(Block):
|
|
block_type: Literal[BlockType.SEND_EMAIL] = BlockType.SEND_EMAIL
|
|
|
|
smtp_host: AWSSecretParameter
|
|
smtp_port: AWSSecretParameter
|
|
smtp_username: AWSSecretParameter
|
|
# if you're using a Gmail account, you need to pass in an app password instead of your regular password
|
|
smtp_password: AWSSecretParameter
|
|
sender: str
|
|
recipients: list[str]
|
|
subject: str
|
|
body: str
|
|
file_attachments: list[str] = []
|
|
|
|
def get_all_parameters(
|
|
self,
|
|
workflow_run_id: str,
|
|
) -> list[PARAMETER_TYPE]:
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
parameters = [self.smtp_host, self.smtp_port, self.smtp_username, self.smtp_password]
|
|
|
|
if self.file_attachments:
|
|
for file_path in self.file_attachments:
|
|
if workflow_run_context.has_parameter(file_path):
|
|
parameters.append(workflow_run_context.get_parameter(file_path))
|
|
|
|
if self.subject and workflow_run_context.has_parameter(self.subject):
|
|
parameters.append(workflow_run_context.get_parameter(self.subject))
|
|
|
|
if self.body and workflow_run_context.has_parameter(self.body):
|
|
parameters.append(workflow_run_context.get_parameter(self.body))
|
|
|
|
return parameters
|
|
|
|
def _decrypt_smtp_parameters(self, workflow_run_context: WorkflowRunContext) -> tuple[str, int, str, str]:
|
|
obfuscated_smtp_host_value = workflow_run_context.get_value(self.smtp_host.key)
|
|
obfuscated_smtp_port_value = workflow_run_context.get_value(self.smtp_port.key)
|
|
obfuscated_smtp_username_value = workflow_run_context.get_value(self.smtp_username.key)
|
|
obfuscated_smtp_password_value = workflow_run_context.get_value(self.smtp_password.key)
|
|
smtp_host_value = workflow_run_context.get_original_secret_value_or_none(obfuscated_smtp_host_value)
|
|
smtp_port_value = workflow_run_context.get_original_secret_value_or_none(obfuscated_smtp_port_value)
|
|
smtp_username_value = workflow_run_context.get_original_secret_value_or_none(obfuscated_smtp_username_value)
|
|
smtp_password_value = workflow_run_context.get_original_secret_value_or_none(obfuscated_smtp_password_value)
|
|
|
|
email_config_problems = []
|
|
if smtp_host_value is None:
|
|
email_config_problems.append("Missing SMTP server")
|
|
if smtp_port_value is None:
|
|
email_config_problems.append("Missing SMTP port")
|
|
elif not smtp_port_value.isdigit():
|
|
email_config_problems.append("SMTP port should be a number")
|
|
if smtp_username_value is None:
|
|
email_config_problems.append("Missing SMTP username")
|
|
if smtp_password_value is None:
|
|
email_config_problems.append("Missing SMTP password")
|
|
|
|
if email_config_problems:
|
|
raise InvalidEmailClientConfiguration(email_config_problems)
|
|
|
|
return smtp_host_value, smtp_port_value, smtp_username_value, smtp_password_value
|
|
|
|
def _get_file_paths(self, workflow_run_context: WorkflowRunContext, workflow_run_id: str) -> list[str]:
|
|
file_paths = []
|
|
for path in self.file_attachments:
|
|
# if the file path is a parameter, get the value from the workflow run context first
|
|
if workflow_run_context.has_parameter(path):
|
|
file_path_parameter_value = workflow_run_context.get_value(path)
|
|
# if the file path is a secret, get the original secret value from the workflow run context
|
|
file_path_parameter_secret_value = workflow_run_context.get_original_secret_value_or_none(
|
|
file_path_parameter_value
|
|
)
|
|
if file_path_parameter_secret_value:
|
|
path = file_path_parameter_secret_value
|
|
else:
|
|
path = file_path_parameter_value
|
|
|
|
if path == SettingsManager.get_settings().WORKFLOW_DOWNLOAD_DIRECTORY_PARAMETER_KEY:
|
|
# if the path is WORKFLOW_DOWNLOAD_DIRECTORY_PARAMETER_KEY, use download directory for the workflow run
|
|
path = str(get_path_for_workflow_download_directory(workflow_run_id).absolute())
|
|
LOG.info(
|
|
"SendEmailBlock: Using download directory for the workflow run",
|
|
workflow_run_id=workflow_run_id,
|
|
file_path=path,
|
|
)
|
|
|
|
# if the file path is a directory, add all files in the directory, skip directories, limit to 10 files
|
|
if os.path.exists(path) and os.path.isdir(path):
|
|
if len(os.listdir(path)) > 10:
|
|
LOG.warning("SendEmailBlock: Too many files in the directory, not attaching to email")
|
|
continue
|
|
for file in os.listdir(path):
|
|
if os.path.isdir(os.path.join(path, file)):
|
|
LOG.warning("SendEmailBlock: Skipping directory", file=file)
|
|
continue
|
|
file_path = os.path.join(path, file)
|
|
file_paths.append(file_path)
|
|
else:
|
|
# covers the case where the file path is a single file, a url, or an S3 uri
|
|
file_paths.append(path)
|
|
|
|
return file_paths
|
|
|
|
async def _download_from_s3(self, s3_uri: str) -> str:
|
|
client = self.get_async_aws_client()
|
|
downloaded_bytes = await client.download_file(uri=s3_uri)
|
|
file_path = NamedTemporaryFile(delete=False)
|
|
file_path.write(downloaded_bytes)
|
|
return file_path.name
|
|
|
|
async def _build_email_message(
|
|
self, workflow_run_context: WorkflowRunContext, workflow_run_id: str
|
|
) -> EmailMessage:
|
|
msg = EmailMessage()
|
|
msg["Subject"] = self.subject + f" - Workflow Run ID: {workflow_run_id}"
|
|
msg["To"] = ", ".join(self.recipients)
|
|
msg["From"] = self.sender
|
|
if self.body and workflow_run_context.has_parameter(self.body) and workflow_run_context.has_value(self.body):
|
|
# We're purposely not decrypting the body parameter value here because we don't want to expose secrets
|
|
body_parameter_value = workflow_run_context.get_value(self.body)
|
|
msg.set_content(body_parameter_value)
|
|
else:
|
|
msg.set_content(self.body)
|
|
|
|
for filename in self._get_file_paths(workflow_run_context, workflow_run_id):
|
|
path = None
|
|
try:
|
|
if filename.startswith("s3://"):
|
|
path = await self._download_from_s3(filename)
|
|
elif filename.startswith("http://") or filename.startswith("https://"):
|
|
path = await download_file(filename)
|
|
else:
|
|
LOG.info("SendEmailBlock: Looking for file locally", filename=filename)
|
|
if not os.path.exists(filename):
|
|
raise FileNotFoundError(f"File not found: {filename}")
|
|
if not os.path.isfile(filename):
|
|
raise IsADirectoryError(f"Path is a directory: {filename}")
|
|
|
|
LOG.info("SendEmailBlock: Found file locally", path=path)
|
|
path = filename
|
|
|
|
if not path:
|
|
raise FileNotFoundError(f"File not found: {filename}")
|
|
|
|
# Guess the content type based on the file's extension. Encoding
|
|
# will be ignored, although we should check for simple things like
|
|
# gzip'd or compressed files.
|
|
kind = filetype.guess(path)
|
|
if kind:
|
|
ctype = kind.mime
|
|
extension = kind.extension
|
|
else:
|
|
# No guess could be made, or the file is encoded (compressed), so
|
|
# use a generic bag-of-bits type.
|
|
ctype = "application/octet-stream"
|
|
extension = None
|
|
|
|
maintype, subtype = ctype.split("/", 1)
|
|
attachment_filename = urlparse(filename).path.replace("/", "_")
|
|
|
|
# Check if the filename has an extension
|
|
if not Path(attachment_filename).suffix:
|
|
# If no extension, guess it based on the MIME type
|
|
if extension:
|
|
attachment_filename += f".{extension}"
|
|
|
|
LOG.info(
|
|
"SendEmailBlock: Adding attachment",
|
|
filename=attachment_filename,
|
|
maintype=maintype,
|
|
subtype=subtype,
|
|
)
|
|
with open(path, "rb") as fp:
|
|
msg.add_attachment(fp.read(), maintype=maintype, subtype=subtype, filename=attachment_filename)
|
|
finally:
|
|
if path:
|
|
os.unlink(path)
|
|
|
|
return msg
|
|
|
|
async def execute(self, workflow_run_id: str, **kwargs: dict) -> BlockResult:
|
|
workflow_run_context = self.get_workflow_run_context(workflow_run_id)
|
|
smtp_host_value, smtp_port_value, smtp_username_value, smtp_password_value = self._decrypt_smtp_parameters(
|
|
workflow_run_context
|
|
)
|
|
|
|
smtp_host = None
|
|
try:
|
|
smtp_host = smtplib.SMTP(smtp_host_value, smtp_port_value)
|
|
LOG.info("SendEmailBlock: Connected to SMTP server")
|
|
smtp_host.starttls()
|
|
smtp_host.login(smtp_username_value, smtp_password_value)
|
|
LOG.info("SendEmailBlock: Logged in to SMTP server")
|
|
message = await self._build_email_message(workflow_run_context, workflow_run_id)
|
|
smtp_host.send_message(message)
|
|
LOG.info("SendEmailBlock: Email sent")
|
|
except Exception as e:
|
|
LOG.error("SendEmailBlock: Failed to send email", exc_info=True)
|
|
if self.output_parameter:
|
|
result_dict = {"success": False, "error": str(e)}
|
|
await workflow_run_context.register_output_parameter_value_post_execution(
|
|
parameter=self.output_parameter,
|
|
value=result_dict,
|
|
)
|
|
await app.DATABASE.create_workflow_run_output_parameter(
|
|
workflow_run_id=workflow_run_id,
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=result_dict,
|
|
)
|
|
return BlockResult(
|
|
success=False, output_parameter=self.output_parameter, output_parameter_value=result_dict
|
|
)
|
|
raise e
|
|
finally:
|
|
if smtp_host:
|
|
smtp_host.quit()
|
|
|
|
result_dict = {"success": True}
|
|
if self.output_parameter:
|
|
await workflow_run_context.register_output_parameter_value_post_execution(
|
|
parameter=self.output_parameter,
|
|
value=result_dict,
|
|
)
|
|
await app.DATABASE.create_workflow_run_output_parameter(
|
|
workflow_run_id=workflow_run_id,
|
|
output_parameter_id=self.output_parameter.output_parameter_id,
|
|
value=result_dict,
|
|
)
|
|
return BlockResult(success=True, output_parameter=self.output_parameter, output_parameter_value=result_dict)
|
|
|
|
return BlockResult(success=True)
|
|
|
|
|
|
BlockSubclasses = Union[
|
|
ForLoopBlock, TaskBlock, CodeBlock, TextPromptBlock, DownloadToS3Block, UploadToS3Block, SendEmailBlock
|
|
]
|
|
BlockTypeVar = Annotated[BlockSubclasses, Field(discriminator="block_type")]
|