Files
unstract/backend/pipeline_v2/pipeline_processor.py
ali 0c5997f9a9 UN-2470 [FEAT] Remove Django dependency from Celery workers with internal APIs (#1494)
* UN-2470 [MISC] Remove Django dependency from Celery workers

This commit introduces a new worker architecture that decouples
Celery workers from Django where possible, enabling support for
gevent/eventlet pool types and reducing worker startup overhead.

Key changes:
- Created separate worker modules (api-deployment, callback, file_processing, general)
- Added internal API endpoints for worker communication
- Implemented Django-free task execution where appropriate
- Added shared utilities and client facades
- Updated container configurations for new worker architecture

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* Fix pre-commit issues: file permissions and ruff errors

Setup the docker for new workers

- Add executable permissions to worker entrypoint files
- Fix import order in namespace package __init__.py
- Remove unused variable api_status in general worker
- Address ruff E402 and F841 errors

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* refactoreed, Dockerfiles,fixes

* flexibility on celery run commands

* added debug logs

* handled filehistory for API

* cleanup

* cleanup

* cloud plugin structure

* minor changes in import plugin

* added notification and logger workers under new worker module

* add docker compatibility for new workers

* handled docker issues

* log consumer worker fixes

* added scheduler worker

* minor env changes

* cleanup the logs

* minor changes in logs

* resolved scheduler worker issues

* cleanup and refactor

* ensuring backward compatibbility to existing wokers

* added configuration internal apis and cache utils

* optimization

* Fix API client singleton pattern to share HTTP sessions

- Fix flawed singleton implementation that was trying to share BaseAPIClient instances
- Now properly shares HTTP sessions between specialized clients
- Eliminates 6x BaseAPIClient initialization by reusing the same underlying session
- Should reduce API deployment orchestration time by ~135ms (from 6 clients to 1 session)
- Added debug logging to verify singleton pattern activation

* cleanup and structuring

* cleanup in callback

* file system connectors  issue

* celery env values changes

* optional gossip

* variables for sync, mingle and gossip

* Fix for file type check

* Task pipeline issue resolving

* api deployement failed response handled

* Task pipline fixes

* updated file history cleanup with active file execution

* pipline status update and workflow ui page execution

* cleanup and resolvinf conflicts

* remove unstract-core from conenctoprs

* Commit uv.lock changes

* uv locks updates

* resolve migration issues

* defer connector-metadtda

* Fix connector migration for production scale

- Add encryption key handling with defer() to prevent decryption failures
- Add final cleanup step to fix duplicate connector names
- Optimize for large datasets with batch processing and bulk operations
- Ensure unique constraint in migration 0004 can be created successfully

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* hitl fixes

* minor fixes on hitl

* api_hub related changes

* dockerfile fixes

* api client cache fixes with actual response class

* fix: tags and llm_profile_id

* optimized clear cache

* cleanup

* enhanced logs

* added more handling on is file dir and added loggers

* cleanup the runplatform script

* internal apis are excempting from csrf

* sonal cloud issues

* sona-cloud issues

* resolving sonar cloud issues

* resolving sonar cloud issues

* Delta: added Batch size fix in workers

* comments addressed

* celery configurational changes for new workers

* fiixes in callback regaurding the pipline type check

* change internal url registry logic

* gitignore changes

* gitignore changes

* addressng pr cmmnets and cleanup the codes

* adding missed profiles for v2

* sonal cloud blocker issues resolved

* imlement otel

* Commit uv.lock changes

* handle execution time and some cleanup

* adding user_data in metadata Pr: https://github.com/Zipstack/unstract/pull/1544

* scheduler backward compatibitlity

* replace user_data with custom_data

* Commit uv.lock changes

* celery worker command issue resolved

* enhance package imports in connectors by changing to lazy imports

* Update runner.py by removing the otel from it

Update runner.py by removing the otel from it

Signed-off-by: ali <117142933+muhammad-ali-e@users.noreply.github.com>

* added delta changes

* handle erro to destination db

* resolve tool instances id validation and hitl queu name in API

* handled direct execution from workflow page to worker and logs

* handle cost logs

* Update health.py

Signed-off-by: Ritwik G <100672805+ritwik-g@users.noreply.github.com>

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* minor log changes

* introducing log consumer scheduler to bulk create, and socket .emit from worker for ws

* Commit uv.lock changes

* time limit or timeout celery config cleanup

* implemented redis client class in worker

* pipline status enum mismatch

* notification worker fixes

* resolve uv lock conflicts

* workflow log fixes

* ws channel name issue resolved. and handling redis down in status tracker, and removing redis keys

* default TTL changed for unified logs

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Signed-off-by: ali <117142933+muhammad-ali-e@users.noreply.github.com>
Signed-off-by: Ritwik G <100672805+ritwik-g@users.noreply.github.com>
Co-authored-by: Claude <noreply@anthropic.com>
Co-authored-by: Ritwik G <100672805+ritwik-g@users.noreply.github.com>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2025-10-03 11:24:07 +05:30

125 lines
4.0 KiB
Python

import logging
from django.utils import timezone
from pipeline_v2.exceptions import InactivePipelineError
from pipeline_v2.models import Pipeline
from pipeline_v2.notification import PipelineNotification
logger = logging.getLogger(__name__)
class PipelineProcessor:
@staticmethod
def initialize_pipeline_sync(pipeline_id: str) -> Pipeline:
"""Fetches and initializes the sync for a pipeline.
Args:
pipeline_id (str): UUID of the pipeline to sync
"""
pipeline: Pipeline = PipelineProcessor.fetch_pipeline(pipeline_id)
pipeline.run_count = pipeline.run_count + 1
return PipelineProcessor._update_pipeline_status(
pipeline=pipeline,
status=Pipeline.PipelineStatus.RESTARTING,
is_end=False,
)
@staticmethod
def fetch_pipeline(pipeline_id: str, check_active: bool = True) -> Pipeline:
"""Retrieves and checks for an active pipeline.
Args:
pipeline_id (str): UUID of the pipeline
check_active (bool): Whether to check if the pipeline is active
Raises:
InactivePipelineError: If an active pipeline is not found
"""
pipeline: Pipeline = Pipeline.objects.get(pk=pipeline_id)
if check_active and not pipeline.is_active():
logger.error(f"Inactive pipeline fetched: {pipeline_id}")
raise InactivePipelineError(pipeline_name=pipeline.pipeline_name)
return pipeline
@classmethod
def get_active_pipeline(cls, pipeline_id: str) -> Pipeline | None:
"""Retrieves a list of active pipelines."""
try:
return cls.fetch_pipeline(pipeline_id, check_active=True)
except Pipeline.DoesNotExist:
return None
@staticmethod
def _update_pipeline_status(
pipeline: Pipeline,
status: tuple[str, str],
is_end: bool,
is_active: bool | None = None,
) -> Pipeline:
"""Updates pipeline status during execution.
Raises:
PipelineSaveError: Exception while saving a pipeline
Returns:
Pipeline: Updated pipeline
"""
if is_end:
pipeline.last_run_time = timezone.now()
if status:
pipeline.last_run_status = status
if is_active is not None:
pipeline.active = is_active
pipeline.save()
return pipeline
@staticmethod
def _send_notification(
pipeline: Pipeline,
execution_id: str | None = None,
error_message: str | None = None,
) -> None:
"""Sends a notification for the pipeline.
Args:
pipeline (Pipeline): Pipeline to send notification for
Returns:
None
"""
pipeline_notification = PipelineNotification(
pipeline=pipeline, execution_id=execution_id, error_message=error_message
)
pipeline_notification.send()
@staticmethod
def update_pipeline(
pipeline_guid: str | None,
status: tuple[str, str],
is_active: bool | None = None,
execution_id: str | None = None,
error_message: str | None = None,
is_end: bool = False,
) -> None:
if not pipeline_guid:
return
# Skip check if we are enabling an inactive pipeline
check_active = not is_active
pipeline: Pipeline = PipelineProcessor.fetch_pipeline(
pipeline_id=pipeline_guid, check_active=check_active
)
pipeline = PipelineProcessor._update_pipeline_status(
pipeline=pipeline, is_end=is_end, status=status, is_active=is_active
)
# Only send notifications if execution_id is provided
# This avoids duplicate notifications when called from workers (who handle notifications separately)
if execution_id:
PipelineProcessor._send_notification(
pipeline=pipeline, execution_id=execution_id, error_message=error_message
)
logger.info(f"Updated pipeline {pipeline_guid} status: {status}")