OXIESEC PANEL
- Current Dir:
/
/
opt
/
alt
/
python37
/
lib
/
python3.7
/
site-packages
/
sentry_sdk
/
integrations
Server IP: 2a02:4780:11:1084:0:327f:3464:10
Upload:
Create Dir:
Name
Size
Modified
Perms
📁
..
-
02/16/2024 09:01:45 PM
rwxr-xr-x
📄
__init__.py
6.36 KB
07/27/2021 02:04:04 PM
rw-r--r--
📁
__pycache__
-
02/16/2024 09:01:45 PM
rwxr-xr-x
📄
_wsgi_common.py
4.65 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
aiohttp.py
7.77 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
argv.py
945 bytes
07/27/2021 02:04:04 PM
rw-r--r--
📄
asgi.py
8.26 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
atexit.py
1.79 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
aws_lambda.py
14.99 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
beam.py
5.53 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
boto3.py
4.05 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
bottle.py
6.06 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
celery.py
9.31 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
chalice.py
4.5 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
dedupe.py
1.14 KB
07/27/2021 02:04:04 PM
rw-r--r--
📁
django
-
02/16/2024 09:01:45 PM
rwxr-xr-x
📄
excepthook.py
2.14 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
executing.py
1.98 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
falcon.py
6.64 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
flask.py
7.3 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
gcp.py
7.84 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
gnu_backtrace.py
2.84 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
httpx.py
2.55 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
logging.py
7.54 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
modules.py
1.36 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
pure_eval.py
4.41 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
pyramid.py
6.91 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
redis.py
3 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
rq.py
4.84 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
sanic.py
8.44 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
serverless.py
1.92 KB
07/27/2021 02:04:04 PM
rw-r--r--
📁
spark
-
02/16/2024 09:01:45 PM
rwxr-xr-x
📄
sqlalchemy.py
2.87 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
stdlib.py
7.25 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
threading.py
2.79 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
tornado.py
7.02 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
trytond.py
1.69 KB
07/27/2021 02:04:04 PM
rw-r--r--
📄
wsgi.py
10.89 KB
07/27/2021 02:04:04 PM
rw-r--r--
Editing: celery.py
Close
from __future__ import absolute_import import sys from sentry_sdk.hub import Hub from sentry_sdk.utils import capture_internal_exceptions, event_from_exception from sentry_sdk.tracing import Transaction from sentry_sdk._compat import reraise from sentry_sdk.integrations import Integration, DidNotEnable from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk._types import MYPY from sentry_sdk._functools import wraps if MYPY: from typing import Any from typing import TypeVar from typing import Callable from typing import Optional from sentry_sdk._types import EventProcessor, Event, Hint, ExcInfo F = TypeVar("F", bound=Callable[..., Any]) try: from celery import VERSION as CELERY_VERSION # type: ignore from celery.exceptions import ( # type: ignore SoftTimeLimitExceeded, Retry, Ignore, Reject, ) except ImportError: raise DidNotEnable("Celery not installed") CELERY_CONTROL_FLOW_EXCEPTIONS = (Retry, Ignore, Reject) class CeleryIntegration(Integration): identifier = "celery" def __init__(self, propagate_traces=True): # type: (bool) -> None self.propagate_traces = propagate_traces @staticmethod def setup_once(): # type: () -> None if CELERY_VERSION < (3,): raise DidNotEnable("Celery 3 or newer required.") import celery.app.trace as trace # type: ignore old_build_tracer = trace.build_tracer def sentry_build_tracer(name, task, *args, **kwargs): # type: (Any, Any, *Any, **Any) -> Any if not getattr(task, "_sentry_is_patched", False): # Need to patch both methods because older celery sometimes # short-circuits to task.run if it thinks it's safe. task.__call__ = _wrap_task_call(task, task.__call__) task.run = _wrap_task_call(task, task.run) # `build_tracer` is apparently called for every task # invocation. Can't wrap every celery task for every invocation # or we will get infinitely nested wrapper functions. task._sentry_is_patched = True return _wrap_tracer(task, old_build_tracer(name, task, *args, **kwargs)) trace.build_tracer = sentry_build_tracer from celery.app.task import Task # type: ignore Task.apply_async = _wrap_apply_async(Task.apply_async) _patch_worker_exit() # This logger logs every status of every task that ran on the worker. # Meaning that every task's breadcrumbs are full of stuff like "Task # <foo> raised unexpected <bar>". ignore_logger("celery.worker.job") ignore_logger("celery.app.trace") # This is stdout/err redirected to a logger, can't deal with this # (need event_level=logging.WARN to reproduce) ignore_logger("celery.redirected") def _wrap_apply_async(f): # type: (F) -> F @wraps(f) def apply_async(*args, **kwargs): # type: (*Any, **Any) -> Any hub = Hub.current integration = hub.get_integration(CeleryIntegration) if integration is not None and integration.propagate_traces: with hub.start_span(op="celery.submit", description=args[0].name) as span: with capture_internal_exceptions(): headers = dict(hub.iter_trace_propagation_headers(span)) if headers: # Note: kwargs can contain headers=None, so no setdefault! # Unsure which backend though. kwarg_headers = kwargs.get("headers") or {} kwarg_headers.update(headers) # https://github.com/celery/celery/issues/4875 # # Need to setdefault the inner headers too since other # tracing tools (dd-trace-py) also employ this exact # workaround and we don't want to break them. kwarg_headers.setdefault("headers", {}).update(headers) kwargs["headers"] = kwarg_headers return f(*args, **kwargs) else: return f(*args, **kwargs) return apply_async # type: ignore def _wrap_tracer(task, f): # type: (Any, F) -> F # Need to wrap tracer for pushing the scope before prerun is sent, and # popping it after postrun is sent. # # This is the reason we don't use signals for hooking in the first place. # Also because in Celery 3, signal dispatch returns early if one handler # crashes. @wraps(f) def _inner(*args, **kwargs): # type: (*Any, **Any) -> Any hub = Hub.current if hub.get_integration(CeleryIntegration) is None: return f(*args, **kwargs) with hub.push_scope() as scope: scope._name = "celery" scope.clear_breadcrumbs() scope.add_event_processor(_make_event_processor(task, *args, **kwargs)) transaction = None # Celery task objects are not a thing to be trusted. Even # something such as attribute access can fail. with capture_internal_exceptions(): transaction = Transaction.continue_from_headers( args[3].get("headers") or {}, op="celery.task", name="unknown celery task", ) transaction.name = task.name transaction.set_status("ok") if transaction is None: return f(*args, **kwargs) with hub.start_transaction( transaction, custom_sampling_context={ "celery_job": { "task": task.name, # for some reason, args[1] is a list if non-empty but a # tuple if empty "args": list(args[1]), "kwargs": args[2], } }, ): return f(*args, **kwargs) return _inner # type: ignore def _wrap_task_call(task, f): # type: (Any, F) -> F # Need to wrap task call because the exception is caught before we get to # see it. Also celery's reported stacktrace is untrustworthy. # functools.wraps is important here because celery-once looks at this # method's name. # https://github.com/getsentry/sentry-python/issues/421 @wraps(f) def _inner(*args, **kwargs): # type: (*Any, **Any) -> Any try: return f(*args, **kwargs) except Exception: exc_info = sys.exc_info() with capture_internal_exceptions(): _capture_exception(task, exc_info) reraise(*exc_info) return _inner # type: ignore def _make_event_processor(task, uuid, args, kwargs, request=None): # type: (Any, Any, Any, Any, Optional[Any]) -> EventProcessor def event_processor(event, hint): # type: (Event, Hint) -> Optional[Event] with capture_internal_exceptions(): tags = event.setdefault("tags", {}) tags["celery_task_id"] = uuid extra = event.setdefault("extra", {}) extra["celery-job"] = { "task_name": task.name, "args": args, "kwargs": kwargs, } if "exc_info" in hint: with capture_internal_exceptions(): if issubclass(hint["exc_info"][0], SoftTimeLimitExceeded): event["fingerprint"] = [ "celery", "SoftTimeLimitExceeded", getattr(task, "name", task), ] return event return event_processor def _capture_exception(task, exc_info): # type: (Any, ExcInfo) -> None hub = Hub.current if hub.get_integration(CeleryIntegration) is None: return if isinstance(exc_info[1], CELERY_CONTROL_FLOW_EXCEPTIONS): # ??? Doesn't map to anything _set_status(hub, "aborted") return _set_status(hub, "internal_error") if hasattr(task, "throws") and isinstance(exc_info[1], task.throws): return # If an integration is there, a client has to be there. client = hub.client # type: Any event, hint = event_from_exception( exc_info, client_options=client.options, mechanism={"type": "celery", "handled": False}, ) hub.capture_event(event, hint=hint) def _set_status(hub, status): # type: (Hub, str) -> None with capture_internal_exceptions(): with hub.configure_scope() as scope: if scope.span is not None: scope.span.set_status(status) def _patch_worker_exit(): # type: () -> None # Need to flush queue before worker shutdown because a crashing worker will # call os._exit from billiard.pool import Worker # type: ignore old_workloop = Worker.workloop def sentry_workloop(*args, **kwargs): # type: (*Any, **Any) -> Any try: return old_workloop(*args, **kwargs) finally: with capture_internal_exceptions(): hub = Hub.current if hub.get_integration(CeleryIntegration) is not None: hub.flush() Worker.workloop = sentry_workloop