Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performance Issue for NestJS Project with Sentry Performance Enabled #15725

Closed
3 tasks done
ReneGreen27 opened this issue Mar 18, 2025 · 15 comments · Fixed by #15867
Closed
3 tasks done

Performance Issue for NestJS Project with Sentry Performance Enabled #15725

ReneGreen27 opened this issue Mar 18, 2025 · 15 comments · Fixed by #15867
Assignees
Labels
Package: nestjs Issues related to the Sentry Nestjs SDK Package: node Issues related to the Sentry Node SDK Package-Meta: OpenTelemetry

Comments

@ReneGreen27
Copy link
Member

ReneGreen27 commented Mar 18, 2025

Is there an existing issue for this?

How do you use Sentry?

Sentry Saas (sentry.io)

Which SDK are you using?

@sentry/browser

SDK Version

9.5.0

Framework Version

NestJS

Link to Sentry event

No response

Reproduction Example/SDK Setup

Mini repro app and details regarding performance included in internal ticket #146790

Steps to Reproduce

clean NestJS framework + Sentry. The route for testing simply returned a small JSON text.

According to the measurement results:

Clean framework without Sentry - 30,000 rps, 3ms average response time.

Clean framework + Sentry - 2,500 rps, 39ms average response time.

I used the following Sentry configuration:

Sentry.init({
  dsn: '',
  defaultIntegrations: false,
  integrations: [
  nodeProfilingIntegration(),
  Sentry.nestIntegration(),
  Sentry.httpIntegration(),
  Sentry.onUncaughtExceptionIntegration(),
  Sentry.onUnhandledRejectionIntegration(),
  ],
  environment: 'local',
  tracesSampleRate: 1,
  profilesSampleRate: 1,
});

Reducing the values of tracesSampleRate and profilesSampleRate to 0.1 slightly improves the situation

main performance loss occurs due to the Sentry.httpIntegration() module

Expected Result

Not impactful to app performance to have sentry performance enabled

Actual Result

Performance affected

@getsantry getsantry bot moved this to Waiting for: Product Owner in GitHub Issues with 👀 3 Mar 18, 2025
@github-actions github-actions bot added the Package: browser Issues related to the Sentry Browser SDK label Mar 18, 2025
@andreiborza
Copy link
Member

Thanks for the easy to follow reproduction repo. Can confirm, I'm observing the same outcomes with the http integration added as the only integration. I also tried production mode with the same results.

When only using the httpIntegration and disabling everything that's configurable inside it, I got up to 27k rps and 3ms avg response time on my machine. As a baseline I'm getting 85k rps tho.

Sentry.httpIntegration({
   breadcrumbs: false,
   spans: false,
   trackIncomingRequestsAsSessions: false,
   disableIncomingRequestSpans: true,
}),

cc @chargome @mydea

@chargome
Copy link
Member

Hey @ReneGreen27, we have several PRs in place that should slightly improve performance, but generally the instrumentation will cause some overhead. Current recommendation is to use the setup posted above by @andreiborza and update to latest versions as we'll try to further optimize this.

@lforst
Copy link
Member

lforst commented Mar 20, 2025

Here's a flamegraph of an extremely simple node.js server load tested:

https://drive.google.com/file/d/19CC_zFcRhVme_exz1gpthNv6fh8tNcKe/view?usp=sharing

const Sentry = require("@sentry/node");

Sentry.init({
  dsn: "....",

  // Add Tracing by setting tracesSampleRate
  // We recommend adjusting this value in production
  tracesSampleRate: 1.0,

  // Set sampling rate for profiling
  // This is relative to tracesSampleRate
  profilesSampleRate: 1.0,

  debug: true,

  environment: "node-test-app",
});

const http = require("http");

const hostname = "127.0.0.1";
const port = 3000;

const server = http.createServer((req, res) => {
  res.statusCode = 200;
  res.setHeader("Content-Type", "text/plain");
  res.end("Hello, World!\n");
});

server.listen(port, hostname, () => {
  console.log(`Server running at http://${hostname}:${port}/`);
});

Some offenders I could make out:

  • httpRequestToRequestData - or maybe drop undefined keys
  • processEvent
  • promiseBuffer is using an enormous amount of self-time
  • normalize
  • otel’s getIncomingRequestAttributes
  • for whatever reason some anonymous function in the SentryHttpInstrumentation
  • processEvent in eventFilters.js
  • createTransactionForOtelSpan
  • I think logger we can ignore, I had debug: true in the test

lforst added a commit that referenced this issue Mar 20, 2025
Ref: #15725

- Only does the options merging once
- Adds short circuiting for event types to avoid doing things multiple
times/unnecessarily

---------

Co-authored-by: Francesco Gringl-Novy <[email protected]>
@AbhiPrasad
Copy link
Member

Found another one - looks like the spanStart hooks we run in integrations can be avoided

Image

I know we're improving dropUndefinedKeys but we should look at if we can avoid adding client hooks if no instrumentation was added.

For example with tedious, we should not even register client.on('spanStart' if tedious is not instrumented (which for vast majority of people it wont be

client.on('spanStart', span => {
const { description, data } = spanToJSON(span);
// Tedius integration always set a span name and `db.system` attribute to `mssql`.
if (!description || data['db.system'] !== 'mssql') {
return;
}
const operation = description.split(' ')[0] || '';
if (TEDIUS_INSTRUMENTED_METHODS.has(operation)) {
span.setAttribute(SEMANTIC_ATTRIBUTE_SENTRY_ORIGIN, 'auto.db.otel.tedious');
}
});

We should also avoid dropUndefinedKeys in spanToJson

return dropUndefinedKeys({
span_id,
trace_id,
data: attributes,
description: name,
parent_span_id: parentSpanId,
start_timestamp: spanTimeInputToSeconds(startTime),
// This is [0,0] by default in OTEL, in which case we want to interpret this as no end time
timestamp: spanTimeInputToSeconds(endTime) || undefined,
status: getStatusMessage(status),
op: attributes[SEMANTIC_ATTRIBUTE_SENTRY_OP],
origin: attributes[SEMANTIC_ATTRIBUTE_SENTRY_ORIGIN] as SpanOrigin | undefined,
links: convertSpanLinksForEnvelope(links),
});

@AbhiPrasad
Copy link
Member

Image

Object.defineProperty is def expensive. We should aim to avoid this as much as possible. I actually see a lot of use cases where we can replace this with a weakmap instead.

addChildSpanToSpan, setContextOnScope, and setCapturedScopesOnSpan

@AbhiPrasad
Copy link
Member

shouldSample

Image

uuid4

Why is getCrypto getter so expensive???

Image

@AbhiPrasad
Copy link
Member

#15767

Image

chargome added a commit that referenced this issue Mar 21, 2025
Updates the implementation of `dropUndefinedKeys`:
- added early returns
- used for loops in case of larger data structures
- handle array case before object case to avoid calls to `isPojo`
- simplified `isPojo` by checking
[Object.prototype.constructor](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Object/constructor#description)


ref #15725
AbhiPrasad added a commit that referenced this issue Mar 21, 2025
ref
#15725 (comment)

Similar to the work done in
#15765 we can avoid
usage of `addNonEnumerableProperty` with usage of a `WeakSet`.
@AbhiPrasad
Copy link
Member

Note: After profiling the changes we made in:

We seem to have made some improvements!

We've dropped dropUndefinedKey usage total time by 50%, isPojo has pretty much disappeared off the list,

Before:

Image

After:

Image

Now we gotta make spanToJson avoid dropUndefinedKeys, and fix SentryError usage:

Image Image

@mydea
Copy link
Member

mydea commented Mar 24, 2025

So dropUndefinedKeys is completely going away 🎉 : #15796

@mydea
Copy link
Member

mydea commented Mar 25, 2025

I wonder what makes SentryError slow 🤔

@lforst
Copy link
Member

lforst commented Mar 25, 2025

I wonder what makes SentryError slow 🤔

AFAIK instantiating errors is actually quite expensive because the runtime has to compute a stack trace. That might be it.

@lforst
Copy link
Member

lforst commented Mar 25, 2025

Maybe if we threw an object with a certain shape instead it wouldn't be so expensive.

mydea added a commit that referenced this issue Mar 25, 2025
This replaces usage of `SentryError` for promise buffer control flow.
Instead, we can use a symbol and just check this directly, we do not
even care about the message here.

ref
#15725 (comment)
mydea added a commit that referenced this issue Mar 25, 2025
#15823)

This PR replaces our usage of `SentryError` for control flow in our
event processing.
Instead, we now throw either an `InternalError` or `DoNotSendEventError`
(which are just POJOs with a symbol, without a stackframe). These two
also allow us to differentiate between these two use cases, which so far
have been kind of combined via the log level, but are really different
things - one is "expected"/configured by a user, the other is unexpected
and more of a warning.

I also removed the handling for `SentryError` from inbound filters, as
this should then become unused.

This also deprecates `SentryError`, it is no longer used and we can
eventually remove it.

ref
#15725 (comment)
mydea added a commit that referenced this issue Mar 26, 2025
…5802)

Related to
#15725 (comment),
this PR is an idea to avoid registering `on('spanStart')` hooks for all
the node libraries that do not have proper hooks, unless they are used.

Today, for OTEL instrumentation that does not provide span hooks, we
fall back to `client.on('spanStart')` to enhance/mutate spans emitted by
the instrumentation. This PR improved this by only registering the
`spanStart` client hook if we detect that the OTEL instrumentation is
actually wrapping something. This avoids us calling a bunch of span
callbacks each time a span is started, when it is not even needed.

For this, a new `instrumentWhenWrapped` utility is added which can be
passed an instrumentation. This works by monkey-patching `_wrap` on the
instrumentation, and ensuring a callback is only called conditionally.

Note: For some (e.g. connect, fastify) we register `spanStart` in the
error handler, which is even better, so there we do not need this logic.
@mydea
Copy link
Member

mydea commented Mar 26, 2025

Improvements that have been made now:

  • Stopped using dropUndefinedKeys altogether
  • Ensure that Node Performance Instrumentation only registers on('spanStart') handlers when they are actually used
  • Stopped using SentryError
  • Short-circuits to event processing integration

mydea added a commit that referenced this issue Mar 26, 2025
Related to
#15725 (comment)

When a span is not a root span, we do not need to do sampling work for
it. By reordering stuff in the `shouldSample` function, we should be
able to save a bunch of operations for the vast majority of (non-root)
spans.

I _think_ this should be just fine!
@AbhiPrasad
Copy link
Member

Took another detailed trace, we've made some good improvements:

Image

Here's whats still left that is notable:

uuid4 is expensive, mainly from generateTraceId usage:

Image

addNonEnumerableProperty - tracked with #15808

Image

processEvent, and normalize visit but hard to tell why

Image

Image

createTransactionForOtelSpan

Image

OTEL sdk-trace-base Tracer.startSpan

Image

inferSpanData - should be fixed with work in #15767

Image

OTEL sanitizeAttributes

Image

createAndFinishSpanForOtelSpan

Image

isValidTraceId in otel span utils

Image

@AbhiPrasad AbhiPrasad added Package: node Issues related to the Sentry Node SDK Package: nestjs Issues related to the Sentry Nestjs SDK Package-Meta: OpenTelemetry and removed Package: browser Issues related to the Sentry Browser SDK labels Mar 26, 2025
@AbhiPrasad
Copy link
Member

Since we've started the initiative we've improved performance by 30-40%.

The next release will contain a majority of these improvements - will leave a comment when that is released.

For now going to close this issue. Further improvements will be tracked by #15861

Aside from upgrading the SDK, adjusting the sample rate will also have a lot of benefits performance-wise. We recommend using that as an avenue to get application performance to something you think is reasonable.

onurtemizkan pushed a commit that referenced this issue Apr 3, 2025
Related to
#15725 (comment)

When a span is not a root span, we do not need to do sampling work for
it. By reordering stuff in the `shouldSample` function, we should be
able to save a bunch of operations for the vast majority of (non-root)
spans.

I _think_ this should be just fine!
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Package: nestjs Issues related to the Sentry Nestjs SDK Package: node Issues related to the Sentry Node SDK Package-Meta: OpenTelemetry
Projects
Archived in project
Development

Successfully merging a pull request may close this issue.

6 participants