-
Notifications
You must be signed in to change notification settings - Fork 1.8k
Including metrics on number of batches handled through persistent queue #4117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
213a1a7 to
baf5330
Compare
Codecov Report
@@ Coverage Diff @@
## main #4117 +/- ##
=======================================
Coverage 87.89% 87.90%
=======================================
Files 174 175 +1
Lines 10198 10200 +2
=======================================
+ Hits 8964 8966 +2
Misses 984 984
Partials 250 250
Continue to review full report at Codecov.
|
baf5330 to
d9c27cc
Compare
|
@tigrannajaryan thank you for merging the part 2. Here comes the last of the series - I thought that it would be good to expose some metrics on how persistent queue is doing |
tigrannajaryan
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, just one question.
| zap.String(zapQueueNameKey, pcs.queueName), zap.Error(err)) | ||
| } | ||
|
|
||
| recordCurrentlyDispatchedBatches(ctx, len(pcs.currentlyDispatchedItems), pcs.queueName) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we only record these new metrics for persistent queue or also for in-memory queue when persistent queue is disabled?
I think they are useful regardless of whether persistence is enabled or no.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, I was wondering about doing that for in-memory queue as well though it seems it's using the old API and metrics would need to be pretty much redone there
bogdandrutu
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@tigrannajaryan I am not sold on this, first we should stop adding "random" metrics (we need a plan of how to name metrics, what is important vs not important to monitor, etc.) because sooner than later collector's own telemetry will need to stabilize, and so far I am the only one dealing with this work.
That's a fair argument. I think BTW, I was thinking about adding some more metrics in the future to persistent queue which would hold info on number of operations underneath |
Sounds good. Perhaps that can also eventually feed into a broader scope of defining semantic conventions for "data processors" in Otel spec, not just the Collector. |
|
Considering the discussion above and idea of leveraging a better strategy for collector metrics overall, I am closing this PR |
Testing: Unit tests updated, manual tests to follow