-
Notifications
You must be signed in to change notification settings - Fork 8.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Alert instances view to display data from event log instead of current instances #57446
Comments
Pinging @elastic/kibana-alerting-services (Team:Alerting Services) |
Seems like an nice "extension" to the current UI would be to show all the alert instances seen over some interval like a day (fixed for now, customizable later) and their most recent start/stop/durations. Here's the current view: Changes would be that you would see more instances - every instance that scheduled actions within the last 24 hours. And we'd start seeing inactive instances, not just active ones. This will likely end up making us add some filtering/sorting - you might want to sort by instance/status/start/duration, and you might want to filter by instance and status, as a separate issue/PR ... We'll have a semantic issue if we find an instance that was resolved within the time period we query over, but not the start of that instance. For these we're back into an "unknown" state, though we do know the minimum amount of time it has been active (the resolved time - the start time of the query we run) - duration could be something like "at least two hours" or "over two hours", kinda thing. |
@mdefazio to provide latest mockup of what the alert instances look like in the alert details page once data is pulled from event log. |
I'd be happy to pair with someone on this; I'm very familiar with the event log :-) |
I took a quick look at the current code to get a mental model of the pieces we need to put together:
Hopefully by then we'll have a fresh design and we can change the table in One problem we need to address is that we would, presumably, calculate the duration by looking back along the api response at how long an instance persists across execution cycles. The problem here is that if the time window specified by the user dictates how many events we fetch from the API then we can only ever evaluate duration as far back as that time window. This means that if, for example, the user specifies they want to see a time frame of "last 15 minutes" then the longest duration any single instance could have is 15 minutes, even if it has been going off for an hour. @mdefazio We need to think about how we might want to express that a specific instance runs all the way back to the edge of the time frame... possibly exceeding it. 🤔 |
|
Notes on the chart: #56280 (comment) |
resolves elastic#57446 Adds a new API (AlertClient and HTTP endpoint) `getAlertStatus()` which returns data calculated from the event log. The data returned in this PR is fairly minimal - just enough to replace the current instance details view data. In the future, we can add the following sorts of things: - alert execution errors - counts of alert execution - sequences of active instance time spans - if we can get the alert SO into the action execution, we could also provide action execution errors
resolves #57446 Adds a new API (AlertClient and HTTP endpoint) `getAlertStatus()` which returns alert data calculated from the event log.
…#68437) resolves elastic#57446 Adds a new API (AlertClient and HTTP endpoint) `getAlertStatus()` which returns alert data calculated from the event log.
We currently display the current alert instances now that #56842 got merged. This issue is the following step which would be to display from history with a start and stop column showing the duration of each instance.
The text was updated successfully, but these errors were encountered: