[ti_anomali] Add Benchmark and Policy Test#17428
[ti_anomali] Add Benchmark and Policy Test#17428mohitjha-elastic merged 2 commits intoelastic:mainfrom
Conversation
|
Pinging @elastic/security-service-integrations (Team:Security-Service Integrations) |
chrisberkhout
left a comment
There was a problem hiding this comment.
Looks good. One minor suggestion in a policy test.
It looks like this has rally benchmarks, system benchmarks and pipeline benchmarks. It would be great if you could remind me of how those benchmark types differ in function and value. Are all three worthwhile here?
packages/ti_anomali/data_stream/intelligence/_dev/test/policy/test-all.yml
Outdated
Show resolved
Hide resolved
Pipeline benchmarking focuses only on the ingest pipeline layer and measures how efficiently events are processed by processors such as grok, dissect, script, and convert. It helps identify event processing latency, and throughput in docs/sec. Its main value is optimizing parsing and enrichment logic without involving indexing or cluster performance. Rally benchmarking evaluates core Elasticsearch indexing and query performance at scale. It measures bulk indexing rate, search latency, disk I/O, and the impact of mappings and shard configurations. This is primarily used for cluster sizing, capacity planning, and understanding how many events per second a given infrastructure can handle. System benchmarking tests the full end-to-end flow—Elastic Agent to ingest pipelines to Elasticsearch indexing—under realistic conditions. It captures overall throughput, resource usage, and bottlenecks across the stack, making it the closest representation of production behavior. Its value lies in validating real integration performance and identifying whether limitations come from the agent, pipelines, or the cluster. |
🚀 Benchmarks reportTo see the full report comment with |
💚 Build Succeeded
History
|
Thanks. So let's merge this. I think pipeline benchmarking is definitely useful for integrations. A lot of our logic is there. The system benchmarking is probably also useful because sometimes the agent may be the bottleneck and that's definitely something that the integration is in charge of. The rally benchmarking seems a bit less relevant because things like sharding, etc. are not really integration concerns. These seem like they would be useful less for monitoring integration quality and more for informing cluster resourcing in actual setups. Where we get value from these might be a topic for future discussion in the team, but for now let's have them all. Thanks for the good work! |
Proposed commit message
Checklist
changelog.ymlfile.How to test this PR locally
Related Issues