-
Notifications
You must be signed in to change notification settings - Fork 185
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
New large-logs-dataset
challenge in elastic/logs
track
#631
Comments
@elastic/es-perf I see the existing |
Note that the new challenge needs to skip deleting a template, The error is:
See 6251688 |
We would like to run an experiment in Rally which uses considerable amount of data. The idea is to be able to fill the disk of an AWS instance with 7.5 TB of storage. Indexing such large amount of data poses at least two challenges, anyway, which are a result of the way the
elastic/logs
Rally track is designed:For our experiment described in an internal Jira ticket, we:
@timestamp
needs to change depending on how much data we need to index per each day (raw_data_volume_per_day
).is4gen.8xlarge
that has 4 x 7.5 TB = 30 TB of storage available. Note that if we assume x10 raw-to-json expansion we would need 75 TB of Json data to have 7.5 TB of raw data. This means that even the instance with the largest storage can't handle the amount of data we need.As a result, benchmarking this scenario is practically impossible because of resource constraints but also because of the time data generation and indexing would require.
So the idea is to adopt the following strategy which we would like to implement in a new challenge part of the
elastic/logs
track:raw_data_volume_per_day
)logging-querying
existing challenge to collect query latenciesFor the use case above where we need to fill the instance with 7.5 TB of raw data it means restoring the snapshot 75 times. We expect:
elastic/logs
track and thelogging-querying
existing challenge.An experiment configured like described above mimics and environment where 75 hosts are logging exactly the same dataset.
Note that the snapshot API is only available in on-prem deployments...which means we need to run the benchmark on-prem.
The text was updated successfully, but these errors were encountered: