You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The big5 workload is available with two sizes of data corpora, 60 GB and a 100 GB. The latter features a more representative timestamp sequence. Larger data corpora would be appropriate for performance testing at scale. This issue is to track generation of such larger corpora.
Initially, a 1 TB corpus will be generated and tested out. OSB scaling and stability will also be relevant in this context. Once this size of corpus can be used effectively, larger corpora, up to 10 TB in size, perhaps with multiple indices will be tackled.
gkamat
changed the title
Generate 500 GB and 1 TB data corpora for the big5 workload
[META] Generate large data corpora (1 to 10 TB) for the big5 workload
May 24, 2024
Description
The
big5
workload is available with two sizes of data corpora, 60 GB and a 100 GB. The latter features a more representative timestamp sequence. Larger data corpora would be appropriate for performance testing at scale. This issue is to track generation of such larger corpora.Initially, a 1 TB corpus will be generated and tested out. OSB scaling and stability will also be relevant in this context. Once this size of corpus can be used effectively, larger corpora, up to 10 TB in size, perhaps with multiple indices will be tackled.
Task Breakdown
big5
workload and its 1 TB data corpus opensearch-benchmark-workloads#296The text was updated successfully, but these errors were encountered: