Skip to content

Latest commit

 

History

History
54 lines (34 loc) · 1.25 KB

README.md

File metadata and controls

54 lines (34 loc) · 1.25 KB

Timeseries Replay

Streaming Data frequently crosses different systems, frameworks and applications. This makes it really hard to architect for and test.

Generating fake data to load test often fails to replicate real life scenarios. Sometimes it maybe useful to take an existing events table with timestamps and replay those at the same rate (or perhaps sped up rate)

See examples folder for some how to use

Current Available Sources

Read:

  • SQLAlchemy compatible sources
  • Parquet File

Write:

  • Folder on Disk
  • Console
  • Kafka

Notes

  • If there are too many messages in parallel can bottleneck

Running Tests

  • TODO add how to trigger test in the right docker container from host commandline

# start containers
docker-compose -f docker_compose/kafka_testing.yml up


# still need to test this
# docker exec -it docker_compose_replay_service_1 /timeseries_replay/pytest

Testing Notes

Running individual tests

# run specific test file
pytest tests/<file>.py

# specific function
pytest tests/test_integration.py::test_runner_large_file

# specific function with logging to console
pytest tests/test_integration.py::test_runner_large_file -o log_cli=True