You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Would memray/pytest-memary suit our needs? Perhaps we could generate a memory footprint report alongside the performance test—it might help identify where the overhead occurs. If so, I'd be happy to help create a POC for it.
Yes, @josix , I think that would be a great start!
An approach that @sbaldassin used in the past was to dynamically create N tasks and subclass Cosmos Operators so they would use psutil and somehow log the mem_info. This example is a function used by PythonOperator, but we could try to use a similar approach in Cosmos:
In Cosmos, we have benchmark scripts that test task throughput using synthetic data and using Airflow local executor:
https://github.com/astronomer/astronomer-cosmos/blob/main/.github/workflows/test.yml#L350
https://github.com/astronomer/astronomer-cosmos/blob/main/tests/perf/test_performance.py
While those are very helpful, we currently don't analyse memory consumption.
It would be great if, from time to time, we could run a performance benchmark in Astro and we could identify the memory footprint per task,
The text was updated successfully, but these errors were encountered: