Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Machine memory usage (JVM+py) #206

Open
tonkolviktor opened this issue Jan 26, 2023 · 1 comment
Open

Machine memory usage (JVM+py) #206

tonkolviktor opened this issue Jan 26, 2023 · 1 comment

Comments

@tonkolviktor
Copy link

Hi,

would it be possible to get the total memory used of a databricks job across all nodes and on the OS level.

As far as I can work out all memory metrics logged are from the JVM. That's nice, however for heavy pyspark jobs it's enough.
For our use-case we actually do not need this data only at the end of the job.
So ideally the following metrics would be available:

  • job_run_id
  • os total memory
  • os used memory
  • node_id

it's related to: #189

@hallihan
Copy link
Contributor

Please reach out to the contact listed in the README if you still need assistance. https://github.com/mspnp/spark-monitoring/tree/l4jv2#monitoring-azure-databricks-in-an-azure-log-analytics-workspace

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants