-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How can I use this code to profile my GC? #6
Comments
Hi guimaluf, thanks for your interest in our work. I'm sorry that it is a little complex to use this profiler, because I developed a number of parsers and analyzers to obtain statistics from task logs, gc logs, CPU logs, etc. Some of them are used to obtain the statistical results as presented in our paper, while others are obsolete. The usage of this profiler is as follows. After running a Spark application, e.g., After that, we use
This formatted statistics records the GC pause time and related memory usage after each young/old GC pause. Finally, we can use the python code in In general, this profiler covers almost all the fine-grained metrics of a Spark application, including the metrics of application, stages, tasks, executors, etc. If you focus on analyzing the GC logs of executors, please refer to the parsers in |
Hi all,
I read the article 'An Experimental Evaluation of GC on Big Data Applications' and I'm willing to reproduce part of it in my setup.
Isn't clear to me how can I use the SparkProfile.jar package. How it will get GC stats, where it will print output, etc.
I would like to thank you for the research and I appreciate any help
The text was updated successfully, but these errors were encountered: