-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
any way you can boost performance of building the process_map ? #13
Comments
I did some work on improving the performance of the data preparation in the There are certainly some parts which could be further optimised. Do debug your performance problem, we would need some more information on where exactly the bottleneck is. Could you try executing your code with the RStudio profiler activated and upload the saved profvis file? There are some options in using bupaR that can lead to performance degradation. |
My data input for the function process_map is a data.frame, eventlog PS: trying to fix my profvis problem, get back with the result as soon as possible. |
I looked at the profvis log you send me: It looks like R is taking most of the time for garbage collection. That suggests you have to little memory to keep the full data (plus the computation). I will compare it with a normal situation tomorrow, but I think this cannot easily be improved except for having more memory.
|
That is probably a good idea since it would take a lot of time to validate the event log. I see that the available memory should not be the issue. Just realised that I forgot to ask, do you use the current development version (installed from Github master) or the CRAN? Since, there are some improvements in the development version. |
I use the CRAN version, I will test the development version. |
After getting a closer look, it was actually the SVG export that takes a lot of time
grf is a diagrammeR graph structure |
Hello! considering the replacement of |
This tool is great, very great actually.
But is it possible to add multi-processing in the process_map function:
Here you can see (in this photo) that during the build of the dataframe nodes and edges, it is using only one processor. If you add the multiprocessing part it will be very fast, and we can deploy this package in our servers.
I have nearly 3 Millions (300 uniques) rows.
The text was updated successfully, but these errors were encountered: