You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We have network limits which linked to the speed of processing (rank window) and size of the graph (onboard memory of GPU/GPUs)
Mainnet will be started with a pretty big rank's calculation window (>=100 blocks) and a small amount of network bandwidth which will provide time to kernel upgrade by the community and also do hardware upgrades by validators.
Now, much time take to prepare data before sending it to GPU and we have only single GPU CUDA PageRank algorithm implementation.
My proposal to starts with stand-alone optimized kernel and redefine data structures during performance research and implementation of multiple GPUs kernel. Then make refactoring of structures in cyberd and migrate to the new kernel.
We have network limits which linked to the speed of processing (rank window) and size of the graph (onboard memory of GPU/GPUs)
Mainnet will be started with a pretty big rank's calculation window (>=100 blocks) and a small amount of network bandwidth which will provide time to kernel upgrade by the community and also do hardware upgrades by validators.
Now, much time take to prepare data before sending it to GPU and we have only single GPU CUDA PageRank algorithm implementation.
My proposal to starts with stand-alone optimized kernel and redefine data structures during performance research and implementation of multiple GPUs kernel. Then make refactoring of structures in cyberd and migrate to the new kernel.
References: #229
Note:
<-----We are here----->
The text was updated successfully, but these errors were encountered: