Skip to content

mschimek/kamsta

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

36 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DOI

KaMSTa

Implementations of our distributed Minimum Spanning Tree (MST) algorithms which we present in our paper:

P. Sanders and M. Schimek. Engineering Massively Parallel MST Algorithms.

published in the proceedings of the 2023 IEEE International Parallel and Distributed Processing Symposium (IPDPS).

If you use this code in the context of an academic publication, we kindly ask you to cite it:

@inproceedings{DBLP:conf/ipps/SandersS23,
  author       = {Peter Sanders and
                  Matthias Schimek},
  title        = {Engineering Massively Parallel {MST} Algorithms},
  booktitle    = {{IEEE} International Parallel and Distributed Processing Symposium,
                  {IPDPS} 2023, St. Petersburg, FL, USA, May 15-19, 2023},
  pages        = {691--701},
  publisher    = {{IEEE}},
  year         = {2023},
  url          = {https://doi.org/10.1109/IPDPS54959.2023.00075},
  doi          = {10.1109/IPDPS54959.2023.00075},
  timestamp    = {Tue, 25 Jul 2023 16:27:14 +0200},
  biburl       = {https://dblp.org/rec/conf/ipps/SandersS23.bib},
  bibsource    = {dblp computer science bibliography, https://dblp.org}
}

Compiling

To compile the code use the following instructions:

  git clone --recursive <https://github.com/mschimek/kamsta.git|[email protected]:mschimek/kamsta.git>
  cmake -B build -DCMAKE_BUILD_TYPE=Release [-DUSE_EXPLICIT_INSTANTIATION=ON]
  cmake --build build --parallel

Usage

To execute the code use the following instructions:

  export OMP_NUM_THREADS=<number threads>
  mpiexec -n <number mpi processes> ./build/benchmarks/mst_benchmarks [kamsta parameters (--help for help)]

See the evaluation directory for generating preconfigured parameter settings similar to the ones we used in our experiments.

Furthermore, it is important that OMP_NUM_THREADS matches the kamsta parameter --threads. If you use multithreading, you should consider allocating enough CPUs per MPI process to avoid performance problems.

Dependencies

Apart from the included submodules we use OpenMP and Thread Building Blocks(TBB).

Notes

Note that due to restrictions in the graph generator, we tested our implementation only with number of threads and number of mpi processes being powers of two. We tested our implementation with GCC 10/11/12 and OpenMPI 4.0.

The code version used in our paper can be found in this release.

If you encounter problems, feel free to contact us or open a pull request.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages