|Phone:||+49 711 685-63594|
|Fax:||+49 711 685-63658|
|Email:||Dominic.Roehm _at_ icp.uni-stuttgart.de|
Institute for Computational Physics
Investigation of the heterogeneous nucleation, i.e. the early stage of crystal growth, in a colloidal model system near a wall by Molecular Dynamics computer simulations. My focus is on the inﬂuence of the hydrodynamic interaction, which is often neglected, since nucleation is considered as a quasi-static process. However, recent experiments have shown, that the kind of thermalization of the sample has a drastic inﬂuence on the nucleation rate. The rate decreases by an order of magnitude, if the sample is molten by a turbulent ﬂow rather than a laminar ﬂow. In our simulations, hydrodynamic interactions are incorporated by coupling the particles to a lattice ﬂuid representing the solvent. Since the computation of the hydrodynamic interaction is still orders of magnitude more expensive than classical interactions, our MD simulation software ESPResSo employs GPUs for the calculation of the lattice ﬂuid, which was realized during my diploma thesis. By this, we can investigate systems, that previously required a small computer cluster, using a single GPU. By using a cluster equipped with GPUs, we want to use this speedup to systematically investigate the inﬂuence of hydrodynamics on the heterogeneous nucleation.
Lattice-Boltzmann-Simulations on GPUs
In coarse-grained Molecular dynamics (MD) simulations of large macromolecules, the number of solvent molecules is normally so large that most of the computation time is spent on the solvent. For this reason one is interested in replacing the solvent by a lattice fluid using the Lattice-Boltzmann (LB) method. The LB method is well known and on large length and timescales it leads to a hydrodynamic ﬂow ﬁeld that satisﬁes the Navier-Stokes equation. If the lattice fluid should be coupled to a conventional MD simulation of the coarse-grained particles, it is necessary to thermalize the fluid. While the MD particles are easily coupled via friction terms to the fluid, the correct thermalization of the lattice fluid requires to switch into mode space, which makes thermalized LB more complex and computationally expensive. However, the LB method is particularly well suited for the highly parallel architecture of graphics processors (GPUs). I am working on a fully thermalized GPU-LB implementation which is coupled to a MD that is running on a conventional CPU using the simulation package ESPResSo . This implementation is on a single NVIDIA GTX480 or C2050 about 50 times faster than on a recent INTEL XEON E5620 quadcore, therefore replacing a full compute rack by a single desktop PC with a highend graphics card. Furthermore, due to communication overhead problems of the LB CPU code, the performance of a single NVIDIA Tesla C2050 can not achieved. Performance measurements using a AMD Opteron CPU cluster (1.9GHz) showed, that even to 96 CPU nodes are up to 12 times slower then a single GPU.
- May. 2011 - ... Doctorate studies at Institute for Computational Physics (University of Stuttgart)
- May. 2011 Diploma in Physics at University of Stuttgart
- Oct. 2005 - May. 2011 Studies of Physics at the University of Stuttgart