Rocks 6.1 – IPoIB

Disclaimer

The instructions/steps given below worked for me (and Michigan Technological University) running Rocks 6.1 (with Service Pack 1, CentOS 6.3 and GE 2011.11p1) – as has been a common practice for several years now, a full version of Operating System was installed. The HPC cluster (wigner) used to prepare this documentation has Mellanox 56 Gb/s FDR InfiniBand switches and ports. Further, it is assumed the eth0 interface is used for private ethernet network and ib0 is the InfiniBand interface. These instructions may very well work for you (or your institution), on Rocks-like or other linux clusters. Please note that if you decide to use these instructions on your machine, you are doing so entirely at your very own discretion and that neither this site, sgowtham.com, nor its author (or Michigan Technological University) is responsible for any/all damage – intellectual and/or otherwise.

Continue reading … “Rocks 6.1 – IPoIB”

Rocks 5.4.2 – Scheduling GPU jobs via SGE

Disclaimer

The instructions/steps given below worked for me (and Michigan Technological University) running Rocks 5.4.2 (with CentOS 5.5 and SGE 6.2u5) – as has been a common practice for several years now, a full version of Operating System was installed. These instructions may very well work for you (or your institution), on Rocks-like or other linux clusters. Please note that if you decide to use these instructions on your machine, you are doing so entirely at your very own discretion and that neither this site, sgowtham.com, nor its author (or Michigan Technological University) is responsible for any/all damage – intellectual and/or otherwise.

Continue reading … “Rocks 5.4.2 – Scheduling GPU jobs via SGE”

SC12 in Salt Lake City

The description of the conference didn’t change – SC12 is still The International Conference for High Performance Computing, Networking, Storage and Analysis, but the location did: Salt Lake City in Utah, known to the historically inclined as the crossroads of the west. I had never been to SLC ever before but the last time I was in Utah was August of 2005 – a momentary step in and out of state border as part of Glenn Canyon Dam tour which, in itself, was part of a 3 day trip to the Grand Canyon.

Continue reading … “SC12 in Salt Lake City”

Rocks 5.4.2 – HPCC 1.4.1 benchmark with GCC 4.1.2

Disclaimer

The instructions/steps given below worked for me (and Michigan Technological University) running Rocks 5.4.2 (with CentOS 5.5) – as has been a common practice for several years now, a full version of Operating System was installed. These instructions may very well work for you (or your institution), on Rocks-like or other linux clusters. Please note that if you decide to use these instructions on your machine, you are doing so entirely at your very own discretion and that neither this site, sgowtham.com, nor its author (or Michigan Technological University) is responsible for any/all damage – intellectual and/or otherwise.

Continue reading … “Rocks 5.4.2 – HPCC 1.4.1 benchmark with GCC 4.1.2”

Rocks 5.4.2 – HPL 2.0 benchmark with GCC 4.1.2

Disclaimer

The instructions/steps given below worked for me (and Michigan Technological University) running Rocks 5.4.2 (with CentOS 5.5) – as has been a common practice for several years now, a full version of Operating System was installed. These instructions may very well work for you (or your institution), on Rocks-like or other linux clusters. Please note that if you decide to use these instructions on your machine, you are doing so entirely at your very own discretion and that neither this site, sgowtham.com, nor its author (or Michigan Technological University) is responsible for any/all damage – intellectual and/or otherwise.

Continue reading … “Rocks 5.4.2 – HPL 2.0 benchmark with GCC 4.1.2”

Rocks 5.4.2 – Ganglia’s gmond Python module for monitoring NVIDIA GPU

Disclaimer

The instructions/steps given below worked for me (and Michigan Technological University) running Rocks 5.4.2 (with CentOS 5.5) – as has been a common practice for several years now, a full version of Operating System was installed. These instructions may very well work for you (or your institution), on Rocks-like or other linux clusters. Please note that if you decide to use these instructions on your machine, you are doing so entirely at your very own discretion and that neither this site, sgowtham.com, nor its author (or Michigan Technological University) is responsible for any/all damage – intellectual and/or otherwise.

Continue reading … “Rocks 5.4.2 – Ganglia’s gmond Python module for monitoring NVIDIA GPU”

RHEL 6.2 – Ganglia’s gmond Python module for monitoring NVIDIA GPU

Disclaimer

The instructions/steps given below worked for me (and Michigan Technological University) running site licensed Red Hat Enterprise Linux 6.2 – as has been a common practice for several years now, a full version of Operating System was installed and all necessary patches/upgrades have been applied. These instructions may very well work for you (or your institution), on Red Hat-like or other linux distributions. Please note that if you decide to use these instructions on your machine, you are doing so entirely at your very own discretion and that neither this site, sgowtham.com, nor its author (or Michigan Technological University) is responsible for any/all damage – intellectual and/or otherwise.

Continue reading … “RHEL 6.2 – Ganglia’s gmond Python module for monitoring NVIDIA GPU”

SC11, fond memories from a fantastic nerd fest

Apart from what I could infer from the title, SC11, The International Conference for High Performance Computing, Networking, Storage and Analysis, that it was being held in a city that I had already been to once before (and had gained some familiarity, with at least parts of it) as part of The Great American Road Trip with dear friend Nils Stenvig (@UPBeaches), that the city had more than its fair share of friendly & familiar faces and that the event would offer the first of opportunities to put a face to many a names that I had known & interacted with for many years, I knew little about SC11 and of what I wanted from it.

Continue reading … “SC11, fond memories from a fantastic nerd fest”

BASH – Wrappers For qstat In NPACI ROCKS 5.2.2

What is ROCKS?

Rocks Cluster Distribution (originally called NPACI Rocks) is a Linux distribution intended for high-performance computing clusters. It was started by NPACI and the SDSC in 2000, and was initially funded in part by an NSF grant (2000-2007) but is currently funded by the followup NSF grant. Rocks was initially based on the Red Hat Linux distribution, however modern versions of Rocks are now based on CentOS, with a modified Anaconda installer that simplifies mass installation onto many computers. Rocks includes many tools (such as MPI) which are not part of CentOS but are integral components that make a group of computers into a cluster. Installations can be customized with additional software packages at install-time by using special user-supplied CDs (called Roll CDs). The Rolls extend the system by integrating seamlessly and automatically into the management and packaging mechanisms used by base software, greatly simplifying installation and configuration of large numbers of computers. Over a dozen Rolls have been created, including the SGE roll, the Condor roll, the Lustre roll, the Java roll, and the ganglia roll.

Continue reading … “BASH – Wrappers For qstat In NPACI ROCKS 5.2.2”

BASH – Wrappers For qstat In NPACI ROCKS

BASH is a free software UNIX shell written for the GNU Project. Its name is an acronym which stands for Bourne-again shell. The name is a pun on the name of the Bourne shell (sh), an early and important UNIX shell written by Stephen Bourne and distributed with Version 7 UNIX circa 1978, and the concept of being born again. BASH was created in 1987 by Brian Fox. In 1990 Chet Ramey became the primary maintainer. BASH is the default shell on most GNU/Linux systems as well as on Mac OS X and it can be run on most UNIX-like operating systems. It has also been ported to Microsoft Windows using the POSIX emulation provided by Cygwin, to MS-DOS by the DJGPP project and to Novell NetWare.

Continue reading … “BASH – Wrappers For qstat In NPACI ROCKS”

SIESTA, SMEAGOL, VASP and BlueGene/L

Well, some of you have been regular readers (and commentors) for my journal in this part of the e-world and such of those have known that I have been in Europe – Jülich, Germany – to be precise. However, while chatting with few of my friends, I noticed an incoherent idea as to why yours-truly is in Germany any way… Some thought I was here on a pleasure trip/vacation, some others thought I had too much money and free time to burn, some more thought I was here for a photographic expedition, while only few have known the real reason…

Continue reading … “SIESTA, SMEAGOL, VASP and BlueGene/L”