This week we announced the formation of the HPC Advisory Council Centers of Excellence. The HPC Advisory Council Centers of Excellence will provide local support for the HPC Advisory Council’s programs, local workshops and conferences, as well as host local computing centers that can be used to extend such activities.
“We are pleased to be named as one the inaugural HPC Advisory Council’s Centers of Excellence, covering HPC research, outreach and educational activities within Europe,” said Hussein Nasser El-Harake at the Swiss National Supercomputing Centre who serves as the Director of the HPC Advisory Council Center of Excellence in Switzerland. “As part of the HPC Advisory Council’s Center of Excellence, we look forward to advancing awareness of the beneficial capabilities of HPC to new users.”
Wanted to let you know that we have extended the high-performance applications best practices to:
1. Extend the applications performance, optimization and profiling guidelines to cover nearly 30 different applications, both commercial and open source – http://www.hpcadvisorycouncil.com/best_practices.php
2. We have added the first case using RoCE (RDMA over Converged Ethernet) to the performance, optimization and profiling guidelines page. It is under the same link as in item 1
3. New – installations guides – for those who asked to get a detailed description on where to get the application from, what is needed to be installed, how to install on a cluster, and how to actually run the application – it is now posted under the HPC|Works subgroup – http://www.hpcadvisorycouncil.com/subgroups_hpc_works.php. We will be focusing on open source applications, which sometime it challenging to really find this info. At the moment we have installations guides for BQCD, Espresso and NAMD, and more will come in the near future.
If you would like to propose new applications to be covered under the performance, optimization and profiling guidelines, or to be added to the installations guides, please let us know via email@example.com.
Recently we have added new systems into out HPC center, and you see the full list at http://www.hpcadvisorycouncil.com/cluster_center.php.
The newest system is the “Vesta” system (and you can see Pak Lui, the HPC Advisory Council HPC Center Manager standing next to it in the picture below). Vesta consist of six Dell™ PowerEdge™ R815 nodes, each with four processors AMD Opteron 6172 (Magny-Cours) which mean 48 Cores per node and 288 cores for the entire system. The networking was provided by Mellanox, and we have plugged two adapters per node (Mellanox ConnectX®-2 40Gb/s InfiniBand adapters). All nodes are connected via Mellanox 36-Port 40Gb/s InfiniBand Switch. Furthermore, each node has 128 GB, 1333 MHz memory to make sure we can really get the highest performance from this system.
Microsoft has provided us with Windows HPC 2008 v3 preview, so we can check the performance gain versus v2 for example. The system is capable of dual boot – Windows and Linux, and is now available for testing. If you would like to get access, just fill the form on the URL above.
In the picture – Pak Lui standing next to Vesta
I want to thank Dell, AMD and Mellanox for providing this system to the council!
Gilad, HPC Advisory Council Chairman
We recently performed an interview with Addison Snell, General Manager at Tabor Research, where we highlighted the council’s activities for the past year, and provided some insight into our future direction.
Gilad Shainer – Chair, HPC Advisory Council
Brian Sparks – Media Relations Director, HPC Advisory Council
Gautam Shah – CEO, Colfax International
Scot Schultz – Senior Strategic Alliance Manager, AMD
Peter Lillian – Senior Product Marketing Manager, Dell
It’s amazing to me what the Council has been able to accomplish in under a year. Sometimes it all flies by so fast that you don’t have time to sit back and try to take it all in. Am I being a little grandiose here? Ya, sure, but a lot of folks from various companies have put in a huge amount of work…and it’s nice to see it all come into fruition where it benefits all members. Thank you everyone for helping the Council become what it is today.
You can find the whole interview here.
Talk with you soon,
Recently we have completed a small refresh in the cluster center. The Cluster Center offers an environment for developing, testing, benchmarking and optimizing products free of charge. The center, located in Sunnyvale, California, provides on-site technical support and enables secure sessions onsite or remotely. The Cluster Center provides a unique ability to access the latest clustering technology, sometimes even before it reaches public availability.
In the last few weeks, we have completed the installation of a Windows HPC Server 2008 cluster, and now it is available for testing (via the Vulcan cluster). We have also received the Scyld ClusterWare™ HPC cluster management solution from Penguin Computing (a member company) and installed it on the Osiris cluster.
Scyld was designed to make the deployment and management of Linux clusters as easy as the deployment and management of a single system. A Scyld ClusterWare cluster consists of a master node and compute nodes. The master node is the central point of control for the entire cluster. Compute nodes appear as attached processor and memory resources. More information on Scyld can be found here.
Adding Scyld to Osiris helps the Council with the best practices research activities that provide guidelines to end-users on how to maximize productivity for various applications using 20 and 40Gb/s InfiniBand 20 or 10 Gigabit Ethernet. I would like to thank Matt Jacobs and Joshua Bernstein from Penguin Computing for their donation and support during the Scyld installation.
Chairman of the HPC Advisory Council
Recently, we have moved the cluster center to a new location. In the new location (Sunnyvale, CA) we have now enough power and space to accommodate more systems and new technologies.
The HPC Advisory Council has received two new systems from Dell, AMD and Mellanox Technologies. The first system is a 24-node Dell™ PowerEdge™ SC 1435 24- cluster, loaded with Quad-Core AMD Opteron™ Model 2382 processors (“Shanghai”) and Mellanox® InfiniBand ConnectX® HCAs and switches. The system has been operational for couple of months already and is being used for the HPC Advisory Council’s Best Practices work and also available for end-user access.
The second one is a Dell™ M1000e blade system. It was just received and will be operation shortly. The system will be used for extending the HPC Advisory Council’s capability to provide resources for end-user benchmarking, HPC outreach, research activities on applications productivity, and bringing green computing to high-performance computing.
On the behalf of the HPC Advisory Council, I would like to thank Dell, AMD and Mellanox Technologies for providing the systems.
Gilad Shainer (HPC Advisory Council Chairman), Brian Sparks (HPC Advisory Council Media Relations Director) and Tong Liu (HPC Advisory Council Cluster Center Manager) at the HPC Advisory Council Cluster Center.
Gilad Shainer with the new M1000e system