Difference between revisions of "SOSCIP GPU"

From oldwiki.scinet.utoronto.ca
Jump to navigation Jump to search
(Created page with "{{Infobox Computer |image=center|300px|thumb |name=P8 |installed=June 2016 |operatingsystem= Linux RHEL 7.2 le / Ubuntu 16.04 le |loginnode= p8t0[1-2] ...")
 
Line 2: Line 2:
 
|image=[[Image:P8_s822.jpg|center|300px|thumb]]
 
|image=[[Image:P8_s822.jpg|center|300px|thumb]]
 
|name=P8  
 
|name=P8  
|installed=June 2016
+
|installed=September 2017
|operatingsystem= Linux RHEL 7.2 le / Ubuntu 16.04 le  
+
|operatingsystem= Ubuntu 16.04 le  
|loginnode= p8t0[1-2] / p8t0[3-4]
+
|loginnode= sgc01
|nnodes= 2x  Power8 with 2x NVIDIA K80,      2x Power 8 with  4x NVIDIA P100
+
|nnodes= 14x Power 8 with  4x NVIDIA P100
 
|rampernode=512 GB
 
|rampernode=512 GB
|corespernode= 2 x 8core (16 physical, 128 SMT)
+
|corespernode= 2 x 10core (20 physical, 160 SMT)
 
|interconnect=Infiniband EDR  
 
|interconnect=Infiniband EDR  
 
|vendorcompilers=xlc/xlf, nvcc
 
|vendorcompilers=xlc/xlf, nvcc
 
}}
 
}}
 +
 +
== SOSCIP ==
 +
 +
The SOSCIP GPU Cluster is a Southern Ontario Smart Computing Innovation Platform ([http://soscip.org/ SOSCIP]) resource located at theUniversity of Toronto's SciNet HPC facility. The SOSCIP  multi-university/industry consortium is funded by the Ontario Government and the Federal Economic Development Agency for Southern Ontario [http://www.research.utoronto.ca/about/our-research-partners/soscip/].
  
 
== Specifications==
 
== Specifications==
  
The P8 Test System consists of  of 4 IBM Power 822LC Servers each with 2x8core 3.25GHz Power8 CPUs and 512GB Ram. Similar to Power 7, the Power 8 utilizes Simultaneous MultiThreading (SMT), but extends the design to 8 threads per core allowing the 16 physical cores to support up to 128 threads.  2 nodes have two NVIDIA Tesla K80 GPUs with CUDA Capability 3.7 (Kepler), consisting of 2xGK210 GPUs each with 12 GB of RAM connected using PCI-E, and 2 others have 4x NVIDIA Tesla P100 GPUs each wit h 16GB of RAM with CUDA Capability 6.0 (Pascal) connected using NVlink.
+
The SOSCIP GPU Cluster consists of  of 14 IBM Power 822LC Servers each with 2x10core 3.25GHz Power8 CPUs and 512GB Ram. Similar to Power 7, the Power 8 utilizes Simultaneous MultiThreading (SMT), but extends the design to 8 threads per core allowing the 20 physical cores to support up to 160 threads.  Each node has 4x NVIDIA Tesla P100 GPUs each with 16GB of RAM with CUDA Capability 6.0 (Pascal) connected using NVlink.
  
 
== Compile/Devel/Test ==
 
== Compile/Devel/Test ==
  
First login via ssh with your scinet account at '''<tt>login.scinet.utoronto.ca</tt>''', and from there you can proceed to '''<tt>p8t0[1-2]</tt>''' for the K80 GPUs and '''<tt>p8t0[3-4]</tt>''' for the Pascal GPUs.
+
Access is provided through the BGQ login node, '''<tt> bgqdev.scinet.utoronto.ca </tt>''' node using ssh, and from there you can proceed to the development node'''<tt>sgc01</tt>'''.
  
== Software for  ==
+
== Software ==
  
 
==== GNU Compilers ====
 
==== GNU Compilers ====
Line 26: Line 30:
 
To load the newer advance toolchain version use:
 
To load the newer advance toolchain version use:
  
For '''<tt>p8t0[1-2]</tt>'''
 
<pre>
 
module load gcc/5.3.1
 
</pre>
 
 
For '''<tt>p8t0[3-4]</tt>'''
 
 
<pre>
 
<pre>
 
module load gcc/6.2.1
 
module load gcc/6.2.1
Line 39: Line 37:
  
 
To load the native IBM xlc/xlc++ compilers
 
To load the native IBM xlc/xlc++ compilers
 
For '''<tt>p8t0[1-2]</tt>'''
 
<pre>
 
module load xlc/13.1.4
 
module load xlf/13.1.4
 
</pre>
 
  
 
For '''<tt>p8t0[3-4]</tt>'''  
 
For '''<tt>p8t0[3-4]</tt>'''  
Line 73: Line 65:
 
==== OpenMPI ====
 
==== OpenMPI ====
  
Currently OpenMPI has been setup on the four nodes connected over QDR Infiniband.
+
Currently OpenMPI has been setup on the four nodes connected over EDR Infiniband.
 
 
For '''<tt>p8t0[1-2]</tt>'''
 
<pre>
 
$ module load openmpi/1.10.3-gcc-5.3.1
 
$ module load openmpi/1.10.3-XL-13_15.1.4
 
</pre>
 
  
For '''<tt>p8t0[3-4]</tt>'''
 
 
<pre>
 
<pre>
 
$ module load openmpi/1.10.3-gcc-6.2.1
 
$ module load openmpi/1.10.3-gcc-6.2.1
 
$ module load openmpi/1.10.3-XL-13_15.1.5
 
$ module load openmpi/1.10.3-XL-13_15.1.5
 
</pre>
 
</pre>
 
==== PE ====
 
 
IBM's Parallel Environment (PE), is available for use with XL compilers using the following
 
 
<pre>
 
$ module pe/xl.perf
 
</pre>
 
 
<pre>
 
mpiexec -n 4 ./a.out
 
</pre>
 
 
documentation is [http://publib.boulder.ibm.com/epubs/pdf/c2372832.pdf here]
 

Revision as of 13:56, 18 August 2017

P8
P8 s822.jpg
Installed September 2017
Operating System Ubuntu 16.04 le
Number of Nodes 14x Power 8 with 4x NVIDIA P100
Interconnect Infiniband EDR
Ram/Node 512 GB
Cores/Node 2 x 10core (20 physical, 160 SMT)
Login/Devel Node sgc01
Vendor Compilers xlc/xlf, nvcc

SOSCIP

The SOSCIP GPU Cluster is a Southern Ontario Smart Computing Innovation Platform (SOSCIP) resource located at theUniversity of Toronto's SciNet HPC facility. The SOSCIP multi-university/industry consortium is funded by the Ontario Government and the Federal Economic Development Agency for Southern Ontario [1].

Specifications

The SOSCIP GPU Cluster consists of of 14 IBM Power 822LC Servers each with 2x10core 3.25GHz Power8 CPUs and 512GB Ram. Similar to Power 7, the Power 8 utilizes Simultaneous MultiThreading (SMT), but extends the design to 8 threads per core allowing the 20 physical cores to support up to 160 threads. Each node has 4x NVIDIA Tesla P100 GPUs each with 16GB of RAM with CUDA Capability 6.0 (Pascal) connected using NVlink.

Compile/Devel/Test

Access is provided through the BGQ login node, bgqdev.scinet.utoronto.ca node using ssh, and from there you can proceed to the development nodesgc01.

Software

GNU Compilers

To load the newer advance toolchain version use:

module load gcc/6.2.1

IBM Compilers

To load the native IBM xlc/xlc++ compilers

For p8t0[3-4]

module load xlc/13.1.5
module load xlf/13.1.5


Driver Version

The current NVIDIA driver version is 361.93

CUDA

The current installed CUDA Tookit is 8.0

module load cuda/8.0

The CUDA driver is installed locally, however the CUDA Toolkit is installed in:

/usr/local/cuda-8.0

OpenMPI

Currently OpenMPI has been setup on the four nodes connected over EDR Infiniband.

$ module load openmpi/1.10.3-gcc-6.2.1
$ module load openmpi/1.10.3-XL-13_15.1.5