GPU Devel Nodes

From oldwiki.scinet.utoronto.ca
Revision as of 09:23, 16 April 2010 by Northrup (talk | contribs) (→‎MPI)
Jump to navigation Jump to search
GPU Development Cluster
GeForce 9800 GT 3qtr low.png
Installed June 2010
Operating System Linux
Interconnect Infiniband,GigE
Ram/Node 48 Gb
Cores/Node 8
Login/Devel Node cell-srv01 (from login.scinet)
Vendor Compilers gcc,nvcc

The Intel nodes have two 2.53GHz 4core Xeon X5550 CPU's with 48GB of RAM per node with 3 containing NVIDIA 9800GT GPUs.

Login

First login via ssh with your scinet account at login.scinet.utoronto.ca, and from there you can proceed to cell-srv01 which is currently the gateway machine.

Compile/Devel/Compute Nodes

Nehalem (x86_64)

You can log into any of 8 nodes cell-srv[01-08] directly however the nodes have differing configurations as follows:

  • cell-srv01 - login node & nfs server, GigE connected
  • cell-srv[02-05] - no GPU, GigE connected
  • cell-srv[06-07] - 1x NVIDIA 9800GT GPU, Infiniband connected
  • cell-srv08 - 2x NVIDIA 9800GT GPU, GigE connected

Software

The same software installed on the GPC is available on ARC using the same modules framework. See here for full details.

Programming Frameworks

Currently there are two programming frameworks to use, NVIDIA's CUDA framework or OpenCL.

CUDA

The current CUDA Toolkit in use is 3.0. To use it just add the following module

module load cuda

The CUDA driver is installed locally, however the CUDA SDK is installed in.

/project/scinet/arc/cuda/

OpenCL

As of 3.0, OpenCL is included in the CUDA Toolkit so loading the CUDA module is all the is required.

Compilers

  • nvcc -- Nvidia compiler

MPI

The GPC MPI packages can be used on this system. See the GPC section on MPI for more details.

Documentation

  • CUDA
    • google "CUDA"
  • OpenCL
    • see above

Further Info

User Codes

Please discuss put any relevant information/problems/best practices you have encountered when using/developing for CUDA and/or OpenCL