Difference between revisions of "Cell Devel Nodes"

From oldwiki.scinet.utoronto.ca
Jump to navigation Jump to search
 
(20 intermediate revisions by the same user not shown)
Line 1: Line 1:
 
{{Infobox Computer
 
{{Infobox Computer
 
|image=[[Image:300px-Cell_Broadband_Engine_Processor.jpg|center|300px|thumb]]
 
|image=[[Image:300px-Cell_Broadband_Engine_Processor.jpg|center|300px|thumb]]
|name=Cell Development Cluster
+
|name=Cell Development
 
|installed=June 2010
 
|installed=June 2010
 
|operatingsystem= Linux
 
|operatingsystem= Linux
Line 12: Line 12:
 
}}
 
}}
  
The Cell Development Cluster is a technology evaluation cluster with a combination of 14 IBM PowerXCell 8i "Cell" nodes and 8 Intel x86_64 "Nehalem" nodes.
+
''' NOTE: THIS SYSTEM IS NO LONGER AVAILABLE '''
The QS22's each have two 3.2GHz "IBM PowerXCell 8i CPU's, where each CPU has 1 Power Processing Unit (PPU) and 8 Synergistic Processing Units (SPU), and
 
32GB of RAM per node. The Intel nodes have two 2.53GHz 4core Xeon X5550 CPU's with 48GB of RAM per node. 
 
  
 +
The Cell Development nodes are IBM QS22's each with two 3.2GHz "IBM PowerXCell 8i CPU's, where each CPU has 1 Power Processing Unit (PPU) and 8 Synergistic Processing Units (SPU), and
 +
32GB of RAM per node.
 
===Login===
 
===Login===
  
First login via ssh with your scinet account at <tt>login.scinet.utoronto.ca</tt>, and from there you can proceed to <tt>cell-srv01</tt> which  
+
First login via ssh with your scinet account at '''<tt>login.scinet.utoronto.ca</tt>''', and from there you can proceed to '''<tt>arc01</tt>''' which  
 
is currently the gateway machine.
 
is currently the gateway machine.
  
===Compile/Devel/Compute Nodes===
+
==Compile/Devel/Compute Nodes==
  
Currently you can log into any of 12 nodes <tt>blade03..blade14</tt> directly to compile/test/run your code.
+
=== Cell ===
 
+
You can log into any of 12 nodes '''<tt>blade[03-14]</tt>''' directly to compile/test/run Cell
===Local Disk===
+
specific or OpenCL codes.
 
 
This test cluster currently cannot see the global <tt>/home</tt> and <tt>/scratch</tt> space so you will have to copy (scp,etc..) your code to a separate local<tt>/home</tt> dedicated for this cluster.  Also initially you will probably want to copy/modify your scinet .bashrc, .bashrc_profile, and .ssh directory onto this system.
 
  
 
==Programming Frameworks==
 
==Programming Frameworks==
Line 47: Line 45:
 
</pre>
 
</pre>
  
==Compilers==
+
===Compilers===
  
 
NOTE: gcc has been replaced, use ppu-gcc for most cases where gcc whould be used
 
NOTE: gcc has been replaced, use ppu-gcc for most cases where gcc whould be used
Line 56: Line 54:
  
 
* '''xlcl''' -- IBM XL OpenCL compiler
 
* '''xlcl''' -- IBM XL OpenCL compiler
 +
 +
===MPI===
 +
 +
Still a work in progress.
 +
  
 
== Documentation ==
 
== Documentation ==
Line 70: Line 73:
 
== Further Info ==
 
== Further Info ==
 
IBM Cell Broadband engine tutorials http://publib.boulder.ibm.com/infocenter/ieduasst/stgv1r0/index.jsp
 
IBM Cell Broadband engine tutorials http://publib.boulder.ibm.com/infocenter/ieduasst/stgv1r0/index.jsp
 +
 +
 +
== User Codes ==
 +
 +
Please discuss put any relevant information/problems/best practices you have encountered when using/developing for CELL and/or OpenCL
 +
 +
=== MHD ===
 +
 +
=== CFD ===

Latest revision as of 00:18, 18 September 2013

Cell Development
300px-Cell Broadband Engine Processor.jpg
Installed June 2010
Operating System Linux
Interconnect Infiniband
Ram/Node 32 Gb
Cores/Node 2 PPU + 16 SPU
Login/Devel Node cell-srv01 (from login.scinet)
Vendor Compilers ppu-gcc, spu-gcc

NOTE: THIS SYSTEM IS NO LONGER AVAILABLE

The Cell Development nodes are IBM QS22's each with two 3.2GHz "IBM PowerXCell 8i CPU's, where each CPU has 1 Power Processing Unit (PPU) and 8 Synergistic Processing Units (SPU), and 32GB of RAM per node.

Login

First login via ssh with your scinet account at login.scinet.utoronto.ca, and from there you can proceed to arc01 which is currently the gateway machine.

Compile/Devel/Compute Nodes

Cell

You can log into any of 12 nodes blade[03-14] directly to compile/test/run Cell specific or OpenCL codes.

Programming Frameworks

Currently there are two programming frameworks to use. The Cell SDK is the most full featured and "supported" but is architecture specific, whereas the OpenCL should allow code portability to other accelerators (like GP-GPU's) however is from IBM alphaworks so should be considered beta.

OpenCL

Examples are in

/usr/share/doc/OpenCL-0.1-ibm

Cell SDK

Demos, examples, and build details available in

/opt/cell/sdk/

Compilers

NOTE: gcc has been replaced, use ppu-gcc for most cases where gcc whould be used

  • ppu-gcc -- GNU GCC PPC compiler
  • spu-gcc -- GNU GCC SPU compiler
  • xlcl -- IBM XL OpenCL compiler

MPI

Still a work in progress.


Documentation

Further Info

IBM Cell Broadband engine tutorials http://publib.boulder.ibm.com/infocenter/ieduasst/stgv1r0/index.jsp


User Codes

Please discuss put any relevant information/problems/best practices you have encountered when using/developing for CELL and/or OpenCL

MHD

CFD