<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
		<id>https://hcl.ucd.ie/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Zhongziming</id>
		<title>HCL - User contributions [en]</title>
		<link rel="self" type="application/atom+xml" href="https://hcl.ucd.ie/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Zhongziming"/>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php/Special:Contributions/Zhongziming"/>
		<updated>2026-04-15T03:53:04Z</updated>
		<subtitle>User contributions</subtitle>
		<generator>MediaWiki 1.27.1</generator>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=772</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=772"/>
				<updated>2012-08-22T22:32:58Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with [http://developer.amd.com/libraries/acml/pages/default.aspx acml] for CPU and [http://developer.nvidia.com/cublas cublas] for GPU&lt;br /&gt;
&lt;br /&gt;
 $ cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 $ mkdir acml_config &lt;br /&gt;
 $ cd acml_config&lt;br /&gt;
 $ ../configure --with-blas=acml&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
 $ mkdir cuda_config &lt;br /&gt;
 $ cd cuda_config&lt;br /&gt;
 $ ../configure --with-blas=cuda&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for [http://www.open-mpi.org/doc/v1.6/man1/mpirun.1.php#sect8 process binding], and appfile tells mpirun what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
- Example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_1d.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_1d.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_1d.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_2d -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_2d -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=771</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=771"/>
				<updated>2012-08-22T11:36:32Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with [http://developer.amd.com/libraries/acml/pages/default.aspx acml] for CPU and [http://developer.nvidia.com/cublas cublas] for GPU&lt;br /&gt;
&lt;br /&gt;
 $ cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 $ mkdir acml_config &lt;br /&gt;
 $ cd acml_config&lt;br /&gt;
 $ ../configure --with-cblas=acml&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
 $ mkdir cuda_config &lt;br /&gt;
 $ cd cuda_config&lt;br /&gt;
 $ ../configure --with-cblas=cuda&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for [http://www.open-mpi.org/doc/v1.6/man1/mpirun.1.php#sect8 process binding], and appfile tells mpirun what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
- Example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_1d.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_1d.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_1d.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_2d -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_2d -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=757</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=757"/>
				<updated>2012-07-16T10:07:28Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with [http://developer.amd.com/libraries/acml/pages/default.aspx acml] for CPU and [http://developer.nvidia.com/cublas cublas] for GPU&lt;br /&gt;
&lt;br /&gt;
 $ cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 $ mkdir acml_config &lt;br /&gt;
 $ cd acml_config&lt;br /&gt;
 $ ../configure --with-cblas=acml&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
 $ mkdir cuda_config &lt;br /&gt;
 $ cd cuda_config&lt;br /&gt;
 $ ../configure --with-cblas=cuda&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for [http://www.open-mpi.org/doc/v1.6/man1/mpirun.1.php#toc8 process binding], and appfile tells mpirun what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- Example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=756</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=756"/>
				<updated>2012-07-12T12:57:07Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with [http://developer.amd.com/libraries/acml/pages/default.aspx acml] for CPU and [http://developer.nvidia.com/cublas cublas] for GPU&lt;br /&gt;
&lt;br /&gt;
 $ cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 $ mkdir acml_config &lt;br /&gt;
 $ cd acml_config&lt;br /&gt;
 $ ../configure --with-cblas=acml&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
 $ mkdir cuda_config &lt;br /&gt;
 $ cd cuda_config&lt;br /&gt;
 $ ../configure --with-cblas=cuda&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for [http://www.open-mpi.org/doc/v1.6/man1/mpirun.1.php#toc8 process binding], and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- Example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- Example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=755</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=755"/>
				<updated>2012-07-12T11:11:53Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with [http://developer.amd.com/libraries/acml/pages/default.aspx acml] for CPU and [http://developer.nvidia.com/cublas cublas] for GPU&lt;br /&gt;
&lt;br /&gt;
 $ cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 $ mkdir acml_config &lt;br /&gt;
 $ cd acml_config&lt;br /&gt;
 $ ../configure --with-cblas=acml&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
 $ mkdir cuda_config &lt;br /&gt;
 $ cd cuda_config&lt;br /&gt;
 $ ../configure --with-cblas=cuda&lt;br /&gt;
 $ make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=754</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=754"/>
				<updated>2012-07-12T11:11:01Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with [http://developer.amd.com/libraries/acml/pages/default.aspx acml] for CPU and [http://developer.nvidia.com/cublas cublas] for GPU&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ../configure --with-cblas=acml&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ../configure --with-cblas=cuda&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=753</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=753"/>
				<updated>2012-07-12T11:10:45Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with [http://developer.amd.com/libraries/acml/pages/default.aspx acml] for CPU and [http://developer.nvidia.com/cublas] for GPU&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ../configure --with-cblas=acml&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ../configure --with-cblas=cuda&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=752</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=752"/>
				<updated>2012-07-12T11:09:29Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with acml for CPU ([http://developer.amd.com/libraries/acml/pages/default.aspx acml]) and cublas for GPU (), and then make&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=751</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=751"/>
				<updated>2012-07-12T11:08:06Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: configuring with acml blas for CPU and cublas for GPU, and then make&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
 make&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=750</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=750"/>
				<updated>2012-07-12T11:06:33Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl) and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=OpenMPI&amp;diff=749</id>
		<title>OpenMPI</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=OpenMPI&amp;diff=749"/>
				<updated>2012-07-12T11:05:59Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;http://www.open-mpi.org/faq/&lt;br /&gt;
&lt;br /&gt;
== MCA parameter files ==&lt;br /&gt;
If you want to permanently use some MCA parameter settings, you can create a file $HOME/.openmpi/mca-params.conf, e.g.:&lt;br /&gt;
&lt;br /&gt;
 cat $HOME/.openmpi/mca-params.conf&lt;br /&gt;
 btl_tcp_if_exclude = lo,eth1&lt;br /&gt;
&lt;br /&gt;
== Running applications on Multiprocessors/Multicores ==&lt;br /&gt;
Process can be bound to specific sockets and cores on nodes by choosing right options of mpirun.&lt;br /&gt;
* [http://www.open-mpi.org/doc/v1.4/man1/mpirun.1.php#sect9 Process binding]&lt;br /&gt;
* [http://www.open-mpi.org/doc/v1.4/man1/mpirun.1.php#sect10 Rankfile]&lt;br /&gt;
&lt;br /&gt;
== PERUSE ==&lt;br /&gt;
[[Media:current_peruse_spec.pdf|PERUSE Specification]]&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=CUDA_SDK&amp;diff=748</id>
		<title>CUDA SDK</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=CUDA_SDK&amp;diff=748"/>
				<updated>2012-07-12T11:04:57Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: Created page with &amp;quot;http://developer.nvidia.com/gpu-computing-sdk&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;http://developer.nvidia.com/gpu-computing-sdk&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=BLAS_LAPACK_ScaLAPACK&amp;diff=747</id>
		<title>BLAS LAPACK ScaLAPACK</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=BLAS_LAPACK_ScaLAPACK&amp;diff=747"/>
				<updated>2012-07-12T11:04:10Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;A de facto standard API for linear algebra [http://en.wikipedia.org/wiki/Basic_Linear_Algebra_Subprograms BLAS]/[http://en.wikipedia.org/wiki/LAPACK LAPACK]&lt;br /&gt;
* Original http://www.netlib.org/blas/ http://www.netlib.org/lapack/ - implemented in Fortran. The libraries can be used in C/C++ (so called Fortran interface to BLAS/LAPACK).&lt;br /&gt;
* ATLAS http://math-atlas.sourceforge.net/ - provides a C interface to BLAS and partially LAPACK. Binary packages: libatlas-[base or platform name, for example sse2]&lt;br /&gt;
* MKL http://software.intel.com/en-us/intel-mkl/ - Intel implementation&lt;br /&gt;
*ACML http://developer.amd.com/libraries/acml/pages/default.aspx&lt;br /&gt;
*CUBLAS http://developer.nvidia.com/cublas&lt;br /&gt;
&lt;br /&gt;
Using the C interface is preferable. [http://www.inf.bv.tum.de/~heisserer/softwarelab04/doc/blas_report.pdf BLAS: overview, installation, usage]&lt;br /&gt;
&lt;br /&gt;
= ScaLAPACK =&lt;br /&gt;
http://www.netlib.org/scalapack/&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=BLAS_LAPACK_ScaLAPACK&amp;diff=746</id>
		<title>BLAS LAPACK ScaLAPACK</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=BLAS_LAPACK_ScaLAPACK&amp;diff=746"/>
				<updated>2012-07-12T11:03:28Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;A de facto standard API for linear algebra [http://en.wikipedia.org/wiki/Basic_Linear_Algebra_Subprograms BLAS]/[http://en.wikipedia.org/wiki/LAPACK LAPACK]&lt;br /&gt;
* Original http://www.netlib.org/blas/ http://www.netlib.org/lapack/ - implemented in Fortran. The libraries can be used in C/C++ (so called Fortran interface to BLAS/LAPACK).&lt;br /&gt;
* ATLAS http://math-atlas.sourceforge.net/ - provides a C interface to BLAS and partially LAPACK. Binary packages: libatlas-[base or platform name, for example sse2]&lt;br /&gt;
* MKL http://software.intel.com/en-us/intel-mkl/ - Intel implementation&lt;br /&gt;
*ACML http://developer.amd.com/libraries/acml/pages/default.aspx&lt;br /&gt;
&lt;br /&gt;
Using the C interface is preferable. [http://www.inf.bv.tum.de/~heisserer/softwarelab04/doc/blas_report.pdf BLAS: overview, installation, usage]&lt;br /&gt;
&lt;br /&gt;
= ScaLAPACK =&lt;br /&gt;
http://www.netlib.org/scalapack/&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=745</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=745"/>
				<updated>2012-07-12T11:02:15Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
- An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=744</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=744"/>
				<updated>2012-07-12T11:01:21Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
- For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
- Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
 An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
- Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
- An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=743</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=743"/>
				<updated>2012-07-12T11:00:46Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
-For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
-Rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
 An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
-An example of an appfile for building functional permanence model (appfile_fpm):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
-Matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
-An example of an appfile for matrix multiplication (appfile_mxm)&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=742</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=742"/>
				<updated>2012-07-12T10:58:56Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
 An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
An example of an appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
An example of an appfile for matrix multiplication&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=741</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=741"/>
				<updated>2012-07-12T10:58:18Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
 For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
 rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
 An example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
 An example of an appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
 matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
 An example of an appfile for matrix multiplication&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=740</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=740"/>
				<updated>2012-07-12T10:56:08Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
** example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
** example of an appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
** matrix size D = N x N, and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
** example of an appfile for matrix multiplication&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=739</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=739"/>
				<updated>2012-07-12T10:54:56Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
** example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
** example of an appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU #&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # CPU #&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
** matrix size D = N x N, N = sqrt(D), and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
** example of an appfile for matrix multiplication&lt;br /&gt;
  # GPU #&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  # CPU #&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=738</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=738"/>
				<updated>2012-07-12T10:54:07Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
** example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
** example of an appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # -----------------------------------------------------------------------------&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
* Data partitioning&lt;br /&gt;
&lt;br /&gt;
** matrix size D = N x N, N = sqrt(D), and machinefile lists the nodes participating in the computing&lt;br /&gt;
 &lt;br /&gt;
  $ fupermod/tools/partitioner -l fupermod/routines/mxm/.libs/libmxm_col.so -D10000 -o N=100 -m machinefile&lt;br /&gt;
&lt;br /&gt;
* Running matrix multiplication&lt;br /&gt;
&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_mxm&lt;br /&gt;
&lt;br /&gt;
** example of an appfile for matrix multiplication&lt;br /&gt;
  # GPU&lt;br /&gt;
  # Assuming fupermod is configured under cublas_config, linking against cublas&lt;br /&gt;
  # -g0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/routines/mxm/mxm_col -k640 -g0 -m machinefile&lt;br /&gt;
  #--------------------------------------------------------------------------------------------------------&lt;br /&gt;
  # CPU&lt;br /&gt;
  # Assuming fupermod is configured under acml_config, linking against acml&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/routines/mxm/mxm_col -k640 -m machinefile&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=737</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=737"/>
				<updated>2012-07-12T10:49:29Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
** example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
** example of a appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  # -----------------------------------------------------------------------------&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=736</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=736"/>
				<updated>2012-07-12T10:48:45Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
** example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
** example of a appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=735</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=735"/>
				<updated>2012-07-12T10:48:30Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
** example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
** example of a appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
  #---------------------------------------------------------------------------------------------------------------------------------------------------------------&lt;br /&gt;
&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=734</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=734"/>
				<updated>2012-07-12T10:47:34Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
** example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
  ...&lt;br /&gt;
&lt;br /&gt;
** example of a appfile for building functional permanence model (FPM):&lt;br /&gt;
  # GPU&lt;br /&gt;
  # e.g. Linking against cublas, and fupermod is configured under cublas_config&lt;br /&gt;
  # suboption g=0 means device 0 is selected for computing&lt;br /&gt;
&lt;br /&gt;
  -host localhost -np 1 $HOME/fupermod/cublas_config/tools/builder -l $HOME/fupermod/cublas_config/routines/mxm/.libs/libmxm_col.so -o k=640,g=0 -U10000 -s10&lt;br /&gt;
&lt;br /&gt;
  #---------------------------------------------------------------------------------------------------------------------------------------------------------------&lt;br /&gt;
&lt;br /&gt;
  # CPU&lt;br /&gt;
  # e.g. Linking against acml, and fupermod is configured under acml_config&lt;br /&gt;
&lt;br /&gt;
  -host localhost -np 47 $HOME/fupermod/acml_config/tools/builder -l $HOME/fupermod/acml_config/routines/mxm/.libs/libmxm_col.so -o k=640 -U10000 -s10&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=733</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=733"/>
				<updated>2012-07-12T10:45:38Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
** rankfile is for processing binding, and appfile tells mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;br /&gt;
&lt;br /&gt;
  example of a rankfile:&lt;br /&gt;
  rank 0=ig.icl.utk.edu slot=0:0&lt;br /&gt;
  rank 1=ig.icl.utk.edu slot=0:1&lt;br /&gt;
   .&lt;br /&gt;
   .&lt;br /&gt;
   .&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=732</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=732"/>
				<updated>2012-07-12T10:44:06Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Building performance model:&lt;br /&gt;
** rankfile is for processing binding, appfile tell mpi what programs to launch&lt;br /&gt;
  $ mpirun -rf rankfile -app appfile_fpm&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=731</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=731"/>
				<updated>2012-07-12T10:42:16Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
* Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
** For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;br /&gt;
&lt;br /&gt;
* Build performance model:&lt;br /&gt;
&lt;br /&gt;
/*&lt;br /&gt;
 * rankfile is for process binding&lt;br /&gt;
 * appfile tells what processes will execute&lt;br /&gt;
 */&lt;br /&gt;
&lt;br /&gt;
$ mpirun -rf rankfile -app appfile_fpm&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=730</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=730"/>
				<updated>2012-07-12T10:40:04Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
**For example: Using acml blas for CPU and cublas for GPU computing&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=729</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=729"/>
				<updated>2012-07-12T10:38:57Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). For example:&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 ** Using acml for CPU compuing&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
** Using cublas for GPU computing */  &lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=728</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=728"/>
				<updated>2012-07-12T10:38:26Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). For example:&lt;br /&gt;
&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 /* Using acml for CPU compuing*/&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 /* Using cublas for GPU computing */  &lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=727</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=727"/>
				<updated>2012-07-12T10:38:06Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
 For example:&lt;br /&gt;
 cd fupermod/&lt;br /&gt;
&lt;br /&gt;
 /* Using acml for CPU compuing*/&lt;br /&gt;
 mkdir acml_config &lt;br /&gt;
 cd acml_config&lt;br /&gt;
 ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
 /* Using cublas for GPU computing */  &lt;br /&gt;
 mkdir cuda_config &lt;br /&gt;
 cd cuda_config&lt;br /&gt;
 ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=726</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=726"/>
				<updated>2012-07-12T10:37:11Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
  For example:&lt;br /&gt;
&lt;br /&gt;
  cd fupermod/&lt;br /&gt;
&lt;br /&gt;
  /* Using acml for CPU compuing*/&lt;br /&gt;
  mkdir acml_config &lt;br /&gt;
  cd acml_config&lt;br /&gt;
  ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
  /* Using cublas for GPU computing */  &lt;br /&gt;
  mkdir cuda_config &lt;br /&gt;
  cd cuda_config&lt;br /&gt;
  ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=725</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=725"/>
				<updated>2012-07-12T10:36:37Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
cd fupermod/&lt;br /&gt;
&lt;br /&gt;
/* Using acml for CPU compuing*/&lt;br /&gt;
mkdir acml_config &lt;br /&gt;
cd acml_config&lt;br /&gt;
./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
/* Using cublas for GPU computing */  &lt;br /&gt;
mkdir cuda_config &lt;br /&gt;
cd cuda_config&lt;br /&gt;
./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=724</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=724"/>
				<updated>2012-07-12T10:35:51Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). &lt;br /&gt;
  For example:&lt;br /&gt;
&lt;br /&gt;
  cd fupermod/&lt;br /&gt;
&lt;br /&gt;
  /* Using acml for CPU compuing*/&lt;br /&gt;
  mkdir acml_config &lt;br /&gt;
  cd acml_config&lt;br /&gt;
  ./configure --with-cblas=acml&lt;br /&gt;
&lt;br /&gt;
  /* Using cublas for GPU computing */  &lt;br /&gt;
  mkdir cuda_config &lt;br /&gt;
  cd cuda_config&lt;br /&gt;
  ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=723</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=723"/>
				<updated>2012-07-12T10:35:31Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two separate directories for configuration with selected CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). For example:&lt;br /&gt;
&lt;br /&gt;
  cd fupermod/&lt;br /&gt;
  /* Using acml for CPU compuing*/&lt;br /&gt;
  mkdir acml_config &lt;br /&gt;
  cd acml_config&lt;br /&gt;
  ./configure --with-cblas=acml&lt;br /&gt;
  /* Using cublas for GPU computing */  &lt;br /&gt;
  mkdir cuda_config &lt;br /&gt;
  cd cuda_config&lt;br /&gt;
  ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=722</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=722"/>
				<updated>2012-07-12T10:34:21Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two directories for configuration with CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). For example:&lt;br /&gt;
&lt;br /&gt;
  cd fupermod/&lt;br /&gt;
  &lt;br /&gt;
  mkdir acml_config /* Using acml for CPU compuing*/&lt;br /&gt;
  cd acml_config&lt;br /&gt;
  ./configure --with-cblas=acml&lt;br /&gt;
    &lt;br /&gt;
  mkdir cuda_config /* Using cublas for GPU computing */&lt;br /&gt;
  cd cuda_config&lt;br /&gt;
  ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=721</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=721"/>
				<updated>2012-07-12T10:33:46Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two directories for configuration with CPU cblas (e.g. gsl, acml, mkl)and GPU cblas (e.g. cublas). For example:&lt;br /&gt;
&lt;br /&gt;
  cd fupermod/&lt;br /&gt;
&lt;br /&gt;
  /* Using acml for CPU compuing*/&lt;br /&gt;
  mkdir acml_config&lt;br /&gt;
  cd acml_config&lt;br /&gt;
  ./configure --with-cblas=acml&lt;br /&gt;
  &lt;br /&gt;
  /* Using cublas for GPU computing */&lt;br /&gt;
  mkdir cuda_config&lt;br /&gt;
  cd cuda_config&lt;br /&gt;
  ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=720</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=720"/>
				<updated>2012-07-12T10:31:41Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid multicore/GPUs node ==&lt;br /&gt;
*Compiling : Create two directories for configuring with cblas (for CPU) cublas (for GPU). For example:&lt;br /&gt;
  /* Using acml for CPU compuing*/&lt;br /&gt;
  mkdir acml_config&lt;br /&gt;
  cd acml_config&lt;br /&gt;
  ./configure --with-cblas=acml&lt;br /&gt;
  &lt;br /&gt;
  /* Using cublas for GPU computing */&lt;br /&gt;
  mkdir cuda_config&lt;br /&gt;
  cd cuda_config&lt;br /&gt;
  ./configure --with-cblas=cuda&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=719</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=719"/>
				<updated>2012-07-12T10:22:34Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: /* Using Fupermod on hybrid node */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid node ==&lt;br /&gt;
*Compiling&lt;br /&gt;
Currently user need to compile the code for CPU and GPU seperately&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=718</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=718"/>
				<updated>2012-07-12T10:20:57Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;br /&gt;
&lt;br /&gt;
== Using Fupermod on hybrid node ==&lt;br /&gt;
*Compiling&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=717</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=717"/>
				<updated>2012-07-12T10:11:59Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Display a list of available GPUs ==&lt;br /&gt;
$ nvidia-smi -L&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=716</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=716"/>
				<updated>2012-07-12T10:06:19Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
  http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Getting the info of GPUs on a node ==&lt;br /&gt;
  nvidia-smi -L&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=715</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=715"/>
				<updated>2012-07-12T10:05:31Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== List of machines ==&lt;br /&gt;
  http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180&lt;br /&gt;
&lt;br /&gt;
== Checking number and type of GPUs ==&lt;br /&gt;
  nvidia-smi -L&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=714</id>
		<title>UTK multicores + GPU</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=UTK_multicores_%2B_GPU&amp;diff=714"/>
				<updated>2012-07-12T10:04:59Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== List of machines ==&lt;br /&gt;
http://icl.cs.utk.edu/iclhelp/custom/index.html?lid=97&amp;amp;slid=180 UTK machines&lt;br /&gt;
&lt;br /&gt;
== Checking number and type of GPUs ==&lt;br /&gt;
nvidia-smi -L&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=596</id>
		<title>Grid5000</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=596"/>
				<updated>2011-05-03T18:35:04Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: /* Setting up new deploy image */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;https://www.grid5000.fr/mediawiki/index.php/Grid5000:Home &lt;br /&gt;
&lt;br /&gt;
[https://www.grid5000.fr/mediawiki/index.php/Grid5000:UserCharter USAGE POLICY]&lt;br /&gt;
&lt;br /&gt;
== Login, job submission, deployment of image  ==&lt;br /&gt;
&lt;br /&gt;
*Select sites and clusters for experiments, using information on the [https://www.grid5000.fr/mediawiki/index.php/Grid5000:Network#Grid.275000_Sites Grid5000 network] and the [https://www.grid5000.fr/mediawiki/index.php/Status Status page] &lt;br /&gt;
*Access is provided via access nodes '''access.SITE.grid5000.fr''' marked [https://www.grid5000.fr/mediawiki/index.php/External_access here] as ''accessible from '''everywhere''' via ssh with '''keyboard-interactive''' authentication method''. As soon as you are on one of the sites, you can directly ssh frontend node of any other site:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
access_$ ssh frontend.SITE2&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*There is no access to Internet from computing nodes (external IPs should be registered on proxy), therefore, download/update your stuff at the access nodes. Several revision control clients are available. &lt;br /&gt;
*Each site has a separate NFS, therefore, to run an application on several sites at once, you need to copy it '''scp, sftp, rsync''' between access or frontend nodes. &lt;br /&gt;
*Jobs are run from the frondend nodes, using a [http://en.wikipedia.org/wiki/OpenPBS PBS]-like system [https://www.grid5000.fr/mediawiki/index.php/Cluster_experiment-OAR2 OAR]. Basic commands: &lt;br /&gt;
**'''oarstat''' - queue status &lt;br /&gt;
**'''oarsub''' - job submission &lt;br /&gt;
**'''oardel''' - job removal&lt;br /&gt;
&lt;br /&gt;
Interactive job on deployed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub -I -t deploy -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; Batch job on installed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub BATCH_FILE -t allow_classic_ssh -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*The image to deploy can be created and loaded with help of a [http://wiki.systemimager.org/index.php/Main_Page Systemimager]-like system [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2 Kadeploy]. Creating: [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2#Tune_an_environment_to_build_another_one:_customize_authentification_parameters described here]&lt;br /&gt;
&lt;br /&gt;
Loading: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ kadeploy3 -a PATH_TO_PRIVATE_IMAGE_DESC -f $OAR_FILE_NODES  &lt;br /&gt;
&amp;lt;/source&amp;gt; A Linux distribution lenny-x64-nfs-2.1 with mc, subversion, autotools, doxygen, MPICH2, GSL, Boost, R, gnuplot, graphviz, X11, evince is available at Orsay /home/nancy/alastovetsky/grid5000. &lt;br /&gt;
&lt;br /&gt;
== Compiling and running MPI applications  ==&lt;br /&gt;
&lt;br /&gt;
*Compilation should be done on one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`) &lt;br /&gt;
*Running MPI applications is described [https://www.grid5000.fr/mediawiki/index.php/Run_MPI_On_Grid%275000 here] &lt;br /&gt;
**mpirun/mpiexec should be run from one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`)&lt;br /&gt;
&lt;br /&gt;
== Setting up new deploy image  ==&lt;br /&gt;
&lt;br /&gt;
 oarsub -I -t deploy -l nodes=1,walltime=12&lt;br /&gt;
 kadeploy3 -e lenny-x64-nfs -f $OAR_FILE_NODES -k&lt;br /&gt;
 ssh root@`head -n 1 $OAR_NODEFILE`&lt;br /&gt;
&lt;br /&gt;
edit /etc/apt/sources.list &lt;br /&gt;
&lt;br /&gt;
 apt-get update&lt;br /&gt;
 apt-get upgrade&lt;br /&gt;
&lt;br /&gt;
 apt-get install libtool autoconf automake mc colorgcc ctags libboost-serialization-dev libboost-graph-dev libatlas-base-dev gfortran vim gdb valgrind screen subversion&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Compiled for sources by us: &lt;br /&gt;
&lt;br /&gt;
* gsl-1.14 (download: ftp://ftp.gnu.org/gnu/gsl/)&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
* mpich2 (download: http://www.mcs.anl.gov/research/projects/mpich2/downloads/index.php?s=downloads)&lt;br /&gt;
 ./configure --enable-shared --enable-sharedlibs=gcc&lt;br /&gt;
 make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Mpich2 installed to:&lt;br /&gt;
 Installing MPE2 include files to /usr/local/include&lt;br /&gt;
 Installing MPE2 libraries to /usr/local/lib&lt;br /&gt;
 Installing MPE2 utility programs to /usr/local/bin&lt;br /&gt;
 Installing MPE2 configuration files to /usr/local/etc&lt;br /&gt;
 Installing MPE2 system utility programs to /usr/local/sbin&lt;br /&gt;
 Installing MPE2 man to /usr/local/share/man&lt;br /&gt;
 Installing MPE2 html to /usr/local/share/doc/&lt;br /&gt;
 Installed MPE2 in /usr/local&lt;br /&gt;
&lt;br /&gt;
* hwloc (and lstopo) (download: http://www.open-mpi.org/software/hwloc/v1.2/)&lt;br /&gt;
compile from sources. To get xml support install libxml2-dev and pkg-config&lt;br /&gt;
 apt-get install libxml2-dev pkg-config&lt;br /&gt;
 tar -xzvf hwloc-1.1.1.tar.gz&lt;br /&gt;
 cd hwloc-1.1.1&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Cleanup &lt;br /&gt;
&lt;br /&gt;
 apt-get clean&lt;br /&gt;
 rm /etc/udev/rules.d/*-persistent-net.rules&lt;br /&gt;
&lt;br /&gt;
Make image &lt;br /&gt;
&lt;br /&gt;
 ssh root@'''node''' tgz-g5k &amp;amp;gt; $HOME/grid5000/'''imagename'''.tgz&lt;br /&gt;
&lt;br /&gt;
make appropriate .env file.&lt;br /&gt;
 kaenv3 -p lenny-x64-nfs -u deploy &amp;gt; lenny-x64-custom-2.3.env&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=595</id>
		<title>Grid5000</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=595"/>
				<updated>2011-05-03T16:51:40Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: /* Setting up new deploy image */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;https://www.grid5000.fr/mediawiki/index.php/Grid5000:Home &lt;br /&gt;
&lt;br /&gt;
[https://www.grid5000.fr/mediawiki/index.php/Grid5000:UserCharter USAGE POLICY]&lt;br /&gt;
&lt;br /&gt;
== Login, job submission, deployment of image  ==&lt;br /&gt;
&lt;br /&gt;
*Select sites and clusters for experiments, using information on the [https://www.grid5000.fr/mediawiki/index.php/Grid5000:Network#Grid.275000_Sites Grid5000 network] and the [https://www.grid5000.fr/mediawiki/index.php/Status Status page] &lt;br /&gt;
*Access is provided via access nodes '''access.SITE.grid5000.fr''' marked [https://www.grid5000.fr/mediawiki/index.php/External_access here] as ''accessible from '''everywhere''' via ssh with '''keyboard-interactive''' authentication method''. As soon as you are on one of the sites, you can directly ssh frontend node of any other site:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
access_$ ssh frontend.SITE2&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*There is no access to Internet from computing nodes (external IPs should be registered on proxy), therefore, download/update your stuff at the access nodes. Several revision control clients are available. &lt;br /&gt;
*Each site has a separate NFS, therefore, to run an application on several sites at once, you need to copy it '''scp, sftp, rsync''' between access or frontend nodes. &lt;br /&gt;
*Jobs are run from the frondend nodes, using a [http://en.wikipedia.org/wiki/OpenPBS PBS]-like system [https://www.grid5000.fr/mediawiki/index.php/Cluster_experiment-OAR2 OAR]. Basic commands: &lt;br /&gt;
**'''oarstat''' - queue status &lt;br /&gt;
**'''oarsub''' - job submission &lt;br /&gt;
**'''oardel''' - job removal&lt;br /&gt;
&lt;br /&gt;
Interactive job on deployed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub -I -t deploy -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; Batch job on installed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub BATCH_FILE -t allow_classic_ssh -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*The image to deploy can be created and loaded with help of a [http://wiki.systemimager.org/index.php/Main_Page Systemimager]-like system [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2 Kadeploy]. Creating: [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2#Tune_an_environment_to_build_another_one:_customize_authentification_parameters described here]&lt;br /&gt;
&lt;br /&gt;
Loading: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ kadeploy3 -a PATH_TO_PRIVATE_IMAGE_DESC -f $OAR_FILE_NODES  &lt;br /&gt;
&amp;lt;/source&amp;gt; A Linux distribution lenny-x64-nfs-2.1 with mc, subversion, autotools, doxygen, MPICH2, GSL, Boost, R, gnuplot, graphviz, X11, evince is available at Orsay /home/nancy/alastovetsky/grid5000. &lt;br /&gt;
&lt;br /&gt;
== Compiling and running MPI applications  ==&lt;br /&gt;
&lt;br /&gt;
*Compilation should be done on one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`) &lt;br /&gt;
*Running MPI applications is described [https://www.grid5000.fr/mediawiki/index.php/Run_MPI_On_Grid%275000 here] &lt;br /&gt;
**mpirun/mpiexec should be run from one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`)&lt;br /&gt;
&lt;br /&gt;
== Setting up new deploy image  ==&lt;br /&gt;
&lt;br /&gt;
 oarsub -I -t deploy -l nodes=1,walltime=12&lt;br /&gt;
 kadeploy3 -e lenny-x64-nfs -f $OAR_FILE_NODES -k&lt;br /&gt;
 ssh root@`head -n 1 $OAR_NODEFILE`&lt;br /&gt;
&lt;br /&gt;
edit /etc/apt/sources.list &lt;br /&gt;
&lt;br /&gt;
 apt-get update&lt;br /&gt;
 apt-get upgrade&lt;br /&gt;
&lt;br /&gt;
 apt-get install libtool autoconf automake mc colorgcc ctags libboost-serialization-dev libboost-graph-dev libatlas-base-dev gfortran vim gdb valgrind screen subversion&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Compiled for sources by us: &lt;br /&gt;
&lt;br /&gt;
* gsl-1.14 (download: ftp://ftp.gnu.org/gnu/gsl/)&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
* mpich2 (download: http://www.mcs.anl.gov/research/projects/mpich2/downloads/index.php?s=downloads)&lt;br /&gt;
 ./configure --enable-shared --enable-sharedlibs=gcc&lt;br /&gt;
 make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Mpich2 installed to:&lt;br /&gt;
 Installing MPE2 include files to /usr/local/include&lt;br /&gt;
 Installing MPE2 libraries to /usr/local/lib&lt;br /&gt;
 Installing MPE2 utility programs to /usr/local/bin&lt;br /&gt;
 Installing MPE2 configuration files to /usr/local/etc&lt;br /&gt;
 Installing MPE2 system utility programs to /usr/local/sbin&lt;br /&gt;
 Installing MPE2 man to /usr/local/share/man&lt;br /&gt;
 Installing MPE2 html to /usr/local/share/doc/&lt;br /&gt;
 Installed MPE2 in /usr/local&lt;br /&gt;
&lt;br /&gt;
* hwloc (and lstopo) (download: http://www.open-mpi.org/software/hwloc/v1.2/)&lt;br /&gt;
compile from sources. To get xml support install libxml2-dev and pkg-config&lt;br /&gt;
 apt-get install libxml2-dev pkg-config&lt;br /&gt;
 tar -xzvf hwloc-1.1.1.tar.gz&lt;br /&gt;
 cd hwloc-1.1.1&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Cleanup &lt;br /&gt;
&lt;br /&gt;
 apt-get clean&lt;br /&gt;
 rm /etc/udev/rules.d/*-persistent-net.rules&lt;br /&gt;
&lt;br /&gt;
Make image &lt;br /&gt;
&lt;br /&gt;
 ssh root@'''node''' tgz-g5k &amp;amp;gt; $HOME/grid5000/'''imagename'''.tgz&lt;br /&gt;
&lt;br /&gt;
make appropriate .env file.&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=594</id>
		<title>Grid5000</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=594"/>
				<updated>2011-05-03T16:43:28Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: /* Setting up new deploy image */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;https://www.grid5000.fr/mediawiki/index.php/Grid5000:Home &lt;br /&gt;
&lt;br /&gt;
[https://www.grid5000.fr/mediawiki/index.php/Grid5000:UserCharter USAGE POLICY]&lt;br /&gt;
&lt;br /&gt;
== Login, job submission, deployment of image  ==&lt;br /&gt;
&lt;br /&gt;
*Select sites and clusters for experiments, using information on the [https://www.grid5000.fr/mediawiki/index.php/Grid5000:Network#Grid.275000_Sites Grid5000 network] and the [https://www.grid5000.fr/mediawiki/index.php/Status Status page] &lt;br /&gt;
*Access is provided via access nodes '''access.SITE.grid5000.fr''' marked [https://www.grid5000.fr/mediawiki/index.php/External_access here] as ''accessible from '''everywhere''' via ssh with '''keyboard-interactive''' authentication method''. As soon as you are on one of the sites, you can directly ssh frontend node of any other site:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
access_$ ssh frontend.SITE2&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*There is no access to Internet from computing nodes (external IPs should be registered on proxy), therefore, download/update your stuff at the access nodes. Several revision control clients are available. &lt;br /&gt;
*Each site has a separate NFS, therefore, to run an application on several sites at once, you need to copy it '''scp, sftp, rsync''' between access or frontend nodes. &lt;br /&gt;
*Jobs are run from the frondend nodes, using a [http://en.wikipedia.org/wiki/OpenPBS PBS]-like system [https://www.grid5000.fr/mediawiki/index.php/Cluster_experiment-OAR2 OAR]. Basic commands: &lt;br /&gt;
**'''oarstat''' - queue status &lt;br /&gt;
**'''oarsub''' - job submission &lt;br /&gt;
**'''oardel''' - job removal&lt;br /&gt;
&lt;br /&gt;
Interactive job on deployed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub -I -t deploy -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; Batch job on installed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub BATCH_FILE -t allow_classic_ssh -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*The image to deploy can be created and loaded with help of a [http://wiki.systemimager.org/index.php/Main_Page Systemimager]-like system [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2 Kadeploy]. Creating: [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2#Tune_an_environment_to_build_another_one:_customize_authentification_parameters described here]&lt;br /&gt;
&lt;br /&gt;
Loading: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ kadeploy3 -a PATH_TO_PRIVATE_IMAGE_DESC -f $OAR_FILE_NODES  &lt;br /&gt;
&amp;lt;/source&amp;gt; A Linux distribution lenny-x64-nfs-2.1 with mc, subversion, autotools, doxygen, MPICH2, GSL, Boost, R, gnuplot, graphviz, X11, evince is available at Orsay /home/nancy/alastovetsky/grid5000. &lt;br /&gt;
&lt;br /&gt;
== Compiling and running MPI applications  ==&lt;br /&gt;
&lt;br /&gt;
*Compilation should be done on one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`) &lt;br /&gt;
*Running MPI applications is described [https://www.grid5000.fr/mediawiki/index.php/Run_MPI_On_Grid%275000 here] &lt;br /&gt;
**mpirun/mpiexec should be run from one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`)&lt;br /&gt;
&lt;br /&gt;
== Setting up new deploy image  ==&lt;br /&gt;
&lt;br /&gt;
 oarsub -I -t deploy -l nodes=1,walltime=12&lt;br /&gt;
 kadeploy3 -e lenny-x64-nfs -f $OAR_FILE_NODES -k&lt;br /&gt;
 ssh root@`head -n 1 $OAR_NODEFILE`&lt;br /&gt;
&lt;br /&gt;
edit /etc/apt/sources.list &lt;br /&gt;
&lt;br /&gt;
 apt-get update&lt;br /&gt;
 apt-get upgrade&lt;br /&gt;
&lt;br /&gt;
 apt-get install libtool autoconf automake mc colorgcc ctags libboost-serialization-dev libboost-graph-dev libatlas-base-dev gfortran vim gdb valgrind screen subversion&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Compiled for sources by us: &lt;br /&gt;
&lt;br /&gt;
* gsl-1.14 (download: ftp://ftp.gnu.org/gnu/gsl/)&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
* mpich2 (download: http://www.mcs.anl.gov/research/projects/mpich2/downloads/index.php?s=downloads)&lt;br /&gt;
 ./configure --enable-shared --enable-sharedlibs=gcc&lt;br /&gt;
 make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Mpich2 installed to:&lt;br /&gt;
 Installing MPE2 include files to /usr/local/include&lt;br /&gt;
 Installing MPE2 libraries to /usr/local/lib&lt;br /&gt;
 Installing MPE2 utility programs to /usr/local/bin&lt;br /&gt;
 Installing MPE2 configuration files to /usr/local/etc&lt;br /&gt;
 Installing MPE2 system utility programs to /usr/local/sbin&lt;br /&gt;
 Installing MPE2 man to /usr/local/share/man&lt;br /&gt;
 Installing MPE2 html to /usr/local/share/doc/&lt;br /&gt;
 Installed MPE2 in /usr/local&lt;br /&gt;
&lt;br /&gt;
* hwloc (and lstopo) (download: http://www.open-mpi.org/software/hwloc/v1.2/)&lt;br /&gt;
compile from sources. To get xml support install libxml2-dev and pkg-config&lt;br /&gt;
 apt-get install libxml2-dev and pkg-config&lt;br /&gt;
 tar -xzvf hwloc-1.1.1.tar.gz&lt;br /&gt;
 cd hwloc-1.1.1&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Cleanup &lt;br /&gt;
&lt;br /&gt;
 apt-get clean&lt;br /&gt;
 rm /etc/udev/rules.d/*-persistent-net.rules&lt;br /&gt;
&lt;br /&gt;
Make image &lt;br /&gt;
&lt;br /&gt;
 ssh root@'''node''' tgz-g5k &amp;amp;gt; $HOME/grid5000/'''imagename'''.tgz&lt;br /&gt;
&lt;br /&gt;
make appropriate .env file.&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	<entry>
		<id>https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=593</id>
		<title>Grid5000</title>
		<link rel="alternate" type="text/html" href="https://hcl.ucd.ie/wiki/index.php?title=Grid5000&amp;diff=593"/>
				<updated>2011-05-03T16:39:38Z</updated>
		
		<summary type="html">&lt;p&gt;Zhongziming: /* Setting up new deploy image */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;https://www.grid5000.fr/mediawiki/index.php/Grid5000:Home &lt;br /&gt;
&lt;br /&gt;
[https://www.grid5000.fr/mediawiki/index.php/Grid5000:UserCharter USAGE POLICY]&lt;br /&gt;
&lt;br /&gt;
== Login, job submission, deployment of image  ==&lt;br /&gt;
&lt;br /&gt;
*Select sites and clusters for experiments, using information on the [https://www.grid5000.fr/mediawiki/index.php/Grid5000:Network#Grid.275000_Sites Grid5000 network] and the [https://www.grid5000.fr/mediawiki/index.php/Status Status page] &lt;br /&gt;
*Access is provided via access nodes '''access.SITE.grid5000.fr''' marked [https://www.grid5000.fr/mediawiki/index.php/External_access here] as ''accessible from '''everywhere''' via ssh with '''keyboard-interactive''' authentication method''. As soon as you are on one of the sites, you can directly ssh frontend node of any other site:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
access_$ ssh frontend.SITE2&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*There is no access to Internet from computing nodes (external IPs should be registered on proxy), therefore, download/update your stuff at the access nodes. Several revision control clients are available. &lt;br /&gt;
*Each site has a separate NFS, therefore, to run an application on several sites at once, you need to copy it '''scp, sftp, rsync''' between access or frontend nodes. &lt;br /&gt;
*Jobs are run from the frondend nodes, using a [http://en.wikipedia.org/wiki/OpenPBS PBS]-like system [https://www.grid5000.fr/mediawiki/index.php/Cluster_experiment-OAR2 OAR]. Basic commands: &lt;br /&gt;
**'''oarstat''' - queue status &lt;br /&gt;
**'''oarsub''' - job submission &lt;br /&gt;
**'''oardel''' - job removal&lt;br /&gt;
&lt;br /&gt;
Interactive job on deployed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub -I -t deploy -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; Batch job on installed images: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ oarsub BATCH_FILE -t allow_classic_ssh -l [/cluster=N/]nodes=N,walltime=HH[:MM[:SS]] [-p 'PROPERTY=&amp;quot;VALUE&amp;quot;']&lt;br /&gt;
&amp;lt;/source&amp;gt; &lt;br /&gt;
&lt;br /&gt;
*The image to deploy can be created and loaded with help of a [http://wiki.systemimager.org/index.php/Main_Page Systemimager]-like system [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2 Kadeploy]. Creating: [https://www.grid5000.fr/mediawiki/index.php/Deploy_environment-OAR2#Tune_an_environment_to_build_another_one:_customize_authentification_parameters described here]&lt;br /&gt;
&lt;br /&gt;
Loading: &amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
fontend_$ kadeploy3 -a PATH_TO_PRIVATE_IMAGE_DESC -f $OAR_FILE_NODES  &lt;br /&gt;
&amp;lt;/source&amp;gt; A Linux distribution lenny-x64-nfs-2.1 with mc, subversion, autotools, doxygen, MPICH2, GSL, Boost, R, gnuplot, graphviz, X11, evince is available at Orsay /home/nancy/alastovetsky/grid5000. &lt;br /&gt;
&lt;br /&gt;
== Compiling and running MPI applications  ==&lt;br /&gt;
&lt;br /&gt;
*Compilation should be done on one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`) &lt;br /&gt;
*Running MPI applications is described [https://www.grid5000.fr/mediawiki/index.php/Run_MPI_On_Grid%275000 here] &lt;br /&gt;
**mpirun/mpiexec should be run from one of the reserved nodes (e.g. ssh `head -n 1 $OAR_NODEFILE`)&lt;br /&gt;
&lt;br /&gt;
== Setting up new deploy image  ==&lt;br /&gt;
&lt;br /&gt;
 oarsub -I -t deploy -l nodes=1,walltime=12&lt;br /&gt;
 kadeploy3 -e lenny-x64-nfs -f $OAR_FILE_NODES -k&lt;br /&gt;
 ssh root@`head -n 1 $OAR_NODEFILE`&lt;br /&gt;
&lt;br /&gt;
edit /etc/apt/sources.list &lt;br /&gt;
&lt;br /&gt;
 apt-get update&lt;br /&gt;
 apt-get upgrade&lt;br /&gt;
&lt;br /&gt;
 apt-get install libtool autoconf automake mc colorgcc ctags libboost-serialization-dev libboost-graph-dev libatlas-base-dev gfortran vim gdb valgrind screen subversion&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Compiled for sources by us: &lt;br /&gt;
&lt;br /&gt;
* gsl-1.14 (download: ftp://ftp.gnu.org/gnu/gsl/)&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
* mpich2 (download: http://www.mcs.anl.gov/research/projects/mpich2/downloads/index.php?s=downloads)&lt;br /&gt;
 ./configure --enable-shared --enable-sharedlibs=gcc&lt;br /&gt;
 make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Mpich2 installed to:&lt;br /&gt;
 Installing MPE2 include files to /usr/local/include&lt;br /&gt;
 Installing MPE2 libraries to /usr/local/lib&lt;br /&gt;
 Installing MPE2 utility programs to /usr/local/bin&lt;br /&gt;
 Installing MPE2 configuration files to /usr/local/etc&lt;br /&gt;
 Installing MPE2 system utility programs to /usr/local/sbin&lt;br /&gt;
 Installing MPE2 man to /usr/local/share/man&lt;br /&gt;
 Installing MPE2 html to /usr/local/share/doc/&lt;br /&gt;
 Installed MPE2 in /usr/local&lt;br /&gt;
&lt;br /&gt;
* hwloc (and lstopo)&lt;br /&gt;
compile from sources. To get xml support install libxml2-dev and pkg-config&lt;br /&gt;
 apt-get install libxml2-dev and pkg-config&lt;br /&gt;
 tar -xzvf hwloc-1.1.1.tar.gz&lt;br /&gt;
 cd hwloc-1.1.1&lt;br /&gt;
 ./configure &amp;amp;&amp;amp; make &amp;amp;&amp;amp; make install&lt;br /&gt;
&lt;br /&gt;
Cleanup &lt;br /&gt;
&lt;br /&gt;
 apt-get clean&lt;br /&gt;
 rm /etc/udev/rules.d/*-persistent-net.rules&lt;br /&gt;
&lt;br /&gt;
Make image &lt;br /&gt;
&lt;br /&gt;
 ssh root@'''node''' tgz-g5k &amp;amp;gt; $HOME/grid5000/'''imagename'''.tgz&lt;br /&gt;
&lt;br /&gt;
make appropriate .env file.&lt;/div&gt;</summary>
		<author><name>Zhongziming</name></author>	</entry>

	</feed>