<html>
<head>
<meta http-equiv="Content-Type" content="text/html; charset=Windows-1252">
<meta name="Generator" content="Microsoft Exchange Server">
<!-- converted from text --><style><!-- .EmailQuote { margin-left: 1pt; padding-left: 4pt; border-left: #800000 2px solid; } --></style>
</head>
<body>
<style>
<!--
@font-face
{font-family:"Cambria Math"}
@font-face
{font-family:Calibri}
p.x_MsoNormal, li.x_MsoNormal, div.x_MsoNormal
{margin:0in;
margin-bottom:.0001pt;
font-size:11.0pt;
font-family:"Calibri",sans-serif}
a:link, span.x_MsoHyperlink
{color:blue;
text-decoration:underline}
a:visited, span.x_MsoHyperlinkFollowed
{color:#954F72;
text-decoration:underline}
.x_MsoChpDefault
{}
@page WordSection1
{margin:70.85pt 56.7pt 70.85pt 56.7pt}
div.x_WordSection1
{}
-->
</style>
<div lang="EN-US" link="blue" vlink="#954F72">
<div class="x_WordSection1">
<p class="x_MsoNormal">Hi,</p>
<p class="x_MsoNormal"> </p>
<p class="x_MsoNormal">I did my OpenNMT-py experiments on both Abel and Taito.</p>
<p class="x_MsoNormal">On Taito, I got training speeds of about 13000 tokens/s, on Abel it was about 4000 tokens/s.</p>
<p class="x_MsoNormal">A colleague who used an independent OpenNMT-py module on Taito-GPU during the summer obtained about 9000 tokens/s with a different dataset.</p>
<p class="x_MsoNormal">I also just started a CPU-only training run on Taito, which got around 1000 tokens/s.</p>
<p class="x_MsoNormal">This leads me to believe that my experiments – at least those on Taito – did use the GPU…</p>
<p class="x_MsoNormal"> </p>
<p class="x_MsoNormal">Best,</p>
<p class="x_MsoNormal">Yves</p>
<p class="x_MsoNormal"> </p>
</div>
<hr tabindex="-1" style="display:inline-block; width:98%">
<div id="x_divRplyFwdMsg" dir="ltr"><font face="Calibri, sans-serif" color="#000000" style="font-size:11pt"><b>From:</b> Stephan Oepen <oe@ifi.uio.no><br>
<b>Sent:</b> Wednesday, November 28, 2018 4:08:46 PM<br>
<b>To:</b> Scherrer, Yves<br>
<b>Cc:</b> Martin Matthiesen; infrastructure<br>
<b>Subject:</b> Re: [NLPL Task Force (A)] OpenNMT installation for NLPL (on Abel)</font>
<div> </div>
</div>
</div>
<font size="2"><span style="font-size:11pt;">
<div class="PlainText">as for the OpenNMT-py experiments, did you do those on Abel or Taito,<br>
or both? using gpus on Taito? in other words, do you believe that<br>
OpenNMT-py (in contrast to PyTorch) works on Taito gpu nodes?<br>
<br>
oe<br>
<br>
On Wed, Nov 28, 2018 at 2:47 PM Scherrer, Yves<br>
<yves.scherrer@helsinki.fi> wrote:<br>
><br>
> Hi,<br>
><br>
><br>
><br>
> I’m following up on this one with a related issue. I am testing PyTorch independently of OpenNMT-py, but cannot get it to run on (Taito-)GPU.<br>
><br>
><br>
><br>
> Specifically, although I was logged in to Taito-GPU, I cannot get the test script described on the Wiki page to return True:<br>
><br>
><br>
><br>
> [GPU-Env lstmtagger]$ srun -n 1 -p gputest --gres=gpu:k80:1 --mem 1G -t 15 --pty python3 /proj/nlpl/software/pytorch/0.4.1/test.py<br>
><br>
> srun: job 32089470 queued and waiting for resources<br>
><br>
> srun: job 32089470 has been allocated resources<br>
><br>
> False<br>
><br>
><br>
><br>
> I also get ‘False’ when running the following script through sbatch:<br>
><br>
><br>
><br>
> #SBATCH -J cudatest<br>
><br>
> #SBATCH -o cudatest.%j.out<br>
><br>
> #SBATCH -e cudatest.%j.err<br>
><br>
> #SBATCH -t 0:05:00<br>
><br>
> #SBATCH -p gputest<br>
><br>
> #SBATCH -N 1<br>
><br>
> #SBATCH --gres=gpu:k80:1<br>
><br>
> #SBATCH --mem=1g<br>
><br>
> module use -a /proj/nlpl/software/modulefiles/<br>
><br>
> module load nlpl-pytorch<br>
><br>
> srun python3 /proj/nlpl/software/pytorch/0.4.1/test.py<br>
><br>
><br>
><br>
> Has there been any change lately? Or am I missing something obvious?<br>
><br>
><br>
><br>
> Best,<br>
><br>
> Yves<br>
><br>
><br>
><br>
><br>
><br>
> ________________________________<br>
> From: Stephan Oepen <oe@ifi.uio.no><br>
> Sent: Wednesday, September 26, 2018 11:10:12 PM<br>
> To: Scherrer, Yves<br>
> Cc: Martin Matthiesen; infrastructure<br>
> Subject: Re: [NLPL Task Force (A)] OpenNMT installation for NLPL (on Abel)<br>
><br>
> hi again,<br>
><br>
> > i actually had a go at my own glibc and PyTorch installations on Taito, but<br>
> > so far gpu support is evasive.<br>
><br>
> actually, with a little more tinkering, i now believe i might have a<br>
> working installation of PyTorch 0.4.1 and OpenNMT-py 0.2.1 on Taito<br>
> too, seemingly functional on both cpu and gpu nodes:<br>
><br>
> [oe@taito-login4 ~]$ module purge<br>
> [oe@taito-login4 ~]$ module load nlpl-opennmt-py<br>
> Loading application python-3.5.3 environment with needed modules<br>
> [oe@taito-login4 ~]$ module list<br>
><br>
> Currently Loaded Modules:<br>
> 1) gcc/5.4.0 2) intelmpi/5.1.3 3) mkl/11.3.2 4) python/3.5.3<br>
> 5) python-env/3.5.3 6) nlpl-pytorch/0.4.1 7) nlpl-opennmt-py/0.2.1<br>
><br>
> [oe@taito-login4 ~]$ type -all python<br>
> python is /proj/nlpl/software/opennmt-py/0.2.1/bin/python<br>
> python is /proj/nlpl/software/pytorch/0.4.1/bin/python<br>
> python is /appl/opt/python/3.5.3-gnu540/bin/python<br>
> python is /usr/bin/python<br>
> [oe@taito-login4 ~]$ python -c "import torch; import onmt;<br>
> print(torch.cuda.is_available());"<br>
> False<br>
><br>
> [oe@taito-login4 ~]$ srun -n 1 -p gputest --gres=gpu:k80:1 --mem 1G -t<br>
> 15 --pty \<br>
> python -c "import torch; import onmt; print(torch.cuda.is_available());"<br>
> True<br>
><br>
> —yves (or joerg), i would have a hard time testing things in much more<br>
> depth. any chance you would have some time to try and replicate the<br>
> validation steps your are currently running on Abel on Taito too?<br>
><br>
> with a sense of accomplishment :-), oe<br>
</div>
</span></font>
</body>
</html>