提交 d477a218 编写于 作者: P PaParaZz1

Deploying to gh-pages from @ bc0102ba 🚀

上级 c08bffdf
......@@ -49,13 +49,13 @@ For DP, the runnable script demo is demonstrated as follows.
.. code-block:: bash
CUDA_VISIBLE_DEVICES=0,1 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py
CUDA_VISIBLE_DEVICES=0,1 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
or (on cluster managed by Slurm)
.. code-block:: bash
srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n1 --ntasks-per-node=1 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py
srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n1 --ntasks-per-node=1 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
......@@ -129,7 +129,7 @@ To enable DDP training in DI-engine existing codes, you just need to add modific
serial_pipeline(main_config, create_config)
.. tip::
The whole example is located in ``dizoo/atari/entry/spaceinvaders_dqn_main_multi_gpu_ddp.py``
The whole example is located in ``dizoo/atari/entry/spaceinvaders_dqn_config_multi_gpu_ddp.py``
3. Execute launch shell script
......@@ -137,10 +137,10 @@ For DDP, the runnable script demo is demonstrated as follows.
.. code-block:: bash
CUDA_VISIBLE_DEVICES=0,1 python -m torch.distributed.launch --nnodes=1 --node_rank=0 --nproc_per_node=2 spaceinvaders_dqn_main_multi_gpu_ddp.py
CUDA_VISIBLE_DEVICES=0,1 python -m torch.distributed.launch --nnodes=1 --node_rank=0 --nproc_per_node=2 spaceinvaders_dqn_config_multi_gpu_ddp.py
Or on cluster managed by Slurm
.. code-block:: bash
srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n2 --ntasks-per-node=2 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py
srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n2 --ntasks-per-node=2 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
......@@ -237,11 +237,11 @@ and then the parameters are synchronized with other GPUs.</p>
</div>
<p>We don’t need to change any other code, just simply encapsulate the policy. Please refer to <code class="docutils literal notranslate"><span class="pre">dizoo/atari/config/serial/spaceinvaders/spaceinvaders_dqn_config_multi_gpu_dp.py</span></code></p>
<p>For DP, the runnable script demo is demonstrated as follows.</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div>
</div>
<p>or (on cluster managed by Slurm)</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n1 --ntasks-per-node<span class="o">=</span><span class="m">1</span> python -u spaceinvaders_dqn_main_multi_gpu_ddp.py
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n1 --ntasks-per-node<span class="o">=</span><span class="m">1</span> python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div>
</div>
</div>
......@@ -306,17 +306,17 @@ DI-engine achieves this with allreduce operator in learner and evaluator, and on
</div>
<div class="admonition tip">
<p class="admonition-title">Tip</p>
<p>The whole example is located in <code class="docutils literal notranslate"><span class="pre">dizoo/atari/entry/spaceinvaders_dqn_main_multi_gpu_ddp.py</span></code></p>
<p>The whole example is located in <code class="docutils literal notranslate"><span class="pre">dizoo/atari/entry/spaceinvaders_dqn_config_multi_gpu_ddp.py</span></code></p>
</div>
<ol class="arabic simple" start="3">
<li><p>Execute launch shell script</p></li>
</ol>
<p>For DDP, the runnable script demo is demonstrated as follows.</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -m torch.distributed.launch --nnodes<span class="o">=</span><span class="m">1</span> --node_rank<span class="o">=</span><span class="m">0</span> --nproc_per_node<span class="o">=</span><span class="m">2</span> spaceinvaders_dqn_main_multi_gpu_ddp.py
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -m torch.distributed.launch --nnodes<span class="o">=</span><span class="m">1</span> --node_rank<span class="o">=</span><span class="m">0</span> --nproc_per_node<span class="o">=</span><span class="m">2</span> spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div>
</div>
<p>Or on cluster managed by Slurm</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n2 --ntasks-per-node<span class="o">=</span><span class="m">2</span> python -u spaceinvaders_dqn_main_multi_gpu_ddp.py
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n2 --ntasks-per-node<span class="o">=</span><span class="m">2</span> python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div>
</div>
</div>
......
此差异已折叠。
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册