提交 d477a218 编写于 作者: P PaParaZz1

Deploying to gh-pages from @ bc0102ba 🚀

上级 c08bffdf
...@@ -49,13 +49,13 @@ For DP, the runnable script demo is demonstrated as follows. ...@@ -49,13 +49,13 @@ For DP, the runnable script demo is demonstrated as follows.
.. code-block:: bash .. code-block:: bash
CUDA_VISIBLE_DEVICES=0,1 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py CUDA_VISIBLE_DEVICES=0,1 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
or (on cluster managed by Slurm) or (on cluster managed by Slurm)
.. code-block:: bash .. code-block:: bash
srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n1 --ntasks-per-node=1 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n1 --ntasks-per-node=1 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
...@@ -129,7 +129,7 @@ To enable DDP training in DI-engine existing codes, you just need to add modific ...@@ -129,7 +129,7 @@ To enable DDP training in DI-engine existing codes, you just need to add modific
serial_pipeline(main_config, create_config) serial_pipeline(main_config, create_config)
.. tip:: .. tip::
The whole example is located in ``dizoo/atari/entry/spaceinvaders_dqn_main_multi_gpu_ddp.py`` The whole example is located in ``dizoo/atari/entry/spaceinvaders_dqn_config_multi_gpu_ddp.py``
3. Execute launch shell script 3. Execute launch shell script
...@@ -137,10 +137,10 @@ For DDP, the runnable script demo is demonstrated as follows. ...@@ -137,10 +137,10 @@ For DDP, the runnable script demo is demonstrated as follows.
.. code-block:: bash .. code-block:: bash
CUDA_VISIBLE_DEVICES=0,1 python -m torch.distributed.launch --nnodes=1 --node_rank=0 --nproc_per_node=2 spaceinvaders_dqn_main_multi_gpu_ddp.py CUDA_VISIBLE_DEVICES=0,1 python -m torch.distributed.launch --nnodes=1 --node_rank=0 --nproc_per_node=2 spaceinvaders_dqn_config_multi_gpu_ddp.py
Or on cluster managed by Slurm Or on cluster managed by Slurm
.. code-block:: bash .. code-block:: bash
srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n2 --ntasks-per-node=2 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py srun -p PARTITION_NAME --mpi=pmi2 --gres=gpu:2 -n2 --ntasks-per-node=2 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
...@@ -237,11 +237,11 @@ and then the parameters are synchronized with other GPUs.</p> ...@@ -237,11 +237,11 @@ and then the parameters are synchronized with other GPUs.</p>
</div> </div>
<p>We don’t need to change any other code, just simply encapsulate the policy. Please refer to <code class="docutils literal notranslate"><span class="pre">dizoo/atari/config/serial/spaceinvaders/spaceinvaders_dqn_config_multi_gpu_dp.py</span></code></p> <p>We don’t need to change any other code, just simply encapsulate the policy. Please refer to <code class="docutils literal notranslate"><span class="pre">dizoo/atari/config/serial/spaceinvaders/spaceinvaders_dqn_config_multi_gpu_dp.py</span></code></p>
<p>For DP, the runnable script demo is demonstrated as follows.</p> <p>For DP, the runnable script demo is demonstrated as follows.</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -u spaceinvaders_dqn_main_multi_gpu_ddp.py <div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div> </pre></div>
</div> </div>
<p>or (on cluster managed by Slurm)</p> <p>or (on cluster managed by Slurm)</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n1 --ntasks-per-node<span class="o">=</span><span class="m">1</span> python -u spaceinvaders_dqn_main_multi_gpu_ddp.py <div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n1 --ntasks-per-node<span class="o">=</span><span class="m">1</span> python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div> </pre></div>
</div> </div>
</div> </div>
...@@ -306,17 +306,17 @@ DI-engine achieves this with allreduce operator in learner and evaluator, and on ...@@ -306,17 +306,17 @@ DI-engine achieves this with allreduce operator in learner and evaluator, and on
</div> </div>
<div class="admonition tip"> <div class="admonition tip">
<p class="admonition-title">Tip</p> <p class="admonition-title">Tip</p>
<p>The whole example is located in <code class="docutils literal notranslate"><span class="pre">dizoo/atari/entry/spaceinvaders_dqn_main_multi_gpu_ddp.py</span></code></p> <p>The whole example is located in <code class="docutils literal notranslate"><span class="pre">dizoo/atari/entry/spaceinvaders_dqn_config_multi_gpu_ddp.py</span></code></p>
</div> </div>
<ol class="arabic simple" start="3"> <ol class="arabic simple" start="3">
<li><p>Execute launch shell script</p></li> <li><p>Execute launch shell script</p></li>
</ol> </ol>
<p>For DDP, the runnable script demo is demonstrated as follows.</p> <p>For DDP, the runnable script demo is demonstrated as follows.</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -m torch.distributed.launch --nnodes<span class="o">=</span><span class="m">1</span> --node_rank<span class="o">=</span><span class="m">0</span> --nproc_per_node<span class="o">=</span><span class="m">2</span> spaceinvaders_dqn_main_multi_gpu_ddp.py <div class="highlight-bash notranslate"><div class="highlight"><pre><span></span><span class="nv">CUDA_VISIBLE_DEVICES</span><span class="o">=</span><span class="m">0</span>,1 python -m torch.distributed.launch --nnodes<span class="o">=</span><span class="m">1</span> --node_rank<span class="o">=</span><span class="m">0</span> --nproc_per_node<span class="o">=</span><span class="m">2</span> spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div> </pre></div>
</div> </div>
<p>Or on cluster managed by Slurm</p> <p>Or on cluster managed by Slurm</p>
<div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n2 --ntasks-per-node<span class="o">=</span><span class="m">2</span> python -u spaceinvaders_dqn_main_multi_gpu_ddp.py <div class="highlight-bash notranslate"><div class="highlight"><pre><span></span>srun -p PARTITION_NAME --mpi<span class="o">=</span>pmi2 --gres<span class="o">=</span>gpu:2 -n2 --ntasks-per-node<span class="o">=</span><span class="m">2</span> python -u spaceinvaders_dqn_config_multi_gpu_ddp.py
</pre></div> </pre></div>
</div> </div>
</div> </div>
......
此差异已折叠。
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册