Skip to content

Commit c60646e

Browse files
author
Jessica Lin
authored
Merge pull request pytorch#458 from pytorch/site-v1.6.0
v1.6.0 docs redo for stable (0.7) vision
2 parents 2e018f0 + c551fb9 commit c60646e

File tree

327 files changed

+5668
-6177
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

327 files changed

+5668
-6177
lines changed

docs/stable/.buildinfo

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
# Sphinx build info version 1
22
# This file hashes the configuration used when building these files. When it is not found, a full rebuild will be done.
3-
config: 93625c989914b71802289037a0f16437
3+
config: 892a2ec27a03fe01b7f360c8920a2882
44
tags: 645f666f9bcd5a90fca523b33c5a78b7

docs/stable/_images/add_histogram.png

-47 KB
Binary file not shown.

docs/stable/_images/add_hparam.png

-62.6 KB
Binary file not shown.

docs/stable/_images/add_image.png

-46 KB
Binary file not shown.

docs/stable/_images/add_images.png

-74.9 KB
Binary file not shown.

docs/stable/_images/add_scalar.png

-44.9 KB
Binary file not shown.

docs/stable/_images/add_scalars.png

-96.8 KB
Binary file not shown.

docs/stable/_modules/index.html

-1
Original file line numberDiff line numberDiff line change
@@ -494,7 +494,6 @@ <h1>All modules for which code is available</h1>
494494
<li><a href="torch/utils/data/distributed.html">torch.utils.data.distributed</a></li>
495495
<li><a href="torch/utils/data/sampler.html">torch.utils.data.sampler</a></li>
496496
<li><a href="torch/utils/mobile_optimizer.html">torch.utils.mobile_optimizer</a></li>
497-
<li><a href="torch/utils/tensorboard/writer.html">torch.utils.tensorboard.writer</a></li>
498497
</ul><li><a href="torchvision.html">torchvision</a></li>
499498
<ul><li><a href="torchvision/datasets/celeba.html">torchvision.datasets.celeba</a></li>
500499
<li><a href="torchvision/datasets/cifar.html">torchvision.datasets.cifar</a></li>

docs/stable/_modules/torch.html

+2-2
Original file line numberDiff line numberDiff line change
@@ -838,9 +838,9 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
838838
<span class="k">del</span> <span class="n">_torch_docs</span><span class="p">,</span> <span class="n">_tensor_docs</span><span class="p">,</span> <span class="n">_storage_docs</span>
839839

840840

841-
<div class="viewcode-block" id="compiled_with_cxx11_abi"><a class="viewcode-back" href="../generated/torch.compiled_with_cxx11_abi.html#torch.compiled_with_cxx11_abi">[docs]</a><span class="k">def</span> <span class="nf">compiled_with_cxx11_abi</span><span class="p">():</span>
841+
<span class="k">def</span> <span class="nf">compiled_with_cxx11_abi</span><span class="p">():</span>
842842
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Returns whether PyTorch was built with _GLIBCXX_USE_CXX11_ABI=1&quot;&quot;&quot;</span>
843-
<span class="k">return</span> <span class="n">_C</span><span class="o">.</span><span class="n">_GLIBCXX_USE_CXX11_ABI</span></div>
843+
<span class="k">return</span> <span class="n">_C</span><span class="o">.</span><span class="n">_GLIBCXX_USE_CXX11_ABI</span>
844844

845845

846846
<span class="c1"># Import the ops &quot;namespace&quot;</span>

docs/stable/_modules/torch/_jit_internal.html

+2-2
Original file line numberDiff line numberDiff line change
@@ -710,7 +710,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
710710
<span class="n">fn</span><span class="o">.</span><span class="n">_torchscript_modifier</span> <span class="o">=</span> <span class="n">FunctionModifiers</span><span class="o">.</span><span class="n">UNUSED</span>
711711
<span class="k">return</span> <span class="n">fn</span></div>
712712

713-
<span class="k">def</span> <span class="nf">ignore</span><span class="p">(</span><span class="n">drop</span><span class="o">=</span><span class="kc">False</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
713+
<div class="viewcode-block" id="ignore"><a class="viewcode-back" href="../../generated/torch.jit.ignore.html#torch.jit.ignore">[docs]</a><span class="k">def</span> <span class="nf">ignore</span><span class="p">(</span><span class="n">drop</span><span class="o">=</span><span class="kc">False</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
714714
<span class="sd">&quot;&quot;&quot;</span>
715715
<span class="sd"> This decorator indicates to the compiler that a function or method should</span>
716716
<span class="sd"> be ignored and left as a Python function. This allows you to leave code in</span>
@@ -801,7 +801,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
801801
<span class="k">else</span><span class="p">:</span>
802802
<span class="n">fn</span><span class="o">.</span><span class="n">_torchscript_modifier</span> <span class="o">=</span> <span class="n">FunctionModifiers</span><span class="o">.</span><span class="n">IGNORE</span>
803803
<span class="k">return</span> <span class="n">fn</span>
804-
<span class="k">return</span> <span class="n">decorator</span>
804+
<span class="k">return</span> <span class="n">decorator</span></div>
805805

806806

807807
<span class="k">def</span> <span class="nf">_copy_to_script_wrapper</span><span class="p">(</span><span class="n">fn</span><span class="p">):</span>

docs/stable/_modules/torch/_lowrank.html

+2-2
Original file line numberDiff line numberDiff line change
@@ -419,7 +419,7 @@ <h1>Source code for torch._lowrank</h1><div class="highlight"><pre>
419419
<span class="k">return</span> <span class="n">Q</span>
420420

421421

422-
<span class="k">def</span> <span class="nf">svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="mi">6</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="mi">2</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="kc">None</span><span class="p">):</span>
422+
<div class="viewcode-block" id="svd_lowrank"><a class="viewcode-back" href="../../generated/torch.svd_lowrank.html#torch.svd_lowrank">[docs]</a><span class="k">def</span> <span class="nf">svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="mi">6</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="mi">2</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="kc">None</span><span class="p">):</span>
423423
<span class="c1"># type: (Tensor, Optional[int], Optional[int], Optional[Tensor]) -&gt; Tuple[Tensor, Tensor, Tensor]</span>
424424
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Return the singular value decomposition ``(U, S, V)`` of a matrix,</span>
425425
<span class="sd"> batches of matrices, or a sparse matrix :math:`A` such that</span>
@@ -464,7 +464,7 @@ <h1>Source code for torch._lowrank</h1><div class="highlight"><pre>
464464
<span class="n">tensor_ops</span> <span class="o">=</span> <span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">M</span><span class="p">)</span>
465465
<span class="k">if</span> <span class="p">(</span><span class="ow">not</span> <span class="nb">set</span><span class="p">(</span><span class="nb">map</span><span class="p">(</span><span class="nb">type</span><span class="p">,</span> <span class="n">tensor_ops</span><span class="p">))</span><span class="o">.</span><span class="n">issubset</span><span class="p">((</span><span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">,</span> <span class="nb">type</span><span class="p">(</span><span class="kc">None</span><span class="p">)))</span> <span class="ow">and</span> <span class="n">has_torch_function</span><span class="p">(</span><span class="n">tensor_ops</span><span class="p">)):</span>
466466
<span class="k">return</span> <span class="n">handle_torch_function</span><span class="p">(</span><span class="n">svd_lowrank</span><span class="p">,</span> <span class="n">tensor_ops</span><span class="p">,</span> <span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="n">M</span><span class="p">)</span>
467-
<span class="k">return</span> <span class="n">_svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="n">M</span><span class="p">)</span>
467+
<span class="k">return</span> <span class="n">_svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="n">M</span><span class="p">)</span></div>
468468

469469

470470
<span class="k">def</span> <span class="nf">_svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="mi">6</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="mi">2</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="kc">None</span><span class="p">):</span>

docs/stable/_modules/torch/autograd/grad_mode.html

+5-5
Original file line numberDiff line numberDiff line change
@@ -369,7 +369,7 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
369369
<span class="k">return</span> <span class="n">generator_context</span>
370370

371371

372-
<span class="k">class</span> <span class="nc">no_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
372+
<div class="viewcode-block" id="no_grad"><a class="viewcode-back" href="../../../autograd.html#torch.autograd.no_grad">[docs]</a><span class="k">class</span> <span class="nc">no_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
373373
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Context-manager that disabled gradient calculation.</span>
374374

375375
<span class="sd"> Disabling gradient calculation is useful for inference, when you are sure</span>
@@ -406,10 +406,10 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
406406
<span class="n">torch</span><span class="o">.</span><span class="n">_C</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="kc">False</span><span class="p">)</span>
407407

408408
<span class="k">def</span> <span class="fm">__exit__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="o">*</span><span class="n">args</span><span class="p">):</span>
409-
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span>
409+
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span></div>
410410

411411

412-
<div class="viewcode-block" id="enable_grad"><a class="viewcode-back" href="../../../generated/torch.enable_grad.html#torch.enable_grad">[docs]</a><span class="k">class</span> <span class="nc">enable_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
412+
<div class="viewcode-block" id="enable_grad"><a class="viewcode-back" href="../../../autograd.html#torch.autograd.enable_grad">[docs]</a><span class="k">class</span> <span class="nc">enable_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
413413
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Context-manager that enables gradient calculation.</span>
414414

415415
<span class="sd"> Enables gradient calculation, if it has been disabled via :class:`~no_grad`</span>
@@ -448,7 +448,7 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
448448
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span></div>
449449

450450

451-
<span class="k">class</span> <span class="nc">set_grad_enabled</span><span class="p">(</span><span class="nb">object</span><span class="p">):</span>
451+
<div class="viewcode-block" id="set_grad_enabled"><a class="viewcode-back" href="../../../autograd.html#torch.autograd.set_grad_enabled">[docs]</a><span class="k">class</span> <span class="nc">set_grad_enabled</span><span class="p">(</span><span class="nb">object</span><span class="p">):</span>
452452
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Context-manager that sets gradient calculation to on or off.</span>
453453

454454
<span class="sd"> ``set_grad_enabled`` will enable or disable grads based on its argument :attr:`mode`.</span>
@@ -493,7 +493,7 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
493493
<span class="k">pass</span>
494494

495495
<span class="k">def</span> <span class="fm">__exit__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="o">*</span><span class="n">args</span><span class="p">):</span>
496-
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span>
496+
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span></div>
497497
</pre></div>
498498

499499
</article>

0 commit comments

Comments
 (0)