Skip to content

Commit d121c4e

Browse files
committed
Generate Python docs from pytorch/pytorch@61a5c77
1 parent 47f18c8 commit d121c4e

File tree

1,979 files changed

+3896
-2198
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,979 files changed

+3896
-2198
lines changed

docs/master/_images/RReLU.png

44 Bytes
Loading

docs/master/_modules/index.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/__config__.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/_jit_internal.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/_lobpcg.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/_lowrank.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/_tensor.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/_tensor_str.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/_vmap_internals.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/amp/autocast_mode.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/ns/_numeric_suite.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/ns/_numeric_suite_fx.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/ns/fx/utils.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/quantization/fake_quantize.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/quantization/fuse_modules.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/quantization/observer.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/quantization/qconfig.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/quantization/quantize.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/quantization/quantize_fx.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/ao/quantization/stubs.html

+1-1
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

docs/master/_modules/torch/autograd.html

+26-8
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@
219219
<div class="pytorch-left-menu-search">
220220

221221
<div class="version">
222-
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+gite10b762 ) &#x25BC</a>
222+
<a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git61a5c77 ) &#x25BC</a>
223223
</div>
224224

225225

@@ -438,7 +438,7 @@ <h1>Source code for torch.autograd</h1><div class="highlight"><pre>
438438
<span class="kn">import</span> <span class="nn">torch</span>
439439
<span class="kn">import</span> <span class="nn">warnings</span>
440440

441-
<span class="kn">from</span> <span class="nn">torch.types</span> <span class="kn">import</span> <span class="n">_TensorOrTensors</span>
441+
<span class="kn">from</span> <span class="nn">torch.types</span> <span class="kn">import</span> <span class="n">_TensorOrTensors</span><span class="p">,</span> <span class="n">_size</span>
442442
<span class="kn">from</span> <span class="nn">typing</span> <span class="kn">import</span> <span class="n">Any</span><span class="p">,</span> <span class="n">Callable</span><span class="p">,</span> <span class="n">List</span><span class="p">,</span> <span class="n">Optional</span><span class="p">,</span> <span class="n">Sequence</span><span class="p">,</span> <span class="n">Tuple</span><span class="p">,</span> <span class="n">Union</span><span class="p">,</span> <span class="n">cast</span>
443443

444444
<span class="kn">from</span> <span class="nn">.variable</span> <span class="kn">import</span> <span class="n">Variable</span>
@@ -455,32 +455,50 @@ <h1>Source code for torch.autograd</h1><div class="highlight"><pre>
455455
<span class="n">__all__</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;Variable&#39;</span><span class="p">,</span> <span class="s1">&#39;Function&#39;</span><span class="p">,</span> <span class="s1">&#39;backward&#39;</span><span class="p">,</span> <span class="s1">&#39;grad_mode&#39;</span><span class="p">]</span>
456456

457457
<span class="n">_OptionalTensor</span> <span class="o">=</span> <span class="n">Optional</span><span class="p">[</span><span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">]</span>
458+
<span class="n">_ShapeorNestedShape</span> <span class="o">=</span> <span class="n">Union</span><span class="p">[</span><span class="n">_size</span><span class="p">,</span> <span class="n">Sequence</span><span class="p">[</span><span class="n">_size</span><span class="p">],</span> <span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">]</span>
459+
460+
461+
<span class="k">def</span> <span class="nf">_calculate_shape</span><span class="p">(</span><span class="n">output</span><span class="p">:</span> <span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">,</span> <span class="n">grad</span><span class="p">:</span> <span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">,</span>
462+
<span class="n">is_grads_batched</span><span class="p">:</span> <span class="nb">bool</span><span class="p">)</span> <span class="o">-&gt;</span> <span class="n">Tuple</span><span class="p">[</span><span class="n">_ShapeorNestedShape</span><span class="p">,</span> <span class="n">_ShapeorNestedShape</span><span class="p">]:</span>
463+
<span class="c1"># is_same_size ensures that both tensors are either nested or non nested</span>
464+
<span class="k">if</span> <span class="n">output</span><span class="o">.</span><span class="n">is_nested</span><span class="p">:</span>
465+
<span class="k">if</span> <span class="n">is_grads_batched</span><span class="p">:</span>
466+
<span class="k">raise</span> <span class="ne">RuntimeError</span><span class="p">(</span><span class="s2">&quot;Batched grads are not supported with Nested Tensor.&quot;</span><span class="p">)</span>
467+
<span class="n">out_shape</span> <span class="o">=</span> <span class="n">output</span><span class="o">.</span><span class="n">_nested_tensor_size</span><span class="p">()</span>
468+
<span class="n">grad_shape</span> <span class="o">=</span> <span class="n">grad</span><span class="o">.</span><span class="n">_nested_tensor_size</span><span class="p">()</span>
469+
470+
<span class="k">return</span> <span class="n">out_shape</span><span class="p">,</span> <span class="n">grad_shape</span>
471+
472+
<span class="n">reg_out_shape</span> <span class="o">=</span> <span class="n">output</span><span class="o">.</span><span class="n">shape</span>
473+
<span class="n">reg_grad_shape</span> <span class="o">=</span> <span class="n">grad</span><span class="o">.</span><span class="n">shape</span> <span class="k">if</span> <span class="ow">not</span> <span class="n">is_grads_batched</span> <span class="k">else</span> <span class="n">grad</span><span class="o">.</span><span class="n">shape</span><span class="p">[</span><span class="mi">1</span><span class="p">:]</span>
474+
<span class="k">return</span> <span class="n">reg_out_shape</span><span class="p">,</span> <span class="n">reg_grad_shape</span>
458475

459476
<span class="k">def</span> <span class="nf">_make_grads</span><span class="p">(</span><span class="n">outputs</span><span class="p">:</span> <span class="n">Sequence</span><span class="p">[</span><span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">],</span> <span class="n">grads</span><span class="p">:</span> <span class="n">Sequence</span><span class="p">[</span><span class="n">_OptionalTensor</span><span class="p">],</span>
460477
<span class="n">is_grads_batched</span><span class="p">:</span> <span class="nb">bool</span><span class="p">)</span> <span class="o">-&gt;</span> <span class="n">Tuple</span><span class="p">[</span><span class="n">_OptionalTensor</span><span class="p">,</span> <span class="o">...</span><span class="p">]:</span>
461478
<span class="n">new_grads</span><span class="p">:</span> <span class="n">List</span><span class="p">[</span><span class="n">_OptionalTensor</span><span class="p">]</span> <span class="o">=</span> <span class="p">[]</span>
462479
<span class="k">for</span> <span class="n">out</span><span class="p">,</span> <span class="n">grad</span> <span class="ow">in</span> <span class="nb">zip</span><span class="p">(</span><span class="n">outputs</span><span class="p">,</span> <span class="n">grads</span><span class="p">):</span>
463480
<span class="k">if</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">grad</span><span class="p">,</span> <span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">):</span>
464-
<span class="n">grad_shape</span> <span class="o">=</span> <span class="n">grad</span><span class="o">.</span><span class="n">shape</span> <span class="k">if</span> <span class="ow">not</span> <span class="n">is_grads_batched</span> <span class="k">else</span> <span class="n">grad</span><span class="o">.</span><span class="n">shape</span><span class="p">[</span><span class="mi">1</span><span class="p">:]</span>
465-
<span class="k">if</span> <span class="ow">not</span> <span class="n">out</span><span class="o">.</span><span class="n">shape</span> <span class="o">==</span> <span class="n">grad_shape</span><span class="p">:</span>
481+
<span class="n">first_grad</span> <span class="o">=</span> <span class="n">grad</span> <span class="k">if</span> <span class="ow">not</span> <span class="n">is_grads_batched</span> <span class="k">else</span> <span class="n">grad</span><span class="p">[</span><span class="mi">0</span><span class="p">]</span>
482+
<span class="k">if</span> <span class="ow">not</span> <span class="n">torch</span><span class="o">.</span><span class="n">is_same_size</span><span class="p">(</span><span class="n">out</span><span class="p">,</span> <span class="n">first_grad</span><span class="p">):</span>
483+
<span class="n">out_shape</span><span class="p">,</span> <span class="n">grad_shape</span> <span class="o">=</span> <span class="n">_calculate_shape</span><span class="p">(</span><span class="n">out</span><span class="p">,</span> <span class="n">first_grad</span><span class="p">,</span> <span class="n">is_grads_batched</span><span class="p">)</span>
466484
<span class="k">if</span> <span class="n">is_grads_batched</span><span class="p">:</span>
467485
<span class="k">raise</span> <span class="ne">RuntimeError</span><span class="p">(</span><span class="s2">&quot;If `is_grads_batched=True`, we interpret the first &quot;</span>
468486
<span class="s2">&quot;dimension of each grad_output as the batch dimension. &quot;</span>
469487
<span class="s2">&quot;The sizes of the remaining dimensions are expected to match &quot;</span>
470488
<span class="s2">&quot;the shape of corresponding output, but a mismatch &quot;</span>
471489
<span class="s2">&quot;was detected: grad_output[&quot;</span>
472490
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">grads</span><span class="o">.</span><span class="n">index</span><span class="p">(</span><span class="n">grad</span><span class="p">))</span> <span class="o">+</span> <span class="s2">&quot;] has a shape of &quot;</span>
473-
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">grad</span><span class="o">.</span><span class="n">shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot; and output[&quot;</span>
491+
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">grad_shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot; and output[&quot;</span>
474492
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">outputs</span><span class="o">.</span><span class="n">index</span><span class="p">(</span><span class="n">out</span><span class="p">))</span> <span class="o">+</span> <span class="s2">&quot;] has a shape of &quot;</span>
475-
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">out</span><span class="o">.</span><span class="n">shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot;. &quot;</span>
493+
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">out_shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot;. &quot;</span>
476494
<span class="s2">&quot;If you only want some tensors in `grad_output` to be considered &quot;</span>
477495
<span class="s2">&quot;batched, consider using vmap.&quot;</span><span class="p">)</span>
478496
<span class="k">else</span><span class="p">:</span>
479497
<span class="k">raise</span> <span class="ne">RuntimeError</span><span class="p">(</span><span class="s2">&quot;Mismatch in shape: grad_output[&quot;</span>
480498
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">grads</span><span class="o">.</span><span class="n">index</span><span class="p">(</span><span class="n">grad</span><span class="p">))</span> <span class="o">+</span> <span class="s2">&quot;] has a shape of &quot;</span>
481-
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">grad</span><span class="o">.</span><span class="n">shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot; and output[&quot;</span>
499+
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">grad_shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot; and output[&quot;</span>
482500
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">outputs</span><span class="o">.</span><span class="n">index</span><span class="p">(</span><span class="n">out</span><span class="p">))</span> <span class="o">+</span> <span class="s2">&quot;] has a shape of &quot;</span>
483-
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">out</span><span class="o">.</span><span class="n">shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot;.&quot;</span><span class="p">)</span>
501+
<span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">out_shape</span><span class="p">)</span> <span class="o">+</span> <span class="s2">&quot;.&quot;</span><span class="p">)</span>
484502
<span class="k">if</span> <span class="n">out</span><span class="o">.</span><span class="n">dtype</span><span class="o">.</span><span class="n">is_complex</span> <span class="o">!=</span> <span class="n">grad</span><span class="o">.</span><span class="n">dtype</span><span class="o">.</span><span class="n">is_complex</span><span class="p">:</span>
485503
<span class="k">raise</span> <span class="ne">RuntimeError</span><span class="p">(</span><span class="s2">&quot;For complex Tensors, both grad_output and output&quot;</span>
486504
<span class="s2">&quot; are required to have the same dtype.&quot;</span>

0 commit comments

Comments
 (0)