|
219 | 219 | <div class="pytorch-left-menu-search">
|
220 | 220 |
|
221 | 221 | <div class="version">
|
222 |
| - <a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git0b25664 ) ▼</a> |
| 222 | + <a href='https://pytorch.org/docs/versions.html'>master (1.13.0a0+git4bb7e14 ) ▼</a> |
223 | 223 | </div>
|
224 | 224 |
|
225 | 225 |
|
@@ -587,10 +587,10 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
|
587 | 587 | <span class="sa">f</span><span class="s2">"Unsupported qscheme </span><span class="si">{</span><span class="bp">self</span><span class="o">.</span><span class="n">qscheme</span><span class="p">()</span><span class="si">}</span><span class="s2"> in deepcopy"</span>
|
588 | 588 | <span class="p">)</span>
|
589 | 589 | <span class="c1"># TODO: Once we decide to break serialization FC, no longer</span>
|
590 |
| - <span class="c1"># need to wrap with _TypedStorage</span> |
| 590 | + <span class="c1"># need to wrap with TypedStorage</span> |
591 | 591 | <span class="n">new_tensor</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">_utils</span><span class="o">.</span><span class="n">_rebuild_qtensor</span><span class="p">(</span>
|
592 |
| - <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">_TypedStorage</span><span class="p">(</span> |
593 |
| - <span class="n">wrap_storage</span><span class="o">=</span><span class="n">new_storage</span><span class="o">.</span><span class="n">_untyped</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span> |
| 592 | + <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">TypedStorage</span><span class="p">(</span> |
| 593 | + <span class="n">wrap_storage</span><span class="o">=</span><span class="n">new_storage</span><span class="o">.</span><span class="n">untyped</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span> |
594 | 594 | <span class="p">),</span>
|
595 | 595 | <span class="bp">self</span><span class="o">.</span><span class="n">storage_offset</span><span class="p">(),</span>
|
596 | 596 | <span class="bp">self</span><span class="o">.</span><span class="n">size</span><span class="p">(),</span>
|
@@ -684,7 +684,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
|
684 | 684 | <span class="k">if</span> <span class="n">has_torch_function_unary</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>
|
685 | 685 | <span class="k">return</span> <span class="n">handle_torch_function</span><span class="p">(</span><span class="n">Tensor</span><span class="o">.</span><span class="n">storage</span><span class="p">,</span> <span class="p">(</span><span class="bp">self</span><span class="p">,),</span> <span class="bp">self</span><span class="p">)</span>
|
686 | 686 |
|
687 |
| - <span class="k">return</span> <span class="n">torch</span><span class="o">.</span><span class="n">_TypedStorage</span><span class="p">(</span><span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">_storage</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span><span class="p">)</span></div> |
| 687 | + <span class="k">return</span> <span class="n">torch</span><span class="o">.</span><span class="n">TypedStorage</span><span class="p">(</span><span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">_storage</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span><span class="p">)</span></div> |
688 | 688 |
|
689 | 689 | <span class="k">def</span> <span class="nf">_reduce_ex_internal</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">proto</span><span class="p">):</span>
|
690 | 690 | <span class="n">check_serializing_named_tensor</span><span class="p">(</span><span class="bp">self</span><span class="p">)</span>
|
@@ -753,10 +753,10 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
|
753 | 753 | <span class="sa">f</span><span class="s2">"Serialization is not supported for tensors of type </span><span class="si">{</span><span class="bp">self</span><span class="o">.</span><span class="n">qscheme</span><span class="p">()</span><span class="si">}</span><span class="s2">"</span>
|
754 | 754 | <span class="p">)</span>
|
755 | 755 | <span class="c1"># TODO: Once we decide to break serialization FC, no longer</span>
|
756 |
| - <span class="c1"># need to wrap with _TypedStorage</span> |
| 756 | + <span class="c1"># need to wrap with TypedStorage</span> |
757 | 757 | <span class="n">args_qtensor</span> <span class="o">=</span> <span class="p">(</span>
|
758 |
| - <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">_TypedStorage</span><span class="p">(</span> |
759 |
| - <span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">storage</span><span class="p">()</span><span class="o">.</span><span class="n">_untyped</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span> |
| 758 | + <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">TypedStorage</span><span class="p">(</span> |
| 759 | + <span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">storage</span><span class="p">()</span><span class="o">.</span><span class="n">untyped</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span> |
760 | 760 | <span class="p">),</span>
|
761 | 761 | <span class="bp">self</span><span class="o">.</span><span class="n">storage_offset</span><span class="p">(),</span>
|
762 | 762 | <span class="nb">tuple</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">size</span><span class="p">()),</span>
|
@@ -811,10 +811,10 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
|
811 | 811 | <span class="k">return</span> <span class="p">(</span><span class="n">torch</span><span class="o">.</span><span class="n">_utils</span><span class="o">.</span><span class="n">_rebuild_wrapper_subclass</span><span class="p">,</span> <span class="n">arg_wrapper_subclass</span><span class="p">)</span>
|
812 | 812 | <span class="k">else</span><span class="p">:</span>
|
813 | 813 | <span class="c1"># TODO: Once we decide to break serialization FC, no longer</span>
|
814 |
| - <span class="c1"># need to wrap with _TypedStorage</span> |
| 814 | + <span class="c1"># need to wrap with TypedStorage</span> |
815 | 815 | <span class="n">args</span> <span class="o">=</span> <span class="p">(</span>
|
816 |
| - <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">_TypedStorage</span><span class="p">(</span> |
817 |
| - <span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">storage</span><span class="p">()</span><span class="o">.</span><span class="n">_untyped</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span> |
| 816 | + <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">TypedStorage</span><span class="p">(</span> |
| 817 | + <span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">storage</span><span class="p">()</span><span class="o">.</span><span class="n">untyped</span><span class="p">(),</span> <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span> |
818 | 818 | <span class="p">),</span>
|
819 | 819 | <span class="bp">self</span><span class="o">.</span><span class="n">storage_offset</span><span class="p">(),</span>
|
820 | 820 | <span class="nb">tuple</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">size</span><span class="p">()),</span>
|
|
0 commit comments