158
158
159
159
160
160
< div class ="version ">
161
- < a href ='http://pytorch.org/docs/versions.html '> 1.7.0a0+fa9ae67 ▼</ a >
161
+ < a href ='http://pytorch.org/docs/versions.html '> 1.7.0a0+079c2ce ▼</ a >
162
162
</ div >
163
163
164
164
227
227
< li class ="toctree-l1 "> < a class ="reference internal " href ="../amp.html "> torch.cuda.amp</ a > </ li >
228
228
< li class ="toctree-l1 "> < a class ="reference internal " href ="../distributed.html "> torch.distributed</ a > </ li >
229
229
< li class ="toctree-l1 "> < a class ="reference internal " href ="../distributions.html "> torch.distributions</ a > </ li >
230
+ < li class ="toctree-l1 "> < a class ="reference internal " href ="../fft.html "> torch.fft</ a > </ li >
230
231
< li class ="toctree-l1 "> < a class ="reference internal " href ="../futures.html "> torch.futures</ a > </ li >
231
232
< li class ="toctree-l1 "> < a class ="reference internal " href ="../hub.html "> torch.hub</ a > </ li >
232
233
< li class ="toctree-l1 "> < a class ="reference internal " href ="../jit.html "> torch.jit</ a > </ li >
234
+ < li class ="toctree-l1 "> < a class ="reference internal " href ="../linalg.html "> torch.linalg</ a > </ li >
233
235
< li class ="toctree-l1 "> < a class ="reference internal " href ="../nn.init.html "> torch.nn.init</ a > </ li >
234
236
< li class ="toctree-l1 "> < a class ="reference internal " href ="../onnx.html "> torch.onnx</ a > </ li >
235
237
< li class ="toctree-l1 "> < a class ="reference internal " href ="../optim.html "> torch.optim</ a > </ li >
@@ -576,13 +578,13 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
576
578
< span class ="k "> return</ span > < span class ="nb "> isinstance</ span > < span class ="p "> (</ span > < span class ="n "> obj</ span > < span class ="p "> ,</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> Tensor</ span > < span class ="p "> )</ span > </ div >
577
579
578
580
579
- < div class =" viewcode-block " id =" is_storage " > < a class =" viewcode-back " href =" ../generated/torch.is_storage.html#torch.is_storage " > [docs] </ a > < span class ="k "> def</ span > < span class ="nf "> is_storage</ span > < span class ="p "> (</ span > < span class ="n "> obj</ span > < span class ="p "> ):</ span >
581
+ < span class ="k "> def</ span > < span class ="nf "> is_storage</ span > < span class ="p "> (</ span > < span class ="n "> obj</ span > < span class ="p "> ):</ span >
580
582
< span class ="sa "> r</ span > < span class ="sd "> """Returns True if `obj` is a PyTorch storage object.</ span >
581
583
582
584
< span class ="sd "> Args:</ span >
583
585
< span class ="sd "> obj (Object): Object to test</ span >
584
586
< span class ="sd "> """</ span >
585
- < span class ="k "> return</ span > < span class ="nb "> type</ span > < span class ="p "> (</ span > < span class ="n "> obj</ span > < span class ="p "> )</ span > < span class ="ow "> in</ span > < span class ="n "> _storage_classes</ span > </ div >
587
+ < span class ="k "> return</ span > < span class ="nb "> type</ span > < span class ="p "> (</ span > < span class ="n "> obj</ span > < span class ="p "> )</ span > < span class ="ow "> in</ span > < span class ="n "> _storage_classes</ span >
586
588
587
589
588
590
< span class ="k "> def</ span > < span class ="nf "> set_default_tensor_type</ span > < span class ="p "> (</ span > < span class ="n "> t</ span > < span class ="p "> ):</ span >
@@ -800,6 +802,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
800
802
< span class ="kn "> import</ span > < span class ="nn "> torch.cuda</ span >
801
803
< span class ="kn "> import</ span > < span class ="nn "> torch.autograd</ span >
802
804
< span class ="kn "> from</ span > < span class ="nn "> torch.autograd</ span > < span class ="kn "> import</ span > < span class ="n "> no_grad</ span > < span class ="p "> ,</ span > < span class ="n "> enable_grad</ span > < span class ="p "> ,</ span > < span class ="n "> set_grad_enabled</ span >
805
+ < span class ="c1 "> # import torch.fft # TODO: enable once torch.fft() is removed</ span >
803
806
< span class ="kn "> import</ span > < span class ="nn "> torch.futures</ span >
804
807
< span class ="kn "> import</ span > < span class ="nn "> torch.nn</ span >
805
808
< span class ="kn "> import</ span > < span class ="nn "> torch.nn.intrinsic</ span >
@@ -810,6 +813,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
810
813
< span class ="kn "> import</ span > < span class ="nn "> torch.utils.backcompat</ span >
811
814
< span class ="kn "> import</ span > < span class ="nn "> torch.onnx</ span >
812
815
< span class ="kn "> import</ span > < span class ="nn "> torch.jit</ span >
816
+ < span class ="kn "> import</ span > < span class ="nn "> torch.linalg</ span >
813
817
< span class ="kn "> import</ span > < span class ="nn "> torch.hub</ span >
814
818
< span class ="kn "> import</ span > < span class ="nn "> torch.random</ span >
815
819
< span class ="kn "> import</ span > < span class ="nn "> torch.distributions</ span >
@@ -831,9 +835,9 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
831
835
< span class ="k "> del</ span > < span class ="n "> _torch_docs</ span > < span class ="p "> ,</ span > < span class ="n "> _tensor_docs</ span > < span class ="p "> ,</ span > < span class ="n "> _storage_docs</ span >
832
836
833
837
834
- < span class ="k "> def</ span > < span class ="nf "> compiled_with_cxx11_abi</ span > < span class ="p "> ():</ span >
838
+ < div class =" viewcode-block " id =" compiled_with_cxx11_abi " > < a class =" viewcode-back " href =" ../generated/torch.compiled_with_cxx11_abi.html#torch.compiled_with_cxx11_abi " > [docs] </ a > < span class ="k "> def</ span > < span class ="nf "> compiled_with_cxx11_abi</ span > < span class ="p "> ():</ span >
835
839
< span class ="sa "> r</ span > < span class ="sd "> """Returns whether PyTorch was built with _GLIBCXX_USE_CXX11_ABI=1"""</ span >
836
- < span class ="k "> return</ span > < span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _GLIBCXX_USE_CXX11_ABI</ span >
840
+ < span class ="k "> return</ span > < span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _GLIBCXX_USE_CXX11_ABI</ span > </ div >
837
841
838
842
839
843
< span class ="c1 "> # Import the ops "namespace"</ span >
0 commit comments