| 1 |
torch.Tensor.addmm_ |
@wwaawwaaee @enkilee #77385
|
paddle.Tensor.addmm_ |
c++下沉或装饰器 {'mat1': 'x', 'mat2': 'y'}; |
| 2 |
torch.Tensor.baddbmm_ |
@youge325 #76578 #77177 #7677 #798
|
paddle.Tensor.baddbmm_ |
c++下沉或装饰器 {'batch1': 'x', 'batch2': 'y'}; |
| 3 |
torch.Tensor.bitwise_and_ |
@youge325 #76705 @ADchampion3
|
paddle.Tensor.bitwise_and_ |
c++下沉或装饰器 {'other': 'y'}; |
| 4 |
torch.Tensor.bitwise_left_shift_ |
@ADchampion3 #76795
|
paddle.Tensor.bitwise_left_shift_ |
c++下沉或装饰器 {'other': 'y'}; |
| 5 |
torch.Tensor.bitwise_right_shift_ |
@ADchampion3 #76795
|
paddle.Tensor.bitwise_right_shift_ |
c++下沉或装饰器 {'other': 'y'}; |
| 6 |
torch.Tensor.bitwise_xor_ |
@Manfredss #76341 #754
|
paddle.Tensor.bitwise_xor_ |
c++下沉或装饰器 {'other': 'y'}; |
| 7 |
torch.Tensor.cauchy_ |
@WHoutstanding
|
paddle.Tensor.cauchy_ |
c++下沉或装饰器 {'median': 'loc', 'sigma': 'scale'}; |
| 8 |
torch.Tensor.copysign_ |
@WHoutstanding
|
paddle.Tensor.copysign_ |
c++下沉或装饰器 {'other': 'y'}; |
| 9 |
torch.Tensor.cumsum_ |
@xunyoyo #77573 #7704 #820
|
paddle.Tensor.cumsum_ |
c++下沉或装饰器 {'dim': 'axis', 'dtype': 'dtype'}; |
| 10 |
torch.Tensor.floor_divide_ |
@CAICAIIs #77556 #818
|
paddle.Tensor.floor_divide_ |
c++下沉或装饰器 {'other': 'y'}; |
| 11 |
torch.Tensor.gcd_ |
@Manfredss #76486 #787 #7670
|
paddle.Tensor.gcd_ |
c++下沉或装饰器 {'other': 'y'}; |
| 12 |
torch.Tensor.geometric_ |
@xunyoyo
|
paddle.Tensor.geometric_ |
c++下沉或装饰器 {'p': 'probs'}; |
| 13 |
torch.Tensor.hypot_ |
@Khoray #78046 #833 #7851
|
paddle.Tensor.hypot_ |
c++下沉或装饰器 {'other': 'y'}; |
| 14 |
torch.Tensor.index_fill_ |
@xunyoyo #77470 #77725
|
paddle.Tensor.index_fill_ |
c++下沉或装饰器 {'dim': 'axis'}; |
| 15 |
torch.Tensor.index_put_ |
@Aidenwu0209 #77792
|
paddle.Tensor.index_put_ |
c++下沉或装饰器 {'values': 'value'}; |
| 16 |
torch.Tensor.lcm_ |
@Manfredss #76486 #787 #7670
|
paddle.Tensor.lcm_ |
c++下沉或装饰器 {'other': 'y'}; |
| 17 |
torch.Tensor.ldexp_ |
@pxzbest #77766 #835
|
paddle.Tensor.ldexp_ |
c++下沉或装饰器 {'other': 'y'}; |
| 18 |
torch.Tensor.lerp_ |
@Aidenwu0209 #77793
|
paddle.Tensor.lerp_ |
c++下沉或装饰器 {'end': 'y'}; |
| 19 |
torch.Tensor.pow_ |
@NJX-njx #78126
|
paddle.Tensor.pow_ |
c++下沉或装饰器 {'exponent': 'y'}; |
| 20 |
torch.Tensor.remainder_ |
@CAICAIIs
|
paddle.Tensor.remainder_ |
c++下沉或装饰器 {'other': 'y'}; |
| 21 |
torch.Tensor.renorm_ |
@enkilee #77708
|
paddle.Tensor.renorm_ |
c++下沉或装饰器 {'dim': 'axis', 'maxnorm': 'max_norm'}; |
| 22 |
torch.Tensor.squeeze_ |
@xunyoyo #77550 #77725
|
paddle.Tensor.squeeze_ |
c++下沉或装饰器 {'dim': 'axis'}; |
| 23 |
torch.Tensor.unsqueeze_ |
@xunyoyo #77551 #77725
|
paddle.Tensor.unsqueeze_ |
c++下沉或装饰器 {'dim': 'axis'}; |
| 24 |
torch.abs_ |
@onenewcode #76431 @algorithm1832 #76994 #77294 #811 #7697
|
paddle.abs_ |
c++下沉或装饰器 {'input': 'x'}; |
| 25 |
torch.acos |
@enkilee #76390 #7654 #775
|
paddle.acos |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 26 |
torch.acosh |
@enkilee #76456 #783 #7657
|
paddle.acosh |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 27 |
torch.addmm |
@enkilee #77385
|
paddle.addmm |
c++下沉或装饰器 {'mat1': 'x', 'mat2': 'y'}; 新增out; |
| 28 |
torch.angle |
@enkilee #77012 #806 #7705
|
paddle.angle |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 29 |
torch.as_strided |
@enkilee #77355 #7705
|
paddle.as_strided |
c++下沉或装饰器 {'input': 'x', 'size': 'shape', 'storage_offset': 'offset'}; |
| 30 |
torch.asin |
@Manfredss #76421 #759 #7643
|
paddle.asin |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 31 |
torch.asinh |
@Manfredss #76736 #785 #7668
|
paddle.asinh |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 32 |
torch.atan |
@Manfredss #76736 #785 #7668
|
paddle.atan |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 33 |
torch.atan2 |
@Manfredss #76736 #785 #7668
|
paddle.atan2 |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 34 |
torch.atanh |
@enkilee #76663 #7666 #786
|
paddle.atanh |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 35 |
torch.baddbmm |
@youge325 #76578 #77177 #7677 #798
|
paddle.baddbmm |
c++下沉或装饰器 {'batch1': 'x', 'batch2': 'y'}; 新增out; |
| 36 |
torch.bernoulli |
@xunyoyo #77567
|
paddle.bernoulli |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 37 |
torch.bincount |
@Carousel126
|
paddle.bincount |
c++下沉或装饰器 {'input': 'x'}; |
| 38 |
torch.bitwise_and |
@youge325 #76705
|
paddle.bitwise_and |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 39 |
torch.bitwise_left_shift |
|
paddle.bitwise_left_shift |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 40 |
torch.bitwise_not |
|
paddle.bitwise_not |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 41 |
torch.bitwise_right_shift |
|
paddle.bitwise_right_shift |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 42 |
torch.bitwise_xor |
@Manfredss
|
paddle.bitwise_xor |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 43 |
torch.block_diag |
|
paddle.block_diag |
c++下沉或装饰器 {'tensors': 'inputs'}; |
| 44 |
torch.broadcast_tensors |
|
paddle.broadcast_tensors |
c++下沉或装饰器 {'tensors': 'input'}; |
| 45 |
torch.cartesian_prod |
@SidusAntares
|
paddle.cartesian_prod |
c++下沉或装饰器 {'tensors': 'x'}; |
| 46 |
torch.cdist |
|
paddle.cdist |
c++下沉或装饰器 {'x1': 'x', 'x2': 'y'}; |
| 47 |
torch.clone |
@liulin1124 #77780 @enkilee #78047
|
paddle.clone |
c++下沉或装饰器 {'input': 'x'}; |
| 48 |
torch.column_stack |
@enkilee #78069
|
paddle.column_stack |
c++下沉或装饰器 {'tensors': 'x'}; 新增out; |
| 49 |
torch.combinations |
|
paddle.combinations |
c++下沉或装饰器 {'input': 'x'}; |
| 50 |
torch.conj |
@enkilee #77079 #7682 #806
|
paddle.conj |
c++下沉或装饰器 {'input': 'x'}; |
| 51 |
torch.copysign |
|
paddle.copysign |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 52 |
torch.cosh |
@algorithm1832 #76429 #7640 #766
|
paddle.cosh |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 53 |
torch.count_nonzero |
@MerlinSMQWQ #78077 #7806 #836
|
paddle.count_nonzero |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; |
| 54 |
torch.cross |
@pxzbest #77691
|
paddle.cross |
c++下沉或装饰器 {'input': 'x', 'other': 'y', 'dim': 'axis'}; 新增out; |
| 55 |
torch.cummax |
@xcq-code #78081 #78183
|
paddle.cummax |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 56 |
torch.cummin |
@xcq-code #78081 #78183
|
paddle.cummin |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 57 |
torch.cumulative_trapezoid |
|
paddle.cumulative_trapezoid |
c++下沉或装饰器 {'dim': 'axis'}; |
| 58 |
torch.diag |
@algorithm1832 #76939 #7675 #792
|
paddle.diag |
c++下沉或装饰器 {'input': 'x', 'diagonal': 'offset'}; 新增out; |
| 59 |
torch.diagflat |
@HaolinGuan7211
|
paddle.diagflat |
c++下沉或装饰器 {'input': 'x'}; |
| 60 |
torch.diagonal_scatter |
|
paddle.diagonal_scatter |
c++下沉或装饰器 {'input': 'x', 'src': 'y', 'dim1': 'axis1', 'dim2': 'axis2'}; |
| 61 |
torch.dist |
|
paddle.dist |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; |
| 62 |
torch.dsplit |
@enkilee #78048
|
paddle.dsplit |
c++下沉或装饰器 {'input': 'x', 'indices': 'num_or_indices', 'sections': 'num_or_indices'}; |
| 63 |
torch.dstack |
@enkilee #78069
|
paddle.dstack |
c++下沉或装饰器 {'tensors': 'x'}; 新增out; |
| 64 |
torch.flip |
|
paddle.flip |
c++下沉或装饰器 {'input': 'x', 'dims': 'axis'}; |
| 65 |
torch.fmax |
|
paddle.fmax |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 66 |
torch.fmin |
|
paddle.fmin |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 67 |
torch.frac |
@algorithm1832 #76671 #7662 #784
|
paddle.frac |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 68 |
torch.frexp |
|
paddle.frexp |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 69 |
torch.gcd |
@Manfredss #76486 #7670 #787
|
paddle.gcd |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 70 |
torch.heaviside |
|
paddle.heaviside |
c++下沉或装饰器 {'input': 'x', 'values': 'y'}; 新增out; |
| 71 |
torch.histogram |
|
paddle.histogram |
c++下沉或装饰器 {'range': ['min', 'max']}; 新增out; |
| 72 |
torch.histogramdd |
|
paddle.histogramdd |
c++下沉或装饰器 {'input': 'x', 'range': 'ranges', 'weight': 'weights'}; |
| 73 |
torch.hsplit |
@enkilee #78048
|
paddle.hsplit |
c++下沉或装饰器 {'input': 'x', 'indices': 'num_or_indices', 'sections': 'num_or_indices'}; |
| 74 |
torch.hstack |
@enkilee #78069
|
paddle.hstack |
c++下沉或装饰器 {'tensors': 'x'}; 新增out; |
| 75 |
torch.hypot |
@Khoray #78046 #833 #7851 @enkilee #78058
|
paddle.hypot |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 76 |
torch.iinfo |
@why1te
|
paddle.iinfo |
c++下沉或装饰器 {'type': 'dtype'}; |
| 77 |
torch.imag |
@liulin1124 #78038
|
paddle.imag |
c++下沉或装饰器 {'input': 'x'}; |
| 78 |
torch.index_add |
|
paddle.index_add |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 79 |
torch.index_fill |
|
paddle.index_fill |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; |
| 80 |
torch.inner |
|
paddle.inner |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 81 |
torch.isneginf |
|
paddle.isneginf |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 82 |
torch.isposinf |
|
paddle.isposinf |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 83 |
torch.isreal |
@Carousel126 #78181 #847
|
paddle.isreal |
c++下沉或装饰器 {'input': 'x'}; |
| 84 |
torch.kron |
|
paddle.kron |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 85 |
torch.kthvalue |
|
paddle.kthvalue |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 86 |
torch.lcm |
@Manfredss #76486 #7670 #787
|
paddle.lcm |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 87 |
torch.ldexp |
|
paddle.ldexp |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 88 |
torch.lgamma |
@cvbret
|
paddle.lgamma |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 89 |
torch.linalg.cholesky |
|
paddle.linalg.cholesky |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 90 |
torch.linalg.cond |
|
paddle.linalg.cond |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 91 |
torch.linalg.det |
|
paddle.linalg.det |
c++下沉或装饰器 {'A': 'x'}; 新增out; |
| 92 |
torch.linalg.eig |
|
paddle.linalg.eig |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 93 |
torch.linalg.eigh |
|
paddle.linalg.eigh |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 94 |
torch.linalg.eigvals |
|
paddle.linalg.eigvals |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 95 |
torch.linalg.eigvalsh |
|
paddle.linalg.eigvalsh |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 96 |
torch.linalg.householder_product |
|
paddle.linalg.householder_product |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 97 |
torch.linalg.inv |
|
paddle.linalg.inv |
c++下沉或装饰器 {'A': 'x'}; 新增out; |
| 98 |
torch.linalg.lstsq |
|
paddle.linalg.lstsq |
c++下沉或装饰器 {'input': 'x', 'b': 'y'}; |
| 99 |
torch.linalg.lu |
|
paddle.linalg.lu |
c++下沉或装饰器 {'A': 'x'}; 新增out; |
| 100 |
torch.linalg.lu_solve |
|
paddle.linalg.lu_solve |
c++下沉或装饰器 {'LU': 'lu', 'B': 'b'}; 新增out; |
| 101 |
torch.linalg.matrix_norm |
|
paddle.linalg.matrix_norm |
c++下沉或装饰器 {'input': 'x', 'ord': 'p', 'dim': 'axis'}; 新增out; |
| 102 |
torch.linalg.matrix_power |
|
paddle.linalg.matrix_power |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 103 |
torch.linalg.matrix_rank |
|
paddle.linalg.matrix_rank |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 104 |
torch.linalg.multi_dot |
|
paddle.linalg.multi_dot |
c++下沉或装饰器 {'tensors': 'x'}; 新增out; |
| 105 |
torch.linalg.pinv |
|
paddle.linalg.pinv |
c++下沉或装饰器 {'input': 'x', 'atol': '', 'rtol': 'rcond'}; 新增out; |
| 106 |
torch.linalg.qr |
|
paddle.linalg.qr |
c++下沉或装饰器 {'A': 'x'}; 新增out; |
| 107 |
torch.linalg.svdvals |
|
paddle.linalg.svdvals |
c++下沉或装饰器 {'A': 'x', 'driver': ''}; 新增out; |
| 108 |
torch.linalg.vecdot |
|
paddle.linalg.vecdot |
c++下沉或装饰器 {'dim': 'axis'}; 新增out; |
| 109 |
torch.log10 |
@zade23 #76381
|
paddle.log10 |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 110 |
torch.log1p |
@zade23 #76381
|
paddle.log1p |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 111 |
torch.logaddexp |
@ADchampion3
|
paddle.logaddexp |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 112 |
torch.logcumsumexp |
@ADchampion3
|
paddle.logcumsumexp |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 113 |
torch.logspace |
@ADchampion3
|
paddle.logspace |
c++下沉或装饰器 {'end': 'stop', 'steps': 'num', 'dtype': 'dtype'}; 新增out; |
| 114 |
torch.masked_fill |
@Aidenwu0209 #77801 #7753 #828
|
paddle.masked_fill |
c++下沉或装饰器 {'input': 'x'}; |
| 115 |
torch.mode |
@SidusAntares
|
paddle.mode |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 116 |
torch.moveaxis |
|
paddle.moveaxis |
c++下沉或装饰器 {'input': 'x'}; |
| 117 |
torch.mv |
|
paddle.mv |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 118 |
torch.nan_to_num |
|
paddle.nan_to_num |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 119 |
torch.nanmean |
|
paddle.nanmean |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 120 |
torch.nanquantile |
|
paddle.nanquantile |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增out; |
| 121 |
torch.nansum |
|
paddle.nansum |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis', 'dtype': 'dtype'}; 新增out; |
| 122 |
torch.neg |
@seasameye #78216
|
paddle.neg |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 123 |
torch.nextafter |
@enkilee #77189
|
paddle.nextafter |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; |
| 124 |
torch.nn.Bilinear |
|
paddle.nn.Bilinear |
c++下沉或装饰器 {'bias': 'bias_attr', 'device': '', 'dtype': ''}; |
| 125 |
torch.nn.CosineSimilarity |
|
paddle.nn.CosineSimilarity |
c++下沉或装饰器 {'dim': 'axis'}; |
| 126 |
torch.nn.Flatten |
|
paddle.nn.Flatten |
c++下沉或装饰器 {'start_dim': 'start_axis', 'end_dim': |
| 127 |
torch.nn.Fold |
|
paddle.nn.Fold |
c++下沉或装饰器 {'output_size': 'output_sizes', 'kernel_size': 'kernel_sizes', 'stride': 'strides', 'padding': 'paddings', 'dilation': 'dilations'}; |
| 128 |
torch.nn.GLU |
|
paddle.nn.GLU |
c++下沉或装饰器 {'dim': 'axis'}; |
| 129 |
torch.nn.GRU |
|
paddle.nn.GRU |
c++下沉或装饰器 {'bias': ['bias_ih_attr', 'bias_hh_attr'], 'device': '', 'dtype': ''}; |
| 130 |
torch.nn.GRUCell |
|
paddle.nn.GRUCell |
c++下沉或装饰器 {'bias': ['bias_ih_attr', 'bias_hh_attr'], 'device': '', 'dtype': ''}; |
| 131 |
torch.nn.GaussianNLLLoss |
|
paddle.nn.GaussianNLLLoss |
c++下沉或装饰器 {'eps': 'epsilon'}; |
| 132 |
torch.nn.Hardshrink |
|
paddle.nn.Hardshrink |
c++下沉或装饰器 {'lambd': 'threshold'}; |
| 133 |
torch.nn.Hardtanh |
|
paddle.nn.Hardtanh |
c++下沉或装饰器 {'min_val': 'min', 'max_val': 'max'}; |
| 134 |
torch.nn.LSTM |
@Manfredss #76915
|
paddle.nn.LSTM |
c++下沉或装饰器 {'bias': ['bias_ih_attr', 'bias_hh_attr'], 'device': '', 'dtype': ''}; |
| 135 |
torch.nn.LSTMCell |
@Manfredss #76915
|
paddle.nn.LSTMCell |
c++下沉或装饰器 {'bias': ['bias_ih_attr', 'bias_hh_attr'], 'device': '', 'dtype': ''}; |
| 136 |
torch.nn.PReLU |
|
paddle.nn.PReLU |
c++下沉或装饰器 {'device': '', 'dtype': ''}; |
| 137 |
torch.nn.PairwiseDistance |
|
paddle.nn.PairwiseDistance |
c++下沉或装饰器 {'eps': 'epsilon'}; |
| 138 |
torch.nn.ParameterList |
|
paddle.nn.ParameterList |
c++下沉或装饰器 {'values': 'parameters'}; |
| 139 |
torch.nn.PoissonNLLLoss |
|
paddle.nn.PoissonNLLLoss |
c++下沉或装饰器 {'eps': 'epsilon'}; |
| 140 |
torch.nn.Softshrink |
|
paddle.nn.Softshrink |
c++下沉或装饰器 {'lambd': 'threshold'}; |
| 141 |
torch.nn.SyncBatchNorm |
|
paddle.nn.SyncBatchNorm |
c++下沉或装饰器 {'eps': 'epsilon', 'affine': ['weight_attr', 'bias_attr'], 'track_running_stats': '', 'process_group': '', 'device': '', 'dtype': ''}; |
| 142 |
torch.nn.TransformerEncoder |
|
paddle.nn.TransformerEncoder |
c++下沉或装饰器 {'enable_nested_tensor': '', 'mask_check': ''}; |
| 143 |
torch.nn.TripletMarginLoss |
|
paddle.nn.TripletMarginLoss |
c++下沉或装饰器 {'eps': 'epsilon'}; |
| 144 |
torch.nn.Unflatten |
|
paddle.nn.Unflatten |
c++下沉或装饰器 {'dim': 'axis', 'unflattened_size': 'shape'}; |
| 145 |
torch.nn.functional.adaptive_avg_pool2d |
|
paddle.nn.functional.adaptive_avg_pool2d |
c++下沉或装饰器 {'input': 'x'}; |
| 146 |
torch.nn.functional.adaptive_avg_pool3d |
|
paddle.nn.functional.adaptive_avg_pool3d |
c++下沉或装饰器 {'input': 'x'}; |
| 147 |
torch.nn.functional.adaptive_max_pool1d |
|
paddle.nn.functional.adaptive_max_pool1d |
c++下沉或装饰器 {'input': 'x', 'return_indices': 'return_mask'}; |
| 148 |
torch.nn.functional.adaptive_max_pool2d |
|
paddle.nn.functional.adaptive_max_pool2d |
c++下沉或装饰器 {'input': 'x', 'return_indices': 'return_mask'}; |
| 149 |
torch.nn.functional.adaptive_max_pool3d |
|
paddle.nn.functional.adaptive_max_pool3d |
c++下沉或装饰器 {'input': 'x', 'return_indices': 'return_mask'}; |
| 150 |
torch.nn.functional.affine_grid |
|
paddle.nn.functional.affine_grid |
c++下沉或装饰器 {'size': 'out_shape'}; |
| 151 |
torch.nn.functional.alpha_dropout |
|
paddle.nn.functional.alpha_dropout |
c++下沉或装饰器 {'input': 'x'}; |
| 152 |
torch.nn.functional.batch_norm |
|
paddle.nn.functional.batch_norm |
c++下沉或装饰器 {'input': 'x', 'eps': 'epsilon'}; |
| 153 |
torch.nn.functional.bilinear |
|
paddle.nn.functional.bilinear |
c++下沉或装饰器 {'input1': 'x1', 'input2': 'x2'}; |
| 154 |
torch.nn.functional.binary_cross_entropy |
@Aidenwu0209 #78132 #7887 #863
|
paddle.nn.functional.binary_cross_entropy |
c++下沉或装饰器 {'target': 'label'}; |
| 155 |
torch.nn.functional.binary_cross_entropy_with_logits |
@Aidenwu0209 #78140 #7842 #846
|
paddle.nn.functional.binary_cross_entropy_with_logits |
c++下沉或装饰器 {'input': 'logit', 'target': 'label'}; |
| 156 |
torch.nn.functional.celu |
|
paddle.nn.functional.celu |
c++下沉或装饰器 {'input': 'x'}; |
| 157 |
torch.nn.functional.channel_shuffle |
|
paddle.nn.functional.channel_shuffle |
c++下沉或装饰器 {'input': 'x'}; |
| 158 |
torch.nn.functional.cosine_embedding_loss |
|
paddle.nn.functional.cosine_embedding_loss |
c++下沉或装饰器 {'target': 'label'}; |
| 159 |
torch.nn.functional.cosine_similarity |
|
paddle.nn.functional.cosine_similarity |
c++下沉或装饰器 {'dim': 'axis'}; |
| 160 |
torch.nn.functional.dropout2d |
|
paddle.nn.functional.dropout2d |
c++下沉或装饰器 {'input': 'x'}; |
| 161 |
torch.nn.functional.dropout3d |
|
paddle.nn.functional.dropout3d |
c++下沉或装饰器 {'input': 'x'}; |
| 162 |
torch.nn.functional.elu |
|
paddle.nn.functional.elu |
c++下沉或装饰器 {'input': 'x'}; |
| 163 |
torch.nn.functional.elu_ |
|
paddle.nn.functional.elu_ |
c++下沉或装饰器 {'input': 'x'}; |
| 164 |
torch.nn.functional.fold |
|
paddle.nn.functional.fold |
c++下沉或装饰器 {'input': 'x', 'output_size': 'output_sizes', 'kernel_size': 'kernel_sizes', 'stride': 'strides', 'padding': 'paddings', 'dilation': 'dilations'}; |
| 165 |
torch.nn.functional.gaussian_nll_loss |
|
paddle.nn.functional.gaussian_nll_loss |
c++下沉或装饰器 {'target': 'label', 'var': 'variance', 'eps': 'epsilon'}; |
| 166 |
torch.nn.functional.glu |
|
paddle.nn.functional.glu |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; |
| 167 |
torch.nn.functional.gumbel_softmax |
|
paddle.nn.functional.gumbel_softmax |
c++下沉或装饰器 {'logits': 'x', 'tau': 'temperature', 'eps': '', 'dim': 'axis'}; |
| 168 |
torch.nn.functional.hardshrink |
|
paddle.nn.functional.hardshrink |
c++下沉或装饰器 {'input': 'x', 'lambd': 'threshold'}; |
| 169 |
torch.nn.functional.hardsigmoid |
|
paddle.nn.functional.hardsigmoid |
c++下沉或装饰器 {'input': 'x'}; |
| 170 |
torch.nn.functional.hardswish |
|
paddle.nn.functional.hardswish |
c++下沉或装饰器 {'input': 'x'}; |
| 171 |
torch.nn.functional.hardtanh |
|
paddle.nn.functional.hardtanh |
c++下沉或装饰器 {'input': 'x', 'min_val': 'min', 'max_val': 'max'}; |
| 172 |
torch.nn.functional.hardtanh_ |
|
paddle.nn.functional.hardtanh_ |
c++下沉或装饰器 {'input': 'x', 'min_val': 'min', 'max_val': 'max'}; |
| 173 |
torch.nn.functional.hinge_embedding_loss |
|
paddle.nn.functional.hinge_embedding_loss |
c++下沉或装饰器 {'target': 'label'}; |
| 174 |
torch.nn.functional.instance_norm |
|
paddle.nn.functional.instance_norm |
c++下沉或装饰器 {'input': 'x'}; |
| 175 |
torch.nn.functional.kl_div |
|
paddle.nn.functional.kl_div |
c++下沉或装饰器 {'target': 'label'}; |
| 176 |
torch.nn.functional.l1_loss |
|
paddle.nn.functional.l1_loss |
c++下沉或装饰器 {'target': 'label'}; |
| 177 |
torch.nn.functional.leaky_relu_ |
@Manfredss #77300 #802
|
paddle.nn.functional.leaky_relu_ |
c++下沉或装饰器 {'input': 'x'}; |
| 178 |
torch.nn.functional.local_response_norm |
|
paddle.nn.functional.local_response_norm |
c++下沉或装饰器 {'input': 'x'}; |
| 179 |
torch.nn.functional.lp_pool1d |
|
paddle.nn.functional.lp_pool1d |
c++下沉或装饰器 {'input': 'x'}; |
| 180 |
torch.nn.functional.lp_pool2d |
|
paddle.nn.functional.lp_pool2d |
c++下沉或装饰器 {'input': 'x'}; |
| 181 |
torch.nn.functional.margin_ranking_loss |
|
paddle.nn.functional.margin_ranking_loss |
c++下沉或装饰器 {'input1': 'input', 'input2': 'other', 'target': 'label'}; |
| 182 |
torch.nn.functional.max_unpool1d |
|
paddle.nn.functional.max_unpool1d |
c++下沉或装饰器 {'input': 'x'}; |
| 183 |
torch.nn.functional.max_unpool2d |
|
paddle.nn.functional.max_unpool2d |
c++下沉或装饰器 {'input': 'x'}; |
| 184 |
torch.nn.functional.max_unpool3d |
|
paddle.nn.functional.max_unpool3d |
c++下沉或装饰器 {'input': 'x'}; |
| 185 |
torch.nn.functional.mish |
@Aidenwu0209 #77923 #7825 #838
|
paddle.nn.functional.mish |
c++下沉或装饰器 {'input': 'x'}; |
| 186 |
torch.nn.functional.mse_loss |
@Aidenwu0209 #78128 #7840 #845
|
paddle.nn.functional.mse_loss |
c++下沉或装饰器 {'target': 'label'}; |
| 187 |
torch.nn.functional.multi_margin_loss |
@Aidenwu0209 #78134
|
paddle.nn.functional.multi_margin_loss |
c++下沉或装饰器 {'target': 'label'}; |
| 188 |
torch.nn.functional.nll_loss |
@Aidenwu0209 #78133 #7841 #844
|
paddle.nn.functional.nll_loss |
c++下沉或装饰器 {'target': 'label'}; |
| 189 |
torch.nn.functional.pairwise_distance |
|
paddle.nn.functional.pairwise_distance |
c++下沉或装饰器 {'x1': 'x', 'x2': 'y', 'eps': 'epsilon'}; |
| 190 |
torch.nn.functional.pixel_unshuffle |
@Aidenwu0209 #77920 #7827 #840
|
paddle.nn.functional.pixel_unshuffle |
c++下沉或装饰器 {'input': 'x'}; |
| 191 |
torch.nn.functional.poisson_nll_loss |
@Aidenwu0209 #77919 #7824 #837
|
paddle.nn.functional.poisson_nll_loss |
c++下沉或装饰器 {'target': 'label', 'eps': 'epsilon'}; |
| 192 |
torch.nn.functional.prelu |
|
paddle.nn.functional.prelu |
c++下沉或装饰器 {'input': 'x'}; |
| 193 |
torch.nn.functional.relu6 |
|
paddle.nn.functional.relu6 |
c++下沉或装饰器 {'input': 'x'}; |
| 194 |
torch.nn.functional.relu_ |
@Manfredss #77300 #802
|
paddle.nn.functional.relu_ |
c++下沉或装饰器 {'input': 'x'}; |
| 195 |
torch.nn.functional.rrelu |
|
paddle.nn.functional.rrelu |
c++下沉或装饰器 {'input': 'x'}; |
| 196 |
torch.nn.functional.selu |
|
paddle.nn.functional.selu |
c++下沉或装饰器 {'input': 'x'}; |
| 197 |
torch.nn.functional.soft_margin_loss |
|
paddle.nn.functional.soft_margin_loss |
c++下沉或装饰器 {'target': 'label'}; |
| 198 |
torch.nn.functional.softshrink |
|
paddle.nn.functional.softshrink |
c++下沉或装饰器 {'input': 'x', 'lambd': 'threshold'}; |
| 199 |
torch.nn.functional.softsign |
|
paddle.nn.functional.softsign |
c++下沉或装饰器 {'input': 'x'}; |
| 200 |
torch.nn.functional.tanhshrink |
|
paddle.nn.functional.tanhshrink |
c++下沉或装饰器 {'input': 'x'}; |
| 201 |
torch.nn.functional.triplet_margin_loss |
|
paddle.nn.functional.triplet_margin_loss |
c++下沉或装饰器 {'anchor': 'input', 'eps': 'epsilon'}; |
| 202 |
torch.nn.functional.triplet_margin_with_distance_loss |
|
paddle.nn.functional.triplet_margin_with_distance_loss |
c++下沉或装饰器 {'anchor': 'input'}; |
| 203 |
torch.nn.utils.clip_grad_value_ |
|
paddle.nn.utils.clip_grad_value_ |
c++下沉或装饰器 {'foreach': ''}; |
| 204 |
torch.nn.utils.remove_weight_norm |
|
paddle.nn.utils.remove_weight_norm |
c++下沉或装饰器 {'module': 'layer'}; |
| 205 |
torch.nn.utils.spectral_norm |
|
paddle.nn.utils.spectral_norm |
c++下沉或装饰器 {'module': 'layer'}; |
| 206 |
torch.nn.utils.weight_norm |
|
paddle.nn.utils.weight_norm |
c++下沉或装饰器 {'module': 'layer'}; |
| 207 |
torch.poisson |
@tjujingzong #77715 #7831 #841
|
paddle.poisson |
c++下沉或装饰器 {'input': 'x'}; |
| 208 |
torch.positive |
|
paddle.positive |
c++下沉或装饰器 {'input': 'x'}; |
| 209 |
torch.rad2deg |
|
paddle.rad2deg |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 210 |
torch.randint_like |
|
paddle.randint_like |
c++下沉或装饰器 {'input': 'x', 'dtype': 'dtype'}; |
| 211 |
torch.real |
@OctoberPrayRain #78198
|
paddle.real |
c++下沉或装饰器 {'input': 'x'}; |
| 212 |
torch.reciprocal |
|
paddle.reciprocal |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 213 |
torch.renorm |
@enkilee #77708
|
paddle.renorm |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis', 'maxnorm': 'max_norm'}; 新增out; |
| 214 |
torch.rot90 |
@Aidenwu0209 #77925 #7826 #839
|
paddle.rot90 |
c++下沉或装饰器 {'input': 'x', 'dims': 'axes'}; |
| 215 |
torch.row_stack |
@enkilee #78069
|
paddle.row_stack |
c++下沉或装饰器 {'tensors': 'x'}; 新增out; |
| 216 |
torch.save |
|
paddle.save |
c++下沉或装饰器 {'f': 'path', 'pickle_module': '', 'pickle_protocol': 'protocol', '_use_new_zipfile_serialization': ''}; |
| 217 |
torch.searchsorted |
|
paddle.searchsorted |
c++下沉或装饰器 {'input': 'values'}; 新增out; |
| 218 |
torch.select_scatter |
@Aidenwu0209 #77922 #7852 #851
|
paddle.select_scatter |
c++下沉或装饰器 {'input': 'x', 'src': 'values', 'dim': 'axis'}; |
| 219 |
torch.set_rng_state |
|
paddle.set_rng_state |
c++下沉或装饰器 {'new_state': 'state_list'}; |
| 220 |
torch.sgn |
|
paddle.sgn |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 221 |
torch.signbit |
|
paddle.signbit |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 222 |
torch.sinh |
@enkilee #76832 #7671 #790
|
paddle.sinh |
c++下沉或装饰器 {'input': 'x'}; 新增out; |
| 223 |
torch.slice_scatter |
|
paddle.slice_scatter |
c++下沉或装饰器 {'input': 'x', 'src': 'value', 'dim': 'axes', 'start': 'starts', 'end': 'ends', 'step': 'strides'}; |
| 224 |
torch.sparse.addmm |
|
paddle.sparse.addmm |
c++下沉或装饰器 {'mat1': 'x', 'mat2': 'y'}; |
| 225 |
torch.take |
|
paddle.take |
c++下沉或装饰器 {'input': 'x'}; |
| 226 |
torch.tensordot |
|
paddle.tensordot |
c++下沉或装饰器 {'a': 'x', 'b': 'y', 'dims': 'axes'}; 新增out; |
| 227 |
torch.tile |
@LiaoYFBH #76784 #789
|
paddle.tile |
c++下沉或装饰器 {'input': 'x', 'dims': 'repeat_times'}; |
| 228 |
torch.trace |
@LiaoYFBH #76785
|
paddle.trace |
c++下沉或装饰器 {'input': 'x'}; |
| 229 |
torch.trapezoid |
|
paddle.trapezoid |
c++下沉或装饰器 {'dim': 'axis'}; |
| 230 |
torch.tril_indices |
|
paddle.tril_indices |
c++下沉或装饰器 {'dtype': 'dtype'}; |
| 231 |
torch.triu_indices |
|
paddle.triu_indices |
c++下沉或装饰器 {'dtype': 'dtype'}; |
| 232 |
torch.unflatten |
|
paddle.unflatten |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis', 'sizes': 'shape'}; |
| 233 |
torch.vander |
|
paddle.vander |
c++下沉或装饰器 {'N': 'n'}; |
| 234 |
torch.vsplit |
@enkilee #78048
|
paddle.vsplit |
c++下沉或装饰器 {'input': 'x', 'indices': 'num_or_indices', 'sections': 'num_or_indices'}; |
| 235 |
torch.vstack |
@enkilee #78069
|
paddle.vstack |
c++下沉或装饰器 {'tensors': 'x'}; 新增out; |
| 236 |
torch.Tensor.addmv_ |
|
paddle.mm |
新增API paddle.Tensor.addmv_; |
| 237 |
torch.Tensor.addr_ |
|
paddle.outer |
新增API paddle.Tensor.addr_; |
| 238 |
torch.Tensor.col_indices |
|
paddle.Tensor.cols |
仅调用方式不一致; 新增API paddle.Tensor.col_indices; |
| 239 |
torch.Tensor.crow_indices |
|
paddle.Tensor.crows |
仅调用方式不一致; 新增API paddle.Tensor.crow_indices; |
| 240 |
torch.Tensor.erf_ |
|
paddle.erf_ |
仅调用方式不一致; 新增API paddle.Tensor.erf_; |
| 241 |
torch.Tensor.expm1_ |
|
paddle.expm1_ |
仅调用方式不一致; 新增API paddle.Tensor.expm1_; |
| 242 |
torch.Tensor.fix_ |
|
paddle.Tensor.trunc_ |
仅调用方式不一致; 新增API paddle.Tensor.fix_; |
| 243 |
torch.Tensor.matrix_exp |
|
paddle.linalg.matrix_exp |
仅调用方式不一致; 新增API paddle.Tensor.matrix_exp; |
| 244 |
torch.Tensor.mvlgamma_ |
|
paddle.Tensor.multigammaln_ |
仅调用方式不一致; 新增API paddle.Tensor.mvlgamma_; |
| 245 |
torch.Tensor.negative_ |
|
paddle.Tensor.neg_ |
仅调用方式不一致; 新增API paddle.Tensor.negative_; |
| 246 |
torch.Tensor.retain_grad |
@BayesianAura #78083 #7847 #848 @wuhaostudio #78083 #7847 #848
|
paddle.Tensor.retain_grads |
仅调用方式不一致; 新增API paddle.Tensor.retain_grad; |
| 247 |
torch.Tensor.sparse_mask |
@BayesianAura #78083 #7847 #848 @wuhaostudio #78083 #7847 #848
|
paddle.sparse.mask_as |
仅调用方式不一致; 新增API paddle.Tensor.sparse_mask; |
| 248 |
torch.Tensor.square_ |
|
paddle.square_ |
仅调用方式不一致; 新增API paddle.Tensor.square_; |
| 249 |
torch.Tensor.to_sparse |
|
paddle.Tensor.to_sparse_coo |
仅调用方式不一致; 新增API paddle.Tensor.to_sparse; |
| 250 |
torch.addmv |
|
paddle.mm |
新增out; 新增API paddle.addmv; |
| 251 |
torch.addr |
|
paddle.outer |
新增out; 新增API paddle.addr; |
| 252 |
torch.autograd.enable_grad |
|
paddle.enable_grad |
仅调用方式不一致; 新增API paddle.autograd.enable_grad; |
| 253 |
torch.autograd.grad_mode.set_grad_enabled |
|
paddle.set_grad_enabled |
仅调用方式不一致; 新增API paddle.autograd.grad_mode.set_grad_enabled; |
| 254 |
torch.backends.cuda.is_built |
|
paddle.device.is_compiled_with_cuda |
仅调用方式不一致; 新增API paddle.backends.cuda.is_built; |
| 255 |
torch.backends.cudnn.version |
|
paddle.device.get_cudnn_version |
仅调用方式不一致; 新增API paddle.backends.cudnn.version; |
| 256 |
torch.distributions.constraints.Constraint |
|
paddle.distribution.constraint.Constraint |
仅调用方式不一致; 新增API paddle.distributions.constraints.Constraint; |
| 257 |
torch.distributions.distribution.Distribution.log_prob |
|
paddle.distribution.Distribution.log_prob |
仅调用方式不一致; 新增API paddle.distributions.distribution.Distribution.log_prob; |
| 258 |
torch.distributions.kl.kl_divergence |
|
paddle.distribution.kl_divergence |
仅调用方式不一致; 新增API paddle.distributions.kl.kl_divergence; |
| 259 |
torch.fix |
|
paddle.trunc |
新增out; 新增API paddle.fix; |
| 260 |
torch.histc |
|
paddle.histogram |
新增out; 新增API paddle.histc; |
| 261 |
torch.special.round |
|
paddle.round |
新增out; 新增API paddle.special.round; |
| 262 |
torch.trunc |
|
paddle.trunc |
新增out; |
| 263 |
torch.Tensor.fmod_ |
|
paddle.Tensor.mod_ |
c++下沉或装饰器 {'other': 'y'}; 新增API paddle.Tensor.fmod_; |
| 264 |
torch.Tensor.hardshrink |
|
paddle.nn.functional.hardshrink |
c++下沉或装饰器 {'lambd': 'threshold'}; 仅调用方式不一致; 新增API paddle.Tensor.hardshrink; |
| 265 |
torch.Tensor.igamma |
|
paddle.Tensor.gammainc |
c++下沉或装饰器 {'other': 'y'}; 新增API paddle.Tensor.igamma; |
| 266 |
torch.Tensor.igamma_ |
|
paddle.Tensor.gammainc_ |
c++下沉或装饰器 {'other': 'y'}; 新增API paddle.Tensor.igamma_; |
| 267 |
torch.Tensor.igammac |
|
paddle.Tensor.gammaincc |
c++下沉或装饰器 {'other': 'y'}; 新增API paddle.Tensor.igammac; |
| 268 |
torch.Tensor.igammac_ |
|
paddle.Tensor.gammaincc_ |
c++下沉或装饰器 {'other': 'y'}; 新增API paddle.Tensor.igammac_; |
| 269 |
torch.Tensor.new_tensor |
|
paddle.to_tensor |
c++下沉或装饰器 {'device': 'place'}; 新增API paddle.Tensor.new_tensor; |
| 270 |
torch.Tensor.orgqr |
|
paddle.Tensor.householder_product |
c++下沉或装饰器 {'input2': 'tau'}; 新增API paddle.Tensor.orgqr; |
| 271 |
torch.Tensor.true_divide_ |
|
paddle.Tensor.divide_ |
c++下沉或装饰器 {'other': 'y'}; 新增API paddle.Tensor.true_divide_; |
| 272 |
torch.alpha_dropout |
|
paddle.nn.functional.alpha_dropout |
c++下沉或装饰器 {'input': 'x', 'train': 'training'}; 新增API paddle.alpha_dropout; |
| 273 |
torch.celu |
|
paddle.nn.functional.celu |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.celu; |
| 274 |
torch.cholesky |
|
paddle.linalg.cholesky |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.cholesky; |
| 275 |
torch.cholesky_inverse |
|
paddle.linalg.cholesky_inverse |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.cholesky_inverse; |
| 276 |
torch.cholesky_solve |
|
paddle.linalg.cholesky_solve |
c++下沉或装饰器 {'input': 'x', 'input2': 'y'}; 新增out; 新增API paddle.cholesky_solve; |
| 277 |
torch.clamp_max |
|
paddle.clip |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.clamp_max; |
| 278 |
torch.clamp_min |
|
paddle.clip |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.clamp_min; |
| 279 |
torch.conj_physical |
|
paddle.conj |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.conj_physical; |
| 280 |
torch.cov |
|
paddle.linalg.cov |
c++下沉或装饰器 {'input': 'x', 'correction': 'ddof'}; 新增API paddle.cov; |
| 281 |
torch.det |
|
paddle.linalg.det |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.det; |
| 282 |
torch.dropout |
|
paddle.nn.functional.dropout |
c++下沉或装饰器 {'input': 'x', 'train': 'training'}; 新增API paddle.dropout; |
| 283 |
torch.fmod |
|
paddle.mod |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; 新增API paddle.fmod; |
| 284 |
torch.istft |
|
paddle.signal.istft |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.istft; |
| 285 |
torch.linalg.cholesky_ex |
|
paddle.linalg.cholesky |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.linalg.cholesky_ex; |
| 286 |
torch.linalg.inv_ex |
|
paddle.linalg.inv |
c++下沉或装饰器 {'A': 'x'}; 新增out; 新增API paddle.linalg.inv_ex; |
| 287 |
torch.linalg.lu_factor |
|
paddle.linalg.lu |
c++下沉或装饰器 {'A': 'x'}; 新增out; 新增API paddle.linalg.lu_factor; |
| 288 |
torch.linalg.solve_triangular |
|
paddle.linalg.triangular_solve |
c++下沉或装饰器 {'input': 'x', 'B': 'y', 'left': 'transpose'}; 新增out; 新增API paddle.linalg.solve_triangular; |
| 289 |
torch.lu |
|
paddle.linalg.lu |
c++下沉或装饰器 {'A': 'x'}; 新增out; 新增API paddle.lu; |
| 290 |
torch.lu_unpack |
|
paddle.linalg.lu_unpack |
c++下沉或装饰器 {'LU_data': 'x', 'LU_pivots': 'y', 'unpack_data': 'unpack_ludata'}; 新增out; 新增API paddle.lu_unpack; |
| 291 |
torch.movedim |
|
paddle.moveaxis |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.movedim; |
| 292 |
torch.mvlgamma |
|
paddle.multigammaln |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.mvlgamma; |
| 293 |
torch.negative |
|
paddle.neg |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.negative; |
| 294 |
torch.nn.BatchNorm1d |
|
paddle.nn.BatchNorm1D |
c++下沉或装饰器 {'eps': 'epsilon', 'affine':['weight_attr', 'bias_attr'], 'track_running_stats': '', 'device': '', 'dtype': ''}; 新增API paddle.nn.BatchNorm1d; |
| 295 |
torch.nn.BatchNorm2d |
|
paddle.nn.BatchNorm2D |
c++下沉或装饰器 {'eps': 'epsilon', 'affine': ['weight_attr', 'bias_attr'], 'track_running_stats': '', 'device': '', 'dtype': ''}; 新增API paddle.nn.BatchNorm2d; |
| 296 |
torch.nn.BatchNorm3d |
|
paddle.nn.BatchNorm3D |
c++下沉或装饰器 {'eps': 'epsilon', 'affine': ['weight_attr', 'bias_attr'], 'track_running_stats': '', 'device': '', 'dtype': ''}; 新增API paddle.nn.BatchNorm3d; |
| 297 |
torch.nn.InstanceNorm1d |
|
paddle.nn.InstanceNorm1D |
c++下沉或装饰器 {'eps': 'epsilon', 'affine': ['weight_attr', 'bias_attr'], 'track_running_stats': '', 'device': '', 'dtype': ''}; 新增API paddle.nn.InstanceNorm1d; |
| 298 |
torch.nn.InstanceNorm2d |
|
paddle.nn.InstanceNorm2D |
c++下沉或装饰器 {'eps': 'epsilon', 'affine': ['weight_attr', 'bias_attr'], 'track_running_stats': '', 'device': '', 'dtype': ''}; 新增API paddle.nn.InstanceNorm2d; |
| 299 |
torch.nn.InstanceNorm3d |
|
paddle.nn.InstanceNorm3D |
c++下沉或装饰器 {'eps': 'epsilon', 'affine': ['weight_attr', 'bias_attr'], 'track_running_stats': '', 'device': '', 'dtype': ''}; 新增API paddle.nn.InstanceNorm3d; |
| 300 |
torch.nn.RNN |
|
paddle.nn.SimpleRNN |
c++下沉或装饰器 {'nonlinearity': 'activation', 'bias': ['bias_ih_attr', 'bias_hh_attr'], 'device': '', 'dtype': ''}; 新增API paddle.nn.RNN; |
| 301 |
torch.nn.functional.conv_transpose1d |
|
paddle.nn.functional.conv1d_transpose |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.nn.functional.conv_transpose1d; |
| 302 |
torch.nn.functional.conv_transpose3d |
|
paddle.nn.functional.conv3d_transpose |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.nn.functional.conv_transpose3d; |
| 303 |
torch.nn.functional.huber_loss |
|
paddle.nn.functional.smooth_l1_loss |
c++下沉或装饰器 {'target': 'label'}; 新增API paddle.nn.functional.huber_loss; |
| 304 |
torch.nn.functional.multilabel_margin_loss |
|
paddle.nn.functional.multi_label_margin_loss |
c++下沉或装饰器 {'target': 'label'}; 新增API paddle.nn.functional.multilabel_margin_loss; |
| 305 |
torch.nn.functional.multilabel_soft_margin_loss |
|
paddle.nn.functional.multi_label_soft_margin_loss |
c++下沉或装饰器 {'target': 'label'}; 新增API paddle.nn.functional.multilabel_soft_margin_loss; |
| 306 |
torch.nn.functional.pdist |
|
paddle.pdist |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.nn.functional.pdist; |
| 307 |
torch.nn.functional.rrelu_ |
|
paddle.nn.functional.rrelu |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.nn.functional.rrelu_; |
| 308 |
torch.nn.functional.softmin |
|
paddle.nn.functional.softmax |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis', '_stacklevel': '', 'dtype': 'dtype'}; 新增API paddle.nn.functional.softmin; |
| 309 |
torch.nn.functional.threshold |
|
paddle.nn.functional.thresholded_relu |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.nn.functional.threshold; |
| 310 |
torch.nn.functional.threshold_ |
|
paddle.nn.functional.thresholded_relu_ |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.nn.functional.threshold_; |
| 311 |
torch.nn.functional.upsample_nearest |
|
paddle.nn.functional.upsample |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.nn.functional.upsample_nearest; |
| 312 |
torch.nn.utils.parametrizations.spectral_norm |
|
paddle.nn.utils.spectral_norm |
c++下沉或装饰器 {'module': 'layer'}; 新增API paddle.nn.utils.parametrizations.spectral_norm; |
| 313 |
torch.nn.utils.parametrizations.weight_norm |
|
paddle.nn.utils.weight_norm |
c++下沉或装饰器 {'module': 'layer'}; 新增API paddle.nn.utils.parametrizations.weight_norm; |
| 314 |
torch.ormqr |
|
paddle.linalg.ormqr |
c++下沉或装饰器 {'input': 'x', 'input2': 'tau', 'input3': 'y'}; 新增out; 新增API paddle.ormqr; |
| 315 |
torch.pca_lowrank |
|
paddle.linalg.pca_lowrank |
c++下沉或装饰器 {'A': 'x'}; 新增API paddle.pca_lowrank; |
| 316 |
torch.pinverse |
|
paddle.linalg.pinv |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.pinverse; |
| 317 |
torch.qr |
|
paddle.linalg.qr |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.qr; |
| 318 |
torch.rrelu |
|
paddle.nn.functional.rrelu |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.rrelu; |
| 319 |
torch.selu |
|
paddle.nn.functional.selu |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.selu; |
| 320 |
torch.set_default_tensor_type |
|
paddle.set_default_dtype |
c++下沉或装饰器 {'t': 'd'}; 新增API paddle.set_default_tensor_type; |
| 321 |
torch.sparse.FloatTensor |
|
paddle.sparse.sparse_coo_tensor |
c++下沉或装饰器 {'size': 'shape', 'device': 'place'}; 新增API paddle.sparse.FloatTensor; |
| 322 |
torch.sparse.mm |
|
paddle.sparse.matmul |
c++下沉或装饰器 {'sparse': 'x', 'dense': 'y'}; 新增API paddle.sparse.mm; |
| 323 |
torch.sparse.softmax |
|
paddle.sparse.nn.functional.softmax |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis'}; 新增API paddle.sparse.softmax; |
| 324 |
torch.sparse_coo_tensor |
|
paddle.sparse.sparse_coo_tensor |
c++下沉或装饰器 {'size': 'shape', 'dtype': 'dtype', 'device': 'place', 'check_invariants': ''}; 新增API paddle.sparse_coo_tensor; |
| 325 |
torch.sparse_csr_tensor |
|
paddle.sparse.sparse_csr_tensor |
c++下沉或装饰器 {'crow_indices': 'crows', 'col_indices': 'cols', 'size': 'shape', 'dtype': 'dtype', 'device': 'place', 'pin_memory': '', 'check_invariants': ''}; 新增API paddle.sparse_csr_tensor; |
| 326 |
torch.special.digamma |
@lllijinggg1110-rgb #78113
|
paddle.digamma |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.digamma; |
| 327 |
torch.special.erf |
|
paddle.erf |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.erf; |
| 328 |
torch.special.erfinv |
|
paddle.erfinv |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.erfinv; |
| 329 |
torch.special.gammainc |
|
paddle.gammainc |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; 新增API paddle.special.gammainc; |
| 330 |
torch.special.gammaincc |
|
paddle.gammaincc |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; 新增API paddle.special.gammaincc; |
| 331 |
torch.special.gammaln |
|
paddle.gammaln |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.gammaln; |
| 332 |
torch.special.i0 |
@enkilee #77211 #7705
|
paddle.i0 |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.i0; |
| 333 |
torch.special.i0e |
@enkilee #77211 #7705
|
paddle.i0e |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.i0e; |
| 334 |
torch.special.i1 |
@enkilee #77218 #7682 #806
|
paddle.i1 |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.i1; |
| 335 |
torch.special.i1e |
@enkilee #77218 #7682 #806
|
paddle.i1e |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.i1e; |
| 336 |
torch.special.log1p |
@enkilee #78092
|
paddle.log1p |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.log1p; |
| 337 |
torch.special.log_softmax |
|
paddle.nn.functional.log_softmax |
c++下沉或装饰器 {'input': 'x', 'dim': 'axis', 'dtype': 'dtype'}; 新增API paddle.special.log_softmax; |
| 338 |
torch.special.multigammaln |
|
paddle.multigammaln |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.multigammaln; |
| 339 |
torch.special.polygamma |
|
paddle.polygamma |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.polygamma; |
| 340 |
torch.special.psi |
|
paddle.digamma |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.psi; |
| 341 |
torch.special.sinc |
|
paddle.sinc |
c++下沉或装饰器 {'input': 'x'}; 新增out; 新增API paddle.special.sinc; |
| 342 |
torch.stft |
|
paddle.signal.stft |
c++下沉或装饰器 {'input': 'x'}; 新增API paddle.stft; |
| 343 |
torch.svd_lowrank |
|
paddle.linalg.svd_lowrank |
c++下沉或装饰器 {'A': 'x'}; 新增API paddle.svd_lowrank; |
| 344 |
torch.symeig |
|
paddle.linalg.eigh |
c++下沉或装饰器 {'input': 'x', 'upper': 'UPLO'}; 新增out; 新增API paddle.symeig; |
| 345 |
torch.triangular_solve |
|
paddle.linalg.triangular_solve |
c++下沉或装饰器 {'input': 'y', 'A': 'x'}; 新增out; 新增API paddle.triangular_solve; |
| 346 |
torch.vdot |
|
paddle.dot |
c++下沉或装饰器 {'input': 'x', 'other': 'y'}; 新增out; 新增API paddle.vdot; |
背景
为了降低新模型(特别是新的大模型)使用飞桨开发或迁移到飞桨的成本,飞桨在3.2版本开展了 API兼容性适配 工作,提升了API针对不同框架写法的自适应能力,针对Pytorch项目,仅需修改接口代码前缀
torch.为paddle.,即可无缝迁移到Paddle。当前已完成了500+个重点API与Pytorch API的无缝兼容工作(包括:API名称、参数名称、参数个数、参数语义),具体名单详见 PyTorch-Paddle API 映射表。但仍有很多Paddle API尚未完成兼容性适配 工作,本次活动旨在对Paddle API进行整体兼容性推全。
任务讲解:https://meeting.tencent.com/cw/KP4g34Y9f3 (第5分钟开始)
示例修改&提交方式&FAQ
#76301 (comment)
认领方式
请大家以 comment 的形式认领任务,如:
任务表
torch.Tensor.addmm_@enkilee
torch.Tensor.baddbmm_torch.Tensor.bitwise_and_@ADchampion3
torch.Tensor.bitwise_left_shift_torch.Tensor.bitwise_right_shift_torch.Tensor.bitwise_xor_torch.Tensor.cauchy_torch.Tensor.copysign_torch.Tensor.cumsum_torch.Tensor.gcd_torch.Tensor.geometric_torch.Tensor.hypot_torch.Tensor.index_fill_torch.Tensor.index_put_torch.Tensor.lcm_torch.Tensor.lerp_torch.Tensor.renorm_torch.abs_@algorithm1832
torch.acostorch.acoshtorch.addmmtorch.angletorch.as_stridedtorch.asintorch.asinhtorch.atantorch.atan2torch.atanhtorch.baddbmmtorch.bincounttorch.bitwise_andtorch.bitwise_left_shifttorch.bitwise_nottorch.bitwise_right_shifttorch.bitwise_xortorch.block_diagtorch.broadcast_tensorstorch.cartesian_prodtorch.cdist@enkilee
torch.conjtorch.copysigntorch.coshtorch.crosstorch.diagtorch.disttorch.fliptorch.fmaxtorch.fmintorch.fractorch.heaviside@enkilee
torch.index_addtorch.index_filltorch.log10torch.log1ptorch.nextaftertorch.nn.CosineSimilaritytorch.nn.Foldtorch.nn.GLUtorch.nn.Hardshrinktorch.nn.PairwiseDistancetorch.nn.Softshrinktorch.nn.functional.adaptive_avg_pool2dtorch.nn.functional.adaptive_avg_pool3dtorch.nn.functional.adaptive_max_pool1dtorch.nn.functional.adaptive_max_pool2dtorch.nn.functional.adaptive_max_pool3dtorch.nn.functional.cosine_similaritytorch.nn.functional.foldtorch.nn.functional.glutorch.nn.functional.hardshrinktorch.nn.functional.leaky_relu_torch.nn.functional.lp_pool1dtorch.nn.functional.lp_pool2dtorch.nn.functional.pairwise_distancetorch.nn.functional.relu_torch.nn.functional.softshrinktorch.reciprocaltorch.renormtorch.searchsortedtorch.sinhtorch.tiletorch.unflatten@wuhaostudio
@wuhaostudio
torch.special.i0torch.special.i0etorch.special.i1torch.special.i1e该任务表正在内部同步开发中,请遵循以下规则:
ChangePrefixMatcher的API(已完成验收)不要认领!看板信息
统计信息