Math friendly syntax
This Section contains the full API doc of the Math-friendly aliases.
Summary
|
Alias for |
|
Alias for |
|
Alias for |
|
Alias for |
Syntax
- pykeops.torch.generic_sum(formula, output, *aliases, **kwargs)[source]
Alias for
torch.Genred
with a “Sum” reduction.- Parameters:
formula (string) – Symbolic KeOps expression, as in
torch.Genred
.output (string) –
An identifier of the form
"AL = TYPE(DIM)"
that specifies the category and dimension of the output variable. Here:AL
is a dummy alphanumerical name.TYPE
is a category. One of:Vi
: indexation by \(i\) along axis 0; reduction is performed along axis 1.Vj
: indexation by \(j\) along axis 1; reduction is performed along axis 0.
DIM
is an integer, the dimension of the output variable; it should be compatible with formula.
*aliases (strings) – List of identifiers, as in
torch.Genred
.
Keyword Args:
- Returns:
A generic reduction that can be called on arbitrary Torch tensors, as documented in
torch.Genred
.
Example
>>> my_conv = generic_sum( # Custom Kernel Density Estimator ... 'Exp(-SqNorm2(x - y))', # Formula ... 'a = Vi(1)', # Output: 1 scalar per line ... 'x = Vi(3)', # 1st input: dim-3 vector per line ... 'y = Vj(3)') # 2nd input: dim-3 vector per line >>> # Apply it to 2d arrays x and y with 3 columns and a (huge) number of lines >>> x = torch.randn(1000000, 3, requires_grad=True).cuda() >>> y = torch.randn(2000000, 3).cuda() >>> a = my_conv(x, y) # a_i = sum_j exp(-|x_i-y_j|^2) >>> print(a.shape) torch.Size([1000000, 1])
- pykeops.torch.generic_logsumexp(formula, output, *aliases, **kwargs)[source]
Alias for
torch.Genred
with a “LogSumExp” reduction.- Parameters:
formula (string) – Scalar-valued symbolic KeOps expression, as in
torch.Genred
.output (string) –
An identifier of the form
"AL = TYPE(1)"
that specifies the category and dimension of the output variable. Here:AL
is a dummy alphanumerical name.TYPE
is a category. One of:Vi
: indexation by \(i\) along axis 0; reduction is performed along axis 1.Vj
: indexation by \(j\) along axis 1; reduction is performed along axis 0.
*aliases (strings) – List of identifiers, as in
torch.Genred
.
Keyword Args:
- Returns:
A generic reduction that can be called on arbitrary Torch tensors, as documented in
torch.Genred
.
Example
Log-likelihood of a Gaussian Mixture Model,
\[\begin{split}a_i~=~f(x_i)~&=~ \log \sum_{j=1}^{N} \exp(-\gamma\cdot\|x_i-y_j\|^2)\cdot b_j \\\\ ~&=~ \log \sum_{j=1}^{N} \exp\big(-\gamma\cdot\|x_i-y_j\|^2 \,+\, \log(b_j) \big).\end{split}\]>>> log_likelihood = generic_logsumexp( ... '(-(g * SqNorm2(x - y))) + b', # Formula ... 'a = Vi(1)', # Output: 1 scalar per line ... 'x = Vi(3)', # 1st input: dim-3 vector per line ... 'y = Vj(3)', # 2nd input: dim-3 vector per line ... 'g = Pm(1)', # 3rd input: vector of size 1 ... 'b = Vj(1)') # 4th input: 1 scalar per line >>> x = torch.randn(1000000, 3, requires_grad=True).cuda() >>> y = torch.randn(2000000, 3).cuda() >>> g = torch.Tensor([.5]).cuda() # Parameter of our GMM >>> b = torch.rand(2000000, 1).cuda() # Positive weights... >>> b = b / b.sum() # Normalized to get a probability measure >>> a = log_likelihood(x, y, g, b.log()) # a_i = log sum_j exp(-g*|x_i-y_j|^2) * b_j >>> print(a.shape) torch.Size([1000000, 1])
- pykeops.torch.generic_argmin(formula, output, *aliases, **kwargs)[source]
Alias for
torch.Genred
with an “ArgMin” reduction.- Parameters:
formula (string) – Scalar-valued symbolic KeOps expression, as in
torch.Genred
.output (string) –
An identifier of the form
"AL = TYPE(1)"
that specifies the category and dimension of the output variable. Here:AL
is a dummy alphanumerical name.TYPE
is a category. One of:Vi
: indexation by \(i\) along axis 0; reduction is performed along axis 1.Vj
: indexation by \(j\) along axis 1; reduction is performed along axis 0.
*aliases (strings) – List of identifiers, as in
torch.Genred
.
Keyword Args:
- Returns:
A generic reduction that can be called on arbitrary Torch tensors, as documented in
torch.Genred
.
Example
Bruteforce nearest neighbor search in dimension 100:
>>> nearest_neighbor = generic_argmin( ... 'SqDist(x, y)', # Formula ... 'a = Vi(1)', # Output: 1 scalar per line ... 'x = Vi(100)', # 1st input: dim-100 vector per line ... 'y = Vj(100)') # 2nd input: dim-100 vector per line >>> x = torch.randn(5, 100) >>> y = torch.randn(20000, 100) >>> a = nearest_neighbor(x, y) >>> print(a) tensor([[ 8761.], [ 2836.], [ 906.], [16130.], [ 3158.]]) >>> dists = (x - y[ a.view(-1).long() ] ).norm(dim=1) # Distance to the nearest neighbor >>> print(dists) tensor([10.5926, 10.9132, 9.9694, 10.1396, 10.1955])
- pykeops.torch.generic_argkmin(formula, output, *aliases, **kwargs)[source]
Alias for
torch.Genred
with an “ArgKMin” reduction.- Parameters:
formula (string) – Scalar-valued symbolic KeOps expression, as in
torch.Genred
.output (string) –
An identifier of the form
"AL = TYPE(K)"
that specifies the category and dimension of the output variable. Here:AL
is a dummy alphanumerical name.TYPE
is a category. One of:Vi
: indexation by \(i\) along axis 0; reduction is performed along axis 1.Vj
: indexation by \(j\) along axis 1; reduction is performed along axis 0.
K
is an integer, the number of values to extract.
*aliases (strings) – List of identifiers, as in
torch.Genred
.
Keyword Args:
- Returns:
A generic reduction that can be called on arbitrary Torch tensors, as documented in
torch.Genred
.
Example
Bruteforce K-nearest neighbors search in dimension 100:
>>> knn = generic_argkmin( ... 'SqDist(x, y)', # Formula ... 'a = Vi(3)', # Output: 3 scalars per line ... 'x = Vi(100)', # 1st input: dim-100 vector per line ... 'y = Vj(100)') # 2nd input: dim-100 vector per line >>> x = torch.randn(5, 100) >>> y = torch.randn(20000, 100) >>> a = knn(x, y) >>> print(a) tensor([[ 9054., 11653., 11614.], [13466., 11903., 14180.], [14164., 8809., 3799.], [ 2092., 3323., 18479.], [14433., 11315., 11841.]]) >>> print( (x - y[ a[:,0].long() ]).norm(dim=1) ) # Distance to the nearest neighbor tensor([10.7933, 10.3235, 10.1218, 11.4919, 10.5100]) >>> print( (x - y[ a[:,1].long() ]).norm(dim=1) ) # Distance to the second neighbor tensor([11.3702, 10.6550, 10.7646, 11.5676, 11.1356]) >>> print( (x - y[ a[:,2].long() ]).norm(dim=1) ) # Distance to the third neighbor tensor([11.3820, 10.6725, 10.8510, 11.6071, 11.1968])