-
Notifications
You must be signed in to change notification settings - Fork 2.1k
Derive logprob of matmul #7542
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Derive logprob of matmul #7542
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,102 @@ | ||
# Copyright 2024 The PyMC Developers | ||
# | ||
# Licensed under the Apache License, Version 2.0 (the "License"); | ||
# you may not use this file except in compliance with the License. | ||
# You may obtain a copy of the License at | ||
# | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# | ||
# Unless required by applicable law or agreed to in writing, software | ||
# distributed under the License is distributed on an "AS IS" BASIS, | ||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
# See the License for the specific language governing permissions and | ||
# limitations under the License. | ||
import pytensor.tensor as pt | ||
|
||
from pytensor.graph.rewriting.basic import node_rewriter | ||
from pytensor.tensor.math import _matrix_matrix_matmul | ||
|
||
from pymc.logprob.abstract import MeasurableBlockwise, MeasurableOp, _logprob, _logprob_helper | ||
from pymc.logprob.rewriting import measurable_ir_rewrites_db | ||
from pymc.logprob.utils import check_potential_measurability, filter_measurable_variables | ||
|
||
|
||
class MeasurableMatMul(MeasurableBlockwise): | ||
"""Measurable matrix multiplication operation.""" | ||
|
||
right_measurable: bool | ||
|
||
def __init__(self, measurable_right: bool, **kwargs): | ||
self.right_measurable = measurable_right | ||
super().__init__(**kwargs) | ||
|
||
|
||
@_logprob.register(MeasurableMatMul) | ||
def logprob_measurable_matmul(op, values, l, r): # noqa: E741 | ||
[y_value] = values | ||
if op.right_measurable: | ||
A, x = l, r | ||
x_value = pt.linalg.solve(A, y_value) | ||
else: | ||
x, A = l, r | ||
x_value = pt.linalg.solve(A.mT, y_value.mT).mT | ||
|
||
x_logp = _logprob_helper(x, x_value) | ||
|
||
# The operation has a support dimensionality of 2 | ||
# We need to reduce it if it's still present in the base logp | ||
if x_logp.type.ndim == x_value.type.ndim: | ||
x_logp = pt.sum(x_logp, axis=(-1, -2)) | ||
elif x_logp.type.ndim == x_value.type.ndim - 1: | ||
x_logp = pt.sum(x_logp, axis=-1) | ||
|
||
_, log_abs_jac_det = pt.linalg.slogdet(A) | ||
|
||
return x_logp - log_abs_jac_det | ||
|
||
|
||
@node_rewriter(tracks=[_matrix_matrix_matmul]) | ||
def find_measurable_matmul(fgraph, node): | ||
"""Find measurable matrix-matrix multiplication operations.""" | ||
if isinstance(node.op, MeasurableOp): | ||
return None | ||
|
||
[out] = node.outputs | ||
[l, r] = node.inputs # noqa: E741 | ||
|
||
# Check that not both a and r are measurable | ||
measurable_inputs = filter_measurable_variables([l, r]) | ||
if len(measurable_inputs) != 1: | ||
return None | ||
|
||
[measurable_input] = measurable_inputs | ||
|
||
# Check the measurable input is not broadcasted | ||
if measurable_input.type.broadcastable[:-2] != out.type.broadcastable[:-2]: | ||
return None | ||
|
||
measurable_right = measurable_input is r | ||
A = l if measurable_right else r | ||
|
||
# Check if the static shape already reveals a non-square matrix, | ||
if ( | ||
A.type.shape[-1] is not None | ||
and A.type.shape[-2] is not None | ||
and A.type.shape[-1] != A.type.shape[-2] | ||
): | ||
return None | ||
|
||
# Check the other input is not potentially measurable | ||
if check_potential_measurability([A]): | ||
return None | ||
|
||
measurable_matmul = MeasurableMatMul(measurable_right=measurable_right, **node.op._props_dict()) | ||
return [measurable_matmul(l, r)] | ||
|
||
|
||
measurable_ir_rewrites_db.register( | ||
find_measurable_matmul.__name__, | ||
find_measurable_matmul, | ||
"basic", | ||
"linalg", | ||
) |
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -41,13 +41,19 @@ | |
from pytensor.graph.rewriting.basic import node_rewriter | ||
from pytensor.tensor import TensorVariable | ||
from pytensor.tensor.basic import Join, MakeVector | ||
from pytensor.tensor.elemwise import DimShuffle | ||
from pytensor.tensor.elemwise import DimShuffle, Elemwise | ||
from pytensor.tensor.random.op import RandomVariable | ||
from pytensor.tensor.random.rewriting import ( | ||
local_dimshuffle_rv_lift, | ||
) | ||
|
||
from pymc.logprob.abstract import MeasurableOp, _logprob, _logprob_helper, promised_valued_rv | ||
from pymc.logprob.abstract import ( | ||
MeasurableOp, | ||
ValuedRV, | ||
_logprob, | ||
_logprob_helper, | ||
promised_valued_rv, | ||
) | ||
from pymc.logprob.rewriting import ( | ||
assume_valued_outputs, | ||
early_measurable_ir_rewrites_db, | ||
|
@@ -57,6 +63,7 @@ | |
from pymc.logprob.utils import ( | ||
check_potential_measurability, | ||
filter_measurable_variables, | ||
get_related_valued_nodes, | ||
replace_rvs_by_values, | ||
) | ||
from pymc.pytensorf import constant_fold | ||
|
@@ -183,6 +190,9 @@ class MeasurableDimShuffle(MeasurableOp, DimShuffle): | |
# find it locally and fails when a new `Op` is initialized | ||
c_func_file = str(DimShuffle.get_path(Path(DimShuffle.c_func_file))) # type: ignore[arg-type] | ||
|
||
def __str__(self): | ||
return f"Measurable{super().__str__()}" | ||
|
||
|
||
@_logprob.register(MeasurableDimShuffle) | ||
def logprob_dimshuffle(op: MeasurableDimShuffle, values, base_var, **kwargs): | ||
|
@@ -215,29 +225,66 @@ def logprob_dimshuffle(op: MeasurableDimShuffle, values, base_var, **kwargs): | |
return raw_logp.dimshuffle(redo_ds) | ||
|
||
|
||
def _elemwise_univariate_chain(fgraph, node) -> bool: | ||
# Check whether only Elemwise operations connect a base univariate RV to the valued node through var. | ||
from pymc.distributions.distribution import SymbolicRandomVariable | ||
from pymc.logprob.transforms import MeasurableTransform | ||
|
||
[inp] = node.inputs | ||
[out] = node.outputs | ||
|
||
def elemwise_root(var: TensorVariable) -> TensorVariable | None: | ||
if isinstance(var.owner.op, RandomVariable | SymbolicRandomVariable): | ||
return var | ||
elif isinstance(var.owner.op, MeasurableTransform): | ||
return elemwise_root(var.owner.inputs[var.owner.op.measurable_input_idx]) | ||
else: | ||
return None | ||
|
||
# Check that the root is a univariate distribution linked by only elemwise operations | ||
root = elemwise_root(inp) | ||
if root is None: | ||
return False | ||
elif root.owner.op.ndim_supp != 0: | ||
# This is still fine if the variable is directly valued | ||
return any(get_related_valued_nodes(fgraph, node)) | ||
|
||
def elemwise_leaf(var: TensorVariable, clients=fgraph.clients) -> bool: | ||
var_clients = clients[var] | ||
if len(var_clients) != 1: | ||
return False | ||
[(client, _)] = var_clients | ||
if isinstance(client.op, ValuedRV): | ||
return True | ||
elif isinstance(client.op, Elemwise) and len(client.outputs) == 1: | ||
return elemwise_leaf(client.outputs[0]) | ||
else: | ||
return False | ||
|
||
# Check that the path to the valued node consists only of elemwise operations | ||
return elemwise_leaf(out) | ||
|
||
|
||
@node_rewriter([DimShuffle]) | ||
def find_measurable_dimshuffles(fgraph, node) -> list[TensorVariable] | None: | ||
r"""Find `Dimshuffle`\s for which a `logprob` can be computed.""" | ||
from pymc.distributions.distribution import SymbolicRandomVariable | ||
|
||
if isinstance(node.op, MeasurableOp): | ||
return None | ||
|
||
if not filter_measurable_variables(node.inputs): | ||
return None | ||
|
||
base_var = node.inputs[0] | ||
# In cases where DimShuffle transposes dimensions, we only apply this rewrite when only Elemwise | ||
# operations separate it from the valued node. Further transformations likely need to know where | ||
# the support axes are for a correct implementation (and thus assume they are the rightmost axes). | ||
# TODO: When we include the support axis as meta information in each intermediate MeasurableVariable, | ||
# we can lift this restriction (see https://github.com/pymc-devs/pymc/issues/6360) | ||
if tuple(node.op.shuffle) != tuple(sorted(node.op.shuffle)) and not _elemwise_univariate_chain( | ||
fgraph, node | ||
): | ||
return None | ||
|
||
|
||
# We can only apply this rewrite directly to `RandomVariable`s, as those are | ||
# the only `Op`s for which we always know the support axis. Other measurable | ||
# variables can have arbitrary support axes (e.g., if they contain separate | ||
# `MeasurableDimShuffle`s). Most measurable variables with `DimShuffle`s | ||
# should still be supported as long as the `DimShuffle`s can be merged/ | ||
# lifted towards the base RandomVariable. | ||
# TODO: If we include the support axis as meta information in each | ||
# intermediate MeasurableVariable, we can lift this restriction. | ||
if not isinstance(base_var.owner.op, RandomVariable | SymbolicRandomVariable): | ||
return None # pragma: no cover | ||
base_var = node.inputs[0] | ||
|
||
measurable_dimshuffle = MeasurableDimShuffle(node.op.input_broadcastable, node.op.new_order)( | ||
base_var | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -320,7 +320,7 @@ def find_negated_var(var): | |
return None | ||
|
||
|
||
def get_related_valued_nodes(node: Apply, fgraph: FunctionGraph) -> list[Apply]: | ||
def get_related_valued_nodes(fgraph: FunctionGraph, node: Apply) -> list[Apply]: | ||
|
||
"""Get all ValuedVars related to the same RV node. | ||
|
||
Returns | ||
|
Uh oh!
There was an error while loading. Please reload this page.