Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
c3347b98
提交
c3347b98
authored
12月 18, 2025
作者:
jessegrabowski
提交者:
Ricardo Vieira
12月 19, 2025
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Implement `SplitDims` and `JoinDims`
上级
9ac65794
隐藏空白字符变更
内嵌
并排
正在显示
3 个修改的文件
包含
368 行增加
和
0 行删除
+368
-0
__init__.py
pytensor/tensor/__init__.py
+7
-0
reshape.py
pytensor/tensor/reshape.py
+264
-0
test_reshape.py
tests/tensor/test_reshape.py
+97
-0
没有找到文件。
pytensor/tensor/__init__.py
浏览文件 @
c3347b98
...
...
@@ -133,6 +133,11 @@ from pytensor.tensor.extra_ops import *
from
pytensor.tensor.interpolate
import
interp
,
interpolate1d
from
pytensor.tensor.math
import
*
from
pytensor.tensor.pad
import
pad
# isort: off
# reshape needs to be imported before shape.reshape, otherwise the tensor.reshape imports fail
from
pytensor.tensor.reshape
import
*
from
pytensor.tensor.shape
import
(
reshape
,
shape
,
...
...
@@ -143,6 +148,8 @@ from pytensor.tensor.shape import (
specify_shape
,
)
# isort: on
# We import as `_shared` instead of `shared` to avoid confusion between
# `pytensor.shared` and `tensor._shared`.
from
pytensor.tensor.sort
import
argsort
,
sort
...
...
pytensor/tensor/reshape.py
0 → 100644
浏览文件 @
c3347b98
from
collections.abc
import
Sequence
from
typing
import
cast
as
type_cast
import
numpy
as
np
from
numpy.lib._array_utils_impl
import
normalize_axis_tuple
from
pytensor
import
Variable
from
pytensor.graph
import
Apply
from
pytensor.graph.op
import
Op
from
pytensor.graph.replace
import
_vectorize_node
from
pytensor.tensor
import
TensorLike
,
as_tensor_variable
from
pytensor.tensor.basic
import
infer_static_shape
from
pytensor.tensor.math
import
prod
from
pytensor.tensor.shape
import
ShapeValueType
from
pytensor.tensor.type
import
tensor
from
pytensor.tensor.variable
import
TensorVariable
class
JoinDims
(
Op
):
__props__
=
(
"start_axis"
,
"n_axes"
,
)
view_map
=
{
0
:
[
0
]}
def
__init__
(
self
,
start_axis
:
int
,
n_axes
:
int
):
if
start_axis
<
0
:
raise
ValueError
(
"JoinDims start_axis must be non-negative"
)
if
n_axes
<
0
:
raise
ValueError
(
"JoinDims n_axes must be non-negative"
)
self
.
start_axis
=
start_axis
self
.
n_axes
=
n_axes
@property
def
axis_range
(
self
):
return
range
(
self
.
start_axis
,
self
.
start_axis
+
self
.
n_axes
)
def
output_shapes
(
self
,
input_shapes
,
joined_shape
):
return
(
*
input_shapes
[:
self
.
start_axis
],
joined_shape
,
*
input_shapes
[
self
.
start_axis
+
self
.
n_axes
:],
)
def
make_node
(
self
,
x
:
Variable
)
->
Apply
:
# type: ignore[override]
x
=
as_tensor_variable
(
x
)
static_shapes
=
x
.
type
.
shape
axis_range
=
self
.
axis_range
joined_shape
=
(
int
(
np
.
prod
([
static_shapes
[
i
]
for
i
in
axis_range
]))
if
all
(
static_shapes
[
i
]
is
not
None
for
i
in
axis_range
)
else
None
)
output_shapes
=
self
.
output_shapes
(
static_shapes
,
joined_shape
)
output_type
=
tensor
(
shape
=
output_shapes
,
dtype
=
x
.
type
.
dtype
)
return
Apply
(
self
,
[
x
],
[
output_type
])
def
infer_shape
(
self
,
fgraph
,
node
,
shapes
):
[
input_shape
]
=
shapes
axis_range
=
self
.
axis_range
joined_shape
=
prod
([
input_shape
[
i
]
for
i
in
axis_range
])
return
[
self
.
output_shapes
(
input_shape
,
joined_shape
)]
def
perform
(
self
,
node
,
inputs
,
outputs
):
(
x
,)
=
inputs
(
out
,)
=
outputs
output_shape
=
(
*
x
.
shape
[:
self
.
start_axis
],
-
1
,
*
x
.
shape
[
self
.
start_axis
+
self
.
n_axes
:],
)
out
[
0
]
=
x
.
reshape
(
output_shape
)
@_vectorize_node.register
(
JoinDims
)
def
_vectorize_joindims
(
op
,
node
,
x
):
[
old_x
]
=
node
.
inputs
batched_ndims
=
x
.
type
.
ndim
-
old_x
.
type
.
ndim
start_axis
=
op
.
start_axis
n_axes
=
op
.
n_axes
return
JoinDims
(
start_axis
+
batched_ndims
,
n_axes
)
.
make_node
(
x
)
def
join_dims
(
x
:
TensorLike
,
axis
:
Sequence
[
int
]
|
int
|
None
=
None
)
->
TensorVariable
:
"""Join consecutive dimensions of a tensor into a single dimension.
Parameters
----------
x : Variable
The input tensor.
axis : int or sequence of int, optional
The dimensions to join. If None, all dimensions are joined.
Returns
-------
joined_x : Variable
The reshaped tensor with joined dimensions.
Examples
--------
>>> import pytensor.tensor as pt
>>> x = pt.tensor("x", shape=(2, 3, 4, 5))
>>> y = pt.join_dims(x, axis=(1, 2))
>>> y.type.shape
(2, 12, 5)
"""
x
=
as_tensor_variable
(
x
)
if
axis
is
None
:
axis
=
list
(
range
(
x
.
ndim
))
elif
isinstance
(
axis
,
int
):
axis
=
[
axis
]
elif
not
isinstance
(
axis
,
list
|
tuple
):
raise
TypeError
(
"axis must be an int, a list/tuple of ints, or None"
)
axis
=
normalize_axis_tuple
(
axis
,
x
.
ndim
)
if
len
(
axis
)
<=
1
:
return
x
# type: ignore[unreachable]
if
np
.
diff
(
axis
)
.
max
()
>
1
:
raise
ValueError
(
f
"join_dims axis must be consecutive, got normalized axis: {axis}"
)
start_axis
=
min
(
axis
)
n_axes
=
len
(
axis
)
return
type_cast
(
TensorVariable
,
JoinDims
(
start_axis
=
start_axis
,
n_axes
=
n_axes
)(
x
),
)
class
SplitDims
(
Op
):
__props__
=
(
"axis"
,)
view_map
=
{
0
:
[
0
]}
def
__init__
(
self
,
axis
:
int
):
if
axis
<
0
:
raise
ValueError
(
"SplitDims axis must be non-negative"
)
self
.
axis
=
axis
def
make_node
(
self
,
x
:
Variable
,
shape
:
Variable
)
->
Apply
:
# type: ignore[override]
if
shape
.
type
.
numpy_dtype
.
kind
not
in
"iu"
:
raise
TypeError
(
"shape must be an integer tensor"
)
x
=
as_tensor_variable
(
x
)
shape
=
as_tensor_variable
(
shape
,
dtype
=
int
,
ndim
=
1
)
axis
=
self
.
axis
_
,
constant_shape
=
infer_static_shape
(
shape
)
output_shapes
=
[
*
x
.
type
.
shape
[:
axis
],
*
constant_shape
,
*
x
.
type
.
shape
[
axis
+
1
:],
]
output
=
tensor
(
shape
=
tuple
(
x
if
isinstance
(
x
,
int
)
else
None
for
x
in
output_shapes
),
dtype
=
x
.
type
.
dtype
,
)
return
Apply
(
self
,
[
x
,
shape
],
[
output
])
def
infer_shape
(
self
,
fgraph
,
node
,
shapes
):
[
input_shape
,
_
]
=
shapes
_
,
shape
=
node
.
inputs
output_shapes
=
list
(
input_shape
)
axis
=
self
.
axis
inferred_shape
=
[
*
output_shapes
[:
axis
],
*
shape
,
*
output_shapes
[
axis
+
1
:]]
return
[
inferred_shape
]
def
perform
(
self
,
node
,
inputs
,
outputs
):
(
x
,
shape
)
=
inputs
(
out
,)
=
outputs
output_shape
=
(
*
x
.
shape
[:
self
.
axis
],
*
shape
,
*
x
.
shape
[
self
.
axis
+
1
:])
out
[
0
]
=
x
.
reshape
(
output_shape
)
@_vectorize_node.register
(
SplitDims
)
def
_vectorize_splitdims
(
op
,
node
,
x
,
shape
):
from
pytensor.tensor.blockwise
import
vectorize_node_fallback
old_x
,
_
=
node
.
inputs
batched_ndims
=
x
.
type
.
ndim
-
old_x
.
type
.
ndim
if
as_tensor_variable
(
shape
)
.
type
.
ndim
!=
1
:
return
vectorize_node_fallback
(
op
,
node
,
x
,
shape
)
axis
=
op
.
axis
return
SplitDims
(
axis
=
axis
+
batched_ndims
)
.
make_node
(
x
,
shape
)
def
split_dims
(
x
:
TensorLike
,
shape
:
ShapeValueType
|
Sequence
[
ShapeValueType
],
axis
:
int
|
None
=
None
,
)
->
TensorVariable
:
"""Split a dimension of a tensor into multiple dimensions.
Parameters
----------
x : TensorLike
The input tensor.
shape : int or sequence of int
The new shape to split the specified dimension into.
axis : int, optional
The dimension to split. If None, the input is assumed to be 1D and axis 0 is used.
Returns
-------
split_x : Variable
The reshaped tensor with split dimensions.
Examples
--------
>>> import pytensor.tensor as pt
>>> x = pt.tensor("x", shape=(6, 4, 6))
>>> y = pt.split_dims(x, shape=(2, 3), axis=0)
>>> y.type.shape
(2, 3, 4, 6)
"""
x
=
as_tensor_variable
(
x
)
if
axis
is
None
:
if
x
.
ndim
!=
1
:
raise
ValueError
(
"split_dims can only be called with axis=None for 1d inputs"
)
axis
=
0
if
isinstance
(
shape
,
int
):
shape
=
[
shape
]
else
:
shape
=
list
(
shape
)
# type: ignore[arg-type]
if
not
shape
:
# If we get an empty shape, there is potentially a dummy dimension at the requested axis. This happens for
# example when splitting a packed tensor that had its dims expanded before packing (e.g. when packing shapes
# (3, ) and (3, 3) to (3, 4)
return
type_cast
(
TensorVariable
,
x
.
squeeze
(
axis
=
axis
))
[
axis
]
=
normalize_axis_tuple
(
axis
,
x
.
ndim
)
# type: ignore[misc]
shape
=
as_tensor_variable
(
shape
,
dtype
=
"int64"
,
ndim
=
1
)
# type: ignore[arg-type]
split_op
=
SplitDims
(
axis
=
axis
)
return
type_cast
(
TensorVariable
,
split_op
(
x
,
shape
))
__all__
=
[
"join_dims"
,
"split_dims"
]
tests/tensor/test_reshape.py
0 → 100644
浏览文件 @
c3347b98
import
numpy
as
np
import
pytest
from
pytensor
import
config
,
function
from
pytensor
import
tensor
as
pt
from
pytensor.graph
import
vectorize_graph
from
pytensor.tensor.reshape
import
(
join_dims
,
split_dims
,
)
def
test_join_dims
():
rng
=
np
.
random
.
default_rng
()
x
=
pt
.
tensor
(
"x"
,
shape
=
(
2
,
3
,
4
,
5
))
assert
join_dims
(
x
,
axis
=
(
0
,
1
))
.
type
.
shape
==
(
6
,
4
,
5
)
assert
join_dims
(
x
,
axis
=
(
1
,
2
))
.
type
.
shape
==
(
2
,
12
,
5
)
assert
join_dims
(
x
,
axis
=
(
-
1
,
-
2
))
.
type
.
shape
==
(
2
,
3
,
20
)
assert
join_dims
(
x
,
axis
=
())
.
type
.
shape
==
(
2
,
3
,
4
,
5
)
assert
join_dims
(
x
,
axis
=
(
2
,))
.
type
.
shape
==
(
2
,
3
,
4
,
5
)
with
pytest
.
raises
(
ValueError
,
match
=
r"join_dims axis must be consecutive, got normalized axis: \(0, 2\)"
,
):
_
=
join_dims
(
x
,
axis
=
(
0
,
2
))
.
type
.
shape
==
(
8
,
3
,
5
)
x_joined
=
join_dims
(
x
,
axis
=
(
1
,
2
))
x_value
=
rng
.
normal
(
size
=
(
2
,
3
,
4
,
5
))
.
astype
(
config
.
floatX
)
fn
=
function
([
x
],
x_joined
,
mode
=
"FAST_COMPILE"
)
x_joined_value
=
fn
(
x_value
)
np
.
testing
.
assert_allclose
(
x_joined_value
,
x_value
.
reshape
(
2
,
12
,
5
))
assert
join_dims
(
x
,
axis
=
(
1
,))
.
eval
({
x
:
x_value
})
.
shape
==
(
2
,
3
,
4
,
5
)
assert
join_dims
(
x
,
axis
=
())
.
eval
({
x
:
x_value
})
.
shape
==
(
2
,
3
,
4
,
5
)
x
=
pt
.
tensor
(
"x"
,
shape
=
(
3
,
5
))
x_joined
=
join_dims
(
x
,
axis
=
(
0
,
1
))
x_batched
=
pt
.
tensor
(
"x_batched"
,
shape
=
(
10
,
3
,
5
))
x_joined_batched
=
vectorize_graph
(
x_joined
,
{
x
:
x_batched
})
assert
x_joined_batched
.
type
.
shape
==
(
10
,
15
)
x_batched_val
=
rng
.
normal
(
size
=
(
10
,
3
,
5
))
.
astype
(
config
.
floatX
)
assert
x_joined_batched
.
eval
({
x_batched
:
x_batched_val
})
.
shape
==
(
10
,
15
)
@pytest.mark.parametrize
(
"axis, shape, expected_shape"
,
[
(
0
,
pt
.
as_tensor
([
2
,
3
]),
(
2
,
3
,
4
,
6
)),
(
2
,
[
2
,
3
],
(
6
,
4
,
2
,
3
)),
(
-
1
,
6
,
(
6
,
4
,
6
)),
],
ids
=
[
"tensor"
,
"list"
,
"integer"
],
)
def
test_split_dims
(
axis
,
shape
,
expected_shape
):
rng
=
np
.
random
.
default_rng
()
x
=
pt
.
tensor
(
"x"
,
shape
=
(
6
,
4
,
6
))
x_split
=
split_dims
(
x
,
axis
=
axis
,
shape
=
shape
)
assert
x_split
.
type
.
shape
==
expected_shape
x_split
=
split_dims
(
x
,
axis
=
axis
,
shape
=
shape
)
x_value
=
rng
.
normal
(
size
=
(
6
,
4
,
6
))
.
astype
(
config
.
floatX
)
fn
=
function
([
x
],
x_split
,
mode
=
"FAST_COMPILE"
)
x_split_value
=
fn
(
x_value
)
np
.
testing
.
assert_allclose
(
x_split_value
,
x_value
.
reshape
(
expected_shape
))
x
=
pt
.
tensor
(
"x"
,
shape
=
(
10
,))
x_split
=
split_dims
(
x
,
shape
=
(
5
,
2
),
axis
=
0
)
x_batched
=
pt
.
tensor
(
"x_batched"
,
shape
=
(
3
,
10
))
x_split_batched
=
vectorize_graph
(
x_split
,
{
x
:
x_batched
})
assert
x_split_batched
.
type
.
shape
==
(
3
,
5
,
2
)
x_batched_val
=
rng
.
normal
(
size
=
(
3
,
10
))
.
astype
(
config
.
floatX
)
assert
x_split_batched
.
eval
({
x_batched
:
x_batched_val
})
.
shape
==
(
3
,
5
,
2
)
def
test_split_size_zero_shape
():
x
=
pt
.
tensor
(
"x"
,
shape
=
(
1
,
4
,
6
))
x_split
=
split_dims
(
x
,
axis
=
0
,
shape
=
pt
.
as_tensor
(
np
.
zeros
((
0
,))))
assert
x_split
.
type
.
shape
==
(
4
,
6
)
x_value
=
np
.
empty
((
1
,
4
,
6
),
dtype
=
config
.
floatX
)
fn
=
function
([
x
],
x_split
,
mode
=
"FAST_COMPILE"
)
x_split_value
=
fn
(
x_value
)
np
.
testing
.
assert_allclose
(
x_split_value
,
x_value
.
squeeze
(
0
))
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论