Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
f42f2163
提交
f42f2163
authored
10月 19, 2020
作者:
Brandon T. Willard
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Replace theano.tensor alias T with tt in theano.gpuarray
上级
43ab4ff0
隐藏空白字符变更
内嵌
并排
正在显示
4 个修改的文件
包含
81 行增加
和
62 行删除
+81
-62
ctc.py
theano/gpuarray/ctc.py
+13
-11
fft.py
theano/gpuarray/fft.py
+25
-20
neighbours.py
theano/gpuarray/neighbours.py
+13
-7
subtensor.py
theano/gpuarray/subtensor.py
+30
-24
没有找到文件。
theano/gpuarray/ctc.py
浏览文件 @
f42f2163
import
os
import
sys
import
theano
import
theano
import
theano.tensor
as
tt
import
theano.tensor.nnet.ctc
from
theano
import
config
,
gof
from
theano
import
config
,
gof
import
theano.tensor
as
T
from
theano.gpuarray.basic_ops
import
(
from
.basic_ops
import
(
gpu_contiguous
,
gpu_contiguous
,
as_gpuarray_variable
,
as_gpuarray_variable
,
infer_context_name
,
infer_context_name
,
gpuarray_helper_inc_dir
,
gpuarray_helper_inc_dir
,
)
)
import
theano.tensor.nnet.ctc
from
theano.gpuarray.type
import
GpuArrayType
,
gpu_context_type
from
.type
import
GpuArrayType
,
gpu_context_type
from
theano.gpuarray.elemwise
import
GpuDimShuffle
from
.elemwise
import
GpuDimShuffle
from
theano.gradient
import
grad_undefined
from
theano.gradient
import
grad_undefined
from
theano.gof
import
local_optimizer
from
theano.gof
import
local_optimizer
from
theano.tensor.opt
import
register_canonicalize
from
theano.tensor.opt
import
register_canonicalize
from
theano.tensor.nnet.ctc
import
ctc_available
from
theano.tensor.nnet.ctc
import
ctc_available
import
os
from
theano.gpuarray
import
pygpu
import
sys
from
.
import
pygpu
class
GpuConnectionistTemporalClassification
(
gof
.
COp
):
class
GpuConnectionistTemporalClassification
(
gof
.
COp
):
...
@@ -104,8 +106,8 @@ class GpuConnectionistTemporalClassification(gof.COp):
...
@@ -104,8 +106,8 @@ class GpuConnectionistTemporalClassification(gof.COp):
t_activations
=
gpu_contiguous
(
t_activations
)
t_activations
=
gpu_contiguous
(
t_activations
)
# Labels and input lengths are always on the CPU
# Labels and input lengths are always on the CPU
t_labels
=
T
.
as_tensor_variable
(
labels
)
t_labels
=
tt
.
as_tensor_variable
(
labels
)
t_input_lengths
=
T
.
as_tensor_variable
(
input_lengths
)
t_input_lengths
=
tt
.
as_tensor_variable
(
input_lengths
)
if
t_activations
.
type
.
dtype
!=
"float32"
:
if
t_activations
.
type
.
dtype
!=
"float32"
:
raise
TypeError
(
"activations must use the float32 type."
)
raise
TypeError
(
"activations must use the float32 type."
)
...
@@ -162,7 +164,7 @@ class GpuConnectionistTemporalClassification(gof.COp):
...
@@ -162,7 +164,7 @@ class GpuConnectionistTemporalClassification(gof.COp):
),
),
new_order
=
(
1
,
0
,
2
),
new_order
=
(
1
,
0
,
2
),
)(
gradients
)
)(
gradients
)
grad_bdot
=
T
.
basic
.
batched_dot
(
grad_op
,
grad_shuffle
)
grad_bdot
=
tt
.
batched_dot
(
grad_op
,
grad_shuffle
)
grad_shuffle_reverse
=
GpuDimShuffle
(
grad_shuffle_reverse
=
GpuDimShuffle
(
input_broadcastable
=
(
input_broadcastable
=
(
False
,
False
,
...
...
theano/gpuarray/fft.py
浏览文件 @
f42f2163
import
numpy
as
np
import
numpy
as
np
import
theano
import
theano
import
theano.tensor
as
tt
from
theano
import
Op
from
theano
import
Op
import
theano.tensor
as
T
from
theano.gradient
import
DisconnectedType
from
theano.gradient
import
DisconnectedType
from
theano.gpuarray.basic_ops
import
(
gpu_contiguous
,
as_gpuarray_variable
,
infer_context_name
,
)
from
theano.gpuarray.type
import
GpuArrayType
from
.basic_ops
import
gpu_contiguous
,
as_gpuarray_variable
,
infer_context_name
from
theano.tensor.fft
import
IRFFTOp
from
.type
import
GpuArrayType
from
theano.gpuarray.opt
import
register_opt
,
op_lifter
,
register_opt2
import
theano.tensor.fft
from
.opt
import
register_opt
,
op_lifter
,
register_opt2
try
:
try
:
import
pygpu
import
pygpu
...
@@ -67,11 +72,11 @@ class CuRFFTOp(Op):
...
@@ -67,11 +72,11 @@ class CuRFFTOp(Op):
# If no shape is provided as input, default to input data shape.
# If no shape is provided as input, default to input data shape.
if
s
is
None
:
if
s
is
None
:
s
=
inp
.
shape
[
1
:]
s
=
inp
.
shape
[
1
:]
s
=
T
.
as_tensor_variable
(
s
)
s
=
tt
.
as_tensor_variable
(
s
)
assert
inp
.
dtype
==
"float32"
assert
inp
.
dtype
==
"float32"
assert
s
.
ndim
==
1
assert
s
.
ndim
==
1
assert
s
.
dtype
in
t
heano
.
tensor
.
integer_dtypes
assert
s
.
dtype
in
t
t
.
integer_dtypes
return
theano
.
Apply
(
self
,
[
inp
,
s
],
[
self
.
output_type
(
inp
)()])
return
theano
.
Apply
(
self
,
[
inp
,
s
],
[
self
.
output_type
(
inp
)()])
...
@@ -153,7 +158,7 @@ class CuRFFTOp(Op):
...
@@ -153,7 +158,7 @@ class CuRFFTOp(Op):
+
[
slice
(
1
,
(
s
[
-
1
]
//
2
)
+
(
s
[
-
1
]
%
2
))]
+
[
slice
(
1
,
(
s
[
-
1
]
//
2
)
+
(
s
[
-
1
]
%
2
))]
+
[
slice
(
None
)]
+
[
slice
(
None
)]
)
)
gout
=
T
.
set_subtensor
(
gout
[
idx
],
gout
[
idx
]
*
0.5
)
gout
=
tt
.
set_subtensor
(
gout
[
idx
],
gout
[
idx
]
*
0.5
)
return
[
cuirfft_op
(
gout
,
s
),
DisconnectedType
()()]
return
[
cuirfft_op
(
gout
,
s
),
DisconnectedType
()()]
def
connection_pattern
(
self
,
node
):
def
connection_pattern
(
self
,
node
):
...
@@ -198,8 +203,8 @@ class CuIRFFTOp(Op):
...
@@ -198,8 +203,8 @@ class CuIRFFTOp(Op):
# If no shape is provided as input, calculate shape assuming even real transform.
# If no shape is provided as input, calculate shape assuming even real transform.
if
s
is
None
:
if
s
is
None
:
s
=
inp
.
shape
[
1
:
-
1
]
s
=
inp
.
shape
[
1
:
-
1
]
s
=
T
.
set_subtensor
(
s
[
-
1
],
(
s
[
-
1
]
-
1
)
*
2
)
s
=
tt
.
set_subtensor
(
s
[
-
1
],
(
s
[
-
1
]
-
1
)
*
2
)
s
=
T
.
as_tensor_variable
(
s
)
s
=
tt
.
as_tensor_variable
(
s
)
assert
inp
.
dtype
==
"float32"
assert
inp
.
dtype
==
"float32"
assert
s
.
ndim
==
1
assert
s
.
ndim
==
1
...
@@ -285,7 +290,7 @@ class CuIRFFTOp(Op):
...
@@ -285,7 +290,7 @@ class CuIRFFTOp(Op):
+
[
slice
(
1
,
(
s
[
-
1
]
//
2
)
+
(
s
[
-
1
]
%
2
))]
+
[
slice
(
1
,
(
s
[
-
1
]
//
2
)
+
(
s
[
-
1
]
%
2
))]
+
[
slice
(
None
)]
+
[
slice
(
None
)]
)
)
gf
=
T
.
set_subtensor
(
gf
[
idx
],
gf
[
idx
]
*
2
)
gf
=
tt
.
set_subtensor
(
gf
[
idx
],
gf
[
idx
]
*
2
)
return
[
gf
,
DisconnectedType
()()]
return
[
gf
,
DisconnectedType
()()]
def
connection_pattern
(
self
,
node
):
def
connection_pattern
(
self
,
node
):
...
@@ -325,7 +330,7 @@ def curfft(inp, norm=None):
...
@@ -325,7 +330,7 @@ def curfft(inp, norm=None):
cond_norm
=
_unitary
(
norm
)
cond_norm
=
_unitary
(
norm
)
scaling
=
1
scaling
=
1
if
cond_norm
==
"ortho"
:
if
cond_norm
==
"ortho"
:
scaling
=
T
.
sqrt
(
s
.
prod
()
.
astype
(
"float32"
))
scaling
=
tt
.
sqrt
(
s
.
prod
()
.
astype
(
"float32"
))
return
curfft_op
(
inp
,
s
)
/
scaling
return
curfft_op
(
inp
,
s
)
/
scaling
...
@@ -364,16 +369,16 @@ def cuirfft(inp, norm=None, is_odd=False):
...
@@ -364,16 +369,16 @@ def cuirfft(inp, norm=None, is_odd=False):
s
=
inp
.
shape
[
1
:
-
1
]
s
=
inp
.
shape
[
1
:
-
1
]
if
is_odd
:
if
is_odd
:
s
=
T
.
set_subtensor
(
s
[
-
1
],
(
s
[
-
1
]
-
1
)
*
2
+
1
)
s
=
tt
.
set_subtensor
(
s
[
-
1
],
(
s
[
-
1
]
-
1
)
*
2
+
1
)
else
:
else
:
s
=
T
.
set_subtensor
(
s
[
-
1
],
(
s
[
-
1
]
-
1
)
*
2
)
s
=
tt
.
set_subtensor
(
s
[
-
1
],
(
s
[
-
1
]
-
1
)
*
2
)
cond_norm
=
_unitary
(
norm
)
cond_norm
=
_unitary
(
norm
)
scaling
=
1
scaling
=
1
if
cond_norm
is
None
:
if
cond_norm
is
None
:
scaling
=
s
.
prod
()
.
astype
(
"float32"
)
scaling
=
s
.
prod
()
.
astype
(
"float32"
)
elif
cond_norm
==
"ortho"
:
elif
cond_norm
==
"ortho"
:
scaling
=
T
.
sqrt
(
s
.
prod
()
.
astype
(
"float32"
))
scaling
=
tt
.
sqrt
(
s
.
prod
()
.
astype
(
"float32"
))
return
cuirfft_op
(
inp
,
s
)
/
scaling
return
cuirfft_op
(
inp
,
s
)
/
scaling
...
@@ -389,13 +394,13 @@ def _unitary(norm):
...
@@ -389,13 +394,13 @@ def _unitary(norm):
if
skcuda_available
:
if
skcuda_available
:
@register_opt
(
"fast_compile"
)
@register_opt
(
"fast_compile"
)
@op_lifter
([
theano
.
tensor
.
fft
.
RFFTOp
])
@op_lifter
([
I
RFFTOp
])
@register_opt2
([
theano
.
tensor
.
fft
.
RFFTOp
],
"fast_compile"
)
@register_opt2
([
I
RFFTOp
],
"fast_compile"
)
def
local_gpua_curfft_op
(
op
,
ctx_name
,
inputs
,
outputs
):
def
local_gpua_curfft_op
(
op
,
ctx_name
,
inputs
,
outputs
):
return
curfft_op
return
curfft_op
@register_opt
(
"fast_compile"
)
@register_opt
(
"fast_compile"
)
@op_lifter
([
theano
.
tensor
.
fft
.
IRFFTOp
])
@op_lifter
([
IRFFTOp
])
@register_opt2
([
theano
.
tensor
.
fft
.
IRFFTOp
],
"fast_compile"
)
@register_opt2
([
IRFFTOp
],
"fast_compile"
)
def
local_gpua_cuirfft_op
(
op
,
ctx_name
,
inputs
,
outputs
):
def
local_gpua_cuirfft_op
(
op
,
ctx_name
,
inputs
,
outputs
):
return
cuirfft_op
return
cuirfft_op
theano/gpuarray/neighbours.py
浏览文件 @
f42f2163
import
theano.tensor
as
tt
from
theano
import
Op
,
Apply
from
theano
import
Op
,
Apply
from
theano.gof
import
ParamsType
from
theano.gof
import
ParamsType
from
theano.tensor.nnet.neighbours
import
Images2Neibs
from
theano.tensor.nnet.neighbours
import
Images2Neibs
import
theano.tensor
as
T
try
:
try
:
from
pygpu
import
gpuarray
from
pygpu
import
gpuarray
except
ImportError
:
except
ImportError
:
pass
pass
from
.basic_ops
import
as_gpuarray_variable
,
GpuKernelBase
,
Kernel
,
infer_context_name
from
theano.gpuarray.basic_ops
import
(
from
.type
import
GpuArrayType
,
gpu_context_type
as_gpuarray_variable
,
GpuKernelBase
,
Kernel
,
infer_context_name
,
)
from
theano.gpuarray.type
import
GpuArrayType
,
gpu_context_type
class
GpuImages2Neibs
(
GpuKernelBase
,
Images2Neibs
,
Op
):
class
GpuImages2Neibs
(
GpuKernelBase
,
Images2Neibs
,
Op
):
...
@@ -25,17 +31,17 @@ class GpuImages2Neibs(GpuKernelBase, Images2Neibs, Op):
...
@@ -25,17 +31,17 @@ class GpuImages2Neibs(GpuKernelBase, Images2Neibs, Op):
def
make_node
(
self
,
ten4
,
neib_shape
,
neib_step
=
None
):
def
make_node
(
self
,
ten4
,
neib_shape
,
neib_step
=
None
):
ten4
=
as_gpuarray_variable
(
ten4
,
infer_context_name
(
ten4
))
ten4
=
as_gpuarray_variable
(
ten4
,
infer_context_name
(
ten4
))
neib_shape
=
T
.
as_tensor_variable
(
neib_shape
)
neib_shape
=
tt
.
as_tensor_variable
(
neib_shape
)
if
neib_step
is
None
:
if
neib_step
is
None
:
neib_step
=
neib_shape
neib_step
=
neib_shape
else
:
else
:
neib_step
=
T
.
as_tensor_variable
(
neib_step
)
neib_step
=
tt
.
as_tensor_variable
(
neib_step
)
assert
ten4
.
ndim
==
4
assert
ten4
.
ndim
==
4
assert
neib_shape
.
ndim
==
1
assert
neib_shape
.
ndim
==
1
assert
neib_step
.
ndim
==
1
assert
neib_step
.
ndim
==
1
assert
neib_shape
.
dtype
in
T
.
integer_dtypes
assert
neib_shape
.
dtype
in
tt
.
integer_dtypes
assert
neib_step
.
dtype
in
T
.
integer_dtypes
assert
neib_step
.
dtype
in
tt
.
integer_dtypes
return
Apply
(
return
Apply
(
self
,
self
,
...
...
theano/gpuarray/subtensor.py
浏览文件 @
f42f2163
import
numpy
as
np
import
numpy
as
np
import
theano.tensor
as
T
import
theano.tensor
as
tt
from
six
import
integer_types
from
six
import
integer_types
from
six.moves
import
StringIO
from
six.moves
import
StringIO
from
theano
import
tensor
,
gof
,
Op
from
theano
import
gof
,
Op
from
theano.gof
import
ParamsType
from
theano.gof
import
ParamsType
from
theano.gradient
import
grad_not_implemented
from
theano.gradient
import
grad_not_implemented
from
theano.tensor.subtensor
import
IncSubtensor
,
Subtensor
,
get_idx_list
from
theano.tensor
import
AllocDiag
from
theano.tensor
import
AllocDiag
from
theano.tensor.subtensor
import
(
IncSubtensor
,
AdvancedSubtensor
,
Subtensor
,
AdvancedIncSubtensor
,
AdvancedSubtensor1
,
get_idx_list
,
)
from
theano.scalar
import
bool
as
bool_t
,
int32
as
int_t
,
uint32
as
size_t
from
theano.scalar
import
bool
as
bool_t
,
int32
as
int_t
,
uint32
as
size_t
try
:
try
:
...
@@ -17,8 +25,8 @@ try:
...
@@ -17,8 +25,8 @@ try:
except
ImportError
:
except
ImportError
:
pass
pass
from
.type
import
GpuArrayType
,
gpu_context_type
from
theano.gpuarray
.type
import
GpuArrayType
,
gpu_context_type
from
.basic_ops
import
(
from
theano.gpuarray
.basic_ops
import
(
as_gpuarray_variable
,
as_gpuarray_variable
,
HideC
,
HideC
,
GpuKernelBase
,
GpuKernelBase
,
...
@@ -52,7 +60,7 @@ class GpuSubtensor(HideC, Subtensor):
...
@@ -52,7 +60,7 @@ class GpuSubtensor(HideC, Subtensor):
def
make_node
(
self
,
x
,
*
inputs
):
def
make_node
(
self
,
x
,
*
inputs
):
ctx_name
=
infer_context_name
(
x
)
ctx_name
=
infer_context_name
(
x
)
rval
=
tensor
.
Subtensor
.
make_node
(
self
,
x
,
*
inputs
)
rval
=
Subtensor
.
make_node
(
self
,
x
,
*
inputs
)
otype
=
GpuArrayType
(
otype
=
GpuArrayType
(
dtype
=
rval
.
outputs
[
0
]
.
type
.
dtype
,
dtype
=
rval
.
outputs
[
0
]
.
type
.
dtype
,
broadcastable
=
rval
.
outputs
[
0
]
.
type
.
broadcastable
,
broadcastable
=
rval
.
outputs
[
0
]
.
type
.
broadcastable
,
...
@@ -226,7 +234,7 @@ class GpuIncSubtensor(IncSubtensor):
...
@@ -226,7 +234,7 @@ class GpuIncSubtensor(IncSubtensor):
-----
-----
The optimization to make this inplace is in tensor/opt.
The optimization to make this inplace is in tensor/opt.
The same optimization handles IncSubtensor and GpuIncSubtensor.
The same optimization handles IncSubtensor and GpuIncSubtensor.
This Op has c_code too; it inherits
tensor.
IncSubtensor's c_code.
This Op has c_code too; it inherits IncSubtensor's c_code.
The helper methods like :meth:`do_type_checking`,
The helper methods like :meth:`do_type_checking`,
:meth:`copy_of_x`, etc. specialize the c_code for this Op.
:meth:`copy_of_x`, etc. specialize the c_code for this Op.
...
@@ -239,7 +247,7 @@ class GpuIncSubtensor(IncSubtensor):
...
@@ -239,7 +247,7 @@ class GpuIncSubtensor(IncSubtensor):
ctx_name
=
infer_context_name
(
x
,
y
)
ctx_name
=
infer_context_name
(
x
,
y
)
x
=
as_gpuarray_variable
(
x
,
ctx_name
)
x
=
as_gpuarray_variable
(
x
,
ctx_name
)
y
=
as_gpuarray_variable
(
y
,
ctx_name
)
y
=
as_gpuarray_variable
(
y
,
ctx_name
)
rval
=
tensor
.
IncSubtensor
.
make_node
(
self
,
x
,
y
,
*
inputs
)
rval
=
IncSubtensor
.
make_node
(
self
,
x
,
y
,
*
inputs
)
ret
=
gof
.
Apply
(
self
,
[
x
,
y
]
+
rval
.
inputs
[
2
:],
[
x
.
type
()])
ret
=
gof
.
Apply
(
self
,
[
x
,
y
]
+
rval
.
inputs
[
2
:],
[
x
.
type
()])
return
ret
return
ret
...
@@ -450,7 +458,7 @@ int sub_setarray(GpuArray *dst, GpuArray *src) {
...
@@ -450,7 +458,7 @@ int sub_setarray(GpuArray *dst, GpuArray *src) {
return
parent_version
+
(
10
,)
return
parent_version
+
(
10
,)
class
GpuAdvancedSubtensor1
(
HideC
,
tensor
.
AdvancedSubtensor1
):
class
GpuAdvancedSubtensor1
(
HideC
,
AdvancedSubtensor1
):
"""
"""
AdvancedSubrensor1 on the GPU.
AdvancedSubrensor1 on the GPU.
"""
"""
...
@@ -461,11 +469,11 @@ class GpuAdvancedSubtensor1(HideC, tensor.AdvancedSubtensor1):
...
@@ -461,11 +469,11 @@ class GpuAdvancedSubtensor1(HideC, tensor.AdvancedSubtensor1):
ctx_name
=
infer_context_name
(
x
,
ilist
)
ctx_name
=
infer_context_name
(
x
,
ilist
)
x_
=
as_gpuarray_variable
(
x
,
ctx_name
)
x_
=
as_gpuarray_variable
(
x
,
ctx_name
)
ilist__
=
t
ensor
.
as_tensor_variable
(
ilist
)
ilist__
=
t
t
.
as_tensor_variable
(
ilist
)
if
ilist__
.
type
.
dtype
not
in
t
ensor
.
integer_dtypes
:
if
ilist__
.
type
.
dtype
not
in
t
t
.
integer_dtypes
:
raise
TypeError
(
"index must be integers"
)
raise
TypeError
(
"index must be integers"
)
if
ilist__
.
type
.
dtype
!=
"int64"
:
if
ilist__
.
type
.
dtype
!=
"int64"
:
ilist__
=
t
ensor
.
cast
(
ilist__
,
"int64"
)
ilist__
=
t
t
.
cast
(
ilist__
,
"int64"
)
ilist_
=
gpu_contiguous
(
as_gpuarray_variable
(
ilist__
,
ctx_name
))
ilist_
=
gpu_contiguous
(
as_gpuarray_variable
(
ilist__
,
ctx_name
))
...
@@ -676,14 +684,14 @@ class BaseGpuAdvancedSubtensor(object):
...
@@ -676,14 +684,14 @@ class BaseGpuAdvancedSubtensor(object):
out
[
0
]
=
o
out
[
0
]
=
o
class
GpuAdvancedSubtensor
(
HideC
,
BaseGpuAdvancedSubtensor
,
tensor
.
AdvancedSubtensor
):
class
GpuAdvancedSubtensor
(
HideC
,
BaseGpuAdvancedSubtensor
,
AdvancedSubtensor
):
"""
"""
AdvancedSubtensor on the GPU.
AdvancedSubtensor on the GPU.
"""
"""
def
make_node
(
self
,
x
,
*
inputs
):
def
make_node
(
self
,
x
,
*
inputs
):
ctx_name
=
infer_context_name
(
x
)
ctx_name
=
infer_context_name
(
x
)
rval
=
tensor
.
AdvancedSubtensor
.
make_node
(
self
,
x
,
*
inputs
)
rval
=
AdvancedSubtensor
.
make_node
(
self
,
x
,
*
inputs
)
otype
=
GpuArrayType
(
otype
=
GpuArrayType
(
dtype
=
rval
.
outputs
[
0
]
.
type
.
dtype
,
dtype
=
rval
.
outputs
[
0
]
.
type
.
dtype
,
broadcastable
=
rval
.
outputs
[
0
]
.
type
.
broadcastable
,
broadcastable
=
rval
.
outputs
[
0
]
.
type
.
broadcastable
,
...
@@ -809,9 +817,7 @@ class BaseGpuAdvancedIncSubtensor(object):
...
@@ -809,9 +817,7 @@ class BaseGpuAdvancedIncSubtensor(object):
out
[
0
]
=
x_
out
[
0
]
=
x_
class
GpuAdvancedIncSubtensor
(
class
GpuAdvancedIncSubtensor
(
HideC
,
BaseGpuAdvancedIncSubtensor
,
AdvancedIncSubtensor
):
HideC
,
BaseGpuAdvancedIncSubtensor
,
tensor
.
AdvancedIncSubtensor
):
"""
"""
Implement AdvancedIncSubtensor on the gpu.
Implement AdvancedIncSubtensor on the gpu.
...
@@ -819,7 +825,7 @@ class GpuAdvancedIncSubtensor(
...
@@ -819,7 +825,7 @@ class GpuAdvancedIncSubtensor(
def
make_node
(
self
,
x
,
y
,
*
inputs
):
def
make_node
(
self
,
x
,
y
,
*
inputs
):
ctx_name
=
infer_context_name
(
x
,
y
)
ctx_name
=
infer_context_name
(
x
,
y
)
rval
=
tensor
.
AdvancedIncSubtensor
.
make_node
(
self
,
x
,
y
,
*
inputs
)
rval
=
AdvancedIncSubtensor
.
make_node
(
self
,
x
,
y
,
*
inputs
)
otype
=
GpuArrayType
(
otype
=
GpuArrayType
(
dtype
=
rval
.
outputs
[
0
]
.
type
.
dtype
,
dtype
=
rval
.
outputs
[
0
]
.
type
.
dtype
,
broadcastable
=
rval
.
outputs
[
0
]
.
type
.
broadcastable
,
broadcastable
=
rval
.
outputs
[
0
]
.
type
.
broadcastable
,
...
@@ -863,11 +869,11 @@ class GpuAdvancedIncSubtensor1(Op):
...
@@ -863,11 +869,11 @@ class GpuAdvancedIncSubtensor1(Op):
ctx_name
=
infer_context_name
(
x
,
y
)
ctx_name
=
infer_context_name
(
x
,
y
)
x_
=
as_gpuarray_variable
(
x
,
ctx_name
)
x_
=
as_gpuarray_variable
(
x
,
ctx_name
)
y_
=
as_gpuarray_variable
(
y
,
ctx_name
)
y_
=
as_gpuarray_variable
(
y
,
ctx_name
)
ilist_
=
t
ensor
.
as_tensor_variable
(
ilist
)
ilist_
=
t
t
.
as_tensor_variable
(
ilist
)
assert
x_
.
type
.
ndim
>=
y_
.
type
.
ndim
assert
x_
.
type
.
ndim
>=
y_
.
type
.
ndim
if
ilist_
.
type
.
dtype
not
in
t
ensor
.
integer_dtypes
:
if
ilist_
.
type
.
dtype
not
in
t
t
.
integer_dtypes
:
raise
TypeError
(
"index must be integers"
)
raise
TypeError
(
"index must be integers"
)
if
ilist_
.
type
.
ndim
!=
1
:
if
ilist_
.
type
.
ndim
!=
1
:
raise
TypeError
(
"index must be vector"
)
raise
TypeError
(
"index must be vector"
)
...
@@ -1106,7 +1112,7 @@ class GpuAdvancedIncSubtensor1_dev20(GpuKernelBase, HideC, GpuAdvancedIncSubtens
...
@@ -1106,7 +1112,7 @@ class GpuAdvancedIncSubtensor1_dev20(GpuKernelBase, HideC, GpuAdvancedIncSubtens
assert
x_
.
type
.
ndim
>=
y_
.
type
.
ndim
assert
x_
.
type
.
ndim
>=
y_
.
type
.
ndim
if
ilist_
.
type
.
dtype
not
in
t
ensor
.
integer_dtypes
:
if
ilist_
.
type
.
dtype
not
in
t
t
.
integer_dtypes
:
raise
TypeError
(
"index must be integers"
)
raise
TypeError
(
"index must be integers"
)
if
ilist_
.
type
.
ndim
!=
1
:
if
ilist_
.
type
.
ndim
!=
1
:
raise
TypeError
(
"index must be vector"
)
raise
TypeError
(
"index must be vector"
)
...
@@ -1437,11 +1443,11 @@ class GpuExtractDiag(Op):
...
@@ -1437,11 +1443,11 @@ class GpuExtractDiag(Op):
# The following logic is inspired by C code of PyArray_Diagonal().
# The following logic is inspired by C code of PyArray_Diagonal().
offset
=
self
.
offset
offset
=
self
.
offset
if
offset
>
0
:
if
offset
>
0
:
diag_size
=
T
.
clip
(
dim2
-
offset
,
0
,
dim1
)
diag_size
=
tt
.
clip
(
dim2
-
offset
,
0
,
dim1
)
elif
offset
<
0
:
elif
offset
<
0
:
diag_size
=
T
.
clip
(
dim1
+
offset
,
0
,
dim2
)
diag_size
=
tt
.
clip
(
dim1
+
offset
,
0
,
dim2
)
else
:
else
:
diag_size
=
T
.
minimum
(
dim1
,
dim2
)
diag_size
=
tt
.
minimum
(
dim1
,
dim2
)
out_shape
.
append
(
diag_size
)
out_shape
.
append
(
diag_size
)
return
[
tuple
(
out_shape
)]
return
[
tuple
(
out_shape
)]
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论