Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
61b6baeb
提交
61b6baeb
authored
7月 12, 2013
作者:
lamblin
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #1453 from nouiz/bouchnic-truedot
Bouchnic truedot rebase
上级
22623909
77c3e858
隐藏空白字符变更
内嵌
并排
正在显示
4 个修改的文件
包含
255 行增加
和
248 行删除
+255
-248
index.txt
doc/library/sparse/index.txt
+24
-5
basic.py
theano/sparse/basic.py
+131
-0
truedot.py
theano/sparse/sandbox/truedot.py
+12
-242
test_basic.py
theano/sparse/tests/test_basic.py
+88
-1
没有找到文件。
doc/library/sparse/index.txt
浏览文件 @
61b6baeb
...
...
@@ -194,14 +194,33 @@ List of Implemented Operations
- Dot Product
- :class:`Dot <theano.sparse.basic.Dot>` and
:func:`dot <theano.sparse.basic.dot>`.
The grad implemented is regular.
- The grad implemented is regular.
- No C code for perform and no C code for grad.
- Return a dense for perform and a dense for grad.
- :class:`StructuredDot <theano.sparse.basic.StructuredDot>`
and :func:`structured_dot <theano.sparse.basic.structured_dot>`.
The grad implemented is structured.
- :class:`SamplingDot <theano.sparse.basic.SamplingDot>` and ``sampling_dot``.
The grad implemented is structured for `p`.
- The grad implemented is structured.
- C code for perform and grad.
- Return a dense for perforn and a sparse for grad.
- :class:`TrueDot <theano.sparse.basic.TrueDot>` and
:func:`true_dot <theano.sparse.basic.true_dot>`.
- The grad implemented is regular.
- No C code for perform and no C code for grad.
- Return a Sparse for perform and a Sparse for grad.
- Flags trough constructor can change the output of
grad to be dense if the second input of the op is dense.
- :class:`SamplingDot <theano.sparse.basic.SamplingDot>` and
``sampling_dot``.
- The grad implemented is structured for `p`.
- Sample of the dot and sample of the gradient.
- C code for perform but not for grad.
- Return sparse for perform and grad.
- :class:`Usmm <theano.sparse.basic.Usmm>` and ``usmm``.
There is no grad implemented for this op.
- This op is the equivalent of gemm for sparse dot.
- There is no grad implemented for this op.
- There is optimization that transform a
:class:`Dot <theano.sparse.basic.Dot>` to ``Usmm`` when possible.
You shouldn't need to insert it yourself.
- Slice Operations
- sparse_variable[N, N], return a tensor scalar.
...
...
theano/sparse/basic.py
浏览文件 @
61b6baeb
...
...
@@ -2600,6 +2600,137 @@ def sqrt(x):
# see decorator for function body
class
TrueDot
(
gof
.
op
.
Op
):
"""Calculate the true dot operation between two matrices.
`TrueDot` is different of `StructuredDot` for sparse matrix
since the grad of `TrueDot` is regular, i.e. not structured.
The parameter `grad_preserves_dense`, controlled by the
constructor, is a boolean flags to controls whether gradients
with respect to inputs are converted to dense matrices when the
corresponding input y is dense (not in a L{SparseVariable} wrapper).
This is generally a good idea when L{Dot} is in the middle of a
larger graph, because the types of gy will match that of y. This
conversion might be inefficient if the gradients are graph outputs
though, hence this mask.
:param x: Sparse matrix for the left operand.
:param y: Sparse or dense matrix for the right operand.
:return: The dot product `x` . `y` in a sparse matrix.
:note:
- The grad implemented is regular, i.e. not structured.
"""
# TODO
# Simplify code by splitting into DotSS and DotSD.
def
__init__
(
self
,
grad_preserves_dense
=
True
):
self
.
grad_preserves_dense
=
grad_preserves_dense
def
__eq__
(
self
,
other
):
return
(
type
(
self
)
==
type
(
other
)
and
self
.
grad_preserves_dense
==
other
.
grad_preserves_dense
)
def
__hash__
(
self
):
return
hash
(
type
(
self
))
^
hash
(
self
.
grad_preserves_dense
)
def
__ne__
(
self
,
other
):
return
not
(
self
==
other
)
def
make_node
(
self
,
x
,
y
):
# NOTE
# Because of trickiness of implementing,
# we assume that the left argument x is a
# SparseVariable (not dense)
if
x
.
type
.
dtype
!=
y
.
type
.
dtype
:
raise
NotImplementedError
()
if
not
_is_sparse_variable
(
x
):
raise
TypeError
(
x
)
# These are the conversions performed by scipy.sparse.dot
if
x
.
type
.
format
==
"csc"
or
x
.
type
.
format
==
"coo"
:
myformat
=
"csc"
elif
x
.
type
.
format
==
"csr"
:
myformat
=
"csr"
else
:
raise
NotImplementedError
()
inputs
=
[
x
,
y
]
# Need to convert? e.g. assparse
outputs
=
[
SparseType
(
dtype
=
x
.
type
.
dtype
,
format
=
myformat
)
.
make_variable
()]
return
gof
.
Apply
(
self
,
inputs
,
outputs
)
def
perform
(
self
,
node
,
inp
,
out_
):
# TODO
# -Verify that output is sufficiently sparse,
# and raise a warning if it is not.
# -Also determine that we are storing the
# output in the best storage format?
x
,
y
=
inp
out
,
=
out_
rval
=
x
.
dot
(
y
)
if
not
scipy
.
sparse
.
issparse
(
rval
):
rval
=
getattr
(
scipy
.
sparse
,
x
.
format
+
'_matrix'
)(
rval
)
out
[
0
]
=
rval
def
grad
(
self
,
(
x
,
y
),
(
gz
,
)):
assert
_is_sparse_variable
(
gz
)
assert
_is_sparse_variable
(
x
)
rval
=
[
true_dot
(
gz
,
y
.
T
),
true_dot
(
x
.
T
,
gz
)]
if
_is_dense_variable
(
y
):
if
self
.
grad_preserves_dense
:
rval
[
1
]
=
dense_from_sparse
(
rval
[
1
])
return
rval
def
infer_shape
(
self
,
node
,
shapes
):
return
[(
shapes
[
0
][
0
],
shapes
[
1
][
1
])]
def
__str__
(
self
):
return
self
.
__class__
.
__name__
def
true_dot
(
x
,
y
,
grad_preserves_dense
=
True
):
"""
Operation for efficiently calculating the dot product when
one or all operands is sparse. Supported format are CSC and CSR.
The output of the operation is sparse.
:param x: Matrix variable.
:param y: Matrix variable.
:param grad_preserves_dense: if True and one on the input is dense,
make the output dense.
:return: The dot product `x`.`y` in a sparse format.
"""
# TODO
# Maybe the triple-transposition formulation
# (when x is dense) is slow. See if there is a
# direct way to do this.
if
hasattr
(
x
,
'getnnz'
):
x
=
as_sparse_variable
(
x
)
if
hasattr
(
y
,
'getnnz'
):
y
=
as_sparse_variable
(
y
)
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
y_is_sparse_variable
=
_is_sparse_variable
(
y
)
if
not
x_is_sparse_variable
and
not
y_is_sparse_variable
:
raise
TypeError
()
if
x_is_sparse_variable
:
return
TrueDot
(
grad_preserves_dense
)(
x
,
y
)
else
:
assert
y_is_sparse_variable
return
transpose
(
TrueDot
(
grad_preserves_dense
)(
y
.
T
,
x
.
T
))
# Dot
class
StructuredDot
(
gof
.
Op
):
"""Structured Dot is like dot, except that only the
...
...
theano/sparse/sandbox/truedot.py
浏览文件 @
61b6baeb
import
unittest
import
theano
import
numpy
import
scipy.sparse
as
sp
from
theano
import
gof
,
tensor
,
compile
from
theano
import
sparse
from
theano
import
gof
,
tensor
,
compile
from
theano.sparse.tests.test_basic
import
eval_outputs
from
theano.sparse.basic
import
_is_sparse_variable
,
_is_dense_variable
,
as_sparse_variable
,
_is_sparse
,
_mtypes
,
_mtype_to_str
from
theano.sparse.basic
import
(
_is_sparse_variable
,
_is_dense_variable
,
as_sparse_variable
,
_is_sparse
,
_mtypes
,
_mtype_to_str
)
from
theano.sparse
import
SparseType
,
dense_from_sparse
,
transpose
###############
#
# TrueDot
#
class
TrueDot
(
gof
.
op
.
Op
):
"""
Attributes:
grad_preserves_dense - a boolean flags [default: True].
grad_preserves_dense controls whether gradients with respect to inputs
are converted to dense matrices when the corresponding input y is
dense (not in a L{SparseVariable} wrapper). This is generally a good idea
when L{Dot} is in the middle of a larger graph, because the types
of gy will match that of y. This conversion might be inefficient if
the gradients are graph outputs though, hence this mask.
from
theano.sparse.tests.test_basic
import
sparse_random_inputs
from
theano.tests
import
unittest_tools
as
utt
from
theano.sparse
import
verify_grad_sparse
@todo: Simplify code by splitting into DotSS and DotSD.
"""
def
__init__
(
self
,
grad_preserves_dense
=
True
):
self
.
grad_preserves_dense
=
grad_preserves_dense
def
__eq__
(
self
,
other
):
return
type
(
self
)
==
type
(
other
)
and
self
.
grad_preserves_dense
==
other
.
grad_preserves_dense
def
__hash__
(
self
):
return
hash
(
self
.
grad_preserves_dense
)
def
__ne__
(
self
,
other
):
return
not
(
self
==
other
)
def
make_node
(
self
,
x
,
y
):
"""
:note: Because of trickiness of implementing, we assume that the left argument x is SparseVariable (not dense)
"""
if
x
.
type
.
dtype
!=
y
.
type
.
dtype
:
raise
NotImplementedError
()
if
not
_is_sparse_variable
(
x
):
raise
TypeError
(
x
)
# These are the conversions performed by scipy.sparse.dot
if
x
.
type
.
format
==
"csc"
or
x
.
type
.
format
==
"coo"
:
myformat
=
"csc"
elif
x
.
type
.
format
==
"csr"
:
myformat
=
"csr"
else
:
raise
NotImplementedError
()
inputs
=
[
x
,
y
]
# Need to convert? e.g. assparse
outputs
=
[
SparseType
(
dtype
=
x
.
type
.
dtype
,
format
=
myformat
)
.
make_variable
()]
return
gof
.
Apply
(
self
,
inputs
,
outputs
)
def
perform
(
self
,
node
,
inp
,
out_
):
"""
@todo: Verify that output is sufficiently sparse, and raise a warning if it is not
@todo: Also determine that we are storing the output in the best storage format?
"""
x
,
y
=
inp
out
,
=
out_
rval
=
x
.
dot
(
y
)
out
[
0
]
=
rval
def
grad
(
self
,
inp
,
grads
):
x
,
y
=
inp
gz
,
=
grads
assert
_is_sparse_variable
(
gz
)
assert
_is_sparse_variable
(
x
)
rval
=
[
true_dot
(
gz
,
y
.
T
),
true_dot
(
x
.
T
,
gz
)]
if
_is_dense_variable
(
y
):
if
self
.
grad_preserves_dense
:
rval
[
1
]
=
dense_from_sparse
(
rval
[
1
])
return
rval
def
true_dot
(
x
,
y
,
grad_preserves_dense
=
True
):
"""
@todo: Maybe the triple-transposition formulation (when x is dense)
is slow. See if there is a direct way to do this.
"""
if
hasattr
(
x
,
'getnnz'
):
x
=
as_sparse_variable
(
x
)
if
hasattr
(
y
,
'getnnz'
):
y
=
as_sparse_variable
(
y
)
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
y_is_sparse_variable
=
_is_sparse_variable
(
y
)
if
not
x_is_sparse_variable
and
not
y_is_sparse_variable
:
raise
TypeError
()
if
x_is_sparse_variable
:
return
TrueDot
(
grad_preserves_dense
)(
x
,
y
)
else
:
assert
y_is_sparse_variable
return
transpose
(
TrueDot
(
grad_preserves_dense
)(
y
.
T
,
x
.
T
))
class
test_true_dot
(
unittest
.
TestCase
):
def
setUp
(
self
):
numpy
.
random
.
seed
(
44
)
def
test_basicSS
(
self
):
for
mtype
in
_mtypes
:
x
=
as_sparse_variable
(
mtype
((
500
,
3
)))
x
.
data
[(
10
,
1
)]
=
1
x
.
data
[(
20
,
2
)]
=
2
self
.
assertTrue
(
_is_sparse_variable
(
x
))
xT
=
x
.
T
self
.
assertTrue
(
_is_sparse_variable
(
xT
))
zop
=
true_dot
(
x
,
xT
)
self
.
assertTrue
(
_is_sparse_variable
(
zop
))
z
=
eval_outputs
([
zop
])
self
.
assertTrue
(
_is_sparse
(
z
))
self
.
assertTrue
(
z
.
shape
==
(
500
,
500
))
self
.
assertTrue
(
type
(
z
)
is
mtype
)
w
=
mtype
((
500
,
500
))
w
[(
10
,
10
)]
=
1
w
[(
20
,
20
)]
=
4
self
.
assertTrue
(
z
.
shape
==
w
.
shape
)
self
.
assertTrue
(
type
(
z
)
==
type
(
w
))
self
.
assertTrue
(
z
.
dtype
==
w
.
dtype
)
#self.assertTrue(z == w)
self
.
assertTrue
(
abs
(
z
-
w
)
.
nnz
==
0
)
z
=
z
.
todense
()
w
=
w
.
todense
()
self
.
assertTrue
((
z
==
w
)
.
all
()
==
True
)
def
test_basicSD
(
self
):
for
mtype
in
_mtypes
:
x
=
as_sparse_variable
(
mtype
((
500
,
3
)))
x
.
data
[(
10
,
1
)]
=
1
x
.
data
[(
20
,
2
)]
=
2
self
.
assertTrue
(
_is_sparse_variable
(
x
))
y
=
tensor
.
as_tensor_variable
([[
1.
,
2
],
[
3
,
4
],
[
2
,
1
]])
self
.
assertTrue
(
_is_dense_variable
(
y
))
zop
=
true_dot
(
x
,
y
)
self
.
assertTrue
(
_is_sparse_variable
(
zop
))
z
=
eval_outputs
([
zop
])
self
.
assertTrue
(
_is_sparse
(
z
))
self
.
assertTrue
(
z
.
shape
==
(
500
,
2
))
self
.
assertTrue
(
type
(
z
)
is
mtype
)
w
=
mtype
((
500
,
2
))
w
[(
10
,
0
)]
=
3.
w
[(
20
,
0
)]
=
4
w
[(
10
,
1
)]
=
4
w
[(
20
,
1
)]
=
2
self
.
assertTrue
(
z
.
shape
==
w
.
shape
)
self
.
assertTrue
(
type
(
z
)
==
type
(
w
))
self
.
assertTrue
(
z
.
dtype
==
w
.
dtype
)
#self.assertTrue(z == w)
self
.
assertTrue
(
abs
(
z
-
w
)
.
nnz
==
0
)
z
=
z
.
todense
()
w
=
w
.
todense
()
self
.
assertTrue
((
z
==
w
)
.
all
()
==
True
)
def
test_basicDS
(
self
):
for
mtype
in
_mtypes
:
x
=
as_sparse_variable
(
mtype
((
500
,
3
)))
x
.
data
[(
10
,
1
)]
=
1
x
.
data
[(
20
,
2
)]
=
2
self
.
assertTrue
(
_is_sparse_variable
(
x
))
y
=
tensor
.
as_tensor_variable
([[
1.
,
2
],
[
3
,
4
],
[
2
,
1
]])
self
.
assertTrue
(
_is_dense_variable
(
y
))
x
.
data
=
x
.
data
.
T
y
.
data
=
y
.
data
.
T
zop
=
true_dot
(
y
,
x
)
zop
=
transpose
(
true_dot
(
y
,
x
))
self
.
assertTrue
(
_is_sparse_variable
(
zop
))
z
=
eval_outputs
([
zop
])
self
.
assertTrue
(
_is_sparse
(
z
))
self
.
assertTrue
(
z
.
shape
==
(
500
,
2
))
# self.assertTrue(type(z) is mtype)
w
=
mtype
((
500
,
2
))
w
[(
10
,
0
)]
=
3.
w
[(
20
,
0
)]
=
4
w
[(
10
,
1
)]
=
4
w
[(
20
,
1
)]
=
2
self
.
assertTrue
(
z
.
shape
==
w
.
shape
)
# Type should switch from csr to csc and vice-versa, so don't perform this test
#self.assertTrue(type(z) == type(w))
self
.
assertTrue
(
z
.
dtype
==
w
.
dtype
)
# Type should switch from csr to csc and vice-versa, so don't perform this test
#self.assertTrue(z == w)
self
.
assertTrue
(
abs
(
z
-
w
)
.
nnz
==
0
)
z
=
z
.
todense
()
w
=
w
.
todense
()
self
.
assertTrue
((
z
==
w
)
.
all
()
==
True
)
def
test_graph_bprop0
(
self
):
for
mtype
in
_mtypes
:
x
=
tensor
.
matrix
(
'x'
)
#TensorType('float64', broadcastable=[False,False], name='x')
w
=
SparseType
(
dtype
=
'float64'
,
format
=
_mtype_to_str
[
mtype
])
.
make_variable
()
xw
=
dense_from_sparse
(
true_dot
(
w
,
x
))
y
=
dense_from_sparse
(
true_dot
(
w
.
T
,
xw
))
diff
=
x
-
y
loss
=
tensor
.
sum
(
tensor
.
sqr
(
diff
))
gw
=
tensor
.
grad
(
loss
,
w
)
trainfn
=
compile
.
function
([
x
,
w
],
[
y
,
loss
,
gw
])
x
=
numpy
.
asarray
([[
1.
,
2
],
[
3
,
4
],
[
2
,
1
]])
w
=
mtype
((
500
,
3
))
w
[(
10
,
1
)]
=
1
w
[(
20
,
2
)]
=
2
lr
=
0.001
y
,
origloss
,
gw
=
trainfn
(
x
,
w
)
for
epoch
in
xrange
(
50
):
y
,
loss
,
gw
=
trainfn
(
x
,
w
)
w
=
w
-
(
lr
*
gw
)
print
loss
self
.
assertTrue
(
origloss
>
loss
)
self
.
assertTrue
(
'1.05191241115'
==
str
(
loss
))
def
test_graph_bprop_rand
(
self
):
for
i
in
range
(
10
):
xorig
=
numpy
.
random
.
rand
(
3
,
2
)
for
mtype
in
_mtypes
:
x
=
tensor
.
matrix
(
'x'
)
w
=
SparseType
(
dtype
=
'float64'
,
format
=
_mtype_to_str
[
mtype
])
.
make_variable
()
xw
=
dense_from_sparse
(
true_dot
(
w
,
x
))
y
=
dense_from_sparse
(
true_dot
(
w
.
T
,
xw
))
diff
=
x
-
y
loss
=
tensor
.
sum
(
tensor
.
sqr
(
diff
))
gw
=
tensor
.
grad
(
loss
,
w
)
trainfn
=
compile
.
function
([
x
,
w
],
[
y
,
loss
,
gw
])
x
=
xorig
w
=
mtype
((
500
,
3
))
w
[(
10
,
1
)]
=
1
w
[(
20
,
2
)]
=
2
lr
=
0.001
y
,
origloss
,
gw
=
trainfn
(
x
,
w
)
for
epoch
in
xrange
(
50
):
y
,
loss
,
gw
=
trainfn
(
x
,
w
)
w
=
w
-
(
lr
*
gw
)
self
.
assertTrue
(
origloss
>
loss
)
# To maintain compatibility
from
theano.sparse.basic
import
TrueDot
,
true_dot
theano/sparse/tests/test_basic.py
浏览文件 @
61b6baeb
...
...
@@ -39,7 +39,8 @@ from theano.sparse import (
SamplingDot
,
sampling_dot
,
Diag
,
diag
,
SquareDiagonal
,
square_diagonal
,
EnsureSortedIndices
,
ensure_sorted_indices
,
clean
,
ConstructSparseFromList
,
construct_sparse_from_list
)
ConstructSparseFromList
,
construct_sparse_from_list
,
TrueDot
,
true_dot
)
# Probability distributions are currently tested in test_sp2.py
#from theano.sparse import (
...
...
@@ -2626,6 +2627,92 @@ class StructuredAddSVTester(unittest.TestCase):
out
.
toarray
())
class
TrueDotTester
(
utt
.
InferShapeTester
):
def
setUp
(
self
):
super
(
TrueDotTester
,
self
)
.
setUp
()
self
.
op
=
true_dot
self
.
op_class
=
TrueDot
def
test_op_ss
(
self
):
for
format
in
sparse
.
sparse_formats
:
for
dtype
in
sparse
.
all_dtypes
:
variable
,
data
=
sparse_random_inputs
(
format
,
shape
=
(
10
,
10
),
out_dtype
=
dtype
,
n
=
2
,
p
=
0.1
)
f
=
theano
.
function
(
variable
,
self
.
op
(
*
variable
))
tested
=
f
(
*
data
)
x
,
y
=
[
m
.
toarray
()
for
m
in
data
]
expected
=
numpy
.
dot
(
x
,
y
)
assert
tested
.
format
==
format
assert
tested
.
dtype
==
expected
.
dtype
tested
=
tested
.
toarray
()
assert
numpy
.
allclose
(
tested
,
expected
)
def
test_op_sd
(
self
):
for
format
in
sparse
.
sparse_formats
:
for
dtype
in
sparse
.
all_dtypes
:
variable
,
data
=
sparse_random_inputs
(
format
,
shape
=
(
10
,
10
),
out_dtype
=
dtype
,
n
=
2
,
p
=
0.1
)
variable
[
1
]
=
tensor
.
TensorType
(
dtype
=
dtype
,
broadcastable
=
(
False
,
False
))()
data
[
1
]
=
data
[
1
]
.
toarray
()
f
=
theano
.
function
(
variable
,
self
.
op
(
*
variable
))
tested
=
f
(
*
data
)
expected
=
numpy
.
dot
(
data
[
0
]
.
toarray
(),
data
[
1
])
assert
tested
.
format
==
format
assert
tested
.
dtype
==
expected
.
dtype
tested
=
tested
.
toarray
()
assert
numpy
.
allclose
(
tested
,
expected
)
def
test_infer_shape
(
self
):
for
format
in
sparse
.
sparse_formats
:
for
dtype
in
sparse
.
all_dtypes
:
(
x
,
),
(
x_value
,
)
=
sparse_random_inputs
(
format
,
shape
=
(
9
,
10
),
out_dtype
=
dtype
,
p
=
0.1
)
(
y
,
),
(
y_value
,
)
=
sparse_random_inputs
(
format
,
shape
=
(
10
,
24
),
out_dtype
=
dtype
,
p
=
0.1
)
variable
=
[
x
,
y
]
data
=
[
x_value
,
y_value
]
self
.
_compile_and_check
(
variable
,
[
self
.
op
(
*
variable
)],
data
,
self
.
op_class
)
def
test_grad
(
self
):
for
format
in
sparse
.
sparse_formats
:
for
dtype
in
sparse
.
float_dtypes
:
(
x
,
),
(
x_value
,
)
=
sparse_random_inputs
(
format
,
shape
=
(
9
,
10
),
out_dtype
=
dtype
,
p
=
0.1
)
(
y
,
),
(
y_value
,
)
=
sparse_random_inputs
(
format
,
shape
=
(
10
,
24
),
out_dtype
=
dtype
,
p
=
0.1
)
variable
=
[
x
,
y
]
data
=
[
x_value
,
y_value
]
verify_grad_sparse
(
self
.
op
,
data
,
structured
=
False
)
class
SamplingDotTester
(
utt
.
InferShapeTester
):
x
=
[
tensor
.
matrix
()
for
t
in
range
(
2
)]
x
.
append
(
sparse
.
csr_matrix
())
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论