Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
d1222d9c
提交
d1222d9c
authored
12月 15, 2014
作者:
Arnaud Bergeron
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Cleanup refresh to make the cleanup optional and avoid taking the lock
if we don't delete anything.
上级
57bea857
隐藏空白字符变更
内嵌
并排
正在显示
3 个修改的文件
包含
254 行增加
和
267 行删除
+254
-267
cmodule.py
theano/gof/cmodule.py
+245
-267
compilelock.py
theano/gof/compilelock.py
+9
-0
updates.py
theano/updates.py
+0
-0
没有找到文件。
theano/gof/cmodule.py
浏览文件 @
d1222d9c
...
...
@@ -627,8 +627,7 @@ class ModuleCache(object):
self
.
stats
[
0
]
+=
1
return
self
.
module_from_name
[
name
]
def
refresh
(
self
,
age_thresh_use
=
None
,
delete_if_problem
=
False
,
no_clean_empty
=
False
):
def
refresh
(
self
,
age_thresh_use
=
None
,
delete_if_problem
=
False
,
cleanup
=
True
):
"""Update cache data by walking the cache directory structure.
Load key.pkl files that have not been loaded yet.
...
...
@@ -638,12 +637,15 @@ class ModuleCache(object):
:param age_thresh_use: Do not use modules olther than this.
Defaults to self.age_thresh_use.
:param delete_if_problem: If True, cache entries that meet one
of those
two conditions are deleted:
- Those for which unpickling the KeyData file fails with
an
unknown exception.
:param delete_if_problem: If True, cache entries that meet one
of those
two conditions are deleted:
- Those for which unpickling the KeyData file fails with
an
unknown exception.
- Duplicated modules, regardless of their age.
:param cleanup: Do a cleanup of the cache removing expired and
broken modules.
:returns: a list of modules of age higher than age_thresh_use.
"""
if
age_thresh_use
is
None
:
...
...
@@ -651,238 +653,244 @@ class ModuleCache(object):
start_time
=
time
.
time
()
too_old_to_use
=
[]
compilelock
.
get_lock
()
try
:
# add entries that are not in the entry_from_key dictionary
time_now
=
time
.
time
()
# Go through directories in alphabetical order to ensure consistent
# behavior.
subdirs
=
sorted
(
os
.
listdir
(
self
.
dirname
))
for
root
in
subdirs
:
root
=
os
.
path
.
join
(
self
.
dirname
,
root
)
key_pkl
=
os
.
path
.
join
(
root
,
'key.pkl'
)
if
key_pkl
in
self
.
loaded_key_pkl
:
continue
if
not
os
.
path
.
isdir
(
root
):
continue
files
=
os
.
listdir
(
root
)
if
(
not
files
and
not
no_clean_empty
)
or
'delete.me'
in
files
:
_rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
"delete.me found in dir"
)
continue
elif
'key.pkl'
in
files
:
to_delete
=
[]
def
rmtree
(
*
args
,
**
kwargs
):
if
cleanup
:
to_delete
.
append
((
args
,
kwargs
))
# add entries that are not in the entry_from_key dictionary
time_now
=
time
.
time
()
# Go through directories in alphabetical order to ensure consistent
# behavior.
subdirs
=
sorted
(
os
.
listdir
(
self
.
dirname
))
for
root
in
subdirs
:
root
=
os
.
path
.
join
(
self
.
dirname
,
root
)
key_pkl
=
os
.
path
.
join
(
root
,
'key.pkl'
)
if
key_pkl
in
self
.
loaded_key_pkl
:
continue
if
not
os
.
path
.
isdir
(
root
):
continue
files
=
os
.
listdir
(
root
)
if
not
files
or
'delete.me'
in
files
:
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
"delete.me found in dir"
)
continue
elif
'key.pkl'
in
files
:
try
:
entry
=
module_name_from_dir
(
root
,
files
=
files
)
except
ValueError
:
# there is a key but no dll!
if
not
root
.
startswith
(
"/tmp"
):
# Under /tmp, file are removed periodically by the
# os. So it is normal that this happens from time
# to time.
_logger
.
warning
(
"ModuleCache.refresh() Found key "
"without dll in cache, deleting it.
%
s"
,
key_pkl
)
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
"missing module file"
,
level
=
logging
.
INFO
)
continue
if
(
time_now
-
last_access_time
(
entry
))
<
age_thresh_use
:
_logger
.
debug
(
'refresh adding
%
s'
,
key_pkl
)
def
unpickle_failure
():
_logger
.
info
(
"ModuleCache.refresh() Failed to "
"unpickle cache file
%
s"
,
key_pkl
)
try
:
entry
=
module_name_from_dir
(
root
,
files
=
files
)
except
ValueError
:
# there is a key but no dll!
if
not
root
.
startswith
(
"/tmp"
):
# Under /tmp, file are removed periodically by the
# os. So it is normal that this happens from time
# to time.
_logger
.
warning
(
"ModuleCache.refresh() Found key "
"without dll in cache, deleting it.
%
s"
,
key_pkl
)
_rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
"missing module file"
,
level
=
logging
.
INFO
)
with
open
(
key_pkl
,
'rb'
)
as
f
:
key_data
=
cPickle
.
load
(
f
)
except
EOFError
:
# Happened once... not sure why (would be worth
# investigating if it ever happens again).
unpickle_failure
()
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'broken cache directory [EOF]'
,
level
=
logging
.
WARNING
)
continue
except
ValueError
:
# This can happen when we have bad config value
# in the cuda.nvcc_compiler.py file.
# We should not hide it here, as this will cause
# an unrelated error to appear.
raise
except
Exception
:
unpickle_failure
()
if
delete_if_problem
:
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'broken cache directory'
,
level
=
logging
.
INFO
)
else
:
# This exception is often triggered by keys
# that contain references to classes that have
# not yet been imported (e.g. when running two
# different Theano-based scripts). They are not
# necessarily broken, but we cannot load them
# here.
pass
continue
if
(
time_now
-
last_access_time
(
entry
))
<
age_thresh_use
:
_logger
.
debug
(
'refresh adding
%
s'
,
key_pkl
)
def
unpickle_failure
():
_logger
.
info
(
"ModuleCache.refresh() Failed to "
"unpickle cache file
%
s"
,
key_pkl
)
try
:
with
open
(
key_pkl
,
'rb'
)
as
f
:
key_data
=
cPickle
.
load
(
f
)
except
EOFError
:
# Happened once... not sure why (would be worth
# investigating if it ever happens again).
unpickle_failure
()
_rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'broken cache directory [EOF]'
,
level
=
logging
.
WARNING
)
continue
except
ValueError
:
# This can happen when we have bad config value
# in the cuda.nvcc_compiler.py file.
# We should not hide it here, as this will cause
# an unrelated error to appear.
raise
except
Exception
:
unpickle_failure
()
if
delete_if_problem
:
_rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'broken cache directory'
,
level
=
logging
.
INFO
)
else
:
# This exception is often triggered by keys
# that contain references to classes that have
# not yet been imported (e.g. when running two
# different Theano-based scripts). They are not
# necessarily broken, but we cannot load them
# here.
pass
continue
if
not
isinstance
(
key_data
,
KeyData
):
# This is some old cache data, that does not fit
# the new cache format. It would be possible to
# update it, but it is not entirely safe since we
# do not know the config options that were used.
# As a result, we delete it instead (which is also
# simpler to implement).
_
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
(
'invalid cache entry format -- this '
'should not happen unless your cache '
'was really old'
),
level
=
logging
.
WARN
)
continue
if
not
isinstance
(
key_data
,
KeyData
):
# This is some old cache data, that does not fit
# the new cache format. It would be possible to
# update it, but it is not entirely safe since we
# do not know the config options that were used.
# As a result, we delete it instead (which is also
# simpler to implement).
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
(
'invalid cache entry format -- this '
'should not happen unless your cache '
'was really old'
),
level
=
logging
.
WARN
)
continue
# Check the path to the module stored in the KeyData
# object matches the path to `entry`. There may be
# a mismatch e.g. due to symlinks, or some directory
# being renamed since last time cache was created.
kd_entry
=
key_data
.
get_entry
()
if
kd_entry
!=
entry
:
if
is_same_entry
(
entry
,
kd_entry
):
# Update KeyData object. Note that we also need
# to update the key_pkl field, because it is
# likely to be incorrect if the entry itself
# was wrong.
key_data
.
entry
=
entry
key_data
.
key_pkl
=
key_pkl
else
:
# This is suspicious. Better get rid of it.
_
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'module file path mismatch'
,
level
=
logging
.
INFO
)
continue
# Check the path to the module stored in the KeyData
# object matches the path to `entry`. There may be
# a mismatch e.g. due to symlinks, or some directory
# being renamed since last time cache was created.
kd_entry
=
key_data
.
get_entry
()
if
kd_entry
!=
entry
:
if
is_same_entry
(
entry
,
kd_entry
):
# Update KeyData object. Note that we also need
# to update the key_pkl field, because it is
# likely to be incorrect if the entry itself
# was wrong.
key_data
.
entry
=
entry
key_data
.
key_pkl
=
key_pkl
else
:
# This is suspicious. Better get rid of it.
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'module file path mismatch'
,
level
=
logging
.
INFO
)
continue
# Find unversioned keys from other processes.
# TODO: check if this can happen at all
to_del
=
[
key
for
key
in
key_data
.
keys
if
not
key
[
0
]]
if
to_del
:
# Find unversioned keys from other processes.
# TODO: check if this can happen at all
to_del
=
[
key
for
key
in
key_data
.
keys
if
not
key
[
0
]]
if
to_del
:
_logger
.
warning
(
"ModuleCache.refresh() Found unversioned "
"key in cache, removing it.
%
s"
,
key_pkl
)
# Since the version is in the module hash, all
# keys should be unversioned.
if
len
(
to_del
)
!=
len
(
key_data
.
keys
):
_logger
.
warning
(
"ModuleCache.refresh() Found unversioned "
"key in cache, removing it.
%
s"
,
key_pkl
)
# Since the version is in the module hash, all
# keys should be unversioned.
if
len
(
to_del
)
!=
len
(
key_data
.
keys
):
_logger
.
warning
(
'Found a mix of unversioned and '
'versioned keys for the same '
'module
%
s'
,
key_pkl
)
_rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
"unversioned key(s) in cache"
,
level
=
logging
.
INFO
)
continue
'Found a mix of unversioned and '
'versioned keys for the same '
'module
%
s'
,
key_pkl
)
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
"unversioned key(s) in cache"
,
level
=
logging
.
INFO
)
continue
mod_hash
=
key_data
.
module_hash
if
mod_hash
in
self
.
module_hash_to_key_data
:
# This may happen when two processes running
# simultaneously compiled the same module, one
# after the other. We delete one once it is old
# enough (to be confident there is no other process
# using it), or if `delete_if_problem` is True.
# Note that it is important to walk through
# directories in alphabetical order so as to make
# sure all new processes only use the first one.
mod_hash
=
key_data
.
module_hash
if
mod_hash
in
self
.
module_hash_to_key_data
:
# This may happen when two processes running
# simultaneously compiled the same module, one
# after the other. We delete one once it is old
# enough (to be confident there is no other process
# using it), or if `delete_if_problem` is True.
# Note that it is important to walk through
# directories in alphabetical order so as to make
# sure all new processes only use the first one.
if
cleanup
:
age
=
time
.
time
()
-
last_access_time
(
entry
)
if
delete_if_problem
or
age
>
self
.
age_thresh_del
:
_
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'duplicated module'
,
level
=
logging
.
DEBUG
)
rmtree
(
root
,
ignore_nocleanup
=
True
,
msg
=
'duplicated module'
,
level
=
logging
.
DEBUG
)
else
:
_logger
.
debug
(
'Found duplicated module not '
'old enough yet to be deleted '
'(age:
%
s):
%
s'
,
age
,
entry
)
continue
continue
# Remember the map from a module's hash to the KeyData
# object associated with it.
self
.
module_hash_to_key_data
[
mod_hash
]
=
key_data
for
key
in
key_data
.
keys
:
if
key
not
in
self
.
entry_from_key
:
self
.
entry_from_key
[
key
]
=
entry
# Assert that we have not already got this
# entry somehow.
assert
entry
not
in
self
.
module_from_name
# Store safe part of versioned keys.
if
key
[
0
]:
self
.
similar_keys
.
setdefault
(
get_safe_part
(
key
),
[])
.
append
(
key
)
else
:
_logger
.
warning
(
"The same cache key is associated to "
"different modules (
%
s and
%
s). This "
"is not supposed to happen! You may "
"need to manually delete your cache "
"directory to fix this."
,
self
.
entry_from_key
[
key
],
entry
)
# Clean up the name space to prevent bug.
if
key_data
.
keys
:
del
key
self
.
loaded_key_pkl
.
add
(
key_pkl
)
else
:
too_old_to_use
.
append
(
entry
)
# If the compilation failed, no key.pkl is in that
# directory, but a mod.* should be there.
# We do nothing here.
# Clean up the name space to prevent bug.
del
root
,
files
,
subdirs
# Remove entries that are not in the filesystem.
items_copy
=
list
(
self
.
module_hash_to_key_data
.
iteritems
())
for
module_hash
,
key_data
in
items_copy
:
entry
=
key_data
.
get_entry
()
try
:
# Test to see that the file is [present and] readable.
open
(
entry
)
.
close
()
gone
=
False
except
IOError
:
gone
=
True
if
gone
:
# Assert that we did not have one of the deleted files
# loaded up and in use.
# If so, it should not have been deleted. This should be
# considered a failure of the OTHER process, that deleted
# it.
if
entry
in
self
.
module_from_name
:
_logger
.
warning
(
"A module that was loaded by this "
"ModuleCache can no longer be read from file "
"
%
s... this could lead to problems."
,
# Remember the map from a module's hash to the KeyData
# object associated with it.
self
.
module_hash_to_key_data
[
mod_hash
]
=
key_data
for
key
in
key_data
.
keys
:
if
key
not
in
self
.
entry_from_key
:
self
.
entry_from_key
[
key
]
=
entry
# Assert that we have not already got this
# entry somehow.
assert
entry
not
in
self
.
module_from_name
# Store safe part of versioned keys.
if
key
[
0
]:
self
.
similar_keys
.
setdefault
(
get_safe_part
(
key
),
[])
.
append
(
key
)
else
:
_logger
.
warning
(
"The same cache key is associated to "
"different modules (
%
s and
%
s). This "
"is not supposed to happen! You may "
"need to manually delete your cache "
"directory to fix this."
,
self
.
entry_from_key
[
key
],
entry
)
del
self
.
module_from_name
[
entry
]
_logger
.
info
(
"deleting ModuleCache entry
%
s"
,
entry
)
key_data
.
delete_keys_from
(
self
.
entry_from_key
)
del
self
.
module_hash_to_key_data
[
module_hash
]
if
key_data
.
keys
and
list
(
key_data
.
keys
)[
0
][
0
]:
# this is a versioned entry, so should have been on
# disk. Something weird happened to cause this, so we
# are responding by printing a warning, removing
# evidence that we ever saw this mystery key.
pkl_file_to_remove
=
key_data
.
key_pkl
if
not
key_data
.
key_pkl
.
startswith
(
"/tmp"
):
# Under /tmp, file are removed periodically by the
# os. So it is normal that this happen from time to
# time.
_logger
.
warning
(
"Removing key file
%
s because the "
"corresponding module is gone from the "
"file system."
,
pkl_file_to_remove
)
self
.
loaded_key_pkl
.
remove
(
pkl_file_to_remove
)
finally
:
compilelock
.
release_lock
()
# Clean up the name space to prevent bug.
if
key_data
.
keys
:
del
key
self
.
loaded_key_pkl
.
add
(
key_pkl
)
else
:
too_old_to_use
.
append
(
entry
)
# If the compilation failed, no key.pkl is in that
# directory, but a mod.* should be there.
# We do nothing here.
# Clean up the name space to prevent bug.
del
root
,
files
,
subdirs
# Remove entries that are not in the filesystem.
items_copy
=
list
(
self
.
module_hash_to_key_data
.
iteritems
())
for
module_hash
,
key_data
in
items_copy
:
entry
=
key_data
.
get_entry
()
try
:
# Test to see that the file is [present and] readable.
open
(
entry
)
.
close
()
gone
=
False
except
IOError
:
gone
=
True
if
gone
:
# Assert that we did not have one of the deleted files
# loaded up and in use.
# If so, it should not have been deleted. This should be
# considered a failure of the OTHER process, that deleted
# it.
if
entry
in
self
.
module_from_name
:
_logger
.
warning
(
"A module that was loaded by this "
"ModuleCache can no longer be read from file "
"
%
s... this could lead to problems."
,
entry
)
del
self
.
module_from_name
[
entry
]
_logger
.
info
(
"deleting ModuleCache entry
%
s"
,
entry
)
key_data
.
delete_keys_from
(
self
.
entry_from_key
)
del
self
.
module_hash_to_key_data
[
module_hash
]
if
key_data
.
keys
and
list
(
key_data
.
keys
)[
0
][
0
]:
# this is a versioned entry, so should have been on
# disk. Something weird happened to cause this, so we
# are responding by printing a warning, removing
# evidence that we ever saw this mystery key.
pkl_file_to_remove
=
key_data
.
key_pkl
if
not
key_data
.
key_pkl
.
startswith
(
"/tmp"
):
# Under /tmp, file are removed periodically by the
# os. So it is normal that this happen from time to
# time.
_logger
.
warning
(
"Removing key file
%
s because the "
"corresponding module is gone from the "
"file system."
,
pkl_file_to_remove
)
self
.
loaded_key_pkl
.
remove
(
pkl_file_to_remove
)
if
to_delete
:
with
compilelock
.
lock_ctx
():
for
a
,
kw
in
to_delete
:
_rmtree
(
*
a
,
**
kw
)
_logger
.
debug
(
'Time needed to refresh cache:
%
s'
,
(
time
.
time
()
-
start_time
))
...
...
@@ -919,18 +927,13 @@ class ModuleCache(object):
"previous one"
)
key_data
=
self
.
module_hash_to_key_data
[
module_hash
]
module
=
self
.
_get_from_key
(
None
,
key_data
)
lock_taken
=
False
try
:
compilelock
.
get_lock
()
lock_taken
=
True
key_data
.
add_key
(
key
,
save_pkl
=
bool
(
key
[
0
]))
key_broken
=
False
except
cPickle
.
PicklingError
:
key_data
.
remove_key
(
key
)
key_broken
=
True
finally
:
if
lock_taken
and
not
keep_lock
:
compilelock
.
release_lock
()
with
compilelock
.
lock_ctx
(
keep_lock
=
keep_lock
):
try
:
key_data
.
add_key
(
key
,
save_pkl
=
bool
(
key
[
0
]))
key_broken
=
False
except
cPickle
.
PicklingError
:
key_data
.
remove_key
(
key
)
key_broken
=
True
if
(
key
[
0
]
and
not
key_broken
and
self
.
check_for_broken_eq
):
self
.
check_key
(
key
,
key_data
.
key_pkl
)
...
...
@@ -1026,16 +1029,11 @@ class ModuleCache(object):
if
module
is
not
None
:
return
module
try
:
# Compile the module since it's not cached
compilelock
.
get_lock
()
lock_taken
=
True
with
compilelock
.
lock_ctx
(
keep_lock
=
keep_lock
):
# Maybe somebody else compiled it for us while we
# where waiting for the lock. Try to load it again
self
.
refresh
(
cleanup
=
False
)
# Need no_clean_empty otherwise it deletes the workdir we
# created above.
self
.
refresh
()
module
=
self
.
_get_from_key
(
key
)
if
module
is
not
None
:
return
module
...
...
@@ -1073,10 +1071,6 @@ class ModuleCache(object):
key_data
=
self
.
_add_to_cache
(
module
,
key
,
module_hash
)
self
.
module_hash_to_key_data
[
module_hash
]
=
key_data
finally
:
# Release lock if needed.
if
not
keep_lock
and
lock_taken
:
compilelock
.
release_lock
()
self
.
stats
[
2
]
+=
1
return
module
...
...
@@ -1159,8 +1153,7 @@ class ModuleCache(object):
else
:
age_thresh_use
=
None
compilelock
.
get_lock
()
try
:
with
compilelock
.
lock_ctx
():
# Update the age of modules that have been accessed by other
# processes and get all module that are too old to use
# (not loaded in self.entry_from_key).
...
...
@@ -1179,9 +1172,6 @@ class ModuleCache(object):
_rmtree
(
parent
,
msg
=
'old cache directory'
,
level
=
logging
.
INFO
,
ignore_nocleanup
=
True
)
finally
:
compilelock
.
release_lock
()
def
clear
(
self
,
unversioned_min_age
=
None
,
clear_base_files
=
False
,
delete_if_problem
=
False
):
"""
...
...
@@ -1198,16 +1188,13 @@ class ModuleCache(object):
:param delete_if_problem: See help of refresh() method.
"""
compilelock
.
get_lock
()
try
:
with
compilelock
.
lock_ctx
():
self
.
clear_old
(
age_thresh_del
=-
1.0
,
delete_if_problem
=
delete_if_problem
)
self
.
clear_unversioned
(
min_age
=
unversioned_min_age
)
if
clear_base_files
:
self
.
clear_base_files
()
finally
:
compilelock
.
release_lock
()
def
clear_base_files
(
self
):
"""
...
...
@@ -1219,8 +1206,7 @@ class ModuleCache(object):
rename them with the '.delete.me' extension, to mark them to be deleted
next time we clear the cache.
"""
compilelock
.
get_lock
()
try
:
with
compilelock
.
lock_ctx
()
for
base_dir
in
(
'cuda_ndarray'
,
'cutils_ext'
,
'lazylinker_ext'
,
'scan_perform'
):
to_delete
=
os
.
path
.
join
(
self
.
dirname
,
base_dir
+
'.delete.me'
)
...
...
@@ -1238,8 +1224,6 @@ class ModuleCache(object):
except
Exception
:
_logger
.
warning
(
'Could not move
%
s to
%
s'
,
to_rename
,
to_delete
)
finally
:
compilelock
.
release_lock
()
def
clear_unversioned
(
self
,
min_age
=
None
):
"""
...
...
@@ -1254,9 +1238,8 @@ class ModuleCache(object):
if
min_age
is
None
:
min_age
=
self
.
age_thresh_del_unversioned
compilelock
.
get_lock
()
all_key_datas
=
self
.
module_hash_to_key_data
.
values
()
try
:
with
compilelock
.
lock_ctx
():
all_key_datas
=
self
.
module_hash_to_key_data
.
values
()
for
key_data
in
all_key_datas
:
if
not
key_data
.
keys
:
# May happen for broken versioned keys.
...
...
@@ -1329,17 +1312,12 @@ class ModuleCache(object):
_rmtree
(
os
.
path
.
join
(
self
.
dirname
,
filename
),
msg
=
'old unversioned'
,
level
=
logging
.
INFO
,
ignore_nocleanup
=
True
)
finally
:
compilelock
.
release_lock
()
def
_on_atexit
(
self
):
# Note: no need to call refresh() since it is called by clear_old().
compilelock
.
get_lock
()
try
:
with
compilelock
.
lock_ctx
()
self
.
clear_old
()
self
.
clear_unversioned
()
finally
:
compilelock
.
release_lock
()
_logger
.
debug
(
'Time spent checking keys:
%
s'
,
self
.
time_spent_in_check_key
)
...
...
theano/gof/compilelock.py
浏览文件 @
d1222d9c
...
...
@@ -7,6 +7,7 @@ import random
import
socket
# only used for gethostname()
import
time
import
logging
import
contextlib
from
theano
import
config
from
theano.configparser
import
AddConfigVar
,
IntParam
...
...
@@ -44,6 +45,14 @@ def force_unlock():
release_lock
()
@contextmanager
def
lock_ctx
(
lock_dir
=
None
,
keep_lock
=
False
,
**
kw
):
get_lock
(
lock_dir
=
lock_dir
,
**
kw
)
yield
if
not
keep_lock
:
release_lock
()
def
get_lock
(
lock_dir
=
None
,
**
kw
):
"""
Obtain lock on compilation directory.
...
...
theano/updates.py
100755 → 100644
浏览文件 @
d1222d9c
File mode changed from 100755 to 100644
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论