Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
erp5-Boxiang
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Hamza
erp5-Boxiang
Commits
0ab7d0ea
Commit
0ab7d0ea
authored
Oct 23, 2012
by
Sebastien Robin
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
fixed indentation to have proper try/except messages
parent
bb451659
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
75 additions
and
75 deletions
+75
-75
erp5/util/testnode/testnode.py
erp5/util/testnode/testnode.py
+75
-75
No files found.
erp5/util/testnode/testnode.py
View file @
0ab7d0ea
...
...
@@ -281,7 +281,7 @@ branch = %(branch)s
log
=
self
.
log
log
(
'Testnode.run, finally close'
)
process_manager
.
killPreviousRun
()
if
0
and
test_result
is
not
None
:
if
test_result
is
not
None
:
try
:
test_result
.
removeWatch
(
log_file_name
)
except
KeyError
:
...
...
@@ -298,84 +298,84 @@ branch = %(branch)s
test_result
=
None
try
:
while
True
:
begin
=
time
.
time
()
portal_url
=
config
[
'test_suite_master_url'
]
portal
=
taskdistribution
.
TaskDistributionTool
(
portal_url
,
logger
=
DummyLogger
(
log
))
test_suite_portal
=
taskdistribution
.
TaskDistributor
(
portal_url
,
logger
=
DummyLogger
(
log
))
test_suite_json
=
test_suite_portal
.
startTestSuite
(
config
[
'test_node_title'
]
)
test_suite_data
=
json
.
loads
(
test_suite_json
)
#Clean-up test suites
self
.
checkOldTestSuite
(
test_suite_data
)
for
test_suite
in
test_suite_data
:
self
.
updateConfigForTestSuite
(
test_suite
)
run_software
=
True
self
.
process_manager
.
supervisord_pid_file
=
os
.
path
.
join
(
config
[
'instance_root'
],
'var'
,
'run'
,
'supervisord.pid'
)
# Write our own software.cfg to use the local repository
vcs_repository_list
=
self
.
constructProfile
()
retry
=
False
retry_software_count
=
0
same_revision
_count
=
0
test_suite_title
=
config
[
'test_suite_title'
]
try
:
try
:
begin
=
time
.
time
()
portal_url
=
config
[
'test_suite_master_url'
]
portal
=
taskdistribution
.
TaskDistributionTool
(
portal_url
,
logger
=
DummyLogger
(
log
))
test_suite_portal
=
taskdistribution
.
TaskDistributor
(
portal_url
,
logger
=
DummyLogger
(
log
)
)
test_suite_json
=
test_suite_portal
.
startTestSuite
(
config
[
'test_node_title'
]
)
test_suite_data
=
json
.
loads
(
test_suite_json
)
#Clean-up test suites
self
.
checkOldTestSuite
(
test_suite_data
)
for
test_suite
in
test_suite_data
:
self
.
updateConfigForTestSuite
(
test_suite
)
run_software
=
True
self
.
process_manager
.
supervisord_pid_file
=
os
.
path
.
join
(
config
[
'instance_root'
],
'var'
,
'run'
,
'supervisord.pid'
)
# Write our own software.cfg to use the local repository
vcs_repository_list
=
self
.
constructProfile
()
retry
=
False
retry_software
_count
=
0
same_revision_count
=
0
test_suite_title
=
config
[
'test_suite_title'
]
# kill processes from previous loop if any
process_manager
.
killPreviousRun
()
full_revision_list
=
self
.
getFullRevisionList
(
revision_dict
)
process_manager
.
killPreviousRun
()
full_revision_list
=
self
.
getFullRevisionList
(
revision_dict
)
# Make sure we have local repository
now
=
time
.
time
()
try
:
if
previous_revision_dict
[
test_suite_title
]
==
revision_dict
[
test_suite_title
]:
log
(
'Same Revision'
)
same_revision_count
+=
1
if
not
(
retry
)
and
same_revision_count
<=
2
:
log
(
'Sleeping a bit since same revision'
)
time
.
sleep
(
DEFAULT_SLEEP_TIMEOUT
)
continue
same_revision_count
=
0
log
(
'Retrying install or checking if previous test was cancelled'
)
except
KeyError
:
pass
retry
=
False
previous_revision_dict
[
test_suite_title
]
=
revision_dict
[
test_suite_title
]
now
=
time
.
time
()
test_result
=
portal
.
createTestResult
(
revision_dict
[
test_suite_title
],[],
config
[
'test_node_title'
],
False
,
test_suite_title
,
config
[
'project_title'
])
remote_test_result_needs_cleanup
=
True
log
(
"testnode, test_result : %r"
%
(
test_result
,
))
if
test_result
is
not
None
:
log_file_name
=
self
.
addWatcher
(
test_result
)
self
.
checkRevision
(
test_result
,
revision_dict
,
previous_revision_dict
,
vcs_repository_list
)
# Now prepare the installation of SlapOS and create instance
status_dict
=
self
.
prepareSlapOS
(
retry_software_count
,
retry
)
# Give some time so computer partitions may start
# as partitions can be of any kind we have and likely will never have
# a reliable way to check if they are up or not ...
time
.
sleep
(
20
)
self
.
runTestSuite
(
revision_dict
,
portal_url
)
test_result
.
removeWatch
(
log_file_name
)
except
SubprocessError
,
e
:
log
(
"SubprocessError"
,
exc_info
=
sys
.
exc_info
())
now
=
time
.
time
()
try
:
if
previous_revision_dict
[
test_suite_title
]
==
revision_dict
[
test_suite_title
]:
log
(
'Same Revision'
)
same_revision_count
+=
1
if
not
(
retry
)
and
same_revision_count
<=
2
:
log
(
'Sleeping a bit since same revision'
)
time
.
sleep
(
DEFAULT_SLEEP_TIMEOUT
)
continue
same_revision_count
=
0
log
(
'Retrying install or checking if previous test was cancelled'
)
except
KeyError
:
pass
retry
=
False
previous_revision_dict
[
test_suite_title
]
=
revision_dict
[
test_suite_title
]
now
=
time
.
time
()
test_result
=
portal
.
createTestResult
(
revision_dict
[
test_suite_title
],[],
config
[
'test_node_title'
],
False
,
test_suite_title
,
config
[
'project_title'
])
remote_test_result_needs_cleanup
=
True
log
(
"testnode, test_result : %r"
%
(
test_result
,
))
if
test_result
is
not
None
:
log_file_name
=
self
.
addWatcher
(
test_result
)
self
.
checkRevision
(
test_result
,
revision_dict
,
previous_revision_dict
,
vcs_repository_list
)
# Now prepare the installation of SlapOS and create instance
status_dict
=
self
.
prepareSlapOS
(
retry_software_count
,
retry
)
# Give some time so computer partitions may start
# as partitions can be of any kind we have and likely will never have
# a reliable way to check if they are up or not ...
time
.
sleep
(
20
)
self
.
runTestSuite
(
revision_dict
,
portal_url
)
test_result
.
removeWatch
(
log_file_name
)
if
remote_test_result_needs_cleanup
:
status_dict
=
e
.
status_dict
or
{}
test_result
.
reportFailure
(
command
=
status_dict
.
get
(
'command'
),
stdout
=
status_dict
.
get
(
'stdout'
),
stderr
=
status_dict
.
get
(
'stderr'
),
)
log
(
"SubprocessError, going to sleep %s"
%
DEFAULT_SLEEP_TIMEOUT
)
time
.
sleep
(
DEFAULT_SLEEP_TIMEOUT
)
continue
except
CancellationError
,
e
:
log
(
"CancellationError"
,
exc_info
=
sys
.
exc_info
())
process_manager
.
under_cancellation
=
False
retry
=
True
continue
except
:
log
(
"erp5testnode exception"
,
exc_info
=
sys
.
exc_info
())
raise
except
SubprocessError
,
e
:
log
(
"SubprocessError"
,
exc_info
=
sys
.
exc_info
())
if
test_result
is
not
None
:
test_result
.
removeWatch
(
log_file_name
)
if
remote_test_result_needs_cleanup
:
status_dict
=
e
.
status_dict
or
{}
test_result
.
reportFailure
(
command
=
status_dict
.
get
(
'command'
),
stdout
=
status_dict
.
get
(
'stdout'
),
stderr
=
status_dict
.
get
(
'stderr'
),
)
log
(
"SubprocessError, going to sleep %s"
%
DEFAULT_SLEEP_TIMEOUT
)
time
.
sleep
(
DEFAULT_SLEEP_TIMEOUT
)
continue
except
CancellationError
,
e
:
log
(
"CancellationError"
,
exc_info
=
sys
.
exc_info
())
process_manager
.
under_cancellation
=
False
retry
=
True
continue
except
:
log
(
"erp5testnode exception"
,
exc_info
=
sys
.
exc_info
())
raise
if
(
now
-
begin
)
<
120
:
time
.
sleep
(
120
-
(
now
-
begin
))
finally
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment