Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
N
neoppod
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Carlos Ramos Carreño
neoppod
Commits
8ba42463
Commit
8ba42463
authored
Nov 27, 2019
by
Julien Muchembled
Browse files
Options
Browse Files
Download
Plain Diff
Merge protocol v0
parents
a33c624c
2b9e14e8
Changes
76
Hide whitespace changes
Inline
Side-by-side
Showing
76 changed files
with
2263 additions
and
2723 deletions
+2263
-2723
CHANGELOG.rst
CHANGELOG.rst
+35
-0
neo/admin/app.py
neo/admin/app.py
+10
-34
neo/admin/handler.py
neo/admin/handler.py
+43
-59
neo/client/__init__.py
neo/client/__init__.py
+7
-0
neo/client/app.py
neo/client/app.py
+27
-28
neo/client/handlers/master.py
neo/client/handlers/master.py
+7
-12
neo/client/transactions.py
neo/client/transactions.py
+1
-1
neo/lib/bootstrap.py
neo/lib/bootstrap.py
+5
-9
neo/lib/connection.py
neo/lib/connection.py
+46
-42
neo/lib/connector.py
neo/lib/connector.py
+7
-8
neo/lib/dispatcher.py
neo/lib/dispatcher.py
+1
-1
neo/lib/handler.py
neo/lib/handler.py
+8
-7
neo/lib/interfaces.py
neo/lib/interfaces.py
+15
-6
neo/lib/logger.py
neo/lib/logger.py
+12
-6
neo/lib/node.py
neo/lib/node.py
+1
-1
neo/lib/protocol.py
neo/lib/protocol.py
+476
-1460
neo/lib/pt.py
neo/lib/pt.py
+26
-26
neo/lib/util.py
neo/lib/util.py
+0
-59
neo/master/app.py
neo/master/app.py
+43
-35
neo/master/backup_app.py
neo/master/backup_app.py
+2
-6
neo/master/handlers/__init__.py
neo/master/handlers/__init__.py
+10
-11
neo/master/handlers/administration.py
neo/master/handlers/administration.py
+78
-37
neo/master/handlers/backup.py
neo/master/handlers/backup.py
+9
-5
neo/master/handlers/client.py
neo/master/handlers/client.py
+4
-0
neo/master/handlers/identification.py
neo/master/handlers/identification.py
+22
-9
neo/master/handlers/master.py
neo/master/handlers/master.py
+11
-9
neo/master/handlers/storage.py
neo/master/handlers/storage.py
+2
-2
neo/master/pt.py
neo/master/pt.py
+32
-25
neo/master/recovery.py
neo/master/recovery.py
+11
-15
neo/neoctl/app.py
neo/neoctl/app.py
+47
-8
neo/neoctl/handler.py
neo/neoctl/handler.py
+4
-2
neo/neoctl/neoctl.py
neo/neoctl/neoctl.py
+9
-3
neo/scripts/neolog.py
neo/scripts/neolog.py
+48
-24
neo/scripts/runner.py
neo/scripts/runner.py
+28
-15
neo/storage/app.py
neo/storage/app.py
+33
-44
neo/storage/checker.py
neo/storage/checker.py
+1
-1
neo/storage/database/importer.py
neo/storage/database/importer.py
+73
-26
neo/storage/database/manager.py
neo/storage/database/manager.py
+65
-77
neo/storage/database/mysqldb.py
neo/storage/database/mysqldb.py
+45
-22
neo/storage/database/sqlite.py
neo/storage/database/sqlite.py
+17
-3
neo/storage/handlers/__init__.py
neo/storage/handlers/__init__.py
+3
-3
neo/storage/handlers/client.py
neo/storage/handlers/client.py
+7
-6
neo/storage/handlers/identification.py
neo/storage/handlers/identification.py
+3
-4
neo/storage/handlers/initialization.py
neo/storage/handlers/initialization.py
+5
-4
neo/storage/handlers/storage.py
neo/storage/handlers/storage.py
+2
-2
neo/storage/replicator.py
neo/storage/replicator.py
+1
-1
neo/storage/transactions.py
neo/storage/transactions.py
+4
-1
neo/tests/__init__.py
neo/tests/__init__.py
+99
-14
neo/tests/functional/__init__.py
neo/tests/functional/__init__.py
+29
-33
neo/tests/functional/testClient.py
neo/tests/functional/testClient.py
+0
-2
neo/tests/functional/testCluster.py
neo/tests/functional/testCluster.py
+0
-3
neo/tests/functional/testMaster.py
neo/tests/functional/testMaster.py
+1
-1
neo/tests/functional/testStorage.py
neo/tests/functional/testStorage.py
+2
-2
neo/tests/master/testClientHandler.py
neo/tests/master/testClientHandler.py
+2
-4
neo/tests/master/testMasterApp.py
neo/tests/master/testMasterApp.py
+0
-1
neo/tests/master/testMasterPT.py
neo/tests/master/testMasterPT.py
+3
-1
neo/tests/master/testRecovery.py
neo/tests/master/testRecovery.py
+0
-105
neo/tests/master/testStorageHandler.py
neo/tests/master/testStorageHandler.py
+2
-3
neo/tests/protocol
neo/tests/protocol
+101
-0
neo/tests/protocol_checker.py
neo/tests/protocol_checker.py
+213
-0
neo/tests/storage/testMasterHandler.py
neo/tests/storage/testMasterHandler.py
+3
-3
neo/tests/storage/testStorageApp.py
neo/tests/storage/testStorageApp.py
+0
-122
neo/tests/storage/testStorageDBTests.py
neo/tests/storage/testStorageDBTests.py
+4
-75
neo/tests/storage/testStorageMySQL.py
neo/tests/storage/testStorageMySQL.py
+3
-3
neo/tests/stress.py
neo/tests/stress.py
+4
-6
neo/tests/testHandler.py
neo/tests/testHandler.py
+1
-8
neo/tests/testUtil.py
neo/tests/testUtil.py
+1
-19
neo/tests/threaded/__init__.py
neo/tests/threaded/__init__.py
+38
-26
neo/tests/threaded/test.py
neo/tests/threaded/test.py
+107
-65
neo/tests/threaded/testImporter.py
neo/tests/threaded/testImporter.py
+119
-32
neo/tests/threaded/testReplication.py
neo/tests/threaded/testReplication.py
+104
-14
neo/tests/zodb/testRecovery.py
neo/tests/zodb/testRecovery.py
+0
-2
setup.py
setup.py
+3
-2
tools/matrix
tools/matrix
+1
-1
tools/perfs
tools/perfs
+1
-1
tools/stress
tools/stress
+76
-16
No files found.
CHANGELOG.rst
View file @
8ba42463
Change History
==============
1.12 (2019-04-28)
-----------------
Most changes in this version focus on the ability to migrate efficiently
and reliably a big ZODB to NEO, which required changes in the protocol.
See testSplitAndMakeResilientUsingClone for an example of scenario.
Better cluster management:
- New --new-nid storage option for fast cloning.
- The number of wanted replicas is now a property of the database, which is
modifiable when the cluster is running, and reported by `neoctl print pt`.
- Better error reporting from the master to neoctl for denied requests.
- tweak: do not touch cells of nodes that are intended to be dropped.
- tweak: do not crash when trying to remove all nodes.
- tweak: new neoctl option to ask the master to simulate.
- neoctl: better display of full partition tables.
- master: reject drop/tweak commands that could lead to unwanted status.
Importer:
- Fix possible data loss on writeback.
- v1.9 broke replication (as source) once the import is finished.
- Speed up startup when the import is already finished.
- Fix closure of ZODB, and also do it when the import is finished.
- Fix hidden "maximum recursion depth exceeded" at startup.
- Fix resumption when using SQLite.
- v1.10 broke resumption when there are new transactions since the import
started.
MySQL:
- Better support of RocksDB by specifying column families.
- Fix handling of connection strings (--database) without credentials.
1.11 (2019-03-11)
-----------------
...
...
neo/admin/app.py
View file @
8ba42463
...
...
@@ -18,10 +18,8 @@ from neo.lib import logging
from
neo.lib.app
import
BaseApplication
,
buildOptionParser
from
neo.lib.connection
import
ListeningConnection
from
neo.lib.exception
import
PrimaryFailure
from
.handler
import
AdminEventHandler
,
MasterEventHandler
,
\
MasterRequestEventHandler
from
.handler
import
AdminEventHandler
,
MasterEventHandler
from
neo.lib.bootstrap
import
BootstrapManager
from
neo.lib.pt
import
PartitionTable
from
neo.lib.protocol
import
ClusterStates
,
Errors
,
NodeTypes
,
Packets
from
neo.lib.debug
import
register
as
registerLiveDebugger
...
...
@@ -36,8 +34,8 @@ class Application(BaseApplication):
cls
.
addCommonServerOptions
(
'admin'
,
'127.0.0.1:9999'
)
_
=
_
.
group
(
'admin'
)
_
.
int
(
'
u'
,
'uu
id'
,
help
=
"specify an
UU
ID to use for this process (testing purpose)"
)
_
.
int
(
'
i'
,
'n
id'
,
help
=
"specify an
N
ID to use for this process (testing purpose)"
)
def
__init__
(
self
,
config
):
super
(
Application
,
self
).
__init__
(
...
...
@@ -53,9 +51,8 @@ class Application(BaseApplication):
# The partition table is initialized after getting the number of
# partitions.
self
.
pt
=
None
self
.
uuid
=
config
.
get
(
'
uu
id'
)
self
.
uuid
=
config
.
get
(
'
n
id'
)
logging
.
node
(
self
.
name
,
self
.
uuid
)
self
.
request_handler
=
MasterRequestEventHandler
(
self
)
self
.
master_event_handler
=
MasterEventHandler
(
self
)
self
.
cluster_state
=
None
self
.
reset
()
...
...
@@ -66,7 +63,6 @@ class Application(BaseApplication):
super
(
Application
,
self
).
close
()
def
reset
(
self
):
self
.
bootstrapped
=
False
self
.
master_conn
=
None
self
.
master_node
=
None
...
...
@@ -117,40 +113,20 @@ class Application(BaseApplication):
self
.
cluster_state
=
None
# search, find, connect and identify to the primary master
bootstrap
=
BootstrapManager
(
self
,
NodeTypes
.
ADMIN
,
self
.
server
)
self
.
master_node
,
self
.
master_conn
,
num_partitions
,
num_replicas
=
\
bootstrap
.
getPrimaryConnection
()
if
self
.
pt
is
None
:
self
.
pt
=
PartitionTable
(
num_partitions
,
num_replicas
)
elif
self
.
pt
.
getPartitions
()
!=
num_partitions
:
# XXX: shouldn't we recover instead of raising ?
raise
RuntimeError
(
'the number of partitions is inconsistent'
)
elif
self
.
pt
.
getReplicas
()
!=
num_replicas
:
# XXX: shouldn't we recover instead of raising ?
raise
RuntimeError
(
'the number of replicas is inconsistent'
)
self
.
master_node
,
self
.
master_conn
=
bootstrap
.
getPrimaryConnection
()
# passive handler
self
.
master_conn
.
setHandler
(
self
.
master_event_handler
)
self
.
master_conn
.
ask
(
Packets
.
AskClusterState
())
self
.
master_conn
.
ask
(
Packets
.
AskPartitionTable
())
def
sendPartitionTable
(
self
,
conn
,
min_offset
,
max_offset
,
uuid
):
# we have a pt
self
.
pt
.
log
()
row_list
=
[]
pt
=
self
.
pt
if
max_offset
==
0
:
max_offset
=
self
.
pt
.
getPartitions
()
max_offset
=
pt
.
getPartitions
()
try
:
for
offset
in
xrange
(
min_offset
,
max_offset
):
row
=
[]
try
:
for
cell
in
self
.
pt
.
getCellList
(
offset
):
if
uuid
is
None
or
cell
.
getUUID
()
==
uuid
:
row
.
append
((
cell
.
getUUID
(),
cell
.
getState
()))
except
TypeError
:
pass
row_list
.
append
((
offset
,
row
))
row_list
=
map
(
pt
.
getRow
,
xrange
(
min_offset
,
max_offset
))
except
IndexError
:
conn
.
send
(
Errors
.
ProtocolError
(
'invalid partition table offset'
))
else
:
conn
.
answer
(
Packets
.
AnswerPartitionList
(
self
.
pt
.
getID
(),
row_list
))
conn
.
answer
(
Packets
.
AnswerPartitionList
(
pt
.
getID
(),
pt
.
getReplicas
(),
row_list
))
neo/admin/handler.py
View file @
8ba42463
...
...
@@ -17,30 +17,49 @@
from
neo.lib
import
logging
,
protocol
from
neo.lib.handler
import
EventHandler
from
neo.lib.protocol
import
uuid_str
,
Packets
from
neo.lib.pt
import
PartitionTable
from
neo.lib.exception
import
PrimaryFailure
def
check_primary_master
(
func
):
def
wrapper
(
self
,
*
args
,
**
kw
):
if
self
.
app
.
bootstrapped
:
return
func
(
self
,
*
args
,
**
kw
)
raise
protocol
.
NotReadyError
(
'Not connected to a primary master.'
)
return
wrapper
def
forward_ask
(
klass
):
return
check_primary_master
(
lambda
self
,
conn
,
*
args
,
**
kw
:
self
.
app
.
master_conn
.
ask
(
klass
(
*
args
,
**
kw
),
conn
=
conn
,
msg_id
=
conn
.
getPeerId
()))
def
AdminEventHandlerType
(
name
,
bases
,
d
):
def
check_primary_master
(
func
):
def
wrapper
(
self
,
*
args
,
**
kw
):
if
self
.
app
.
master_conn
is
not
None
:
return
func
(
self
,
*
args
,
**
kw
)
raise
protocol
.
NotReadyError
(
'Not connected to a primary master.'
)
return
wrapper
def
forward_ask
(
klass
):
return
lambda
self
,
conn
,
*
args
:
self
.
app
.
master_conn
.
ask
(
klass
(
*
args
),
conn
=
conn
,
msg_id
=
conn
.
getPeerId
())
del
d
[
'__metaclass__'
]
for
x
in
(
Packets
.
AddPendingNodes
,
Packets
.
AskLastIDs
,
Packets
.
AskLastTransaction
,
Packets
.
AskRecovery
,
Packets
.
CheckReplicas
,
Packets
.
Repair
,
Packets
.
SetClusterState
,
Packets
.
SetNodeState
,
Packets
.
SetNumReplicas
,
Packets
.
Truncate
,
Packets
.
TweakPartitionTable
,
):
d
[
x
.
handler_method_name
]
=
forward_ask
(
x
)
return
type
(
name
,
bases
,
{
k
:
v
if
k
[
0
]
==
'_'
else
check_primary_master
(
v
)
for
k
,
v
in
d
.
iteritems
()})
class
AdminEventHandler
(
EventHandler
):
"""This class deals with events for administrating cluster."""
@
check_primary_master
__metaclass__
=
AdminEventHandlerType
def
askPartitionList
(
self
,
conn
,
min_offset
,
max_offset
,
uuid
):
logging
.
info
(
"ask partition list from %s to %s for %s"
,
min_offset
,
max_offset
,
uuid_str
(
uuid
))
self
.
app
.
sendPartitionTable
(
conn
,
min_offset
,
max_offset
,
uuid
)
@
check_primary_master
def
askNodeList
(
self
,
conn
,
node_type
):
if
node_type
is
None
:
node_type
=
'all'
...
...
@@ -53,36 +72,22 @@ class AdminEventHandler(EventHandler):
p
=
Packets
.
AnswerNodeList
(
node_information_list
)
conn
.
answer
(
p
)
@
check_primary_master
def
askClusterState
(
self
,
conn
):
conn
.
answer
(
Packets
.
AnswerClusterState
(
self
.
app
.
cluster_state
))
@
check_primary_master
def
askPrimary
(
self
,
conn
):
master_node
=
self
.
app
.
master_node
conn
.
answer
(
Packets
.
AnswerPrimary
(
master_node
.
getUUID
()))
@
check_primary_master
def
flushLog
(
self
,
conn
):
self
.
app
.
master_conn
.
send
(
Packets
.
FlushLog
())
super
(
AdminEventHandler
,
self
).
flushLog
(
conn
)
askLastIDs
=
forward_ask
(
Packets
.
AskLastIDs
)
askLastTransaction
=
forward_ask
(
Packets
.
AskLastTransaction
)
addPendingNodes
=
forward_ask
(
Packets
.
AddPendingNodes
)
askRecovery
=
forward_ask
(
Packets
.
AskRecovery
)
tweakPartitionTable
=
forward_ask
(
Packets
.
TweakPartitionTable
)
setClusterState
=
forward_ask
(
Packets
.
SetClusterState
)
setNodeState
=
forward_ask
(
Packets
.
SetNodeState
)
checkReplicas
=
forward_ask
(
Packets
.
CheckReplicas
)
truncate
=
forward_ask
(
Packets
.
Truncate
)
repair
=
forward_ask
(
Packets
.
Repair
)
class
MasterEventHandler
(
EventHandler
):
""" This class is just used to dispatch message to right handler"""
def
_connectionLost
(
self
,
conn
):
def
connectionClosed
(
self
,
conn
):
app
=
self
.
app
if
app
.
listening_conn
:
# if running
assert
app
.
master_conn
in
(
conn
,
None
)
...
...
@@ -91,42 +96,21 @@ class MasterEventHandler(EventHandler):
app
.
uuid
=
None
raise
PrimaryFailure
def
connectionFailed
(
self
,
conn
):
self
.
_connectionLost
(
conn
)
def
connectionClosed
(
self
,
conn
):
self
.
_connectionLost
(
conn
)
def
dispatch
(
self
,
conn
,
packet
,
kw
=
{}):
if
'conn'
in
kw
:
# expected answer
if
packet
.
isResponse
():
packet
.
setId
(
kw
[
'msg_id'
])
kw
[
'conn'
].
answer
(
packet
)
else
:
self
.
app
.
request_handler
.
dispatch
(
conn
,
packet
,
kw
)
else
:
# unexpected answers and notifications
forward
=
kw
.
get
(
'conn'
)
if
forward
is
None
:
super
(
MasterEventHandler
,
self
).
dispatch
(
conn
,
packet
,
kw
)
else
:
forward
.
send
(
packet
,
kw
[
'msg_id'
])
def
answerClusterState
(
self
,
conn
,
state
):
self
.
app
.
cluster_state
=
state
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
cell_list
):
self
.
app
.
pt
.
update
(
ptid
,
cell_list
,
self
.
app
.
nm
)
def
answerPartitionTable
(
self
,
conn
,
ptid
,
row_list
):
self
.
app
.
pt
.
load
(
ptid
,
row_list
,
self
.
app
.
nm
)
self
.
app
.
bootstrapped
=
True
def
sendPartitionTable
(
self
,
conn
,
ptid
,
row_list
):
if
self
.
app
.
bootstrapped
:
self
.
app
.
pt
.
load
(
ptid
,
row_list
,
self
.
app
.
nm
)
def
notifyClusterInformation
(
self
,
conn
,
cluster_state
):
self
.
app
.
cluster_state
=
cluster_state
notifyClusterInformation
=
answerClusterState
def
sendPartitionTable
(
self
,
conn
,
ptid
,
num_replicas
,
row_list
):
pt
=
self
.
app
.
pt
=
object
.
__new__
(
PartitionTable
)
pt
.
load
(
ptid
,
num_replicas
,
row_list
,
self
.
app
.
nm
)
class
MasterRequestEventHandler
(
EventHandler
):
""" This class handle all answer from primary master node"""
# XXX: to be deleted ?
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
num_replicas
,
cell_list
):
self
.
app
.
pt
.
update
(
ptid
,
num_replicas
,
cell_list
,
self
.
app
.
nm
)
neo/client/__init__.py
View file @
8ba42463
...
...
@@ -13,6 +13,13 @@
##############################################################################
def
patch
():
# For msgpack & Py2/ZODB5.
try
:
from
zodbpickle
import
binary
binary
.
_pack
=
bytes
.
__str__
except
ImportError
:
pass
from
hashlib
import
md5
from
ZODB.Connection
import
Connection
...
...
neo/client/app.py
View file @
8ba42463
...
...
@@ -18,6 +18,7 @@ import heapq
import
random
import
time
from
collections
import
defaultdict
try
:
from
ZODB._compat
import
dumps
,
loads
,
_protocol
except
ImportError
:
...
...
@@ -76,11 +77,11 @@ class Application(ThreadedApplication):
self
.
primary_master_node
=
None
self
.
trying_master_node
=
None
# no self-assigned
UU
ID, primary master will supply us one
# no self-assigned
N
ID, primary master will supply us one
self
.
_cache
=
ClientCache
()
if
cache_size
is
None
else
\
ClientCache
(
max_size
=
cache_size
)
self
.
_loading
=
defaultdict
(
lambda
:
(
Lock
(),
[]))
self
.
new_oid
_list
=
()
self
.
new_oid
s
=
()
self
.
last_oid
=
'
\
0
'
*
8
self
.
storage_event_handler
=
storage
.
StorageEventHandler
(
self
)
self
.
storage_bootstrap_handler
=
storage
.
StorageBootstrapHandler
(
self
)
...
...
@@ -181,7 +182,7 @@ class Application(ThreadedApplication):
with
self
.
_connecting_to_master_node
:
result
=
self
.
master_conn
if
result
is
None
:
self
.
new_oid
_list
=
()
self
.
new_oid
s
=
()
result
=
self
.
master_conn
=
self
.
_connectToPrimaryNode
()
return
result
...
...
@@ -220,8 +221,8 @@ class Application(ThreadedApplication):
self
.
notifications_handler
,
node
=
node
,
dispatcher
=
self
.
dispatcher
)
p
=
Packets
.
RequestIdentification
(
NodeTypes
.
CLIENT
,
self
.
uuid
,
None
,
self
.
name
,
(),
None
)
p
=
Packets
.
RequestIdentification
(
NodeTypes
.
CLIENT
,
self
.
uuid
,
None
,
self
.
name
,
None
,
(),
()
)
try
:
ask
(
conn
,
p
,
handler
=
handler
)
except
ConnectionClosed
:
...
...
@@ -238,7 +239,6 @@ class Application(ThreadedApplication):
# operational. Might raise ConnectionClosed so that the new
# primary can be looked-up again.
logging
.
info
(
'Initializing from master'
)
ask
(
conn
,
Packets
.
AskPartitionTable
(),
handler
=
handler
)
ask
(
conn
,
Packets
.
AskLastTransaction
(),
handler
=
handler
)
if
self
.
pt
.
operational
():
break
...
...
@@ -264,7 +264,7 @@ class Application(ThreadedApplication):
conn
=
MTClientConnection
(
self
,
self
.
storage_event_handler
,
node
,
dispatcher
=
self
.
dispatcher
)
p
=
Packets
.
RequestIdentification
(
NodeTypes
.
CLIENT
,
self
.
uuid
,
None
,
self
.
name
,
(),
self
.
id_timestamp
)
self
.
uuid
,
None
,
self
.
name
,
self
.
id_timestamp
,
(),
()
)
try
:
self
.
_ask
(
conn
,
p
,
handler
=
self
.
storage_bootstrap_handler
)
except
ConnectionClosed
:
...
...
@@ -306,15 +306,19 @@ class Application(ThreadedApplication):
"""Get a new OID."""
self
.
_oid_lock_acquire
()
try
:
if
not
self
.
new_oid_list
:
for
oid
in
self
.
new_oids
:
break
else
:
# Get new oid list from master node
# we manage a list of oid here to prevent
# from asking too many time new oid one by one
# from master node
self
.
_askPrimary
(
Packets
.
AskNewOIDs
(
100
))
if
not
self
.
new_oid_list
:
for
oid
in
self
.
new_oids
:
break
else
:
raise
NEOStorageError
(
'new_oid failed'
)
self
.
last_oid
=
oid
=
self
.
new_oid_list
.
pop
()
self
.
last_oid
=
oid
return
oid
finally
:
self
.
_oid_lock_release
()
...
...
@@ -611,7 +615,7 @@ class Application(ThreadedApplication):
# user and description are cast to str in case they're unicode.
# BBB: This is not required anymore with recent ZODB.
packet
=
Packets
.
AskStoreTransaction
(
ttid
,
str
(
transaction
.
user
),
str
(
transaction
.
description
),
ext
,
txn_context
.
cache_dict
)
str
(
transaction
.
description
),
ext
,
list
(
txn_context
.
cache_dict
)
)
queue
=
txn_context
.
queue
conn_dict
=
txn_context
.
conn_dict
# Ask in parallel all involved storage nodes to commit object metadata.
...
...
@@ -696,7 +700,7 @@ class Application(ThreadedApplication):
else
:
try
:
notify
(
Packets
.
AbortTransaction
(
txn_context
.
ttid
,
txn_context
.
conn_dict
))
list
(
txn_context
.
conn_dict
)
))
except
ConnectionClosed
:
pass
# We don't need to flush queue, as it won't be reused by future
...
...
@@ -731,7 +735,8 @@ class Application(ThreadedApplication):
for
oid
in
checked_list
:
del
cache_dict
[
oid
]
ttid
=
txn_context
.
ttid
p
=
Packets
.
AskFinishTransaction
(
ttid
,
cache_dict
,
checked_list
)
p
=
Packets
.
AskFinishTransaction
(
ttid
,
list
(
cache_dict
),
checked_list
)
try
:
tid
=
self
.
_askPrimary
(
p
,
cache_dict
=
cache_dict
,
callback
=
f
)
assert
tid
...
...
@@ -765,17 +770,11 @@ class Application(ThreadedApplication):
def
undo
(
self
,
undone_tid
,
txn
):
txn_context
=
self
.
_txn_container
.
get
(
txn
)
txn_info
,
txn_ext
=
self
.
_getTransactionInformation
(
undone_tid
)
txn_oid_list
=
txn_info
[
'oids'
]
# Regroup objects per partition, to ask a minimum set of storage.
partition_oid_dict
=
{}
for
oid
in
txn_oid_list
:
partition
=
self
.
pt
.
getPartition
(
oid
)
try
:
oid_list
=
partition_oid_dict
[
partition
]
except
KeyError
:
oid_list
=
partition_oid_dict
[
partition
]
=
[]
oid_list
.
append
(
oid
)
partition_oid_dict
=
defaultdict
(
list
)
for
oid
in
txn_info
[
'oids'
]:
partition_oid_dict
[
self
.
pt
.
getPartition
(
oid
)].
append
(
oid
)
# Ask storage the undo serial (serial at which object's previous data
# is)
...
...
@@ -817,8 +816,8 @@ class Application(ThreadedApplication):
raise
UndoError
(
'non-undoable transaction'
)
# Send undo data to all storage nodes.
for
oid
in
txn_oid_list
:
current_serial
,
undo_serial
,
is_current
=
undo_object_tid_dict
[
oid
]
for
oid
,
(
current_serial
,
undo_serial
,
is_current
)
in
\
undo_object_tid_dict
.
iteritems
():
if
is_current
:
data
=
None
else
:
...
...
@@ -852,7 +851,7 @@ class Application(ThreadedApplication):
self
.
_store
(
txn_context
,
oid
,
current_serial
,
data
,
undo_serial
)
self
.
waitStoreResponses
(
txn_context
)
return
None
,
txn_oid_list
return
None
,
list
(
undo_object_tid_dict
)
def
_getTransactionInformation
(
self
,
tid
):
return
self
.
_askStorageForRead
(
tid
,
...
...
@@ -933,9 +932,9 @@ class Application(ThreadedApplication):
for
serial
,
size
in
self
.
_askStorageForRead
(
oid
,
packet
):
txn_info
,
txn_ext
=
self
.
_getTransactionInformation
(
serial
)
# create history dict
txn_info
.
pop
(
'id'
)
txn_info
.
pop
(
'oids'
)
txn_info
.
pop
(
'packed'
)
del
txn_info
[
'id'
]
del
txn_info
[
'oids'
]
del
txn_info
[
'packed'
]
txn_info
[
'tid'
]
=
serial
txn_info
[
'version'
]
=
''
txn_info
[
'size'
]
=
size
...
...
neo/client/handlers/master.py
View file @
8ba42463
...
...
@@ -26,10 +26,6 @@ from ..exception import NEOStorageError
class
PrimaryBootstrapHandler
(
AnswerBaseHandler
):
""" Bootstrap handler used when looking for the primary master """
def
answerPartitionTable
(
self
,
conn
,
ptid
,
row_list
):
assert
row_list
self
.
app
.
pt
.
load
(
ptid
,
row_list
,
self
.
app
.
nm
)
def
answerLastTransaction
(
*
args
):
pass
...
...
@@ -42,9 +38,6 @@ class PrimaryNotificationsHandler(MTEventHandler):
except
PrimaryElected
,
e
:
self
.
app
.
primary_master_node
,
=
e
.
args
def
_acceptIdentification
(
self
,
node
,
num_partitions
,
num_replicas
):
self
.
app
.
pt
=
PartitionTable
(
num_partitions
,
num_replicas
)
def
answerLastTransaction
(
self
,
conn
,
ltid
):
app
=
self
.
app
app_last_tid
=
app
.
__dict__
.
get
(
'last_tid'
,
''
)
...
...
@@ -131,9 +124,12 @@ class PrimaryNotificationsHandler(MTEventHandler):
if
db
is
not
None
:
db
.
invalidate
(
tid
,
oid_list
)
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
cell_list
):
if
self
.
app
.
pt
.
filled
():
self
.
app
.
pt
.
update
(
ptid
,
cell_list
,
self
.
app
.
nm
)
def
sendPartitionTable
(
self
,
conn
,
ptid
,
num_replicas
,
row_list
):
pt
=
self
.
app
.
pt
=
object
.
__new__
(
PartitionTable
)
pt
.
load
(
ptid
,
num_replicas
,
row_list
,
self
.
app
.
nm
)
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
num_replicas
,
cell_list
):
self
.
app
.
pt
.
update
(
ptid
,
num_replicas
,
cell_list
,
self
.
app
.
nm
)
def
notifyNodeInformation
(
self
,
conn
,
timestamp
,
node_list
):
super
(
PrimaryNotificationsHandler
,
self
).
notifyNodeInformation
(
...
...
@@ -161,8 +157,7 @@ class PrimaryAnswersHandler(AnswerBaseHandler):
self
.
app
.
setHandlerData
(
ttid
)
def
answerNewOIDs
(
self
,
conn
,
oid_list
):
oid_list
.
reverse
()
self
.
app
.
new_oid_list
=
oid_list
self
.
app
.
new_oids
=
iter
(
oid_list
)
def
incompleteTransaction
(
self
,
conn
,
message
):
raise
NEOStorageError
(
"storage nodes for which vote failed can not be"
...
...
neo/client/transactions.py
View file @
8ba42463
...
...
@@ -26,7 +26,7 @@ from .exception import NEOStorageError
class
_WakeupPacket
(
object
):
handler_method_name
=
'pong'
decode
=
tuple
_args
=
()
getId
=
int
class
Transaction
(
object
):
...
...
neo/lib/bootstrap.py
View file @
8ba42463
...
...
@@ -26,7 +26,7 @@ class BootstrapManager(EventHandler):
Manage the bootstrap stage, lookup for the primary master then connect to it
"""
def
__init__
(
self
,
app
,
node_type
,
server
=
None
,
devpath
=
()):
def
__init__
(
self
,
app
,
node_type
,
server
=
None
,
devpath
=
()
,
new_nid
=
()
):
"""
Manage the bootstrap stage of a non-master node, it lookup for the
primary master node, connect to it then returns when the master node
...
...
@@ -34,9 +34,8 @@ class BootstrapManager(EventHandler):
"""
self
.
server
=
server
self
.
devpath
=
devpath
self
.
new_nid
=
new_nid
self
.
node_type
=
node_type
self
.
num_replicas
=
None
self
.
num_partitions
=
None
app
.
nm
.
reset
()
uuid
=
property
(
lambda
self
:
self
.
app
.
uuid
)
...
...
@@ -44,7 +43,7 @@ class BootstrapManager(EventHandler):
def
connectionCompleted
(
self
,
conn
):
EventHandler
.
connectionCompleted
(
self
,
conn
)
conn
.
ask
(
Packets
.
RequestIdentification
(
self
.
node_type
,
self
.
uuid
,
self
.
server
,
self
.
app
.
name
,
self
.
devpath
,
None
))
self
.
server
,
self
.
app
.
name
,
None
,
self
.
devpath
,
self
.
new_nid
))
def
connectionFailed
(
self
,
conn
):
EventHandler
.
connectionFailed
(
self
,
conn
)
...
...
@@ -53,10 +52,8 @@ class BootstrapManager(EventHandler):
def
connectionLost
(
self
,
conn
,
new_state
):
self
.
current
=
None
def
_acceptIdentification
(
self
,
node
,
num_partitions
,
num_replicas
):
def
_acceptIdentification
(
self
,
node
):
assert
self
.
current
is
node
,
(
self
.
current
,
node
)
self
.
num_partitions
=
num_partitions
self
.
num_replicas
=
num_replicas
def
getPrimaryConnection
(
self
):
"""
...
...
@@ -73,8 +70,7 @@ class BootstrapManager(EventHandler):
try
:
while
self
.
current
:
if
self
.
current
.
isIdentified
():
return
(
self
.
current
,
self
.
current
.
getConnection
(),
self
.
num_partitions
,
self
.
num_replicas
)
return
self
.
current
,
self
.
current
.
getConnection
()
poll
(
1
)
except
PrimaryElected
,
e
:
if
self
.
current
:
...
...
neo/lib/connection.py
View file @
8ba42463
...
...
@@ -16,12 +16,19 @@
from
functools
import
wraps
from
time
import
time
import
msgpack
from
msgpack.exceptions
import
UnpackValueError
from
.
import
attributeTracker
,
logging
from
.connector
import
ConnectorException
,
ConnectorDelayedConnection
from
.locking
import
RLock
from
.protocol
import
uuid_str
,
Errors
,
PacketMalformedError
,
Packets
from
.util
import
dummy_read_buffer
,
ReadBuffer
from
.protocol
import
uuid_str
,
Errors
,
PacketMalformedError
,
Packets
,
\
Unpacker
@
apply
class
dummy_read_buffer
(
msgpack
.
Unpacker
):
def
feed
(
self
,
_
):
pass
class
ConnectionClosed
(
Exception
):
pass
...
...
@@ -209,7 +216,7 @@ class BaseConnection(object):
def
_getReprInfo
(
self
):
r
=
[
(
'
uu
id'
,
uuid_str
(
self
.
getUUID
())),
(
'
n
id'
,
uuid_str
(
self
.
getUUID
())),
(
'address'
,
(
'[%s]:%s'
if
':'
in
self
.
addr
[
0
]
else
'%s:%s'
)
%
self
.
addr
if
self
.
addr
else
'?'
),
(
'handler'
,
self
.
getHandler
()),
...
...
@@ -291,7 +298,7 @@ class ListeningConnection(BaseConnection):
# message.
else
:
conn
.
_connected
()
self
.
em
.
addWriter
(
conn
)
# for
ENCODED_VERSION
self
.
em
.
addWriter
(
conn
)
# for
HANDSHAKE_PACKET
def
getAddress
(
self
):
return
self
.
connector
.
getAddress
()
...
...
@@ -310,12 +317,12 @@ class Connection(BaseConnection):
client
=
False
server
=
False
peer_id
=
None
_
parser_state
=
None
_
total_unpacked
=
0
_timeout
=
None
def
__init__
(
self
,
event_manager
,
*
args
,
**
kw
):
BaseConnection
.
__init__
(
self
,
event_manager
,
*
args
,
**
kw
)
self
.
read_buf
=
ReadBuff
er
()
self
.
read_buf
=
Unpack
er
()
self
.
cur_id
=
0
self
.
aborted
=
False
self
.
uuid
=
None
...
...
@@ -425,42 +432,39 @@ class Connection(BaseConnection):
self
.
_closure
()
def
_parse
(
self
):
read
=
self
.
read_buf
.
read
version
=
read
(
4
)
if
version
is
None
:
return
from
.protocol
import
(
ENCODED_VERSION
,
MAX_PACKET_SIZE
,
PACKET_HEADER_FORMAT
,
Packets
)
if
version
!=
ENCODED_VERSION
:
logging
.
warning
(
'Protocol version mismatch with %r'
,
self
)
from
.protocol
import
HANDSHAKE_PACKET
,
MAGIC_SIZE
,
Packets
read_buf
=
self
.
read_buf
handshake
=
read_buf
.
read_bytes
(
len
(
HANDSHAKE_PACKET
))
if
handshake
!=
HANDSHAKE_PACKET
:
if
HANDSHAKE_PACKET
.
startswith
(
handshake
):
# unlikely so tested last
# Not enough data and there's no API to know it in advance.
# Put it back.
read_buf
.
feed
(
handshake
)
return
if
HANDSHAKE_PACKET
.
startswith
(
handshake
[:
MAGIC_SIZE
]):
logging
.
warning
(
'Protocol version mismatch with %r'
,
self
)
else
:
logging
.
debug
(
'Rejecting non-NEO %r'
,
self
)
raise
ConnectorException
header_size
=
PACKET_HEADER_FORMAT
.
size
unpack
=
PACKET_HEADER_FORMAT
.
unpack
read_next
=
read_buf
.
next
read_pos
=
read_buf
.
tell
def
parse
():
state
=
self
.
_parser_state
if
state
is
None
:
header
=
read
(
header_size
)
if
header
is
None
:
return
msg_id
,
msg_type
,
msg_len
=
unpack
(
header
)
try
:
packet_klass
=
Packets
[
msg_type
]
except
KeyError
:
raise
PacketMalformedError
(
'Unknown packet type'
)
if
msg_len
>
MAX_PACKET_SIZE
:
raise
PacketMalformedError
(
'message too big (%d)'
%
msg_len
)
else
:
msg_id
,
packet_klass
,
msg_len
=
state
data
=
read
(
msg_len
)
if
data
is
None
:
# Not enough.
if
state
is
None
:
self
.
_parser_state
=
msg_id
,
packet_klass
,
msg_len
else
:
self
.
_parser_state
=
None
packet
=
packet_klass
()
packet
.
setContent
(
msg_id
,
data
)
return
packet
try
:
msg_id
,
msg_type
,
args
=
read_next
()
except
StopIteration
:
return
except
UnpackValueError
as
e
:
raise
PacketMalformedError
(
str
(
e
))
try
:
packet_klass
=
Packets
[
msg_type
]
except
KeyError
:
raise
PacketMalformedError
(
'Unknown packet type'
)
pos
=
read_pos
()
packet
=
packet_klass
(
*
args
)
packet
.
setId
(
msg_id
)
packet
.
size
=
pos
-
self
.
_total_unpacked
self
.
_total_unpacked
=
pos
return
packet
self
.
_parse
=
parse
return
parse
()
...
...
@@ -513,7 +517,7 @@ class Connection(BaseConnection):
def
close
(
self
):
if
self
.
connector
is
None
:
assert
self
.
_on_close
is
None
assert
not
self
.
read_buf
assert
not
self
.
read_buf
.
read_bytes
(
1
)
assert
not
self
.
isPending
()
return
# process the network events with the last registered handler to
...
...
@@ -524,7 +528,7 @@ class Connection(BaseConnection):
if
self
.
_on_close
is
not
None
:
self
.
_on_close
()
self
.
_on_close
=
None
self
.
read_buf
.
clear
()
self
.
read_buf
=
dummy_read_buffer
try
:
if
self
.
connecting
:
handler
.
connectionFailed
(
self
)
...
...
neo/lib/connector.py
View file @
8ba42463
...
...
@@ -19,7 +19,7 @@ import ssl
import
errno
from
time
import
time
from
.
import
logging
from
.protocol
import
ENCODED_VERSION
from
.protocol
import
HANDSHAKE_PACKET
# Global connector registry.
# Fill by calling registerConnectorHandler.
...
...
@@ -74,15 +74,14 @@ class SocketConnector(object):
s
.
setsockopt
(
socket
.
SOL_SOCKET
,
socket
.
SO_KEEPALIVE
,
1
)
# disable Nagle algorithm to reduce latency
s
.
setsockopt
(
socket
.
IPPROTO_TCP
,
socket
.
TCP_NODELAY
,
1
)
self
.
queued
=
[
ENCODED_VERSION
]
self
.
queue_size
=
len
(
ENCODED_VERSION
)
self
.
queued
=
[
HANDSHAKE_PACKET
]
self
.
queue_size
=
len
(
HANDSHAKE_PACKET
)
return
self
def
queue
(
self
,
data
):
was_empty
=
not
self
.
queued
self
.
queued
+=
data
for
data
in
data
:
self
.
queue_size
+=
len
(
data
)
self
.
queued
.
append
(
data
)
self
.
queue_size
+=
len
(
data
)
return
was_empty
def
_error
(
self
,
op
,
exc
=
None
):
...
...
@@ -172,7 +171,7 @@ class SocketConnector(object):
except
socket
.
error
,
e
:
self
.
_error
(
'recv'
,
e
)
if
data
:
read_buf
.
appen
d
(
data
)
read_buf
.
fee
d
(
data
)
return
self
.
_error
(
'recv'
)
...
...
@@ -278,7 +277,7 @@ class _SSL:
def
receive
(
self
,
read_buf
):
try
:
while
1
:
read_buf
.
appen
d
(
self
.
socket
.
recv
(
4096
))
read_buf
.
fee
d
(
self
.
socket
.
recv
(
4096
))
except
ssl
.
SSLWantReadError
:
pass
except
socket
.
error
,
e
:
...
...
neo/lib/dispatcher.py
View file @
8ba42463
...
...
@@ -23,7 +23,7 @@ NOBODY = []
class
_ConnectionClosed
(
object
):
handler_method_name
=
'connectionClosed'
decode
=
tuple
_args
=
()
class
getId
(
object
):
def
__eq__
(
self
,
other
):
...
...
neo/lib/handler.py
View file @
8ba42463
...
...
@@ -26,6 +26,9 @@ from .protocol import (NodeStates, NodeTypes, Packets, uuid_str,
from
.util
import
cached_property
class
AnswerDenied
(
Exception
):
"""Helper exception to stop packet processing and answer a Denied error"""
class
DelayEvent
(
Exception
):
pass
...
...
@@ -68,7 +71,7 @@ class EventHandler(object):
method
=
getattr
(
self
,
packet
.
handler_method_name
)
except
AttributeError
:
raise
UnexpectedPacketError
(
'no handler found'
)
args
=
packet
.
decode
()
or
()
args
=
packet
.
_args
method
(
conn
,
*
args
,
**
kw
)
except
DelayEvent
,
e
:
assert
not
kw
,
kw
...
...
@@ -76,9 +79,6 @@ class EventHandler(object):
except
UnexpectedPacketError
,
e
:
if
not
conn
.
isClosed
():
self
.
__unexpectedPacket
(
conn
,
packet
,
*
e
.
args
)
except
PacketMalformedError
,
e
:
logging
.
error
(
'malformed packet from %r: %s'
,
conn
,
e
)
conn
.
close
()
except
NotReadyError
,
message
:
if
not
conn
.
isClosed
():
if
not
message
.
args
:
...
...
@@ -98,6 +98,8 @@ class EventHandler(object):
%
(
m
.
im_class
.
__module__
,
m
.
im_class
.
__name__
,
m
.
__name__
)))
except
NonReadableCell
,
e
:
conn
.
answer
(
Errors
.
NonReadableCell
())
except
AnswerDenied
,
e
:
conn
.
answer
(
Errors
.
Denied
(
str
(
e
)))
except
AssertionError
:
e
=
sys
.
exc_info
()
try
:
...
...
@@ -160,8 +162,7 @@ class EventHandler(object):
def
_acceptIdentification
(
*
args
):
pass
def
acceptIdentification
(
self
,
conn
,
node_type
,
uuid
,
num_partitions
,
num_replicas
,
your_uuid
):
def
acceptIdentification
(
self
,
conn
,
node_type
,
uuid
,
your_uuid
):
app
=
self
.
app
node
=
app
.
nm
.
getByAddress
(
conn
.
getAddress
())
assert
node
.
getConnection
()
is
conn
,
(
node
.
getConnection
(),
conn
)
...
...
@@ -180,7 +181,7 @@ class EventHandler(object):
elif
node
.
getUUID
()
!=
uuid
or
app
.
uuid
!=
your_uuid
!=
None
:
raise
ProtocolError
(
'invalid uuids'
)
node
.
setIdentified
()
self
.
_acceptIdentification
(
node
,
num_partitions
,
num_replicas
)
self
.
_acceptIdentification
(
node
)
return
conn
.
close
()
...
...
neo/lib/interfaces.py
View file @
8ba42463
...
...
@@ -74,7 +74,7 @@ def implements(obj, ignore=()):
assert
not
wrong_signature
,
wrong_signature
return
obj
def
_s
et_code
(
func
):
def
_s
tub
(
func
):
args
,
varargs
,
varkw
,
_
=
inspect
.
getargspec
(
func
)
if
varargs
:
args
.
append
(
"*"
+
varargs
)
...
...
@@ -82,16 +82,25 @@ def _set_code(func):
args
.
append
(
"**"
+
varkw
)
exec
"def %s(%s): raise NotImplementedError
\
n
f = %s"
%
(
func
.
__name__
,
","
.
join
(
args
),
func
.
__name__
)
func
.
func_code
=
f
.
func_code
return
f
def
abstract
(
func
):
_set_code
(
func
)
func
.
__abstract__
=
1
return
func
f
=
_stub
(
func
)
f
.
__abstract__
=
1
f
.
__defaults__
=
func
.
__defaults__
f
.
__doc__
=
func
.
__doc__
return
f
def
requires
(
*
args
):
for
func
in
args
:
_set_code
(
func
)
# Tolerate useless abstract decoration on required method (e.g. it
# simplifies the implementation of a fallback decorator), but remove
# marker since it does not need to be implemented if it's required
# by a method that is overridden.
try
:
del
func
.
__abstract__
except
AttributeError
:
func
.
__code__
=
_stub
(
func
).
__code__
def
decorator
(
func
):
func
.
__requires__
=
args
return
func
...
...
neo/lib/logger.py
View file @
8ba42463
...
...
@@ -152,7 +152,8 @@ class NEOLogger(Logger):
def
_setup
(
self
,
filename
=
None
,
reset
=
False
):
from
.
import
protocol
as
p
global
uuid_str
global
packb
,
uuid_str
packb
=
p
.
packb
uuid_str
=
p
.
uuid_str
if
self
.
_db
is
not
None
:
self
.
_db
.
close
()
...
...
@@ -250,7 +251,7 @@ class NEOLogger(Logger):
'>'
if
r
.
outgoing
else
'<'
,
uuid_str
(
r
.
uuid
),
ip
,
port
)
msg
=
r
.
msg
if
msg
is
not
None
:
msg
=
buffer
(
msg
)
msg
=
buffer
(
msg
if
type
(
msg
)
is
bytes
else
packb
(
msg
)
)
q
=
"INSERT INTO packet VALUES (?,?,?,?,?,?)"
x
=
[
r
.
created
,
nid
,
r
.
msg_id
,
r
.
code
,
peer
,
msg
]
else
:
...
...
@@ -299,9 +300,14 @@ class NEOLogger(Logger):
def
packet
(
self
,
connection
,
packet
,
outgoing
):
if
self
.
_db
is
not
None
:
body
=
packet
.
_body
if
self
.
_max_packet
and
self
.
_max_packet
<
len
(
body
):
body
=
None
if
self
.
_max_packet
and
self
.
_max_packet
<
packet
.
size
:
args
=
None
else
:
args
=
packet
.
_args
try
:
hash
(
args
)
except
TypeError
:
args
=
packb
(
args
)
self
.
_queue
(
PacketRecord
(
created
=
time
(),
msg_id
=
packet
.
_id
,
...
...
@@ -309,7 +315,7 @@ class NEOLogger(Logger):
outgoing
=
outgoing
,
uuid
=
connection
.
getUUID
(),
addr
=
connection
.
getAddress
(),
msg
=
body
))
msg
=
args
))
def
node
(
self
,
*
cluster_nid
):
name
=
self
.
name
and
str
(
self
.
name
)
...
...
neo/lib/node.py
View file @
8ba42463
...
...
@@ -486,7 +486,7 @@ class NodeManager(EventQueue):
# For the first notification, we receive a full list of nodes from
# the master. Remove all unknown nodes from a previous connection.
for
node
in
self
.
_node_set
.
difference
(
added_list
):
if
app
.
pt
.
dropNode
(
node
):
if
not
node
.
isStorage
()
or
app
.
pt
.
dropNode
(
node
):
self
.
remove
(
node
)
self
.
log
()
self
.
executeQueuedEvents
()
...
...
neo/lib/protocol.py
View file @
8ba42463
...
...
@@ -14,27 +14,63 @@
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import
sys
import
traceback
from
cStringIO
import
StringIO
from
struct
import
Struct
import
threading
from
functools
import
partial
from
msgpack
import
packb
# The protocol version must be increased whenever upgrading a node may require
# to upgrade other nodes. It is encoded as a 4-bytes big-endian integer and
# the high order byte 0 is different from TLS Handshake (0x16).
PROTOCOL_VERSION
=
5
ENCODED_VERSION
=
Struct
(
'!L'
).
pack
(
PROTOCOL_VERSION
)
# to upgrade other nodes.
PROTOCOL_VERSION
=
0
# By encoding the handshake packet with msgpack, the whole NEO stream can be
# decoded with msgpack. The first byte is 0x92, which is different from TLS
# Handshake (0x16).
HANDSHAKE_PACKET
=
packb
((
'NEO'
,
PROTOCOL_VERSION
))
# Used to distinguish non-NEO stream from version mismatch.
MAGIC_SIZE
=
len
(
HANDSHAKE_PACKET
)
-
len
(
packb
(
PROTOCOL_VERSION
))
# Avoid memory errors on corrupted data.
MAX_PACKET_SIZE
=
0x4000000
PACKET_HEADER_FORMAT
=
Struct
(
'!LHL'
)
RESPONSE_MASK
=
0x8000
# Avoid some memory errors on corrupted data.
# Before we use msgpack, we limited the size of a whole packet. That's not
# possible anymore because the size is not known in advance. Packets bigger
# than the buffer size are possible (e.g. a huge list of small items) and for
# that we could compare the stream position (Unpacker.tell); it's not worth it.
UNPACK_BUFFER_SIZE
=
0x4000000
@
apply
def
Unpacker
():
global
registerExtType
,
packb
from
msgpack
import
ExtType
,
unpackb
,
Packer
,
Unpacker
ext_type_dict
=
[]
kw
=
dict
(
use_bin_type
=
True
)
pack_ext
=
Packer
(
**
kw
).
pack
def
registerExtType
(
getstate
,
make
):
code
=
len
(
ext_type_dict
)
ext_type_dict
.
append
(
lambda
data
:
make
(
unpackb
(
data
,
use_list
=
False
)))
return
lambda
obj
:
ExtType
(
code
,
pack_ext
(
getstate
(
obj
)))
iterable_types
=
set
,
tuple
def
default
(
obj
):
try
:
pack
=
obj
.
_pack
except
AttributeError
:
assert
type
(
obj
)
in
iterable_types
,
type
(
obj
)
return
list
(
obj
)
return
pack
()
lock
=
threading
.
Lock
()
pack
=
Packer
(
default
,
strict_types
=
True
,
**
kw
).
pack
def
packb
(
obj
):
with
lock
:
# in case that 'default' is called
return
pack
(
obj
)
return
partial
(
Unpacker
,
use_list
=
False
,
max_buffer_size
=
UNPACK_BUFFER_SIZE
,
ext_hook
=
lambda
code
,
data
:
ext_type_dict
[
code
](
data
))
class
Enum
(
tuple
):
class
Item
(
int
):
__slots__
=
'_name'
,
'_enum'
__slots__
=
'_name'
,
'_enum'
,
'_pack'
def
__str__
(
self
):
return
self
.
_name
def
__repr__
(
self
):
...
...
@@ -49,30 +85,38 @@ class Enum(tuple):
names
=
func
.
func_code
.
co_names
self
=
tuple
.
__new__
(
cls
,
map
(
cls
.
Item
,
xrange
(
len
(
names
))))
self
.
_name
=
func
.
__name__
pack
=
registerExtType
(
int
,
self
.
__getitem__
)
for
item
,
name
in
zip
(
self
,
names
):
setattr
(
self
,
name
,
item
)
item
.
_name
=
name
item
.
_enum
=
self
item
.
_pack
=
(
lambda
x
:
lambda
:
x
)(
pack
(
item
))
return
self
def
__repr__
(
self
):
return
"<Enum %s>"
%
self
.
_name
# The order of extension type is important.
# Enum types first, sorted alphabetically.
@
Enum
def
ErrorCodes
():
ACK
NOT_READY
OID_NOT_FOUND
TID_NOT_FOUND
OID_DOES_NOT_EXIST
PROTOCOL_ERROR
REPLICATION_ERROR
CHECKING_ERROR
BACKEND_NOT_IMPLEMENTED
NON_READABLE_CELL
READ_ONLY_ACCESS
INCOMPLETE_TRANSACTION
def
CellStates
():
# Write-only cell. Last transactions are missing because storage is/was down
# for a while, or because it is new for the partition. It usually becomes
# UP_TO_DATE when replication is done.
OUT_OF_DATE
# Normal state: cell is writable/readable, and it isn't planned to drop it.
UP_TO_DATE
# Same as UP_TO_DATE, except that it will be discarded as soon as another
# node finishes to replicate it. It means a partition is moved from 1 node
# to another. It is also discarded immediately if out-of-date.
FEEDING
# A check revealed that data differs from other replicas. Cell is neither
# readable nor writable.
CORRUPTED
# Not really a state: only used in network packets to tell storages to drop
# partitions.
DISCARDED
@
Enum
def
ClusterStates
():
...
...
@@ -107,11 +151,20 @@ def ClusterStates():
STOPPING_BACKUP
@
Enum
def
NodeTypes
():
MASTER
STORAGE
CLIENT
ADMIN
def
ErrorCodes
():
ACK
DENIED
NOT_READY
OID_NOT_FOUND
TID_NOT_FOUND
OID_DOES_NOT_EXIST
PROTOCOL_ERROR
REPLICATION_ERROR
CHECKING_ERROR
BACKEND_NOT_IMPLEMENTED
NON_READABLE_CELL
READ_ONLY_ACCESS
INCOMPLETE_TRANSACTION
@
Enum
def
NodeStates
():
...
...
@@ -121,23 +174,11 @@ def NodeStates():
PENDING
@
Enum
def
CellStates
():
# Write-only cell. Last transactions are missing because storage is/was down
# for a while, or because it is new for the partition. It usually becomes
# UP_TO_DATE when replication is done.
OUT_OF_DATE
# Normal state: cell is writable/readable, and it isn't planned to drop it.
UP_TO_DATE
# Same as UP_TO_DATE, except that it will be discarded as soon as another
# node finishes to replicate it. It means a partition is moved from 1 node
# to another. It is also discarded immediately if out-of-date.
FEEDING
# A check revealed that data differs from other replicas. Cell is neither
# readable nor writable.
CORRUPTED
# Not really a state: only used in network packets to tell storages to drop
# partitions.
DISCARDED
def
NodeTypes
():
MASTER
STORAGE
CLIENT
ADMIN
# used for logging
node_state_prefix_dict
=
{
...
...
@@ -212,45 +253,22 @@ class NonReadableCell(Exception):
On such event, the client must retry, preferably another cell.
"""
class
Packet
(
object
):
"""
Base class for any packet definition. The _fmt class attribute must be
defined for any non-empty packet.
Base class for any packet definition.
"""
_ignore_when_closed
=
False
_request
=
None
_answer
=
None
_body
=
None
_code
=
None
_fmt
=
None
_id
=
None
nodelay
=
True
poll_thread
=
False
def
__init__
(
self
,
*
args
):
assert
self
.
_code
is
not
None
,
"Packet class not registered"
if
args
:
buf
=
StringIO
()
self
.
_fmt
.
encode
(
buf
.
write
,
args
)
self
.
_body
=
buf
.
getvalue
()
else
:
self
.
_body
=
''
def
decode
(
self
):
assert
self
.
_body
is
not
None
if
self
.
_fmt
is
None
:
return
()
buf
=
StringIO
(
self
.
_body
)
try
:
return
self
.
_fmt
.
decode
(
buf
.
read
)
except
ParseError
,
msg
:
name
=
self
.
__class__
.
__name__
raise
PacketMalformedError
(
"%s fail (%s)"
%
(
name
,
msg
))
def
setContent
(
self
,
msg_id
,
body
):
""" Register the packet content for future decoding """
self
.
_id
=
msg_id
self
.
_body
=
body
self
.
_args
=
args
def
setId
(
self
,
value
):
self
.
_id
=
value
...
...
@@ -259,14 +277,11 @@ class Packet(object):
assert
self
.
_id
is
not
None
,
"No identifier applied on the packet"
return
self
.
_id
def
encode
(
self
):
def
encode
(
self
,
packb
=
packb
):
""" Encode a packet as a string to send it over the network """
content
=
self
.
_body
return
(
PACKET_HEADER_FORMAT
.
pack
(
self
.
_id
,
self
.
_code
,
len
(
content
)),
content
)
def
__len__
(
self
):
return
PACKET_HEADER_FORMAT
.
size
+
len
(
self
.
_body
)
r
=
packb
((
self
.
_id
,
self
.
_code
,
self
.
_args
))
self
.
size
=
len
(
r
)
return
r
def
__repr__
(
self
):
return
'%s[%r]'
%
(
self
.
__class__
.
__name__
,
self
.
_id
)
...
...
@@ -279,10 +294,10 @@ class Packet(object):
return
self
.
_code
==
other
.
_code
def
isError
(
self
):
return
isinstance
(
self
,
Error
)
return
self
.
_code
==
RESPONSE_MASK
def
isResponse
(
self
):
return
self
.
_code
&
RESPONSE_MASK
==
RESPONSE_MASK
return
self
.
_code
&
RESPONSE_MASK
def
getAnswerClass
(
self
):
return
self
.
_answer
...
...
@@ -294,1530 +309,530 @@ class Packet(object):
"""
return
self
.
_ignore_when_closed
class
ParseError
(
Exception
):
"""
An exception that encapsulate another and build the 'path' of the
packet item that generate the error.
"""
def
__init__
(
self
,
item
,
trace
):
Exception
.
__init__
(
self
)
self
.
_trace
=
trace
self
.
_items
=
[
item
]
def
append
(
self
,
item
):
self
.
_items
.
append
(
item
)
def
__repr__
(
self
):
chain
=
'/'
.
join
([
item
.
getName
()
for
item
in
reversed
(
self
.
_items
)])
return
'at %s:
\
n
%s'
%
(
chain
,
self
.
_trace
)
__str__
=
__repr__
# packet parsers
class
PItem
(
object
):
"""
Base class for any packet item, _encode and _decode must be overridden
by subclasses.
"""
def
__init__
(
self
,
name
):
self
.
_name
=
name
def
__repr__
(
self
):
return
self
.
__class__
.
__name__
def
getName
(
self
):
return
self
.
_name
def
_trace
(
self
,
method
,
*
args
):
try
:
return
method
(
*
args
)
except
ParseError
,
e
:
# trace and forward exception
e
.
append
(
self
)
raise
except
Exception
:
# original exception, encapsulate it
trace
=
''
.
join
(
traceback
.
format_exception
(
*
sys
.
exc_info
())[
2
:])
raise
ParseError
(
self
,
trace
)
def
encode
(
self
,
writer
,
items
):
return
self
.
_trace
(
self
.
_encode
,
writer
,
items
)
def
decode
(
self
,
reader
):
return
self
.
_trace
(
self
.
_decode
,
reader
)
def
_encode
(
self
,
writer
,
items
):
raise
NotImplementedError
,
self
.
__class__
.
__name__
def
_decode
(
self
,
reader
):
raise
NotImplementedError
,
self
.
__class__
.
__name__
class
PacketRegistryFactory
(
dict
):
class
PStruct
(
PItem
):
"""
Aggregate other items
"""
def
__init__
(
self
,
name
,
*
items
):
PItem
.
__init__
(
self
,
name
)
self
.
_items
=
items
def
_encode
(
self
,
writer
,
items
):
assert
len
(
self
.
_items
)
==
len
(
items
),
(
items
,
self
.
_items
)
for
item
,
value
in
zip
(
self
.
_items
,
items
):
item
.
encode
(
writer
,
value
)
def
_decode
(
self
,
reader
):
return
tuple
([
item
.
decode
(
reader
)
for
item
in
self
.
_items
])
class
PStructItem
(
PItem
):
"""
A single value encoded with struct
"""
def
__init__
(
self
,
name
):
PItem
.
__init__
(
self
,
name
)
struct
=
Struct
(
self
.
_fmt
)
self
.
pack
=
struct
.
pack
self
.
unpack
=
struct
.
unpack
self
.
size
=
struct
.
size
def
_encode
(
self
,
writer
,
value
):
writer
(
self
.
pack
(
value
))
def
_decode
(
self
,
reader
):
return
self
.
unpack
(
reader
(
self
.
size
))[
0
]
class
PStructItemOrNone
(
PStructItem
):
def
_encode
(
self
,
writer
,
value
):
return
writer
(
self
.
_None
if
value
is
None
else
self
.
pack
(
value
))
def
_decode
(
self
,
reader
):
value
=
reader
(
self
.
size
)
return
None
if
value
==
self
.
_None
else
self
.
unpack
(
value
)[
0
]
class
POption
(
PStruct
):
def
_encode
(
self
,
writer
,
value
):
if
value
is
None
:
writer
(
'
\
0
'
)
def
__call__
(
self
,
name
,
base
,
d
):
for
k
,
v
in
d
.
items
():
if
isinstance
(
v
,
type
)
and
issubclass
(
v
,
Packet
):
v
.
__name__
=
k
v
.
handler_method_name
=
k
[
0
].
lower
()
+
k
[
1
:]
# this builds a "singleton"
return
type
(
'PacketRegistry'
,
base
,
d
)(
self
)
def
register
(
self
,
doc
,
ignore_when_closed
=
None
,
request
=
False
,
error
=
False
,
_base
=
(
Packet
,),
**
kw
):
""" Register a packet in the packet registry """
code
=
len
(
self
)
if
doc
is
None
:
self
[
code
]
=
None
return
# None registered only to skip a code number (for compatibility)
if
error
and
not
request
:
assert
not
code
code
=
RESPONSE_MASK
kw
.
update
(
__doc__
=
doc
,
_code
=
code
)
packet
=
type
(
''
,
_base
,
kw
)
# register the request
self
[
code
]
=
packet
if
request
:
if
ignore_when_closed
is
None
:
# By default, on a closed connection:
# - request: ignore
# - answer: keep
# - notification: keep
packet
.
_ignore_when_closed
=
True
else
:
assert
ignore_when_closed
is
False
if
error
:
packet
.
_answer
=
self
[
RESPONSE_MASK
]
else
:
# build a class for the answer
code
|=
RESPONSE_MASK
kw
[
'_code'
]
=
code
answer
=
packet
.
_answer
=
self
[
code
]
=
type
(
''
,
_base
,
kw
)
return
packet
,
answer
else
:
writer
(
'
\
1
'
)
PStruct
.
_encode
(
self
,
writer
,
value
)
assert
ignore_when_closed
is
None
return
packet
def
_decode
(
self
,
reader
):
if
'
\
0
\
1
'
.
index
(
reader
(
1
)):
return
PStruct
.
_decode
(
self
,
reader
)
class
P
List
(
PStructItem
):
class
P
ackets
(
dict
):
"""
A list of homogeneous items
Packet registry that checks packet code uniqueness and provides an index
"""
_fmt
=
'!L'
__metaclass__
=
PacketRegistryFactory
()
notify
=
__metaclass__
.
register
request
=
partial
(
notify
,
request
=
True
)
def
__init__
(
self
,
name
,
item
):
PStructItem
.
__init__
(
self
,
name
)
self
.
_item
=
item
Error
=
notify
(
"""
Error is a special type of message, because this can be sent against
any other message, even if such a message does not expect a reply
usually.
def
_encode
(
self
,
writer
,
items
):
writer
(
self
.
pack
(
len
(
items
)))
item
=
self
.
_item
for
value
in
items
:
item
.
encode
(
writer
,
value
)
:nodes: * -> *
"""
,
error
=
True
)
def
_decode
(
self
,
reader
):
length
=
self
.
unpack
(
reader
(
self
.
size
))[
0
]
item
=
self
.
_item
return
[
item
.
decode
(
reader
)
for
_
in
xrange
(
length
)]
RequestIdentification
,
AcceptIdentification
=
request
(
"""
Request a node identification. This must be the first packet for any
connection.
class
PDict
(
PStructItem
):
"""
A dictionary with custom key and value formats
"""
_fmt
=
'!L'
def
__init__
(
self
,
name
,
key
,
value
):
PStructItem
.
__init__
(
self
,
name
)
self
.
_key
=
key
self
.
_value
=
value
def
_encode
(
self
,
writer
,
item
):
assert
isinstance
(
item
,
dict
),
(
type
(
item
),
item
)
writer
(
self
.
pack
(
len
(
item
)))
key
,
value
=
self
.
_key
,
self
.
_value
for
k
,
v
in
item
.
iteritems
():
key
.
encode
(
writer
,
k
)
value
.
encode
(
writer
,
v
)
def
_decode
(
self
,
reader
):
length
=
self
.
unpack
(
reader
(
self
.
size
))[
0
]
key
,
value
=
self
.
_key
,
self
.
_value
new_dict
=
{}
for
_
in
xrange
(
length
):
k
=
key
.
decode
(
reader
)
v
=
value
.
decode
(
reader
)
new_dict
[
k
]
=
v
return
new_dict
class
PEnum
(
PStructItem
):
"""
Encapsulate an enumeration value
"""
_fmt
=
'b'
:nodes: * -> *
"""
,
poll_thread
=
True
)
def
__init__
(
self
,
name
,
enum
):
PStructItem
.
__init__
(
self
,
name
)
self
.
_enum
=
enum
Ping
,
Pong
=
request
(
"""
Empty request used as network barrier.
def
_encode
(
self
,
writer
,
item
):
if
item
is
None
:
item
=
-
1
writer
(
self
.
pack
(
item
))
:nodes: * -> *
"""
)
def
_decode
(
self
,
reader
):
code
=
self
.
unpack
(
reader
(
self
.
size
))[
0
]
if
code
==
-
1
:
return
None
try
:
return
self
.
_enum
[
code
]
except
KeyError
:
enum
=
self
.
_enum
.
__class__
.
__name__
raise
ValueError
,
'Invalid code for %s enum: %r'
%
(
enum
,
code
)
CloseClient
=
notify
(
"""
Tell peer that it can close the connection if it has finished with us.
class
PString
(
PStructItem
):
"""
A variable-length string
"""
_fmt
=
'!L'
:nodes: * -> *
"""
)
def
_encode
(
self
,
writer
,
value
):
writer
(
self
.
pack
(
len
(
value
)))
writer
(
value
)
AskPrimary
,
AnswerPrimary
=
request
(
"""
Ask node identier of the current primary master.
def
_decode
(
self
,
reader
):
length
=
self
.
unpack
(
reader
(
self
.
size
))[
0
]
return
reader
(
length
)
:nodes: ctl -> A
"""
)
class
PAddress
(
PString
):
"""
An host address (IPv4/IPv6)
"""
NotPrimaryMaster
=
notify
(
"""
Notify peer that I'm not the primary master. Attach any extra
information to help the peer joining the cluster.
def
__init__
(
self
,
name
):
PString
.
__init__
(
self
,
name
)
self
.
_port
=
Struct
(
'!H'
)
:nodes: SM -> *
"""
)
def
_encode
(
self
,
writer
,
address
):
if
address
:
host
,
port
=
address
PString
.
_encode
(
self
,
writer
,
host
)
writer
(
self
.
_port
.
pack
(
port
))
else
:
PString
.
_encode
(
self
,
writer
,
''
)
NotifyNodeInformation
=
notify
(
"""
Notify information about one or more nodes.
def
_decode
(
self
,
reader
):
host
=
PString
.
_decode
(
self
,
reader
)
if
host
:
p
=
self
.
_port
return
host
,
p
.
unpack
(
reader
(
p
.
size
))[
0
]
:nodes: M -> *
"""
)
class
PBoolean
(
PStructItem
):
"""
A boolean value, encoded as a single byte
"""
_fmt
=
'!?'
AskRecovery
,
AnswerRecovery
=
request
(
"""
Ask storage nodes data needed by master to recover.
Reused by `neoctl print ids`.
class
PNumber
(
PStructItem
):
"""
A integer number (4-bytes length)
"""
_fmt
=
'!L'
:nodes: M -> S; ctl -> A -> M
"""
)
class
PIndex
(
PStructItem
):
"""
A big integer to defined indexes in a huge list.
"""
_fmt
=
'!Q'
AskLastIDs
,
AnswerLastIDs
=
request
(
"""
Ask the last OID/TID so that a master can initialize its
TransactionManager. Reused by `neoctl print ids`.
class
PPTID
(
PStructItemOrNone
):
"""
A None value means an invalid PTID
"""
_fmt
=
'!Q'
_None
=
Struct
(
_fmt
).
pack
(
0
)
:nodes: M -> S; ctl -> A -> M
"""
)
class
PChecksum
(
PItem
):
"""
A hash (SHA1)
"""
def
_encode
(
self
,
writer
,
checksum
):
assert
len
(
checksum
)
==
20
,
(
len
(
checksum
),
checksum
)
writer
(
checksum
)
AskPartitionTable
,
AnswerPartitionTable
=
request
(
"""
Ask storage node the remaining data needed by master to recover.
def
_decode
(
self
,
reader
):
return
reader
(
20
)
:nodes: M -> S
"""
)
class
PSignedNull
(
PStructItemOrNone
):
_fmt
=
'!l'
_None
=
Struct
(
_fmt
).
pack
(
0
)
SendPartitionTable
=
notify
(
"""
Send the full partition table to admin/client/storage nodes on
connection.
class
PUUID
(
PSignedNull
):
"""
An UUID (node identifier, 4-bytes signed integer)
"""
:nodes: M -> A, C, S
"""
)
class
PTID
(
PItem
):
"""
A transaction identifier
"""
def
_encode
(
self
,
writer
,
tid
):
if
tid
is
None
:
tid
=
INVALID_TID
assert
len
(
tid
)
==
8
,
(
len
(
tid
),
tid
)
writer
(
tid
)
def
_decode
(
self
,
reader
):
tid
=
reader
(
8
)
if
tid
==
INVALID_TID
:
tid
=
None
return
tid
# same definition, for now
POID
=
PTID
class
PFloat
(
PStructItemOrNone
):
"""
A float number (8-bytes length)
"""
_fmt
=
'!d'
_None
=
'
\
xff
'
*
8
# common definitions
PFEmpty
=
PStruct
(
'no_content'
)
PFNodeType
=
PEnum
(
'type'
,
NodeTypes
)
PFNodeState
=
PEnum
(
'state'
,
NodeStates
)
PFCellState
=
PEnum
(
'state'
,
CellStates
)
PFNodeList
=
PList
(
'node_list'
,
PStruct
(
'node'
,
PFNodeType
,
PAddress
(
'address'
),
PUUID
(
'uuid'
),
PFNodeState
,
PFloat
(
'id_timestamp'
),
),
)
PFCellList
=
PList
(
'cell_list'
,
PStruct
(
'cell'
,
PUUID
(
'uuid'
),
PFCellState
,
),
)
PFRowList
=
PList
(
'row_list'
,
PStruct
(
'row'
,
PNumber
(
'offset'
),
PFCellList
,
),
)
PFHistoryList
=
PList
(
'history_list'
,
PStruct
(
'history_entry'
,
PTID
(
'serial'
),
PNumber
(
'size'
),
),
)
PFUUIDList
=
PList
(
'uuid_list'
,
PUUID
(
'uuid'
),
)
PFTidList
=
PList
(
'tid_list'
,
PTID
(
'tid'
),
)
PFOidList
=
PList
(
'oid_list'
,
POID
(
'oid'
),
)
# packets definition
class
Error
(
Packet
):
"""
Error is a special type of message, because this can be sent against
any other message, even if such a message does not expect a reply
usually.
NotifyPartitionChanges
=
notify
(
"""
Notify about changes in the partition table.
:nodes: * -> *
"""
_fmt
=
PStruct
(
'error'
,
PNumber
(
'code'
),
PString
(
'message'
),
)
:nodes: M -> *
"""
)
class
Ping
(
Packet
):
"""
Empty request used as network barri
er.
StartOperation
=
notify
(
"""
Tell a storage node to start operation. Before this message,
it must only communicate with the primary mast
er.
:nodes: * -> *
"""
_answer
=
PFEmpty
:nodes: M -> S
"""
)
class
CloseClient
(
Packet
):
"""
Tell peer that it can close the connection if it has finished with us
.
StopOperation
=
notify
(
"""
Notify that the cluster is not operational anymore.
Any operation between nodes must be aborted
.
:nodes: * -> *
"""
:nodes: M -> S, C
"""
)
class
RequestIdentification
(
Packet
):
"""
Request a node identification. This must be the first packet for any
connection.
AskUnfinishedTransactions
,
AnswerUnfinishedTransactions
=
request
(
"""
Ask unfinished transactions, which will be replicated
when they're finished.
:nodes: * -> *
"""
poll_thread
=
True
_fmt
=
PStruct
(
'request_identification'
,
PFNodeType
,
PUUID
(
'uuid'
),
PAddress
(
'address'
),
PString
(
'name'
),
PList
(
'devpath'
,
PString
(
'devid'
)),
PFloat
(
'id_timestamp'
),
)
_answer
=
PStruct
(
'accept_identification'
,
PFNodeType
,
PUUID
(
'my_uuid'
),
PNumber
(
'num_partitions'
),
PNumber
(
'num_replicas'
),
PUUID
(
'your_uuid'
),
)
class
PrimaryMaster
(
Packet
):
"""
Ask node identier of the current primary master.
:nodes: S -> M
"""
)
:nodes: ctl -> A
"""
_answer
=
PStruct
(
'answer_primary'
,
PUUID
(
'primary_uuid'
),
)
AskLockedTransactions
,
AnswerLockedTransactions
=
request
(
"""
Ask locked transactions to replay committed transactions
that haven't been unlocked.
class
NotPrimaryMaster
(
Packet
):
"""
Notify peer that I'm not the primary master. Attach any extra information
to help the peer joining the cluster.
:nodes: M -> S
"""
)
:nodes: SM -> *
"""
_fmt
=
PStruct
(
'not_primary_master'
,
PSignedNull
(
'primary'
),
PList
(
'known_master_list'
,
PAddress
(
'address'
),
),
)
class
Recovery
(
Packet
):
"""
Ask storage nodes data needed by master to recover.
Reused by `neoctl print ids`.
AskFinalTID
,
AnswerFinalTID
=
request
(
"""
Return final tid if ttid has been committed, to recover from certain
failures during tpc_finish.
:nodes: M -> S; ctl -> A -> M
"""
_answer
=
PStruct
(
'answer_recovery'
,
PPTID
(
'ptid'
),
PTID
(
'backup_tid'
),
PTID
(
'truncate_tid'
),
)
:nodes: M -> S; C -> M, S
"""
)
class
LastIDs
(
Packet
):
"""
Ask the last OID/TID so that a master can initialize its TransactionManager.
Reused by `neoctl print ids`.
ValidateTransaction
=
notify
(
"""
Do replay a committed transaction that was not unlocked.
:nodes: M -> S; ctl -> A -> M
"""
_answer
=
PStruct
(
'answer_last_ids'
,
POID
(
'last_oid'
),
PTID
(
'last_tid'
),
)
:nodes: M -> S
"""
)
class
PartitionTable
(
Packet
):
"""
Ask storage node the remaining data needed by master to recover.
This is also how the clients get the full partition table on connection.
AskBeginTransaction
,
AnswerBeginTransaction
=
request
(
"""
Ask to begin a new transaction. This maps to `tpc_begin`.
:nodes: M -> S; C -> M
"""
_answer
=
PStruct
(
'answer_partition_table'
,
PPTID
(
'ptid'
),
PFRowList
,
)
:nodes: C -> M
"""
)
class
NotifyPartitionTable
(
Packet
):
"""
Send the full partition table to admin/storage nodes on conne
ction.
FailedVote
=
request
(
"""
Report storage nodes for which vote failed.
True is returned if it's still possible to finish the transa
ction.
:nodes: M -> A, S
"""
_fmt
=
PStruct
(
'send_partition_table'
,
PPTID
(
'ptid'
),
PFRowList
,
)
:nodes: C -> M
"""
,
error
=
True
)
class
PartitionChanges
(
Packet
):
"""
Notify about changes in the partition table
.
AskFinishTransaction
,
AnswerTransactionFinished
=
request
(
"""
Finish a transaction. Return the TID of the committed transaction.
This maps to `tpc_finish`
.
:nodes: M -> *
"""
_fmt
=
PStruct
(
'notify_partition_changes'
,
PPTID
(
'ptid'
),
PList
(
'cell_list'
,
PStruct
(
'cell'
,
PNumber
(
'offset'
),
PUUID
(
'uuid'
),
PFCellState
,
),
),
)
class
StartOperation
(
Packet
):
"""
Tell a storage node to start operation. Before this message, it must only
communicate with the primary master.
:nodes: C -> M
"""
,
ignore_when_closed
=
False
,
poll_thread
=
True
)
:nodes: M -> S
"""
_fmt
=
PStruct
(
'start_operation'
,
# XXX: Is this boolean needed ? Maybe this
# can be deduced from cluster state.
PBoolean
(
'backup'
),
)
AskLockInformation
,
AnswerInformationLocked
=
request
(
"""
Commit a transaction. The new data is read-locked.
class
StopOperation
(
Packet
):
"""
Notify that the cluster is not operational anymore. Any operation between
nodes must be aborted.
:nodes: M -> S
"""
,
ignore_when_closed
=
False
)
:nodes: M -> S, C
"""
InvalidateObjects
=
notify
(
"""
Notify about a new transaction modifying objects,
invalidating client caches.
class
UnfinishedTransactions
(
Packet
):
"""
Ask unfinished transactions, which will be replicated when they're finished.
:nodes: M -> C
"""
)
:nodes: S -> M
"""
_fmt
=
PStruct
(
'ask_unfinished_transactions'
,
PList
(
'row_list'
,
PNumber
(
'offset'
),
),
)
_answer
=
PStruct
(
'answer_unfinished_transactions'
,
PTID
(
'max_tid'
),
PList
(
'tid_list'
,
PTID
(
'unfinished_tid'
),
),
)
class
LockedTransactions
(
Packet
):
"""
Ask locked transactions to replay committed transactions that haven't been
unlocked.
NotifyUnlockInformation
=
notify
(
"""
Notify about a successfully committed transaction. The new data can be
unlocked.
:nodes: M -> S
"""
_answer
=
PStruct
(
'answer_locked_transactions'
,
PDict
(
'tid_dict'
,
PTID
(
'ttid'
),
PTID
(
'tid'
),
),
)
class
FinalTID
(
Packet
):
"""
Return final tid if ttid has been committed, to recover from certain
failures during tpc_finish.
:nodes: M -> S
"""
)
:nodes: M -> S; C -> M, S
"""
_fmt
=
PStruct
(
'final_tid'
,
PTID
(
'ttid'
),
)
AskNewOIDs
,
AnswerNewOIDs
=
request
(
"""
Ask new OIDs to create objects.
_answer
=
PStruct
(
'final_tid'
,
PTID
(
'tid'
),
)
:nodes: C -> M
"""
)
class
ValidateTransaction
(
Packet
):
"""
Do replay a committed transaction that was not unlocked.
NotifyDeadlock
=
notify
(
"""
Ask master to generate a new TTID that will be used by the client to
solve a deadlock by rebasing the transaction on top of concurrent
changes.
:nodes: M -> S
"""
_fmt
=
PStruct
(
'validate_transaction'
,
PTID
(
'ttid'
),
PTID
(
'tid'
),
)
:nodes: S -> M -> C
"""
)
class
BeginTransaction
(
Packet
):
"""
Ask to begin a new transaction. This maps to `tpc_begin`.
AskRebaseTransaction
,
AnswerRebaseTransaction
=
request
(
"""
Rebase a transaction to solve a deadlock.
:nodes: C -> M
"""
_fmt
=
PStruct
(
'ask_begin_transaction'
,
PTID
(
'tid'
),
)
:nodes: C -> S
"""
)
_answer
=
PStruct
(
'answer_begin_transaction'
,
PTID
(
'tid'
),
)
AskRebaseObject
,
AnswerRebaseObject
=
request
(
"""
Rebase an object change to solve a deadlock.
class
FailedVote
(
Packet
):
"""
Report storage nodes for which vote failed.
True is returned if it's still possible to finish the transaction.
:nodes: C -> S
:nodes: C -> M
"""
_fmt
=
PStruct
(
'failed_vote'
,
PTID
(
'tid'
),
PFUUIDList
,
)
XXX: It is a request packet to simplify the implementation. For more
efficiency, this should be turned into a notification, and the
RebaseTransaction should answered once all objects are rebased
(so that the client can still wait on something).
"""
,
data_path
=
(
1
,
0
,
2
,
0
))
_answer
=
Error
AskStoreObject
,
AnswerStoreObject
=
request
(
"""
Ask to create/modify an object. This maps to `store`.
class
FinishTransaction
(
Packet
):
"""
Finish a transaction. Return the TID of the committed transaction.
This maps to `tpc_finish`.
As for IStorage, 'serial' is ZERO_TID for new objects.
:nodes: C -> M
"""
poll_thread
=
True
_fmt
=
PStruct
(
'ask_finish_transaction'
,
PTID
(
'tid'
),
PFOidList
,
PList
(
'checked_list'
,
POID
(
'oid'
),
),
)
_answer
=
PStruct
(
'answer_information_locked'
,
PTID
(
'ttid'
),
PTID
(
'tid'
),
)
class
NotifyTransactionFinished
(
Packet
):
"""
Notify that a transaction blocking a replication is now finished.
:nodes: C -> S
"""
,
data_path
=
(
0
,
2
))
:nodes: M -> S
"""
_fmt
=
PStruct
(
'notify_transaction_finished'
,
PTID
(
'ttid'
),
PTID
(
'max_tid'
),
)
AbortTransaction
=
notify
(
"""
Abort a transaction. This maps to `tpc_abort`.
class
LockInformation
(
Packet
):
"""
Commit a transaction. The new data is read-locked.
:nodes: C -> S; C -> M -> S
"""
)
:nodes: M -> S
"""
_fmt
=
PStruct
(
'ask_lock_informations'
,
PTID
(
'ttid'
),
PTID
(
'tid'
),
)
AskStoreTransaction
,
AnswerStoreTransaction
=
request
(
"""
Ask to store a transaction. Implies vote.
_answer
=
PStruct
(
'answer_information_locked'
,
PTID
(
'ttid'
),
)
:nodes: C -> S
"""
)
class
InvalidateObjects
(
Packet
):
"""
Notify about a new transaction modifying objects,
invalidating client caches.
AskVoteTransaction
,
AnswerVoteTransaction
=
request
(
"""
Ask to vote a transaction.
:nodes: M -> C
"""
_fmt
=
PStruct
(
'ask_finish_transaction'
,
PTID
(
'tid'
),
PFOidList
,
)
:nodes: C -> S
"""
)
class
UnlockInformation
(
Packet
):
"""
Notify about a successfully committed transaction. The new data can be
unlocked.
AskObject
,
AnswerObject
=
request
(
"""
Ask a stored object by its OID, optionally at/before a specific tid.
This maps to `load/loadBefore/loadSerial`.
:nodes: M -> S
"""
_fmt
=
PStruct
(
'notify_unlock_information'
,
PTID
(
'ttid'
),
)
:nodes: C -> S
"""
,
data_path
=
(
1
,
3
))
class
GenerateOIDs
(
Packet
):
"""
Ask new OIDs to create objects
.
AskTIDs
,
AnswerTIDs
=
request
(
"""
Ask for TIDs between a range of offsets. The order of TIDs is
descending, and the range is [first, last). This maps to `undoLog`
.
:nodes: C -> M
"""
_fmt
=
PStruct
(
'ask_new_oids'
,
PNumber
(
'num_oids'
),
)
:nodes: C -> S
"""
)
_answer
=
PStruct
(
'answer_new_oids'
,
PFOidList
,
)
AskTransactionInformation
,
AnswerTransactionInformation
=
request
(
"""
Ask for transaction metadata.
class
Deadlock
(
Packet
):
"""
Ask master to generate a new TTID that will be used by the client to solve
a deadlock by rebasing the transaction on top of concurrent changes.
:nodes: C -> S
"""
)
:nodes: S -> M -> C
"""
_fmt
=
PStruct
(
'notify_deadlock'
,
PTID
(
'ttid'
),
PTID
(
'locking_tid'
),
)
class
RebaseTransaction
(
Packet
):
"""
Rebase a transaction to solve a deadlock.
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_rebase_transaction'
,
PTID
(
'ttid'
),
PTID
(
'locking_tid'
),
)
AskObjectHistory
,
AnswerObjectHistory
=
request
(
"""
Ask history information for a given object. The order of serials is
descending, and the range is [first, last]. This maps to `history`.
_answer
=
PStruct
(
'answer_rebase_transaction'
,
PFOidList
,
)
:nodes: C -> S
"""
)
class
RebaseObject
(
Packet
):
"""
Rebase an object change to solve a deadlock.
AskPartitionList
,
AnswerPartitionList
=
request
(
"""
Ask information about partitions.
:nodes: C -> S
:nodes: ctl -> A
"""
)
XXX: It is a request packet to simplify the implementation. For more
efficiency, this should be turned into a notification, and the
RebaseTransaction should answered once all objects are rebased
(so that the client can still wait on something).
"""
_fmt
=
PStruct
(
'ask_rebase_object'
,
PTID
(
'ttid'
),
PTID
(
'oid'
),
)
_answer
=
PStruct
(
'answer_rebase_object'
,
POption
(
'conflict'
,
PTID
(
'serial'
),
PTID
(
'conflict_serial'
),
POption
(
'data'
,
PBoolean
(
'compression'
),
PChecksum
(
'checksum'
),
PString
(
'data'
),
),
)
)
class
StoreObject
(
Packet
):
"""
Ask to create/modify an object. This maps to `store`.
AskNodeList
,
AnswerNodeList
=
request
(
"""
Ask information about nodes.
As for IStorage, 'serial' is ZERO_TID for new objects.
:nodes: ctl -> A
"""
)
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_store_object'
,
POID
(
'oid'
),
PTID
(
'serial'
),
PBoolean
(
'compression'
),
PChecksum
(
'checksum'
),
PString
(
'data'
),
PTID
(
'data_serial'
),
PTID
(
'tid'
),
)
_answer
=
PStruct
(
'answer_store_object'
,
PTID
(
'conflict'
),
)
class
AbortTransaction
(
Packet
):
"""
Abort a transaction. This maps to `tpc_abort`.
:nodes: C -> S; C -> M -> S
"""
_fmt
=
PStruct
(
'abort_transaction'
,
PTID
(
'tid'
),
PFUUIDList
,
# unused for * -> S
)
class
StoreTransaction
(
Packet
):
"""
Ask to store a transaction. Implies vote.
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_store_transaction'
,
PTID
(
'tid'
),
PString
(
'user'
),
PString
(
'description'
),
PString
(
'extension'
),
PFOidList
,
)
_answer
=
PFEmpty
class
VoteTransaction
(
Packet
):
"""
Ask to vote a transaction.
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_vote_transaction'
,
PTID
(
'tid'
),
)
_answer
=
PFEmpty
SetNodeState
=
request
(
"""
Change the state of a node.
class
GetObject
(
Packet
):
"""
Ask a stored object by its OID, optionally at/before a specific tid.
This maps to `load/loadBefore/loadSerial`.
:nodes: ctl -> A -> M
"""
,
error
=
True
,
ignore_when_closed
=
False
)
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_object'
,
POID
(
'oid'
),
PTID
(
'at'
),
PTID
(
'before'
),
)
_answer
=
PStruct
(
'answer_object'
,
POID
(
'oid'
),
PTID
(
'serial_start'
),
PTID
(
'serial_end'
),
PBoolean
(
'compression'
),
PChecksum
(
'checksum'
),
PString
(
'data'
),
PTID
(
'data_serial'
),
)
class
TIDList
(
Packet
):
"""
Ask for TIDs between a range of offsets. The order of TIDs is descending,
and the range is [first, last). This maps to `undoLog`.
AddPendingNodes
=
request
(
"""
Mark given pending nodes as running, for future inclusion when tweaking
the partition table.
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_tids'
,
PIndex
(
'first'
),
PIndex
(
'last'
),
PNumber
(
'partition'
),
)
:nodes: ctl -> A -> M
"""
,
error
=
True
,
ignore_when_closed
=
False
)
_answer
=
PStruct
(
'answer_tids'
,
PFTidList
,
)
TweakPartitionTable
,
AnswerTweakPartitionTable
=
request
(
"""
Ask the master to balance the partition table, optionally excluding
specific nodes in anticipation of removing them.
class
TIDListFrom
(
Packet
):
"""
Ask for length TIDs starting at min_tid. The order of TIDs is ascending.
Used by `iterator`.
:nodes: ctl -> A -> M
"""
)
:nodes: C -> S
"""
_fmt
=
PStruct
(
'tid_list_from'
,
PTID
(
'min_tid'
),
PTID
(
'max_tid'
),
PNumber
(
'length'
),
PNumber
(
'partition'
),
)
_answer
=
PStruct
(
'answer_tids'
,
PFTidList
,
)
class
TransactionInformation
(
Packet
):
"""
Ask for transaction metadata.
SetNumReplicas
=
request
(
"""
Set the number of replicas.
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_transaction_information'
,
PTID
(
'tid'
),
)
_answer
=
PStruct
(
'answer_transaction_information'
,
PTID
(
'tid'
),
PString
(
'user'
),
PString
(
'description'
),
PString
(
'extension'
),
PBoolean
(
'packed'
),
PFOidList
,
)
class
ObjectHistory
(
Packet
):
"""
Ask history information for a given object. The order of serials is
descending, and the range is [first, last]. This maps to `history`.
:nodes: ctl -> A -> M
"""
,
error
=
True
,
ignore_when_closed
=
False
)
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_object_history'
,
POID
(
'oid'
),
PIndex
(
'first'
),
PIndex
(
'last'
),
)
_answer
=
PStruct
(
'answer_object_history'
,
POID
(
'oid'
),
PFHistoryList
,
)
class
PartitionList
(
Packet
):
"""
Ask information about partitions.
SetClusterState
=
request
(
"""
Set the cluster state.
:nodes: ctl -> A
"""
_fmt
=
PStruct
(
'ask_partition_list'
,
PNumber
(
'min_offset'
),
PNumber
(
'max_offset'
),
PUUID
(
'uuid'
),
)
_answer
=
PStruct
(
'answer_partition_list'
,
PPTID
(
'ptid'
),
PFRowList
,
)
class
NodeList
(
Packet
):
"""
Ask information about nodes.
:nodes: ctl -> A -> M
"""
,
error
=
True
,
ignore_when_closed
=
False
)
:nodes: ctl -> A
"""
_fmt
=
PStruct
(
'ask_node_list'
,
PFNodeType
,
)
Repair
=
request
(
"""
Ask storage nodes to repair their databases.
_answer
=
PStruct
(
'answer_node_list'
,
PFNodeList
,
)
:nodes: ctl -> A -> M
"""
,
error
=
True
)
class
SetNodeState
(
Packet
):
"""
Change the state of a nod
e.
NotifyRepair
=
notify
(
"""
Repair is translated to this message, asking a specific storage node to
repair its databas
e.
:nodes: ctl -> A -> M
"""
_fmt
=
PStruct
(
'set_node_state'
,
PUUID
(
'uuid'
),
PFNodeState
,
)
:nodes: M -> S
"""
)
_answer
=
Error
NotifyClusterInformation
=
notify
(
"""
Notify about a cluster state change.
class
AddPendingNodes
(
Packet
):
"""
Mark given pending nodes as running, for future inclusion when tweaking
the partition table.
:nodes: M -> *
"""
)
:nodes: ctl -> A -> M
"""
_fmt
=
PStruct
(
'add_pending_nodes'
,
PFUUIDList
,
)
AskClusterState
,
AnswerClusterState
=
request
(
"""
Ask the state of the cluster
_answer
=
Error
:nodes: ctl -> A; A -> M
"""
)
class
TweakPartitionTable
(
Packet
):
"""
Ask the master to balance the partition table, optionally excluding
specific nodes in anticipation of removing them.
AskObjectUndoSerial
,
AnswerObjectUndoSerial
=
request
(
"""
Ask storage the serial where object data is when undoing given
transaction, for a list of OIDs.
:nodes: ctl -> A -> M
"""
_fmt
=
PStruct
(
'tweak_partition_table'
,
PFUUIDList
,
)
_answer
=
Error
class
NotifyNodeInformation
(
Packet
):
"""
Notify information about one or more nodes.
:nodes: M -> *
"""
_fmt
=
PStruct
(
'notify_node_informations'
,
PFloat
(
'id_timestamp'
),
PFNodeList
,
)
class
SetClusterState
(
Packet
):
"""
Set the cluster state.
:nodes: ctl -> A -> M
"""
_fmt
=
PStruct
(
'set_cluster_state'
,
PEnum
(
'state'
,
ClusterStates
),
)
_answer
=
Error
class
Repair
(
Packet
):
"""
Ask storage nodes to repair their databases.
:nodes: ctl -> A -> M
"""
_flags
=
map
(
PBoolean
,
(
'dry_run'
,
# 'prune_orphan' (commented because it's the only option for the moment)
))
_fmt
=
PStruct
(
'repair'
,
PFUUIDList
,
*
_flags
)
Answer a dict mapping oids to 3-tuples:
current_serial (TID)
The latest serial visible to the undoing transaction.
undo_serial (TID)
Where undone data is (tid at which data is before given undo).
is_current (bool)
If current_serial's data is current on storage.
_answer
=
Error
:nodes: C -> S
"""
)
class
RepairOne
(
Packet
):
"""
Repair is translated to this message, asking a specific storage node to
repair its database.
AskTIDsFrom
,
AnswerTIDsFrom
=
request
(
"""
Ask for length TIDs starting at min_tid. The order of TIDs is ascending.
Used by `iterator`.
:nodes: M -> S
"""
_fmt
=
PStruct
(
'repair'
,
*
Repair
.
_flags
)
:nodes: C -> S
"""
)
class
ClusterInformation
(
Packet
):
"""
Notify about a cluster state change.
AskPack
,
AnswerPack
=
request
(
"""
Request a pack at given TID.
:nodes: M -> *
"""
_fmt
=
PStruct
(
'notify_cluster_information'
,
PEnum
(
'state'
,
ClusterStates
),
)
:nodes: C -> M -> S
"""
,
ignore_when_closed
=
False
)
class
ClusterState
(
Packet
):
"""
Ask the state of the cluster
CheckReplicas
=
request
(
"""
Ask the cluster to search for mismatches between replicas, metadata
only, and optionally within a specific range. Reference nodes can be
specified.
:nodes: ctl -> A;
A -> M
"""
:nodes: ctl ->
A -> M
"""
,
error
=
True
)
_answer
=
PStruct
(
'answer_cluster_state'
,
PEnum
(
'state'
,
ClusterStates
),
)
CheckPartition
=
notify
(
"""
Ask a storage node to compare a partition with all other nodes.
Like for CheckReplicas, only metadata are checked, optionally within a
specific range. A reference node can be specified.
class
ObjectUndoSerial
(
Packet
):
"""
Ask storage the serial where object data is when undoing given transaction,
for a list of OIDs.
:nodes: M -> S
"""
)
object_tid_dict has the following format:
key: oid
value: 3-tuple
current_serial (TID)
The latest serial visible to the undoing transaction.
undo_serial (TID)
Where undone data is (tid at which data is before given undo).
is_current (bool)
If current_serial's data is current on storage.
AskCheckTIDRange
,
AnswerCheckTIDRange
=
request
(
"""
Ask some stats about a range of transactions.
Used to know if there are differences between a replicating node and
reference node.
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_undo_transaction'
,
PTID
(
'tid'
),
PTID
(
'ltid'
),
PTID
(
'undone_tid'
),
PFOidList
,
)
_answer
=
PStruct
(
'answer_undo_transaction'
,
PDict
(
'object_tid_dict'
,
POID
(
'oid'
),
PStruct
(
'object_tid_value'
,
PTID
(
'current_serial'
),
PTID
(
'undo_serial'
),
PBoolean
(
'is_current'
),
),
),
)
class
CheckCurrentSerial
(
Packet
):
"""
Check if given serial is current for the given oid, and lock it so that
this state is not altered until transaction ends.
This maps to `checkCurrentSerialInTransaction`.
:nodes: S -> S
"""
)
:nodes: C -> S
"""
_fmt
=
PStruct
(
'ask_check_current_serial'
,
PTID
(
'tid'
),
POID
(
'oid'
),
PTID
(
'serial'
),
)
AskCheckSerialRange
,
AnswerCheckSerialRange
=
request
(
"""
Ask some stats about a range of object history.
Used to know if there are differences between a replicating node and
reference node.
_answer
=
StoreObject
.
_answer
:nodes: S -> S
"""
)
class
Pack
(
Packet
):
"""
Request a pack at given TID.
NotifyPartitionCorrupted
=
notify
(
"""
Notify that mismatches were found while check replicas for a partition.
:nodes: C -> M -> S
"""
_fmt
=
PStruct
(
'ask_pack'
,
PTID
(
'tid'
),
)
:nodes: S -> M
"""
)
_answer
=
PStruct
(
'answer_pack'
,
PBoolean
(
'status'
),
)
NotifyReady
=
notify
(
"""
Notify that we're ready to serve requests.
class
CheckReplicas
(
Packet
):
"""
Ask the cluster to search for mismatches between replicas, metadata only,
and optionally within a specific range. Reference nodes can be specified.
:nodes: S -> M
"""
)
:nodes: ctl -> A -> M
"""
_fmt
=
PStruct
(
'check_replicas'
,
PDict
(
'partition_dict'
,
PNumber
(
'partition'
),
PUUID
(
'source'
),
),
PTID
(
'min_tid'
),
PTID
(
'max_tid'
),
)
_answer
=
Error
class
CheckPartition
(
Packet
):
"""
Ask a storage node to compare a partition with all other nodes.
Like for CheckReplicas, only metadata are checked, optionally within a
specific range. A reference node can be specified.
AskLastTransaction
,
AnswerLastTransaction
=
request
(
"""
Ask last committed TID.
:nodes: M -> S
"""
_fmt
=
PStruct
(
'check_partition'
,
PNumber
(
'partition'
),
PStruct
(
'source'
,
PString
(
'upstream_name'
),
PAddress
(
'address'
),
),
PTID
(
'min_tid'
),
PTID
(
'max_tid'
),
)
class
CheckTIDRange
(
Packet
):
"""
Ask some stats about a range of transactions.
Used to know if there are differences between a replicating node and
reference node.
:nodes: C -> M; ctl -> A -> M
"""
,
poll_thread
=
True
)
:nodes: S -> S
"""
_fmt
=
PStruct
(
'ask_check_tid_range'
,
PNumber
(
'partition'
),
PNumber
(
'length'
),
PTID
(
'min_tid'
),
PTID
(
'max_tid'
),
)
_answer
=
PStruct
(
'answer_check_tid_range'
,
PNumber
(
'count'
),
PChecksum
(
'checksum'
),
PTID
(
'max_tid'
),
)
class
CheckSerialRange
(
Packet
):
"""
Ask some stats about a range of object history.
Used to know if there are differences between a replicating node and
reference node.
AskCheckCurrentSerial
,
AnswerCheckCurrentSerial
=
request
(
"""
Check if given serial is current for the given oid, and lock it so that
this state is not altered until transaction ends.
This maps to `checkCurrentSerialInTransaction`.
:nodes: S -> S
"""
_fmt
=
PStruct
(
'ask_check_serial_range'
,
PNumber
(
'partition'
),
PNumber
(
'length'
),
PTID
(
'min_tid'
),
PTID
(
'max_tid'
),
POID
(
'min_oid'
),
)
_answer
=
PStruct
(
'answer_check_serial_range'
,
PNumber
(
'count'
),
PChecksum
(
'tid_checksum'
),
PTID
(
'max_tid'
),
PChecksum
(
'oid_checksum'
),
POID
(
'max_oid'
),
)
class
PartitionCorrupted
(
Packet
):
"""
Notify that mismatches were found while check replicas for a partition.
:nodes: C -> S
"""
)
:nodes: S -> M
"""
_fmt
=
PStruct
(
'partition_corrupted'
,
PNumber
(
'partition'
),
PList
(
'cell_list'
,
PUUID
(
'uuid'
),
),
)
class
LastTransaction
(
Packet
):
"""
Ask last committed TID.
NotifyTransactionFinished
=
notify
(
"""
Notify that a transaction blocking a replication is now finished.
:nodes: C -> M; ctl -> A -> M
"""
poll_thread
=
True
:nodes: M -> S
"""
)
_answer
=
PStruct
(
'answer_last_transaction'
,
PTID
(
'tid'
),
)
Replicate
=
notify
(
"""
Notify a storage node to replicate partitions up to given 'tid'
and from given sources.
class
NotifyReady
(
Packet
):
"""
Notify that we're ready to serve requests.
args: tid, upstream_name, {partition: address}
- upstream_name: replicate from an upstream cluster
- address: address of the source storage node, or None if there's
no new data up to 'tid' for the given partition
:nodes: S -> M
"""
:nodes: M -> S
"""
)
class
FetchTransactions
(
Packet
):
"""
Ask a storage node to send all transaction data we don't have,
and reply with the list of transactions we should not have.
NotifyReplicationDone
=
notify
(
"""
Notify the master node that a partition has been successfully
replicated from a storage to another.
:nodes: S -> S
"""
_fmt
=
PStruct
(
'ask_transaction_list'
,
PNumber
(
'partition'
),
PNumber
(
'length'
),
PTID
(
'min_tid'
),
PTID
(
'max_tid'
),
PFTidList
,
# already known transactions
)
_answer
=
PStruct
(
'answer_transaction_list'
,
PTID
(
'pack_tid'
),
PTID
(
'next_tid'
),
PFTidList
,
# transactions to delete
)
class
AddTransaction
(
Packet
):
"""
Send metadata of a transaction to a node that do not have them.
:nodes: S -> M
"""
)
:nodes: S -> S
"""
nodelay
=
False
_fmt
=
PStruct
(
'add_transaction'
,
PTID
(
'tid'
),
PString
(
'user'
),
PString
(
'description'
),
PString
(
'extension'
),
PBoolean
(
'packed'
),
PTID
(
'ttid'
),
PFOidList
,
)
class
FetchObjects
(
Packet
):
"""
Ask a storage node to send object records we don't have,
and reply with the list of records we should not have.
AskFetchTransactions
,
AnswerFetchTransactions
=
request
(
"""
Ask a storage node to send all transaction data we don't have,
and reply with the list of transactions we should not have.
:nodes: S -> S
"""
_fmt
=
PStruct
(
'ask_object_list'
,
PNumber
(
'partition'
),
PNumber
(
'length'
),
PTID
(
'min_tid'
),
PTID
(
'max_tid'
),
POID
(
'min_oid'
),
PDict
(
'object_dict'
,
# already known objects
PTID
(
'serial'
),
PFOidList
,
),
)
_answer
=
PStruct
(
'answer_object_list'
,
PTID
(
'pack_tid'
),
PTID
(
'next_tid'
),
POID
(
'next_oid'
),
PDict
(
'object_dict'
,
# objects to delete
PTID
(
'serial'
),
PFOidList
,
),
)
class
AddObject
(
Packet
):
"""
Send an object record to a node that do not have it.
:nodes: S -> S
"""
)
:nodes: S -> S
"""
nodelay
=
False
_fmt
=
PStruct
(
'add_object'
,
POID
(
'oid'
),
PTID
(
'serial'
),
PBoolean
(
'compression'
),
PChecksum
(
'checksum'
),
PString
(
'data'
),
PTID
(
'data_serial'
),
)
class
Replicate
(
Packet
):
"""
Notify a storage node to replicate partitions up to given 'tid'
and from given sources.
AskFetchObjects
,
AnswerFetchObjects
=
request
(
"""
Ask a storage node to send object records we don't have,
and reply with the list of records we should not have.
- upstream_name: replicate from an upstream cluster
- address: address of the source storage node, or None if there's no new
data up to 'tid' for the given partition
:nodes: S -> S
"""
)
:nodes: M -> S
"""
_fmt
=
PStruct
(
'replicate'
,
PTID
(
'tid'
),
PString
(
'upstream_name'
),
PDict
(
'source_dict'
,
PNumber
(
'partition'
),
PAddress
(
'address'
),
)
)
class
ReplicationDone
(
Packet
):
"""
Notify the master node that a partition has been successfully replicated
from a storage to another.
AddTransaction
=
notify
(
"""
Send metadata of a transaction to a node that does not have them.
:nodes: S -> M
"""
_fmt
=
PStruct
(
'notify_replication_done'
,
PNumber
(
'offset'
),
PTID
(
'tid'
),
)
:nodes: S -> S
"""
,
nodelay
=
False
)
class
Truncate
(
Packet
):
"""
Request DB to be truncated. Also used to leave backup mode.
AddObject
=
notify
(
"""
Send an object record to a node that does not have it.
:nodes: ctl -> A -> M; M -> S
"""
_fmt
=
PStruct
(
'truncate'
,
PTID
(
'tid'
),
)
:nodes: S -> S
"""
,
nodelay
=
False
,
data_path
=
(
0
,
2
))
_answer
=
Error
Truncate
=
request
(
"""
Request DB to be truncated. Also used to leave backup mode.
class
FlushLog
(
Packet
):
"""
Request all nodes to flush their logs.
:nodes: ctl -> A -> M; M -> S
"""
,
error
=
True
)
:nodes: ctl -> A -> M -> *
"""
FlushLog
=
notify
(
"""
Request all nodes to flush their logs.
:nodes: ctl -> A -> M -> *
"""
)
_next_code
=
0
def
register
(
request
,
ignore_when_closed
=
None
):
""" Register a packet in the packet registry """
global
_next_code
code
=
_next_code
assert
code
<
RESPONSE_MASK
_next_code
=
code
+
1
if
request
is
Error
:
code
|=
RESPONSE_MASK
# register the request
request
.
_code
=
code
answer
=
request
.
_answer
if
ignore_when_closed
is
None
:
# By default, on a closed connection:
# - request: ignore
# - answer: keep
# - notification: keep
ignore_when_closed
=
answer
is
not
None
request
.
_ignore_when_closed
=
ignore_when_closed
if
answer
in
(
Error
,
None
):
return
request
# build a class for the answer
answer
=
type
(
'Answer'
+
request
.
__name__
,
(
Packet
,
),
{})
answer
.
_fmt
=
request
.
_answer
answer
.
poll_thread
=
request
.
poll_thread
answer
.
_request
=
request
assert
answer
.
_code
is
None
,
"Answer of %s is already used"
%
(
request
,
)
answer
.
_code
=
code
|
RESPONSE_MASK
request
.
_answer
=
answer
return
request
,
answer
del
notify
,
request
class
Packets
(
dict
):
"""
Packet registry that checks packet code uniqueness and provides an index
"""
def
__metaclass__
(
name
,
base
,
d
):
# this builds a "singleton"
cls
=
type
(
'PacketRegistry'
,
base
,
d
)()
for
k
,
v
in
d
.
iteritems
():
if
isinstance
(
v
,
type
)
and
issubclass
(
v
,
Packet
):
v
.
handler_method_name
=
k
[
0
].
lower
()
+
k
[
1
:]
cls
[
v
.
_code
]
=
v
return
cls
Error
=
register
(
Error
)
RequestIdentification
,
AcceptIdentification
=
register
(
RequestIdentification
,
ignore_when_closed
=
True
)
Ping
,
Pong
=
register
(
Ping
)
CloseClient
=
register
(
CloseClient
)
AskPrimary
,
AnswerPrimary
=
register
(
PrimaryMaster
)
NotPrimaryMaster
=
register
(
NotPrimaryMaster
)
NotifyNodeInformation
=
register
(
NotifyNodeInformation
)
AskRecovery
,
AnswerRecovery
=
register
(
Recovery
)
AskLastIDs
,
AnswerLastIDs
=
register
(
LastIDs
)
AskPartitionTable
,
AnswerPartitionTable
=
register
(
PartitionTable
)
SendPartitionTable
=
register
(
NotifyPartitionTable
)
NotifyPartitionChanges
=
register
(
PartitionChanges
)
StartOperation
=
register
(
StartOperation
)
StopOperation
=
register
(
StopOperation
)
AskUnfinishedTransactions
,
AnswerUnfinishedTransactions
=
register
(
UnfinishedTransactions
)
AskLockedTransactions
,
AnswerLockedTransactions
=
register
(
LockedTransactions
)
AskFinalTID
,
AnswerFinalTID
=
register
(
FinalTID
)
ValidateTransaction
=
register
(
ValidateTransaction
)
AskBeginTransaction
,
AnswerBeginTransaction
=
register
(
BeginTransaction
)
FailedVote
=
register
(
FailedVote
)
AskFinishTransaction
,
AnswerTransactionFinished
=
register
(
FinishTransaction
,
ignore_when_closed
=
False
)
AskLockInformation
,
AnswerInformationLocked
=
register
(
LockInformation
,
ignore_when_closed
=
False
)
InvalidateObjects
=
register
(
InvalidateObjects
)
NotifyUnlockInformation
=
register
(
UnlockInformation
)
AskNewOIDs
,
AnswerNewOIDs
=
register
(
GenerateOIDs
)
NotifyDeadlock
=
register
(
Deadlock
)
AskRebaseTransaction
,
AnswerRebaseTransaction
=
register
(
RebaseTransaction
)
AskRebaseObject
,
AnswerRebaseObject
=
register
(
RebaseObject
)
AskStoreObject
,
AnswerStoreObject
=
register
(
StoreObject
)
AbortTransaction
=
register
(
AbortTransaction
)
AskStoreTransaction
,
AnswerStoreTransaction
=
register
(
StoreTransaction
)
AskVoteTransaction
,
AnswerVoteTransaction
=
register
(
VoteTransaction
)
AskObject
,
AnswerObject
=
register
(
GetObject
)
AskTIDs
,
AnswerTIDs
=
register
(
TIDList
)
AskTransactionInformation
,
AnswerTransactionInformation
=
register
(
TransactionInformation
)
AskObjectHistory
,
AnswerObjectHistory
=
register
(
ObjectHistory
)
AskPartitionList
,
AnswerPartitionList
=
register
(
PartitionList
)
AskNodeList
,
AnswerNodeList
=
register
(
NodeList
)
SetNodeState
=
register
(
SetNodeState
,
ignore_when_closed
=
False
)
AddPendingNodes
=
register
(
AddPendingNodes
,
ignore_when_closed
=
False
)
TweakPartitionTable
=
register
(
TweakPartitionTable
,
ignore_when_closed
=
False
)
SetClusterState
=
register
(
SetClusterState
,
ignore_when_closed
=
False
)
Repair
=
register
(
Repair
)
NotifyRepair
=
register
(
RepairOne
)
NotifyClusterInformation
=
register
(
ClusterInformation
)
AskClusterState
,
AnswerClusterState
=
register
(
ClusterState
)
AskObjectUndoSerial
,
AnswerObjectUndoSerial
=
register
(
ObjectUndoSerial
)
AskTIDsFrom
,
AnswerTIDsFrom
=
register
(
TIDListFrom
)
AskPack
,
AnswerPack
=
register
(
Pack
,
ignore_when_closed
=
False
)
CheckReplicas
=
register
(
CheckReplicas
)
CheckPartition
=
register
(
CheckPartition
)
AskCheckTIDRange
,
AnswerCheckTIDRange
=
register
(
CheckTIDRange
)
AskCheckSerialRange
,
AnswerCheckSerialRange
=
register
(
CheckSerialRange
)
NotifyPartitionCorrupted
=
register
(
PartitionCorrupted
)
NotifyReady
=
register
(
NotifyReady
)
AskLastTransaction
,
AnswerLastTransaction
=
register
(
LastTransaction
)
AskCheckCurrentSerial
,
AnswerCheckCurrentSerial
=
register
(
CheckCurrentSerial
)
NotifyTransactionFinished
=
register
(
NotifyTransactionFinished
)
Replicate
=
register
(
Replicate
)
NotifyReplicationDone
=
register
(
ReplicationDone
)
AskFetchTransactions
,
AnswerFetchTransactions
=
register
(
FetchTransactions
)
AskFetchObjects
,
AnswerFetchObjects
=
register
(
FetchObjects
)
AddTransaction
=
register
(
AddTransaction
)
AddObject
=
register
(
AddObject
)
Truncate
=
register
(
Truncate
)
FlushLog
=
register
(
FlushLog
)
def
Errors
():
registry_dict
=
{}
handler_method_name_dict
=
{}
Error
=
Packets
.
Error
def
register_error
(
code
):
return
lambda
self
,
message
=
''
:
Error
(
code
,
message
)
for
error
in
ErrorCodes
:
...
...
@@ -1836,19 +851,20 @@ from operator import itemgetter
def
formatNodeList
(
node_list
,
prefix
=
''
,
_sort_key
=
itemgetter
(
2
)):
if
node_list
:
node_list
.
sort
(
key
=
_sort_key
)
node_list
=
[(
uuid_str
(
uuid
),
str
(
node_type
),
(
'[%s]:%s'
if
':'
in
addr
[
0
]
else
'%s:%s'
)
%
addr
if
addr
else
''
,
str
(
state
),
str
(
id_timestamp
and
datetime
.
utcfromtimestamp
(
id_timestamp
)))
for
node_type
,
addr
,
uuid
,
state
,
id_timestamp
in
node_list
]
for
node_type
,
addr
,
uuid
,
state
,
id_timestamp
in
sorted
(
node_list
,
key
=
_sort_key
)]
t
=
''
.
join
(
'%%-%us | '
%
max
(
len
(
x
[
i
])
for
x
in
node_list
)
for
i
in
xrange
(
len
(
node_list
[
0
])
-
1
))
return
map
((
prefix
+
t
+
'%s'
).
__mod__
,
node_list
)
return
()
NotifyNodeInformation
.
_neolog
=
staticmethod
(
lambda
timestamp
,
node_list
:
Packets
.
NotifyNodeInformation
.
_neolog
=
staticmethod
(
lambda
timestamp
,
node_list
:
((
timestamp
,),
formatNodeList
(
node_list
,
' ! '
)))
Error
.
_neolog
=
staticmethod
(
lambda
*
args
:
((),
(
"%s (%s)"
%
args
,)))
Packets
.
Error
.
_neolog
=
staticmethod
(
lambda
*
args
:
((),
(
"%s (%s)"
%
args
,)))
neo/lib/pt.py
View file @
8ba42463
...
...
@@ -86,15 +86,9 @@ class PartitionTable(object):
'a cell became non-readable whereas all cells were readable'
def
__init__
(
self
,
num_partitions
,
num_replicas
):
self
.
_id
=
None
self
.
np
=
num_partitions
self
.
nr
=
num_replicas
self
.
num_filled_rows
=
0
# Note: don't use [[]] * num_partition construct, as it duplicates
# instance *references*, so the outer list contains really just one
# inner list instance.
self
.
partition_list
=
[[]
for
_
in
xrange
(
num_partitions
)]
self
.
count_dict
=
{}
self
.
clear
()
def
getID
(
self
):
return
self
.
_id
...
...
@@ -113,7 +107,16 @@ class PartitionTable(object):
# instance *references*, so the outer list contains really just one
# inner list instance.
self
.
partition_list
=
[[]
for
_
in
xrange
(
self
.
np
)]
self
.
count_dict
.
clear
()
self
.
count_dict
=
{}
def
addNodeList
(
self
,
node_list
):
"""Add nodes"""
added_list
=
[]
for
node
in
node_list
:
if
node
not
in
self
.
count_dict
:
self
.
count_dict
[
node
]
=
0
added_list
.
append
(
node
)
return
added_list
def
getAssignedPartitionList
(
self
,
uuid
):
""" Return the partition assigned to the specified UUID """
...
...
@@ -203,31 +206,31 @@ class PartitionTable(object):
del
self
.
count_dict
[
node
]
return
not
count
def
load
(
self
,
ptid
,
row_list
,
nm
):
def
_load
(
self
,
ptid
,
num_replicas
,
row_list
,
getByUUID
):
self
.
__init__
(
len
(
row_list
),
num_replicas
)
self
.
_id
=
ptid
for
offset
,
row
in
enumerate
(
row_list
):
for
uuid
,
state
in
row
:
node
=
getByUUID
(
uuid
)
self
.
_setCell
(
offset
,
node
,
state
)
def
load
(
self
,
ptid
,
num_replicas
,
row_list
,
nm
):
"""
Load the partition table with the specified PTID, discard all previous
content.
"""
self
.
clear
()
self
.
_id
=
ptid
for
offset
,
row
in
row_list
:
if
offset
>=
self
.
getPartitions
():
raise
IndexError
for
uuid
,
state
in
row
:
node
=
nm
.
getByUUID
(
uuid
)
# the node must be known by the node manager
assert
node
is
not
None
self
.
_setCell
(
offset
,
node
,
state
)
self
.
_load
(
ptid
,
num_replicas
,
row_list
,
nm
.
getByUUID
)
logging
.
debug
(
'partition table loaded (ptid=%s)'
,
ptid
)
self
.
log
()
def
update
(
self
,
ptid
,
cell_list
,
nm
):
def
update
(
self
,
ptid
,
num_replicas
,
cell_list
,
nm
):
"""
Update the partition with the cell list supplied. If a node
is not known, it is created in the node manager and set as unavailable
"""
assert
self
.
_id
<
ptid
,
(
self
.
_id
,
ptid
)
self
.
_id
=
ptid
self
.
nr
=
num_replicas
readable_list
=
[]
for
row
in
self
.
partition_list
:
if
not
all
(
cell
.
isReadable
()
for
cell
in
row
):
...
...
@@ -310,14 +313,11 @@ class PartitionTable(object):
return
True
def
getRow
(
self
,
offset
):
row
=
self
.
partition_list
[
offset
]
if
row
is
None
:
return
[]
return
[(
cell
.
getUUID
(),
cell
.
getState
())
for
cell
in
row
]
return
[(
cell
.
getUUID
(),
cell
.
getState
())
for
cell
in
self
.
partition_list
[
offset
]]
def
getRowList
(
self
):
getRow
=
self
.
getRow
return
[(
x
,
getRow
(
x
))
for
x
in
xrange
(
self
.
np
)]
return
map
(
self
.
getRow
,
xrange
(
self
.
np
))
class
MTPartitionTable
(
PartitionTable
):
""" Thread-safe aware version of the partition table, override only methods
...
...
neo/lib/util.py
View file @
8ba42463
...
...
@@ -166,65 +166,6 @@ def parseMasterList(masters):
return
map
(
parseNodeAddress
,
masters
.
split
())
class
ReadBuffer
(
object
):
"""
Implementation of a lazy buffer. Main purpose if to reduce useless
copies of data by storing chunks and join them only when the requested
size is available.
TODO: For better performance, use:
- socket.recv_into (64kiB blocks)
- struct.unpack_from
- and a circular buffer of dynamic size (initial size:
twice the length passed to socket.recv_into ?)
"""
def
__init__
(
self
):
self
.
size
=
0
self
.
content
=
deque
()
def
append
(
self
,
data
):
""" Append some data and compute the new buffer size """
self
.
size
+=
len
(
data
)
self
.
content
.
append
(
data
)
def
__len__
(
self
):
""" Return the current buffer size """
return
self
.
size
def
read
(
self
,
size
):
""" Read and consume size bytes """
if
self
.
size
<
size
:
return
None
self
.
size
-=
size
chunk_list
=
[]
pop_chunk
=
self
.
content
.
popleft
append_data
=
chunk_list
.
append
to_read
=
size
# select required chunks
while
to_read
>
0
:
chunk_data
=
pop_chunk
()
to_read
-=
len
(
chunk_data
)
append_data
(
chunk_data
)
if
to_read
<
0
:
# too many bytes consumed, cut the last chunk
last_chunk
=
chunk_list
[
-
1
]
keep
,
let
=
last_chunk
[:
to_read
],
last_chunk
[
to_read
:]
self
.
content
.
appendleft
(
let
)
chunk_list
[
-
1
]
=
keep
# join all chunks (one copy)
data
=
''
.
join
(
chunk_list
)
assert
len
(
data
)
==
size
return
data
def
clear
(
self
):
""" Erase all buffer content """
self
.
size
=
0
self
.
content
.
clear
()
dummy_read_buffer
=
ReadBuffer
()
dummy_read_buffer
.
append
=
lambda
_
:
None
class
cached_property
(
object
):
"""
A property that is only computed once per instance and then replaces itself
...
...
neo/master/app.py
View file @
8ba42463
...
...
@@ -16,6 +16,7 @@
import
sys
from
collections
import
defaultdict
from
functools
import
partial
from
time
import
time
from
neo.lib
import
logging
,
util
...
...
@@ -76,13 +77,11 @@ class Application(BaseApplication):
@
classmethod
def
_buildOptionParser
(
cls
):
_
=
cls
.
option_parser
_
.
description
=
"NEO Master node"
parser
=
cls
.
option_parser
parser
.
description
=
"NEO Master node"
cls
.
addCommonServerOptions
(
'master'
,
'127.0.0.1:10000'
,
''
)
_
=
_
.
group
(
'master'
)
_
.
int
(
'r'
,
'replicas'
,
default
=
0
,
help
=
"replicas number"
)
_
.
int
(
'p'
,
'partitions'
,
default
=
100
,
help
=
"partitions number"
)
_
=
parser
.
group
(
'master'
)
_
.
int
(
'A'
,
'autostart'
,
help
=
"minimum number of pending storage nodes to automatically"
" start new cluster (to avoid unwanted recreation of the"
...
...
@@ -91,8 +90,12 @@ class Application(BaseApplication):
help
=
'the name of cluster to backup'
)
_
(
'M'
,
'upstream-masters'
,
parse
=
util
.
parseMasterList
,
help
=
'list of master nodes in the cluster to backup'
)
_
.
int
(
'u'
,
'uuid'
,
help
=
"specify an UUID to use for this process (testing purpose)"
)
_
.
int
(
'i'
,
'nid'
,
help
=
"specify an NID to use for this process (testing purpose)"
)
_
=
parser
.
group
(
'database creation'
)
_
.
int
(
'r'
,
'replicas'
,
default
=
0
,
help
=
"replicas number"
)
_
.
int
(
'p'
,
'partitions'
,
default
=
100
,
help
=
"partitions number"
)
def
__init__
(
self
,
config
):
super
(
Application
,
self
).
__init__
(
...
...
@@ -108,7 +111,7 @@ class Application(BaseApplication):
for
master_address
in
config
[
'masters'
]:
self
.
nm
.
createMaster
(
address
=
master_address
)
self
.
_node
=
self
.
nm
.
createMaster
(
address
=
self
.
server
,
uuid
=
config
.
get
(
'
uu
id'
))
uuid
=
config
.
get
(
'
n
id'
))
logging
.
node
(
self
.
name
,
self
.
uuid
)
logging
.
debug
(
'IP address is %s, port is %d'
,
*
self
.
server
)
...
...
@@ -117,14 +120,14 @@ class Application(BaseApplication):
replicas
=
config
[
'replicas'
]
partitions
=
config
[
'partitions'
]
if
replicas
<
0
:
raise
RuntimeError
,
'replicas must be a positive integer'
sys
.
exit
(
'replicas must be a positive integer'
)
if
partitions
<=
0
:
raise
RuntimeError
,
'partitions must be more than zero'
self
.
pt
=
PartitionTable
(
partitions
,
replicas
)
sys
.
exit
(
'partitions must be more than zero'
)
logging
.
info
(
'Configuration:'
)
logging
.
info
(
'Partitions: %d'
,
partitions
)
logging
.
info
(
'Replicas : %d'
,
replicas
)
logging
.
info
(
'Name : %s'
,
self
.
name
)
self
.
newPartitionTable
=
partial
(
PartitionTable
,
partitions
,
replicas
)
self
.
listening_conn
=
None
self
.
cluster_state
=
None
...
...
@@ -196,7 +199,7 @@ class Application(BaseApplication):
node_dict
[
NodeTypes
.
MASTER
].
append
(
node_info
)
return
node_dict
def
broadcastNodesInformation
(
self
,
node_list
,
exclude
=
None
):
def
broadcastNodesInformation
(
self
,
node_list
):
"""
Broadcast changes for a set a nodes
Send only one packet per connection to reduce bandwidth
...
...
@@ -209,20 +212,26 @@ class Application(BaseApplication):
# We don't skip pending storage nodes because we don't send them
# the full list of nodes when they're added, and it's also quite
# useful to notify them about new masters.
if
node_list
and
node
is
not
exclude
:
if
node_list
:
node
.
send
(
Packets
.
NotifyNodeInformation
(
now
,
node_list
))
def
broadcastPartitionChanges
(
self
,
cell_list
):
def
broadcastPartitionChanges
(
self
,
cell_list
,
num_replicas
=
None
):
"""Broadcast a Notify Partition Changes packet."""
if
cell_list
:
ptid
=
self
.
pt
.
setNextID
()
self
.
pt
.
logUpdated
()
packet
=
Packets
.
NotifyPartitionChanges
(
ptid
,
cell_list
)
for
node
in
self
.
nm
.
getIdentifiedList
():
# As for broadcastNodesInformation, we don't send the full PT
# when pending storage nodes are added, so keep them notified.
if
not
node
.
isMaster
():
node
.
send
(
packet
)
pt
=
self
.
pt
if
num_replicas
is
not
None
:
pt
.
setReplicas
(
num_replicas
)
elif
cell_list
:
num_replicas
=
pt
.
getReplicas
()
else
:
return
packet
=
Packets
.
NotifyPartitionChanges
(
pt
.
setNextID
(),
num_replicas
,
cell_list
)
pt
.
logUpdated
()
for
node
in
self
.
nm
.
getIdentifiedList
():
# As for broadcastNodesInformation, we don't send the full PT
# when pending storage nodes are added, so keep them notified.
if
not
node
.
isMaster
():
node
.
send
(
packet
)
def
provideService
(
self
):
"""
...
...
@@ -437,16 +446,7 @@ class Application(BaseApplication):
conn
.
send
(
notification_packet
)
elif
conn
.
isServer
():
continue
if
node
.
isClient
():
if
state
==
ClusterStates
.
RUNNING
:
handler
=
self
.
client_service_handler
elif
state
==
ClusterStates
.
BACKINGUP
:
handler
=
self
.
client_ro_service_handler
else
:
if
state
!=
ClusterStates
.
STOPPING
:
conn
.
abort
()
continue
elif
node
.
isMaster
():
if
node
.
isMaster
():
if
state
==
ClusterStates
.
RECOVERING
:
handler
=
self
.
election_handler
else
:
...
...
@@ -454,10 +454,16 @@ class Application(BaseApplication):
elif
node
.
isStorage
()
and
storage_handler
:
handler
=
storage_handler
else
:
# There's a single handler type for admins.
# Client can't change handler without being first disconnected.
assert
state
in
(
ClusterStates
.
STOPPING
,
ClusterStates
.
STOPPING_BACKUP
,
)
or
not
node
.
isClient
(),
(
state
,
node
)
continue
# keep handler
if
type
(
handler
)
is
not
type
(
conn
.
getLastHandler
()):
conn
.
setHandler
(
handler
)
handler
.
connectionComplet
ed
(
conn
,
new
=
False
)
handler
.
handlerSwitch
ed
(
conn
,
new
=
False
)
self
.
cluster_state
=
state
def
getNewUUID
(
self
,
uuid
,
address
,
node_type
):
...
...
@@ -578,7 +584,9 @@ class Application(BaseApplication):
self
.
tm
.
executeQueuedEvents
()
def
startStorage
(
self
,
node
):
node
.
send
(
Packets
.
StartOperation
(
self
.
backup_tid
))
# XXX: Is this boolean 'backup' field needed ?
# Maybe this can be deduced from cluster state.
node
.
send
(
Packets
.
StartOperation
(
bool
(
self
.
backup_tid
)))
uuid
=
node
.
getUUID
()
assert
uuid
not
in
self
.
storage_starting_set
if
uuid
not
in
self
.
storage_ready_dict
:
...
...
neo/master/backup_app.py
View file @
8ba42463
...
...
@@ -65,6 +65,7 @@ There is no conflict of node id between the 2 clusters:
class
BackupApplication
(
object
):
pt
=
None
server
=
None
# like in BaseApplication
uuid
=
None
def
__init__
(
self
,
app
,
name
,
master_addresses
):
...
...
@@ -111,17 +112,12 @@ class BackupApplication(object):
else
:
break
poll
(
1
)
node
,
conn
,
num_partitions
,
num_replicas
=
\
bootstrap
.
getPrimaryConnection
()
node
,
conn
=
bootstrap
.
getPrimaryConnection
()
try
:
app
.
changeClusterState
(
ClusterStates
.
BACKINGUP
)
del
bootstrap
,
node
if
num_partitions
!=
pt
.
getPartitions
():
raise
RuntimeError
(
"inconsistent number of partitions"
)
self
.
ignore_invalidations
=
True
self
.
pt
=
PartitionTable
(
num_partitions
,
num_replicas
)
conn
.
setHandler
(
BackupHandler
(
self
))
conn
.
ask
(
Packets
.
AskPartitionTable
())
conn
.
ask
(
Packets
.
AskLastTransaction
())
# debug variable to log how big 'tid_list' can be.
self
.
debug_tid_count
=
0
...
...
neo/master/handlers/__init__.py
View file @
8ba42463
...
...
@@ -23,10 +23,6 @@ from neo.lib.protocol import Packets
class
MasterHandler
(
EventHandler
):
"""This class implements a generic part of the event handlers."""
def
connectionCompleted
(
self
,
conn
,
new
=
None
):
if
new
is
None
:
super
(
MasterHandler
,
self
).
connectionCompleted
(
conn
)
def
connectionLost
(
self
,
conn
,
new_state
=
None
):
if
self
.
app
.
listening_conn
:
# if running
self
.
_connectionLost
(
conn
)
...
...
@@ -59,17 +55,20 @@ class MasterHandler(EventHandler):
+
app
.
getNodeInformationDict
(
node_list
)[
node
.
getType
()])
conn
.
send
(
Packets
.
NotifyNodeInformation
(
monotonic_time
(),
node_list
))
def
askPartitionTable
(
self
,
conn
):
def
handlerSwitched
(
self
,
conn
,
new
):
pt
=
self
.
app
.
pt
conn
.
answer
(
Packets
.
AnswerPartitionTable
(
pt
.
getID
(),
pt
.
getRowList
()))
# Except storages during recovery and secondary masters, all nodes
# receives the full partition table as soon as they're identified.
# It is also sent in 2 other cases:
# - to admins during recovery, whenever a newer PT is loaded;
# - to storage when switching from recovery to verification.
# After that, non-master nodes only receive incremental updates.
conn
.
send
(
Packets
.
SendPartitionTable
(
pt
.
getID
(),
pt
.
getReplicas
(),
pt
.
getRowList
()))
class
BaseServiceHandler
(
MasterHandler
):
"""This class deals with events for a service phase."""
def
connectionCompleted
(
self
,
conn
,
new
):
pt
=
self
.
app
.
pt
conn
.
send
(
Packets
.
SendPartitionTable
(
pt
.
getID
(),
pt
.
getRowList
()))
"""Common handler class for storage nodes."""
def
connectionLost
(
self
,
conn
,
new_state
):
app
=
self
.
app
...
...
neo/master/handlers/administration.py
View file @
8ba42463
...
...
@@ -15,14 +15,16 @@
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import
random
from
functools
import
wraps
from
.
import
MasterHandler
from
..app
import
monotonic_time
,
StateChangedException
from
neo.lib
import
logging
from
neo.lib.exception
import
StoppedOperation
from
neo.lib.handler
import
AnswerDenied
from
neo.lib.pt
import
PartitionTableException
from
neo.lib.protocol
import
ClusterStates
,
Errors
,
\
NodeStates
,
NodeTypes
,
Packets
,
ProtocolError
,
uuid_str
NodeStates
,
NodeTypes
,
Packets
,
uuid_str
from
neo.lib.util
import
dump
CLUSTER_STATE_WORKFLOW
=
{
...
...
@@ -38,9 +40,25 @@ NODE_STATE_WORKFLOW = {
NodeTypes
.
STORAGE
:
(
NodeStates
.
DOWN
,
NodeStates
.
UNKNOWN
),
}
def
check_state
(
*
states
):
def
decorator
(
wrapped
):
def
wrapper
(
self
,
*
args
):
state
=
self
.
app
.
getClusterState
()
if
state
not
in
states
:
raise
AnswerDenied
(
'%s RPC can not be used in %s state'
%
(
wrapped
.
__name__
,
state
))
wrapped
(
self
,
*
args
)
return
wraps
(
wrapped
)(
wrapper
)
return
decorator
class
AdministrationHandler
(
MasterHandler
):
"""This class deals with messages from the admin node only"""
def
handlerSwitched
(
self
,
conn
,
new
):
assert
new
super
(
AdministrationHandler
,
self
).
handlerSwitched
(
conn
,
new
)
def
connectionLost
(
self
,
conn
,
new_state
):
node
=
self
.
app
.
nm
.
getByUUID
(
conn
.
getUUID
())
if
node
is
not
None
:
...
...
@@ -58,30 +76,28 @@ class AdministrationHandler(MasterHandler):
# check request
try
:
if
app
.
cluster_state
not
in
CLUSTER_STATE_WORKFLOW
[
state
]:
raise
ProtocolError
(
'Can not switch to this state'
)
raise
AnswerDenied
(
'Can not switch to this state'
)
except
KeyError
:
if
state
!=
ClusterStates
.
STOPPING
:
raise
ProtocolError
(
'Invalid state requested'
)
raise
AnswerDenied
(
'Invalid state requested'
)
# change state
if
state
==
ClusterStates
.
VERIFYING
:
storage_list
=
app
.
nm
.
getStorageList
(
only_identified
=
True
)
if
not
storage_list
:
raise
ProtocolError
(
'Cannot exit recovery without any '
'storage node'
)
raise
AnswerDenied
(
'
Cannot exit recovery without any
storage node'
)
for
node
in
storage_list
:
assert
node
.
isPending
(),
node
if
node
.
getConnection
().
isPending
():
# XXX: It's wrong to use ProtocolError here. We must reply
# less aggressively because the admin has no way to
# know that there's still pending activity.
raise
ProtocolError
(
'Cannot exit recovery now: node %r is '
'entering cluster'
%
(
node
,
))
raise
AnswerDenied
(
'Cannot exit recovery now: node %r is entering cluster'
%
node
,)
app
.
_startup_allowed
=
True
state
=
app
.
cluster_state
elif
state
==
ClusterStates
.
STARTING_BACKUP
:
if
app
.
tm
.
hasPending
()
or
app
.
nm
.
getClientList
(
True
):
raise
ProtocolError
(
"Can not switch to %s state with pending"
raise
AnswerDenied
(
"Can not switch to %s state with pending"
" transactions or connected clients"
%
state
)
conn
.
answer
(
Errors
.
Ack
(
'Cluster state changed'
))
...
...
@@ -93,21 +109,24 @@ class AdministrationHandler(MasterHandler):
app
=
self
.
app
node
=
app
.
nm
.
getByUUID
(
uuid
)
if
node
is
None
:
raise
ProtocolError
(
'unknown node'
)
raise
AnswerDenied
(
'unknown node'
)
if
state
not
in
NODE_STATE_WORKFLOW
.
get
(
node
.
getType
(),
()):
raise
ProtocolError
(
'can not switch node to this state'
)
raise
AnswerDenied
(
'can not switch node to %s state'
%
state
)
if
uuid
==
app
.
uuid
:
raise
ProtocolError
(
'can not kill primary master node'
)
raise
AnswerDenied
(
'can not kill primary master node'
)
state_changed
=
state
!=
node
.
getState
()
message
=
(
'state changed'
if
state_changed
else
'node already in %s state'
%
state
)
if
node
.
isStorage
():
keep
=
state
==
NodeStates
.
DOWN
if
node
.
isRunning
()
and
not
keep
:
raise
AnswerDenied
(
"a running node must be stopped before removal"
)
try
:
cell_list
=
app
.
pt
.
dropNodeList
([
node
],
keep
)
except
PartitionTableException
,
e
:
raise
ProtocolError
(
str
(
e
))
raise
AnswerDenied
(
str
(
e
))
node
.
setState
(
state
)
if
node
.
isConnected
():
# notify itself so it can shutdown
...
...
@@ -134,16 +153,17 @@ class AdministrationHandler(MasterHandler):
monotonic_time
(),
[
node
.
asTuple
()]))
app
.
broadcastNodesInformation
([
node
])
# XXX: Would it be safe to allow more states ?
__change_pt_rpc
=
check_state
(
ClusterStates
.
RUNNING
,
ClusterStates
.
STARTING_BACKUP
,
ClusterStates
.
BACKINGUP
)
@
__change_pt_rpc
def
addPendingNodes
(
self
,
conn
,
uuid_list
):
uuids
=
', '
.
join
(
map
(
uuid_str
,
uuid_list
))
logging
.
debug
(
'Add nodes %s'
,
uuids
)
app
=
self
.
app
state
=
app
.
getClusterState
()
# XXX: Would it be safe to allow more states ?
if
state
not
in
(
ClusterStates
.
RUNNING
,
ClusterStates
.
STARTING_BACKUP
,
ClusterStates
.
BACKINGUP
):
raise
ProtocolError
(
'Can not add nodes in %s state'
%
state
)
# take all pending nodes
node_list
=
list
(
app
.
pt
.
addNodeList
(
node
for
node
in
app
.
nm
.
getStorageList
()
...
...
@@ -165,31 +185,50 @@ class AdministrationHandler(MasterHandler):
for
uuid
in
uuid_list
:
node
=
getByUUID
(
uuid
)
if
node
is
None
or
not
(
node
.
isStorage
()
and
node
.
isIdentified
()):
raise
ProtocolError
(
"invalid storage node %s"
%
uuid_str
(
uuid
))
raise
AnswerDenied
(
"invalid storage node %s"
%
uuid_str
(
uuid
))
node_list
.
append
(
node
)
repair
=
Packets
.
NotifyRepair
(
*
args
)
for
node
in
node_list
:
node
.
send
(
repair
)
conn
.
answer
(
Errors
.
Ack
(
''
))
def
tweakPartitionTable
(
self
,
conn
,
uuid_list
):
app
=
self
.
app
state
=
app
.
getClusterState
()
# XXX: Would it be safe to allow more states ?
if
state
not
in
(
ClusterStates
.
RUNNING
,
ClusterStates
.
STARTING_BACKUP
,
ClusterStates
.
BACKINGUP
):
raise
ProtocolError
(
'Can not tweak partition table in %s state'
%
state
)
app
.
broadcastPartitionChanges
(
app
.
pt
.
tweak
([
node
for
node
in
app
.
nm
.
getStorageList
()
if
node
.
getUUID
()
in
uuid_list
or
not
node
.
isRunning
()]))
@
__change_pt_rpc
def
setNumReplicas
(
self
,
conn
,
num_replicas
):
self
.
app
.
broadcastPartitionChanges
((),
num_replicas
)
conn
.
answer
(
Errors
.
Ack
(
''
))
def
truncate
(
self
,
conn
,
tid
):
@
__change_pt_rpc
def
tweakPartitionTable
(
self
,
conn
,
dry_run
,
uuid_list
):
app
=
self
.
app
if
app
.
cluster_state
!=
ClusterStates
.
RUNNING
:
raise
ProtocolError
(
'Can not truncate in this state'
)
drop_list
=
[]
for
node
in
app
.
nm
.
getStorageList
():
if
node
.
getUUID
()
in
uuid_list
or
node
.
isPending
():
drop_list
.
append
(
node
)
elif
not
node
.
isRunning
():
drop_list
.
append
(
node
)
raise
AnswerDenied
(
'tweak: down nodes must be listed explicitly'
)
if
dry_run
:
pt
=
object
.
__new__
(
app
.
pt
.
__class__
)
new_nodes
=
pt
.
load
(
app
.
pt
.
getID
(),
app
.
pt
.
getReplicas
(),
app
.
pt
.
getRowList
(),
app
.
nm
)
assert
not
new_nodes
pt
.
addNodeList
(
node
for
node
,
count
in
app
.
pt
.
count_dict
.
iteritems
()
if
not
count
)
else
:
pt
=
app
.
pt
try
:
changed_list
=
pt
.
tweak
(
drop_list
)
except
PartitionTableException
,
e
:
raise
AnswerDenied
(
str
(
e
))
if
not
dry_run
:
app
.
broadcastPartitionChanges
(
changed_list
)
conn
.
answer
(
Packets
.
AnswerTweakPartitionTable
(
bool
(
changed_list
),
pt
.
getRowList
()))
@
check_state
(
ClusterStates
.
RUNNING
)
def
truncate
(
self
,
conn
,
tid
):
conn
.
answer
(
Errors
.
Ack
(
''
))
raise
StoppedOperation
(
tid
)
...
...
@@ -237,3 +276,5 @@ class AdministrationHandler(MasterHandler):
node
.
send
(
Packets
.
CheckPartition
(
offset
,
source
,
min_tid
,
max_tid
))
conn
.
answer
(
Errors
.
Ack
(
''
))
del
__change_pt_rpc
neo/master/handlers/backup.py
View file @
8ba42463
...
...
@@ -17,6 +17,7 @@
from
neo.lib.exception
import
PrimaryFailure
from
neo.lib.handler
import
EventHandler
from
neo.lib.protocol
import
ZERO_TID
from
neo.lib.pt
import
PartitionTable
class
BackupHandler
(
EventHandler
):
"""Handler dedicated to upstream master during BACKINGUP state"""
...
...
@@ -25,12 +26,15 @@ class BackupHandler(EventHandler):
if
self
.
app
.
app
.
listening_conn
:
# if running
raise
PrimaryFailure
(
'connection lost'
)
def
answerPartitionTable
(
self
,
conn
,
ptid
,
row_list
):
self
.
app
.
pt
.
load
(
ptid
,
row_list
,
self
.
app
.
nm
)
def
sendPartitionTable
(
self
,
conn
,
ptid
,
num_replicas
,
row_list
):
app
=
self
.
app
pt
=
app
.
pt
=
object
.
__new__
(
PartitionTable
)
pt
.
load
(
ptid
,
num_replicas
,
row_list
,
self
.
app
.
nm
)
if
pt
.
getPartitions
()
!=
app
.
app
.
pt
.
getPartitions
():
raise
RuntimeError
(
"inconsistent number of partitions"
)
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
cell_list
):
if
self
.
app
.
pt
.
filled
():
self
.
app
.
pt
.
update
(
ptid
,
cell_list
,
self
.
app
.
nm
)
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
num_replicas
,
cell_list
):
self
.
app
.
pt
.
update
(
ptid
,
num_replicas
,
cell_list
,
self
.
app
.
nm
)
def
answerLastTransaction
(
self
,
conn
,
tid
):
app
=
self
.
app
...
...
neo/master/handlers/client.py
View file @
8ba42463
...
...
@@ -22,6 +22,10 @@ from . import MasterHandler
class
ClientServiceHandler
(
MasterHandler
):
""" Handler dedicated to client during service state """
def
handlerSwitched
(
self
,
conn
,
new
):
assert
new
super
(
ClientServiceHandler
,
self
).
handlerSwitched
(
conn
,
new
)
def
_connectionLost
(
self
,
conn
):
# cancel its transactions and forgot the node
app
=
self
.
app
...
...
neo/master/handlers/identification.py
View file @
8ba42463
...
...
@@ -17,14 +17,14 @@
from
neo.lib
import
logging
from
neo.lib.exception
import
PrimaryElected
from
neo.lib.handler
import
EventHandler
from
neo.lib.protocol
import
C
lusterStates
,
NodeStates
,
NodeTyp
es
,
\
NotReadyError
,
Packets
,
ProtocolError
,
uuid_str
from
neo.lib.protocol
import
C
ellStates
,
ClusterStates
,
NodeStat
es
,
\
No
deTypes
,
No
tReadyError
,
Packets
,
ProtocolError
,
uuid_str
from
..app
import
monotonic_time
class
IdentificationHandler
(
EventHandler
):
def
requestIdentification
(
self
,
conn
,
node_type
,
uuid
,
address
,
name
,
devpath
,
id_timestamp
):
address
,
name
,
id_timestamp
,
devpath
,
new_nid
):
app
=
self
.
app
self
.
checkClusterName
(
name
)
if
address
==
app
.
server
:
...
...
@@ -77,6 +77,16 @@ class IdentificationHandler(EventHandler):
manager
=
app
state
,
handler
=
manager
.
identifyStorageNode
(
uuid
is
not
None
and
node
is
not
None
)
if
not
address
:
if
app
.
cluster_state
==
ClusterStates
.
RECOVERING
:
raise
NotReadyError
if
uuid
or
not
new_nid
:
raise
ProtocolError
state
=
NodeStates
.
DOWN
# We'll let the storage node close the connection. If we
# aborted it at the end of the method, BootstrapManager
# (which is used by storage nodes) could see the closure
# and try to reconnect to a master.
human_readable_node_type
=
' storage (%s) '
%
(
state
,
)
elif
node_type
==
NodeTypes
.
MASTER
:
if
app
.
election
:
...
...
@@ -105,24 +115,27 @@ class IdentificationHandler(EventHandler):
node
.
devpath
=
tuple
(
devpath
)
node
.
id_timestamp
=
monotonic_time
()
node
.
setState
(
state
)
app
.
broadcastNodesInformation
([
node
])
if
new_nid
:
changed_list
=
[]
for
offset
in
new_nid
:
changed_list
.
append
((
offset
,
uuid
,
CellStates
.
OUT_OF_DATE
))
app
.
pt
.
_setCell
(
offset
,
node
,
CellStates
.
OUT_OF_DATE
)
app
.
broadcastPartitionChanges
(
changed_list
)
conn
.
setHandler
(
handler
)
node
.
setConnection
(
conn
,
not
node
.
isIdentified
())
app
.
broadcastNodesInformation
([
node
],
node
)
conn
.
answer
(
Packets
.
AcceptIdentification
(
NodeTypes
.
MASTER
,
app
.
uuid
,
app
.
pt
.
getPartitions
(),
app
.
pt
.
getReplicas
(),
uuid
))
handler
.
_notifyNodeInformation
(
conn
)
handler
.
connectionCompleted
(
conn
,
True
)
handler
.
handlerSwitched
(
conn
,
True
)
class
SecondaryIdentificationHandler
(
EventHandler
):
def
requestIdentification
(
self
,
conn
,
node_type
,
uuid
,
address
,
name
,
devpath
,
id_timestamp
):
address
,
name
,
id_timestamp
,
devpath
,
new_nid
):
app
=
self
.
app
self
.
checkClusterName
(
name
)
if
address
==
app
.
server
:
...
...
neo/master/handlers/master.py
View file @
8ba42463
...
...
@@ -23,6 +23,9 @@ from neo.lib.protocol import ClusterStates, NodeStates, NodeTypes, Packets
class
SecondaryHandler
(
MasterHandler
):
"""Handler used by primary to handle secondary masters"""
def
handlerSwitched
(
self
,
conn
,
new
):
pass
def
_connectionLost
(
self
,
conn
):
app
=
self
.
app
node
=
app
.
nm
.
getByUUID
(
conn
.
getUUID
())
...
...
@@ -30,21 +33,20 @@ class SecondaryHandler(MasterHandler):
app
.
broadcastNodesInformation
([
node
])
class
ElectionHandler
(
Master
Handler
):
class
ElectionHandler
(
Secondary
Handler
):
"""Handler used by primary to handle secondary masters during election"""
def
connectionCompleted
(
self
,
conn
,
new
=
None
):
if
new
is
None
:
super
(
ElectionHandler
,
self
).
connectionCompleted
(
conn
)
app
=
self
.
app
conn
.
ask
(
Packets
.
RequestIdentification
(
NodeTypes
.
MASTER
,
app
.
uuid
,
app
.
server
,
app
.
name
,
(),
app
.
election
))
def
connectionCompleted
(
self
,
conn
):
super
(
ElectionHandler
,
self
).
connectionCompleted
(
conn
)
app
=
self
.
app
conn
.
ask
(
Packets
.
RequestIdentification
(
NodeTypes
.
MASTER
,
app
.
uuid
,
app
.
server
,
app
.
name
,
app
.
election
,
(),
()))
def
connectionFailed
(
self
,
conn
):
super
(
ElectionHandler
,
self
).
connectionFailed
(
conn
)
self
.
connectionLost
(
conn
)
def
_acceptIdentification
(
self
,
node
,
*
args
):
def
_acceptIdentification
(
self
,
node
):
raise
PrimaryElected
(
node
)
def
_connectionLost
(
self
,
*
args
):
...
...
@@ -66,7 +68,7 @@ class ElectionHandler(MasterHandler):
class
PrimaryHandler
(
ElectionHandler
):
"""Handler used by secondaries to handle primary master"""
def
_acceptIdentification
(
self
,
node
,
num_partitions
,
num_replicas
):
def
_acceptIdentification
(
self
,
node
):
assert
self
.
app
.
primary_master
is
node
,
(
self
.
app
.
primary_master
,
node
)
def
_connectionLost
(
self
,
conn
):
...
...
neo/master/handlers/storage.py
View file @
8ba42463
...
...
@@ -26,10 +26,10 @@ from . import BaseServiceHandler
class
StorageServiceHandler
(
BaseServiceHandler
):
""" Handler dedicated to storages during service state """
def
connectionComplet
ed
(
self
,
conn
,
new
):
def
handlerSwitch
ed
(
self
,
conn
,
new
):
app
=
self
.
app
if
new
:
super
(
StorageServiceHandler
,
self
).
connectionComplet
ed
(
conn
,
new
)
super
(
StorageServiceHandler
,
self
).
handlerSwitch
ed
(
conn
,
new
)
node
=
app
.
nm
.
getByUUID
(
conn
.
getUUID
())
if
node
.
isRunning
():
# node may be PENDING
app
.
startStorage
(
node
)
...
...
neo/master/pt.py
View file @
8ba42463
...
...
@@ -56,6 +56,10 @@ class PartitionTable(neo.lib.pt.PartitionTable):
self
.
_id
+=
1
return
self
.
_id
def
setReplicas
(
self
,
num_replicas
):
assert
num_replicas
>=
0
,
num_replicas
self
.
nr
=
num_replicas
def
make
(
self
,
node_list
):
"""Make a new partition table from scratch."""
assert
self
.
_id
is
None
and
node_list
,
(
self
.
_id
,
node_list
)
...
...
@@ -108,26 +112,19 @@ class PartitionTable(neo.lib.pt.PartitionTable):
self
.
num_filled_rows
=
len
(
filter
(
None
,
self
.
partition_list
))
return
change_list
def
load
(
self
,
ptid
,
row_list
,
nm
):
def
load
(
self
,
ptid
,
num_replicas
,
row_list
,
nm
):
"""
Load a partition table from a storage node during the recovery.
Return the new storage nodes registered
"""
# check offsets
for
offset
,
_row
in
row_list
:
if
offset
>=
self
.
getPartitions
():
raise
IndexError
,
offset
# store the partition table
self
.
clear
()
self
.
_id
=
ptid
new_nodes
=
[]
for
offset
,
row
in
row_list
:
for
uuid
,
state
in
row
:
node
=
nm
.
getByUUID
(
uuid
)
if
node
is
None
:
node
=
nm
.
createStorage
(
uuid
=
uuid
)
new_nodes
.
append
(
node
.
asTuple
())
self
.
_setCell
(
offset
,
node
,
state
)
def
getByUUID
(
nid
)
:
node
=
nm
.
getByUUID
(
nid
)
if
node
is
None
:
node
=
nm
.
createStorage
(
uuid
=
nid
)
new_nodes
.
append
(
node
.
asTuple
()
)
return
node
self
.
_load
(
ptid
,
num_replicas
,
row_list
,
getByUUID
)
return
new_nodes
def
setUpToDate
(
self
,
node
,
offset
):
...
...
@@ -166,15 +163,6 @@ class PartitionTable(neo.lib.pt.PartitionTable):
return
cell_list
def
addNodeList
(
self
,
node_list
):
"""Add nodes"""
added_list
=
[]
for
node
in
node_list
:
if
node
not
in
self
.
count_dict
:
self
.
count_dict
[
node
]
=
0
added_list
.
append
(
node
)
return
added_list
def
tweak
(
self
,
drop_list
=
()):
"""Optimize partition table
...
...
@@ -183,7 +171,8 @@ class PartitionTable(neo.lib.pt.PartitionTable):
few readable cells, some cells are instead marked as FEEDING. This is
a preliminary step to drop these nodes, otherwise the partition table
could become non-operational.
- Other nodes must have the same number of cells, off by 1.
In fact, the code touching these cells is disabled (see NOTE below).
- Other nodes must have the same number of non-feeding cells, off by 1.
- When a transaction creates new objects (oids are roughly allocated
sequentially), we expect better performance by maximizing the number
of involved nodes (i.e. parallelizing writes).
...
...
@@ -232,6 +221,8 @@ class PartitionTable(neo.lib.pt.PartitionTable):
# Collect some data in a usable form for the rest of the method.
node_list
=
{
node
:
{}
for
node
in
self
.
count_dict
if
node
not
in
drop_list
}
if
not
node_list
:
raise
neo
.
lib
.
pt
.
PartitionTableException
(
"Can't remove all nodes."
)
drop_list
=
defaultdict
(
list
)
for
offset
,
row
in
enumerate
(
self
.
partition_list
):
for
cell
in
row
:
...
...
@@ -420,6 +411,22 @@ class PartitionTable(neo.lib.pt.PartitionTable):
outdated_list
[
offset
]
-=
1
for
offset
,
cell
in
cell_dict
.
iteritems
():
discard_list
[
offset
].
append
(
cell
)
# NOTE: The following line disables the next 2 lines, which actually
# causes cells in drop_list to be discarded, now or later;
# drop_list could be renamed into ignore_list.
# 1. Deleting data partition per partition is a lot of work, so
# why ask nodes in drop_list to do that when the goal is
# simply to trash the whole underlying database?
# 2. By excluding nodes from a tweak, it becomes possible to have
# parts of the partition table that are tweaked differently.
# This may require to temporarily change the number of
# replicas for the part being tweaked. In the future, this
# number may be specified in the 'tweak' command, to avoid
# race conditions with setUpToDate().
# Overall, a common use case is when importing a ZODB to NEO,
# to keep the initial importing node up until the database is
# split and replicated to the final nodes.
drop_list
=
{}
for
offset
,
drop_list
in
drop_list
.
iteritems
():
discard_list
[
offset
]
+=
drop_list
# We have sorted cells to discard in order to first deallocate nodes
...
...
neo/master/recovery.py
View file @
8ba42463
...
...
@@ -28,7 +28,7 @@ class RecoveryManager(MasterHandler):
def
__init__
(
self
,
app
):
# The target node's uuid to request next.
self
.
target_ptid
=
None
self
.
target_ptid
=
0
self
.
ask_pt
=
[]
self
.
backup_tid_dict
=
{}
self
.
truncate_dict
=
{}
...
...
@@ -52,9 +52,8 @@ class RecoveryManager(MasterHandler):
"""
logging
.
info
(
'begin the recovery of the status'
)
app
=
self
.
app
pt
=
app
.
pt
pt
=
app
.
pt
=
app
.
newPartitionTable
()
app
.
changeClusterState
(
ClusterStates
.
RECOVERING
)
pt
.
clear
()
self
.
try_secondary
=
True
...
...
@@ -113,7 +112,7 @@ class RecoveryManager(MasterHandler):
for
node
in
node_list
:
conn
=
node
.
getConnection
()
conn
.
send
(
truncate
)
self
.
connectionComplet
ed
(
conn
,
False
)
self
.
handlerSwitch
ed
(
conn
,
False
)
continue
node_list
=
pt
.
getConnectedNodeList
()
break
...
...
@@ -140,12 +139,12 @@ class RecoveryManager(MasterHandler):
logging
.
info
(
'creating a new partition table'
)
pt
.
make
(
node_list
)
self
.
_notifyAdmins
(
Packets
.
SendPartitionTable
(
pt
.
getID
(),
pt
.
getRowList
()))
pt
.
getID
(),
pt
.
getR
eplicas
(),
pt
.
getR
owList
()))
else
:
cell_list
=
pt
.
outdate
()
if
cell_list
:
self
.
_notifyAdmins
(
Packets
.
NotifyPartitionChanges
(
pt
.
setNextID
(),
cell_list
))
pt
.
setNextID
(),
pt
.
getReplicas
(),
cell_list
))
if
app
.
backup_tid
:
pt
.
setBackupTidDict
(
self
.
backup_tid_dict
)
app
.
backup_tid
=
pt
.
getBackupTid
()
...
...
@@ -175,16 +174,16 @@ class RecoveryManager(MasterHandler):
if
node
is
None
or
node
.
getState
()
==
new_state
:
return
node
.
setState
(
new_state
)
# broadcast to all so that admin nodes gets informed
self
.
app
.
broadcastNodesInformation
([
node
])
def
connectionComplet
ed
(
self
,
conn
,
new
):
def
handlerSwitch
ed
(
self
,
conn
,
new
):
# ask the last IDs to perform the recovery
conn
.
ask
(
Packets
.
AskRecovery
())
def
answerRecovery
(
self
,
conn
,
ptid
,
backup_tid
,
truncate_tid
):
uuid
=
conn
.
getUUID
()
if
self
.
target_ptid
<=
ptid
:
# ptid is None if the node has an empty partition table.
if
ptid
and
self
.
target_ptid
<=
ptid
:
# Maybe a newer partition table.
if
self
.
target_ptid
==
ptid
and
self
.
ask_pt
:
# Another node is already asked.
...
...
@@ -197,17 +196,14 @@ class RecoveryManager(MasterHandler):
self
.
backup_tid_dict
[
uuid
]
=
backup_tid
self
.
truncate_dict
[
uuid
]
=
truncate_tid
def
answerPartitionTable
(
self
,
conn
,
ptid
,
row_list
):
def
answerPartitionTable
(
self
,
conn
,
ptid
,
num_replicas
,
row_list
):
# If this is not from a target node, ignore it.
if
ptid
==
self
.
target_ptid
:
app
=
self
.
app
try
:
new_nodes
=
app
.
pt
.
load
(
ptid
,
row_list
,
app
.
nm
)
except
IndexError
:
raise
ProtocolError
(
'Invalid offset'
)
new_nodes
=
app
.
pt
.
load
(
ptid
,
num_replicas
,
row_list
,
app
.
nm
)
self
.
_notifyAdmins
(
Packets
.
NotifyNodeInformation
(
monotonic_time
(),
new_nodes
),
Packets
.
SendPartitionTable
(
ptid
,
row_list
))
Packets
.
SendPartitionTable
(
ptid
,
num_replicas
,
row_list
))
self
.
ask_pt
=
()
uuid
=
conn
.
getUUID
()
app
.
backup_tid
=
self
.
backup_tid_dict
[
uuid
]
...
...
neo/neoctl/app.py
View file @
8ba42463
...
...
@@ -16,9 +16,11 @@
import
sys
from
.neoctl
import
NeoCTL
,
NotReadyException
from
neo.lib.node
import
NodeManager
from
neo.lib.pt
import
PartitionTable
from
neo.lib.util
import
p64
,
u64
,
tidFromTime
,
timeStringFromTID
from
neo.lib.protocol
import
uuid_str
,
formatNodeList
,
\
ClusterStates
,
NodeTypes
,
UUID_NAMESPACES
,
ZERO_TID
ClusterStates
,
Node
States
,
Node
Types
,
UUID_NAMESPACES
,
ZERO_TID
action_dict
=
{
'print'
:
{
...
...
@@ -30,6 +32,7 @@ action_dict = {
},
'set'
:
{
'cluster'
:
'setClusterState'
,
'replicas'
:
'setNumReplicas'
,
},
'check'
:
'checkReplicas'
,
'start'
:
'startCluster'
,
...
...
@@ -46,6 +49,11 @@ uuid_int = (lambda ns: lambda uuid:
(
ns
[
uuid
[
0
]]
<<
24
)
+
int
(
uuid
[
1
:])
)({
str
(
k
)[
0
]:
v
for
k
,
v
in
UUID_NAMESPACES
.
iteritems
()})
class
dummy_app
:
id_timestamp
=
uuid
=
0
class
TerminalNeoCTL
(
object
):
def
__init__
(
self
,
*
args
,
**
kw
):
self
.
neoctl
=
NeoCTL
(
*
args
,
**
kw
)
...
...
@@ -67,6 +75,15 @@ class TerminalNeoCTL(object):
asNode
=
staticmethod
(
uuid_int
)
def
formatPartitionTable
(
self
,
row_list
):
nm
=
NodeManager
()
nm
.
update
(
dummy_app
,
1
,
self
.
neoctl
.
getNodeList
(
node_type
=
NodeTypes
.
STORAGE
))
pt
=
object
.
__new__
(
PartitionTable
)
pt
.
_load
(
None
,
None
,
row_list
,
nm
.
getByUUID
)
pt
.
addNodeList
(
nm
.
getByStateList
(
NodeStates
.
RUNNING
))
return
'
\
n
'
.
join
(
line
[
4
:]
for
line
in
pt
.
_format
())
def
formatRowList
(
self
,
row_list
):
return
'
\
n
'
.
join
(
'%03d |%s'
%
(
offset
,
''
.
join
(
' %s - %s |'
%
(
uuid_str
(
uuid
),
state
)
...
...
@@ -105,10 +122,12 @@ class TerminalNeoCTL(object):
max_offset
=
int
(
max_offset
)
if
node
is
not
None
:
node
=
self
.
asNode
(
node
)
ptid
,
row_list
=
self
.
neoctl
.
getPartitionRowList
(
ptid
,
num_replicas
,
row_list
=
self
.
neoctl
.
getPartitionRowList
(
min_offset
=
min_offset
,
max_offset
=
max_offset
,
node
=
node
)
# TODO: return ptid
return
self
.
formatRowList
(
row_list
)
return
'# ptid: %s, replicas: %s
\
n
%s'
%
(
ptid
,
num_replicas
,
self
.
formatRowList
(
enumerate
(
row_list
,
min_offset
))
if
min_offset
or
max_offset
else
self
.
formatPartitionTable
(
row_list
))
def
getNodeList
(
self
,
params
):
"""
...
...
@@ -140,6 +159,18 @@ class TerminalNeoCTL(object):
assert
len
(
params
)
==
1
return
self
.
neoctl
.
setClusterState
(
self
.
asClusterState
(
params
[
0
]))
def
setNumReplicas
(
self
,
params
):
"""
Set number of replicas.
Parameters: nr
nr: positive number (0 means no redundancy)
"""
assert
len
(
params
)
==
1
nr
=
int
(
params
[
0
])
if
nr
<
0
:
sys
.
exit
(
'invalid number of replicas'
)
return
self
.
neoctl
.
setNumReplicas
(
nr
)
def
startCluster
(
self
,
params
):
"""
Starts cluster operation after a startup.
...
...
@@ -167,10 +198,18 @@ class TerminalNeoCTL(object):
def
tweakPartitionTable
(
self
,
params
):
"""
Optimize partition table.
No partition will be assigned to specified storage nodes.
Parameters: [node [...]]
No change is done to the specified/down storage nodes and they don't
count as replicas. The purpose of listing nodes is usually to drop
them once the data is replicated to other nodes.
Parameters: [-n] [node [...]]
-n: dry run
"""
return
self
.
neoctl
.
tweakPartitionTable
(
map
(
self
.
asNode
,
params
))
dry_run
=
params
[
0
]
==
'-n'
changed
,
row_list
=
self
.
neoctl
.
tweakPartitionTable
(
map
(
self
.
asNode
,
params
[
dry_run
:]),
dry_run
)
if
changed
:
return
self
.
formatPartitionTable
(
row_list
)
return
'No change done.'
def
killNode
(
self
,
params
):
"""
...
...
@@ -203,7 +242,7 @@ class TerminalNeoCTL(object):
node: if "all", ask all connected storage nodes to repair,
otherwise, only the given list of storage nodes.
"""
dry_run
=
"01"
.
index
(
params
.
pop
(
0
))
dry_run
=
bool
(
"01"
.
index
(
params
.
pop
(
0
)
))
return
self
.
neoctl
.
repair
(
self
.
_getStorageList
(
params
),
dry_run
)
def
truncate
(
self
,
params
):
...
...
neo/neoctl/handler.py
View file @
8ba42463
...
...
@@ -14,6 +14,7 @@
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import
sys
from
neo.lib.handler
import
EventHandler
from
neo.lib.protocol
import
ErrorCodes
,
Packets
...
...
@@ -44,8 +45,8 @@ class CommandEventHandler(EventHandler):
def
ack
(
self
,
conn
,
msg
):
self
.
__respond
((
Packets
.
Error
,
ErrorCodes
.
ACK
,
msg
))
def
protocolError
(
self
,
conn
,
msg
):
s
elf
.
__respond
((
Packets
.
Error
,
ErrorCodes
.
PROTOCOL_ERROR
,
msg
)
)
def
denied
(
self
,
conn
,
msg
):
s
ys
.
exit
(
msg
)
def
notReady
(
self
,
conn
,
msg
):
self
.
__respond
((
Packets
.
Error
,
ErrorCodes
.
NOT_READY
,
msg
))
...
...
@@ -62,3 +63,4 @@ class CommandEventHandler(EventHandler):
answerLastIDs
=
__answer
(
Packets
.
AnswerLastIDs
)
answerLastTransaction
=
__answer
(
Packets
.
AnswerLastTransaction
)
answerRecovery
=
__answer
(
Packets
.
AnswerRecovery
)
answerTweakPartitionTable
=
__answer
(
Packets
.
AnswerTweakPartitionTable
)
neo/neoctl/neoctl.py
View file @
8ba42463
...
...
@@ -91,8 +91,14 @@ class NeoCTL(BaseApplication):
raise
RuntimeError
(
response
)
return
response
[
2
]
def
tweakPartitionTable
(
self
,
uuid_list
=
()):
response
=
self
.
__ask
(
Packets
.
TweakPartitionTable
(
uuid_list
))
def
tweakPartitionTable
(
self
,
uuid_list
=
(),
dry_run
=
False
):
response
=
self
.
__ask
(
Packets
.
TweakPartitionTable
(
dry_run
,
uuid_list
))
if
response
[
0
]
!=
Packets
.
AnswerTweakPartitionTable
:
raise
RuntimeError
(
response
)
return
response
[
1
:]
def
setNumReplicas
(
self
,
nr
):
response
=
self
.
__ask
(
Packets
.
SetNumReplicas
(
nr
))
if
response
[
0
]
!=
Packets
.
Error
or
response
[
1
]
!=
ErrorCodes
.
ACK
:
raise
RuntimeError
(
response
)
return
response
[
2
]
...
...
@@ -163,7 +169,7 @@ class NeoCTL(BaseApplication):
response
=
self
.
__ask
(
packet
)
if
response
[
0
]
!=
Packets
.
AnswerPartitionList
:
raise
RuntimeError
(
response
)
return
response
[
1
:
3
]
# ptid, row_list
return
response
[
1
:
]
def
startCluster
(
self
):
"""
...
...
neo/scripts/neolog.py
View file @
8ba42463
...
...
@@ -157,27 +157,49 @@ class Log(object):
for
x
in
'uuid_str'
,
'Packets'
,
'PacketMalformedError'
:
setattr
(
self
,
x
,
g
[
x
])
x
=
{}
try
:
Unpacker
=
g
[
'Unpacker'
]
except
KeyError
:
unpackb
=
None
else
:
from
msgpack
import
ExtraData
,
UnpackException
def
unpackb
(
data
):
u
=
Unpacker
()
u
.
feed
(
data
)
data
=
u
.
unpack
()
if
u
.
read_bytes
(
1
):
raise
ExtraData
return
data
self
.
PacketMalformedError
=
UnpackException
self
.
unpackb
=
unpackb
if
self
.
_decode
>
1
:
PStruct
=
g
[
'PStruct'
]
PBoolean
=
g
[
'PBoolean'
]
def
hasData
(
item
):
items
=
item
.
_items
for
i
,
item
in
enumerate
(
items
):
if
isinstance
(
item
,
PStruct
):
j
=
hasData
(
item
)
if
j
:
return
(
i
,)
+
j
elif
(
isinstance
(
item
,
PBoolean
)
and
item
.
_name
==
'compression'
and
i
+
2
<
len
(
items
)
and
items
[
i
+
2
].
_name
==
'data'
):
return
i
,
for
p
in
self
.
Packets
.
itervalues
():
if
p
.
_fmt
is
not
None
:
path
=
hasData
(
p
.
_fmt
)
if
path
:
assert
not
hasattr
(
p
,
'_neolog'
),
p
x
[
p
.
_code
]
=
path
try
:
PStruct
=
g
[
'PStruct'
]
except
KeyError
:
for
p
in
self
.
Packets
.
itervalues
():
data_path
=
getattr
(
p
,
'data_path'
,
(
None
,))
if
p
.
_code
>>
15
==
data_path
[
0
]:
x
[
p
.
_code
]
=
data_path
[
1
:]
else
:
PBoolean
=
g
[
'PBoolean'
]
def
hasData
(
item
):
items
=
item
.
_items
for
i
,
item
in
enumerate
(
items
):
if
isinstance
(
item
,
PStruct
):
j
=
hasData
(
item
)
if
j
:
return
(
i
,)
+
j
elif
(
isinstance
(
item
,
PBoolean
)
and
item
.
_name
==
'compression'
and
i
+
2
<
len
(
items
)
and
items
[
i
+
2
].
_name
==
'data'
):
return
i
,
for
p
in
self
.
Packets
.
itervalues
():
if
p
.
_fmt
is
not
None
:
path
=
hasData
(
p
.
_fmt
)
if
path
:
assert
not
hasattr
(
p
,
'_neolog'
),
p
x
[
p
.
_code
]
=
path
self
.
_getDataPath
=
x
.
get
try
:
...
...
@@ -215,11 +237,13 @@ class Log(object):
if
body
is
not
None
:
log
=
getattr
(
p
,
'_neolog'
,
None
)
if
log
or
self
.
_decode
:
p
=
p
()
p
.
_id
=
msg_id
p
.
_body
=
body
try
:
args
=
p
.
decode
()
if
self
.
unpackb
:
args
=
self
.
unpackb
(
body
)
else
:
p
=
p
()
p
.
_body
=
body
args
=
p
.
decode
()
except
self
.
PacketMalformedError
:
msg
.
append
(
"Can't decode packet"
)
else
:
...
...
neo/scripts/runner.py
View file @
8ba42463
...
...
@@ -51,13 +51,11 @@ UNIT_TEST_MODULES = [
'
neo
.
tests
.
master
.
testClientHandler
',
'
neo
.
tests
.
master
.
testMasterApp
',
'
neo
.
tests
.
master
.
testMasterPT
',
'
neo
.
tests
.
master
.
testRecovery
',
'
neo
.
tests
.
master
.
testStorageHandler
',
'
neo
.
tests
.
master
.
testTransactions
',
# storage application
'
neo
.
tests
.
storage
.
testClientHandler
',
'
neo
.
tests
.
storage
.
testMasterHandler
',
'
neo
.
tests
.
storage
.
testStorageApp
',
'
neo
.
tests
.
storage
.
testStorage
' + os.getenv('
NEO_TESTS_ADAPTER
', '
SQLite
'),
'
neo
.
tests
.
storage
.
testTransactions
',
# client application
...
...
@@ -298,6 +296,13 @@ class TestRunner(BenchmarkRunner):
x('
-
S
', '
--
stop
-
on
-
success
', action='
store_true
', default=None,
help='
Opposite
of
--
stop
-
on
-
error
:
stop
as
soon
as
a
test
'
'
passes
.
Details
about
errors
are
not
printed
at
exit
.
')
x = parser.add_mutually_exclusive_group().add_argument
x('
-
p
', '
--
dump
-
protocol
', const=True,
dest='
protocol
', action='
store_const
',
help='
Dump
schema
of
protocol
instead
of
checking
it
.
')
x('
-
P
', '
--
no
-
check
-
protocol
', const=False,
dest='
protocol
', action='
store_const
',
help='
Do
not
check
schema
of
protocol
.
')
_('
-
r', '
--
readable
-
tid
', action='
store_true
',
help='
Change
master
behaviour
to
generate
readable
TIDs
for
easier
'
'
debugging
(
rather
than
from
current
time
).
')
...
...
@@ -347,6 +352,7 @@ Environment Variables:
coverage = args.coverage,
cov_unit = args.cov_unit,
only = args.only,
protocol = args.protocol,
stop_on_success = args.stop_on_success,
readable_tid = args.readable_tid,
)
...
...
@@ -374,19 +380,26 @@ Environment Variables:
self.__coverage.save()
del self.__coverage
orig(self, success)
try:
for _ in xrange(config.loop):
if config.unit:
runner.run('
Unit
tests
', UNIT_TEST_MODULES, only)
if config.functional:
runner.run('
Functional
tests
', FUNC_TEST_MODULES, only)
if config.zodb:
runner.run('
ZODB
tests
', ZODB_TEST_MODULES, only)
except KeyboardInterrupt:
config['
mail_to
'] = None
traceback.print_exc()
except StopOnSuccess:
pass
if config.protocol is False:
from contextlib import nested
protocol_checker = nested()
else:
from neo.tests.protocol_checker import protocolChecker
protocol_checker = protocolChecker(config.protocol)
with protocol_checker:
try:
for _ in xrange(config.loop):
if config.unit:
runner.run('
Unit
tests
', UNIT_TEST_MODULES, only)
if config.functional:
runner.run('
Functional
tests
', FUNC_TEST_MODULES, only)
if config.zodb:
runner.run('
ZODB
tests
', ZODB_TEST_MODULES, only)
except KeyboardInterrupt:
config['
mail_to
'] = None
traceback.print_exc()
except StopOnSuccess:
pass
if config.coverage:
coverage.stop()
if coverage.neotestrunner:
...
...
neo/storage/app.py
View file @
8ba42463
...
...
@@ -63,11 +63,16 @@ class Application(BaseApplication):
help
=
"do not delete data of discarded cells, which is useful for"
" big databases because the current implementation is"
" inefficient (this option should disappear in the future)"
)
_
.
bool
(
'new-nid'
,
help
=
"request a new NID from a cluster that is already"
" operational, update the database with the new NID and exit,"
" which makes easier to quickly set up a replica by copying"
" the database of another node while it was stopped"
)
_
=
parser
.
group
(
'database creation'
)
_
.
int
(
'
u'
,
'uu
id'
,
help
=
"specify an
UU
ID to use for this process. Previously"
" assigned
UU
ID takes precedence (i.e. you should"
_
.
int
(
'
i'
,
'n
id'
,
help
=
"specify an
N
ID to use for this process. Previously"
" assigned
N
ID takes precedence (i.e. you should"
" always use reset with this switch)"
)
_
(
'e'
,
'engine'
,
help
=
"database engine (MySQL only)"
)
_
.
bool
(
'dedup'
,
...
...
@@ -118,10 +123,16 @@ class Application(BaseApplication):
self
.
loadConfiguration
()
self
.
devpath
=
self
.
dm
.
getTopologyPath
()
# force node uuid from command line argument, for testing purpose only
if
'uuid'
in
config
:
self
.
uuid
=
config
[
'uuid'
]
logging
.
node
(
self
.
name
,
self
.
uuid
)
if
config
.
get
(
'new_nid'
):
self
.
new_nid
=
[
x
[
0
]
for
x
in
self
.
dm
.
iterAssignedCells
()]
if
not
self
.
new_nid
:
sys
.
exit
(
'database is empty'
)
self
.
uuid
=
None
else
:
self
.
new_nid
=
()
if
'nid'
in
config
:
# for testing purpose only
self
.
uuid
=
config
[
'nid'
]
logging
.
node
(
self
.
name
,
self
.
uuid
)
registerLiveDebugger
(
on_log
=
self
.
log
)
...
...
@@ -158,36 +169,27 @@ class Application(BaseApplication):
# load configuration
self
.
uuid
=
dm
.
getUUID
()
logging
.
node
(
self
.
name
,
self
.
uuid
)
num_partitions
=
dm
.
getNumPartitions
()
num_replicas
=
dm
.
getNumReplicas
()
ptid
=
dm
.
getPTID
()
# check partition table configuration
if
num_partitions
is
not
None
and
num_replicas
is
not
None
:
if
num_partitions
<=
0
:
raise
RuntimeError
,
'partitions must be more than zero'
# create a partition table
self
.
pt
=
PartitionTable
(
num_partitions
,
num_replicas
)
logging
.
info
(
'Configuration loaded:'
)
logging
.
info
(
'PTID : %s'
,
dump
(
ptid
))
logging
.
info
(
'PTID : %s'
,
dump
(
dm
.
getPTID
()
))
logging
.
info
(
'Name : %s'
,
self
.
name
)
logging
.
info
(
'Partitions: %s'
,
num_partitions
)
logging
.
info
(
'Replicas : %s'
,
num_replicas
)
def
loadPartitionTable
(
self
):
"""Load a partition table from the database."""
self
.
pt
.
clear
()
ptid
=
self
.
dm
.
getPTID
()
if
ptid
is
None
:
self
.
pt
=
PartitionTable
(
0
,
0
)
return
cell
_list
=
[]
row
_list
=
[]
for
offset
,
uuid
,
state
in
self
.
dm
.
getPartitionTable
():
while
len
(
row_list
)
<=
offset
:
row_list
.
append
([])
# register unknown nodes
if
self
.
nm
.
getByUUID
(
uuid
)
is
None
:
self
.
nm
.
createStorage
(
uuid
=
uuid
)
cell_list
.
append
((
offset
,
uuid
,
CellStates
[
state
]))
self
.
pt
.
update
(
ptid
,
cell_list
,
self
.
nm
)
row_list
[
offset
].
append
((
uuid
,
CellStates
[
state
]))
self
.
pt
=
object
.
__new__
(
PartitionTable
)
self
.
pt
.
load
(
ptid
,
self
.
dm
.
getNumReplicas
(),
row_list
,
self
.
nm
)
def
run
(
self
):
try
:
...
...
@@ -247,29 +249,16 @@ class Application(BaseApplication):
Note that I do not accept any connection from non-master nodes
at this stage."""
pt
=
self
.
pt
# search, find, connect and identify to the primary master
bootstrap
=
BootstrapManager
(
self
,
NodeTypes
.
STORAGE
,
self
.
server
,
self
.
devpath
)
self
.
master_node
,
self
.
master_conn
,
num_partitions
,
num_replicas
=
\
bootstrap
.
getPrimaryConnection
()
bootstrap
=
BootstrapManager
(
self
,
NodeTypes
.
STORAGE
,
None
if
self
.
new_nid
else
self
.
server
,
self
.
devpath
,
self
.
new_nid
)
self
.
master_node
,
self
.
master_conn
=
bootstrap
.
getPrimaryConnection
()
self
.
dm
.
setUUID
(
self
.
uuid
)
# Reload a partition table from the database. This is necessary
# when a previous primary master died while sending a partition
# table, because the table might be incomplete.
if
pt
is
not
None
:
self
.
loadPartitionTable
()
if
num_partitions
!=
pt
.
getPartitions
():
raise
RuntimeError
(
'the number of partitions is inconsistent'
)
if
pt
is
None
or
pt
.
getReplicas
()
!=
num_replicas
:
# changing number of replicas is not an issue
self
.
dm
.
setNumPartitions
(
num_partitions
)
self
.
dm
.
setNumReplicas
(
num_replicas
)
self
.
pt
=
PartitionTable
(
num_partitions
,
num_replicas
)
self
.
loadPartitionTable
()
# Reload a partition table from the database,
# in case that we're in RECOVERING phase.
self
.
loadPartitionTable
()
def
initialize
(
self
):
logging
.
debug
(
'initializing...'
)
...
...
neo/storage/checker.py
View file @
8ba42463
...
...
@@ -51,7 +51,7 @@ class Checker(object):
else
:
conn
=
ClientConnection
(
app
,
StorageOperationHandler
(
app
),
node
)
conn
.
ask
(
Packets
.
RequestIdentification
(
NodeTypes
.
STORAGE
,
uuid
,
app
.
server
,
name
,
(),
app
.
id_timestamp
))
uuid
,
app
.
server
,
name
,
app
.
id_timestamp
,
(),
()
))
self
.
conn_dict
[
conn
]
=
node
.
isIdentified
()
conn_set
=
set
(
self
.
conn_dict
)
conn_set
.
discard
(
None
)
...
...
neo/storage/database/importer.py
View file @
8ba42463
...
...
@@ -33,7 +33,7 @@ from ZODB.FileStorage import FileStorage
from
..app
import
option_defaults
from
.
import
buildDatabaseManager
,
DatabaseFailure
from
.manager
import
DatabaseManager
from
.manager
import
DatabaseManager
,
Fallback
from
neo.lib
import
compress
,
logging
,
patch
,
util
from
neo.lib.interfaces
import
implements
from
neo.lib.protocol
import
BackendNotImplemented
,
MAX_TID
...
...
@@ -216,7 +216,7 @@ class ZODB(object):
self
.
_connect
=
_connect
config
=
section
.
config
if
'read_only'
in
config
.
getSectionAttributes
():
has_next_oid
=
config
.
read_only
=
hasattr
(
self
,
'next_oid'
)
has_next_oid
=
config
.
read_only
=
'next_oid'
in
self
.
__dict__
if
not
has_next_oid
:
import
gc
# This will reopen read-only as soon as we know the last oid.
...
...
@@ -378,8 +378,8 @@ class ImporterDatabaseManager(DatabaseManager):
conf
=
self
.
_conf
db
=
self
.
db
=
buildDatabaseManager
(
conf
[
'adapter'
],
(
conf
[
'database'
],
conf
.
get
(
'engine'
),
conf
[
'wait'
]))
for
x
in
"""getConfiguration _setConfiguration
setNumPartitions
query erase getPartitionTable
_
iterAssignedCells
for
x
in
"""getConfiguration _setConfiguration
_getMaxPartition
query erase getPartitionTable iterAssignedCells
updateCellTID getUnfinishedTIDDict dropUnfinishedData
abortTransaction storeTransaction lockTransaction
loadData storeData getOrphanList _pruneData deferCommit
...
...
@@ -396,9 +396,16 @@ class ImporterDatabaseManager(DatabaseManager):
self
.
_writeback
.
committed
()
self
.
commit
=
db
.
commit
=
commit
def
_updateReadable
(
self
):
def
_updateReadable
(
*
_
):
raise
AssertionError
def
setUUID
(
self
,
nid
):
old_nid
=
self
.
getUUID
()
if
old_nid
:
assert
old_nid
==
nid
,
(
old_nid
,
nid
)
else
:
self
.
setConfiguration
(
'nid'
,
str
(
nid
))
def
changePartitionTable
(
self
,
*
args
,
**
kw
):
self
.
db
.
changePartitionTable
(
*
args
,
**
kw
)
if
self
.
_writeback
:
...
...
@@ -413,7 +420,7 @@ class ImporterDatabaseManager(DatabaseManager):
if
self
.
_writeback
:
self
.
_writeback
.
close
()
self
.
db
.
close
()
if
isinstance
(
self
.
zodb
,
list
):
# _setup called
if
isinstance
(
self
.
zodb
,
tuple
):
# _setup called
for
zodb
in
self
.
zodb
:
zodb
.
close
()
...
...
@@ -436,9 +443,13 @@ class ImporterDatabaseManager(DatabaseManager):
self
.
zodb_ltid
=
max
(
x
.
ltid
for
x
in
self
.
zodb
)
zodb
=
self
.
zodb
[
-
1
]
self
.
zodb_loid
=
zodb
.
shift_oid
+
zodb
.
next_oid
-
1
self
.
zodb_tid
=
self
.
db
.
getLastTID
(
self
.
zodb_ltid
)
or
0
if
callable
(
self
.
_import
):
self
.
_import
=
self
.
_import
()
self
.
zodb_tid
=
self
.
_getMaxPartition
()
is
not
None
and
\
self
.
db
.
getLastTID
(
self
.
zodb_ltid
)
or
0
if
callable
(
self
.
_import
):
# XXX: why ?
if
self
.
zodb_tid
==
self
.
zodb_ltid
:
self
.
_finished
()
else
:
self
.
_import
=
self
.
_import
()
def
doOperation
(
self
,
app
):
if
self
.
_import
:
...
...
@@ -498,12 +509,19 @@ class ImporterDatabaseManager(DatabaseManager):
if
process
:
process
.
join
()
self
.
commit
()
self
.
_finished
()
def
_finished
(
self
):
logging
.
warning
(
"All data are imported. You should change"
" your configuration to use the native backend and restart."
)
self
.
_import
=
None
for
x
in
"""getObject getReplicationTIDList getReplicationObjectList
_fetchObject _getDataTID getLastObjectTID
"""
.
split
():
setattr
(
self
,
x
,
getattr
(
self
.
db
,
x
))
for
zodb
in
self
.
zodb
:
zodb
.
close
()
self
.
zodb
=
None
def
_iter_zodb
(
self
,
zodb_list
):
util
.
setproctitle
(
'neostorage: import'
)
...
...
@@ -556,6 +574,19 @@ class ImporterDatabaseManager(DatabaseManager):
return
(
max
(
tid
,
util
.
p64
(
self
.
zodb_ltid
)),
max
(
oid
,
util
.
p64
(
self
.
zodb_loid
)))
def
_getObject
(
self
,
oid
,
tid
=
None
,
before_tid
=
None
):
p64
=
util
.
p64
r
=
self
.
getObject
(
p64
(
oid
),
None
if
tid
is
None
else
p64
(
tid
),
None
if
before_tid
is
None
else
p64
(
before_tid
))
if
r
:
serial
,
next_serial
,
compression
,
checksum
,
data
,
data_serial
=
r
u64
=
util
.
u64
return
(
u64
(
serial
),
next_serial
and
u64
(
next_serial
),
compression
,
checksum
,
data
,
data_serial
and
u64
(
data_serial
))
def
getObject
(
self
,
oid
,
tid
=
None
,
before_tid
=
None
):
u64
=
util
.
u64
u_oid
=
u64
(
oid
)
...
...
@@ -623,7 +654,11 @@ class ImporterDatabaseManager(DatabaseManager):
def
_deleteRange
(
self
,
partition
,
min_tid
=
None
,
max_tid
=
None
):
# Even if everything is imported, we can't truncate below
# because it would import again if we restart with this backend.
if
min_tid
<
self
.
zodb_ltid
:
# This is also incompatible with writeback, because ZODB has
# no API to truncate.
if
min_tid
<
self
.
zodb_ltid
or
self
.
_writeback
:
# XXX: That's late to fail. The master should ask storage nodes
# whether truncation is possible before going further.
raise
NotImplementedError
self
.
db
.
_deleteRange
(
partition
,
min_tid
,
max_tid
)
...
...
@@ -667,6 +702,12 @@ class ImporterDatabaseManager(DatabaseManager):
length
,
partition
)
return
r
def
_fetchObject
(
*
_
):
raise
AssertionError
getLastObjectTID
=
Fallback
.
getLastObjectTID
.
__func__
_getDataTID
=
Fallback
.
_getDataTID
.
__func__
def
getObjectHistory
(
self
,
*
args
,
**
kw
):
raise
BackendNotImplemented
(
self
.
getObjectHistory
)
...
...
@@ -678,6 +719,7 @@ class WriteBack(object):
_changed
=
False
_process
=
None
chunk_size
=
100
def
__init__
(
self
,
db
,
storage
):
self
.
_db
=
db
...
...
@@ -705,7 +747,7 @@ class WriteBack(object):
self
.
_event
=
Event
()
self
.
_idle
=
Event
()
self
.
_stop
=
Event
()
self
.
_np
=
self
.
_db
.
getNumPartitions
()
self
.
_np
=
1
+
self
.
_db
.
_getMaxPartition
()
self
.
_db
=
cPickle
.
dumps
(
self
.
_db
,
2
)
self
.
_process
=
Process
(
target
=
self
.
_run
)
self
.
_process
.
daemon
=
True
...
...
@@ -737,7 +779,6 @@ class WriteBack(object):
def
iterator
(
self
):
db
=
self
.
_db
np
=
self
.
_np
chunk_size
=
max
(
2
,
1000
//
np
)
offset_list
=
xrange
(
np
)
while
1
:
with
db
:
...
...
@@ -748,23 +789,26 @@ class WriteBack(object):
if
np
==
len
(
db
.
_readable_set
):
while
1
:
tid_list
=
[]
loop
=
False
max_tid
=
MAX_TID
for
offset
in
offset_list
:
x
=
db
.
getReplicationTIDList
(
self
.
min_tid
,
MAX_TID
,
chunk_size
,
offset
)
self
.
min_tid
,
max_tid
,
self
.
chunk_size
,
offset
)
tid_list
+=
x
if
len
(
x
)
==
chunk_size
:
loop
=
True
if
tid_list
:
tid_list
.
sort
()
for
tid
in
tid_list
:
if
self
.
_stop
.
is_set
():
return
yield
TransactionRecord
(
db
,
tid
)
if
len
(
x
)
==
self
.
chunk_size
:
max_tid
=
x
[
-
1
]
if
not
tid_list
:
break
tid_list
.
sort
()
for
tid
in
tid_list
:
if
self
.
_stop
.
is_set
():
return
yield
TransactionRecord
(
db
,
tid
)
if
tid
==
max_tid
:
break
else
:
self
.
min_tid
=
util
.
add64
(
tid
,
1
)
if
loop
:
continue
break
break
self
.
min_tid
=
util
.
add64
(
tid
,
1
)
if
not
self
.
_event
.
is_set
():
self
.
_idle
.
set
()
self
.
_event
.
wait
()
...
...
@@ -785,7 +829,10 @@ class TransactionRecord(BaseStorage.TransactionRecord):
def
__iter__
(
self
):
tid
=
self
.
tid
for
oid
in
self
.
_oid_list
:
_
,
compression
,
_
,
data
,
data_tid
=
self
.
_db
.
fetchObject
(
oid
,
tid
)
r
=
self
.
_db
.
fetchObject
(
oid
,
tid
)
if
r
is
None
:
# checkCurrentSerialInTransaction
continue
_
,
compression
,
_
,
data
,
data_tid
=
r
if
data
is
not
None
:
data
=
compress
.
decompress_list
[
compression
](
data
)
yield
BaseStorage
.
DataRecord
(
oid
,
tid
,
data
,
data_tid
)
neo/storage/database/manager.py
View file @
8ba42463
...
...
@@ -26,22 +26,9 @@ from . import DatabaseFailure
READABLE
=
CellStates
.
UP_TO_DATE
,
CellStates
.
FEEDING
def
lazymethod
(
func
):
def
getter
(
self
):
cls
=
self
.
__class__
name
=
func
.
__name__
assert
name
not
in
cls
.
__dict__
setattr
(
cls
,
name
,
func
(
self
))
return
getattr
(
self
,
name
)
return
property
(
getter
,
doc
=
func
.
__doc__
)
def
fallback
(
func
):
def
warn
(
self
):
logging
.
info
(
"Fallback to generic/slow implementation of %s."
" It should be overridden by backend storage (%s)."
,
func
.
__name__
,
self
.
__class__
.
__name__
)
return
func
return
lazymethod
(
wraps
(
func
)(
warn
))
setattr
(
Fallback
,
func
.
__name__
,
func
)
return
abstract
(
func
)
def
splitOIDField
(
tid
,
oids
):
if
len
(
oids
)
%
8
:
...
...
@@ -52,6 +39,9 @@ def splitOIDField(tid, oids):
class
CreationUndone
(
Exception
):
pass
class
Fallback
(
object
):
pass
class
DatabaseManager
(
object
):
"""This class only describes an interface for database managers."""
...
...
@@ -102,25 +92,24 @@ class DatabaseManager(object):
finally
:
db
.
close
()
_cached_attr_list
=
(
'_readable_set'
,
'_getPartition'
,
'_getReadablePartition'
)
def
__getattr__
(
self
,
attr
):
if
attr
in
(
'_readable_set'
,
'_getPartition'
,
'_getReadablePartition'
)
:
if
attr
in
self
.
_cached_attr_list
:
self
.
_updateReadable
()
return
self
.
__getattribute__
(
attr
)
def
_partitionTableChanged
(
self
):
try
:
del
(
self
.
_readable_set
,
self
.
_getPartition
,
self
.
_getReadablePartition
)
except
AttributeError
:
pass
def
__enter__
(
self
):
assert
not
self
.
LOCK
,
"not a secondary connection"
# XXX: All config caching should be done in this class,
# rather than in backend classes.
self
.
_config
.
clear
()
self
.
_partitionTableChanged
()
try
:
for
attr
in
self
.
_cached_attr_list
:
delattr
(
self
,
attr
)
except
AttributeError
:
pass
def
__exit__
(
self
,
t
,
v
,
tb
):
if
v
is
None
:
...
...
@@ -180,6 +169,10 @@ class DatabaseManager(object):
def
erase
(
self
):
""""""
def
restore
(
self
,
dump
):
# for tests
self
.
erase
()
self
.
_restore
(
dump
)
def
_setup
(
self
,
dedup
=
False
):
"""To be overridden by the backend to set up a database
...
...
@@ -271,6 +264,18 @@ class DatabaseManager(object):
def
_setConfiguration
(
self
,
key
,
value
):
""""""
def
_changePartitionTable
(
self
,
cell_list
,
reset
=
False
):
"""Change a part of a partition table. The list of cells is
a tuple of tuples, each of which consists of an offset (row ID),
the NID of a storage node, and a cell state. If reset is True,
existing data is first thrown away.
"""
def
_getPartitionTable
(
self
):
"""Return a whole partition table as a sequence of rows. Each row
is again a tuple of an offset (row ID), the NID of a storage
node, and a cell state."""
def
getUUID
(
self
):
"""
Load a NID from a database.
...
...
@@ -279,26 +284,19 @@ class DatabaseManager(object):
if
nid
is
not
None
:
return
int
(
nid
)
@
requires
(
_changePartitionTable
,
_getPartitionTable
)
def
setUUID
(
self
,
nid
):
"""
Store a NID into a database.
"""
self
.
setConfiguration
(
'nid'
,
str
(
nid
))
def
getNumPartitions
(
self
):
"""
Load the number of partitions from a database.
"""
n
=
self
.
getConfiguration
(
'partitions'
)
if
n
is
not
None
:
return
int
(
n
)
def
setNumPartitions
(
self
,
num_partitions
):
"""
Store the number of partitions into a database.
"""
self
.
setConfiguration
(
'partitions'
,
num_partitions
)
self
.
_partitionTableChanged
()
old_nid
=
self
.
getUUID
()
if
nid
!=
old_nid
:
if
old_nid
:
self
.
_changePartitionTable
((
offset
,
x
,
tid
)
for
offset
,
x
,
tid
in
self
.
_getPartitionTable
()
if
x
==
old_nid
for
x
,
tid
in
((
x
,
None
),
(
nid
,
tid
)))
self
.
setConfiguration
(
'nid'
,
str
(
nid
))
def
getNumReplicas
(
self
):
"""
...
...
@@ -308,12 +306,6 @@ class DatabaseManager(object):
if
n
is
not
None
:
return
int
(
n
)
def
setNumReplicas
(
self
,
num_replicas
):
"""
Store the number of replicas into a database.
"""
self
.
setConfiguration
(
'replicas'
,
num_replicas
)
def
getName
(
self
):
"""
Load a name from a database.
...
...
@@ -374,8 +366,9 @@ class DatabaseManager(object):
tids are in unpacked format.
"""
if
self
.
getNumPartitions
():
return
max
(
map
(
self
.
_getLastTID
,
self
.
_readable_set
))
x
=
self
.
_readable_set
if
x
:
return
max
(
self
.
_getLastTID
(
x
,
max_tid
)
for
x
in
x
)
def
_getLastIDs
(
self
,
partition
):
"""Return max(tid) & max(oid) for objects of given partition
...
...
@@ -395,7 +388,7 @@ class DatabaseManager(object):
x
=
self
.
_readable_set
if
x
:
tid
,
oid
=
zip
(
*
map
(
self
.
_getLastIDs
,
x
))
tid
=
max
(
self
.
getLastTID
(
None
),
max
(
tid
))
tid
=
max
(
self
.
getLastTID
(),
max
(
tid
))
oid
=
max
(
oid
)
return
(
None
if
tid
is
None
else
util
.
p64
(
tid
),
None
if
oid
is
None
else
util
.
p64
(
oid
))
...
...
@@ -490,6 +483,7 @@ class DatabaseManager(object):
None
if
data_serial
is
None
else
util
.
p64
(
data_serial
))
@
fallback
@
requires
(
_getObject
)
def
_fetchObject
(
self
,
oid
,
tid
):
"""Specialized version of _getObject, for replication"""
r
=
self
.
_getObject
(
oid
,
tid
)
...
...
@@ -511,13 +505,8 @@ class DatabaseManager(object):
return
(
util
.
p64
(
serial
),
compression
,
checksum
,
data
,
None
if
data_serial
is
None
else
util
.
p64
(
data_serial
))
def
_getPartitionTable
(
self
):
"""Return a whole partition table as a sequence of rows. Each row
is again a tuple of an offset (row ID), the NID of a storage
node, and a cell state."""
@
requires
(
_getPartitionTable
)
def
_
iterAssignedCells
(
self
):
def
iterAssignedCells
(
self
):
my_nid
=
self
.
getUUID
()
return
((
offset
,
tid
)
for
offset
,
nid
,
tid
in
self
.
_getPartitionTable
()
if
my_nid
==
nid
)
...
...
@@ -537,24 +526,19 @@ class DatabaseManager(object):
finally
:
readable_set
.
remove
(
offset
)
def
_changePartitionTable
(
self
,
cell_list
,
reset
=
False
):
"""Change a part of a partition table. The list of cells is
a tuple of tuples, each of which consists of an offset (row ID),
the NID of a storage node, and a cell state. If reset is True,
existing data is first thrown away.
def
_getDataLastId
(
self
,
partition
):
"""
"""
def
_get
DataLastId
(
self
,
partition
):
def
_get
MaxPartition
(
self
):
"""
"""
@
requires
(
_getDataLastId
)
def
_updateReadable
(
self
):
try
:
readable_set
=
self
.
__dict__
[
'_readable_set'
]
except
KeyError
:
@
requires
(
_getDataLastId
,
_getMaxPartition
)
def
_updateReadable
(
self
,
reset
=
True
):
if
reset
:
readable_set
=
self
.
_readable_set
=
set
()
np
=
self
.
getNumPartitions
()
np
=
1
+
self
.
_getMaxPartition
()
def
_getPartition
(
x
,
np
=
np
):
return
x
%
np
def
_getReadablePartition
(
x
,
np
=
np
,
r
=
readable_set
):
...
...
@@ -569,14 +553,15 @@ class DatabaseManager(object):
i
=
self
.
_getDataLastId
(
p
)
d
.
append
(
p
<<
48
if
i
is
None
else
i
+
1
)
else
:
readable_set
=
self
.
_readable_set
readable_set
.
clear
()
readable_set
.
update
(
x
[
0
]
for
x
in
self
.
_
iterAssignedCells
()
readable_set
.
update
(
x
[
0
]
for
x
in
self
.
iterAssignedCells
()
if
-
x
[
1
]
in
READABLE
)
@
requires
(
_changePartitionTable
,
_getLastIDs
,
_getLastTID
)
def
changePartitionTable
(
self
,
ptid
,
cell_list
,
reset
=
False
):
def
changePartitionTable
(
self
,
ptid
,
num_replicas
,
cell_list
,
reset
=
False
):
my_nid
=
self
.
getUUID
()
pt
=
dict
(
self
.
_
iterAssignedCells
())
pt
=
dict
(
self
.
iterAssignedCells
())
# In backup mode, the last transactions of a readable cell may be
# incomplete.
backup_tid
=
self
.
getBackupTID
()
...
...
@@ -595,13 +580,14 @@ class DatabaseManager(object):
outofdate_tid
(
offset
)))
for
offset
,
nid
,
state
in
cell_list
]
self
.
_changePartitionTable
(
cell_list
,
reset
)
self
.
_updateReadable
()
self
.
_updateReadable
(
reset
)
assert
isinstance
(
ptid
,
(
int
,
long
)),
ptid
self
.
_setConfiguration
(
'ptid'
,
str
(
ptid
))
self
.
_setConfiguration
(
'replicas'
,
str
(
num_replicas
))
@
requires
(
_changePartitionTable
)
def
updateCellTID
(
self
,
partition
,
tid
):
t
,
=
(
t
for
p
,
t
in
self
.
_
iterAssignedCells
()
if
p
==
partition
)
t
,
=
(
t
for
p
,
t
in
self
.
iterAssignedCells
()
if
p
==
partition
)
if
t
<
0
:
return
tid
=
util
.
u64
(
tid
)
...
...
@@ -623,7 +609,7 @@ class DatabaseManager(object):
next_tid
=
util
.
u64
(
backup_tid
)
if
next_tid
:
next_tid
+=
1
for
offset
,
tid
in
self
.
_
iterAssignedCells
():
for
offset
,
tid
in
self
.
iterAssignedCells
():
if
tid
>=
0
:
# OUT_OF_DATE
yield
offset
,
p64
(
tid
and
tid
+
1
)
elif
-
tid
in
READABLE
:
...
...
@@ -743,6 +729,7 @@ class DatabaseManager(object):
return
self
.
_pruneData
(
data_id_list
)
@
fallback
@
requires
(
_getObject
)
def
_getDataTID
(
self
,
oid
,
tid
=
None
,
before_tid
=
None
):
"""
Return a 2-tuple:
...
...
@@ -809,6 +796,9 @@ class DatabaseManager(object):
oid
,
current_tid
)
return
current_tid
,
current_tid
return
current_tid
,
tid
found_undone_tid
,
undone_data_tid
=
getDataTID
(
tid
=
undone_tid
)
if
found_undone_tid
is
None
:
return
if
transaction_object
:
try
:
current_tid
=
current_data_tid
=
u64
(
transaction_object
[
2
])
...
...
@@ -818,8 +808,6 @@ class DatabaseManager(object):
current_tid
,
current_data_tid
=
getDataTID
(
before_tid
=
ltid
)
if
current_tid
is
None
:
return
None
,
None
,
False
found_undone_tid
,
undone_data_tid
=
getDataTID
(
tid
=
undone_tid
)
assert
found_undone_tid
is
not
None
,
(
oid
,
undone_tid
)
is_current
=
undone_data_tid
in
(
current_data_tid
,
tid
)
# Load object data as it was before given transaction.
# It can be None, in which case it means we are undoing object
...
...
@@ -865,7 +853,7 @@ class DatabaseManager(object):
assert
tid
,
tid
cell_list
=
[]
my_nid
=
self
.
getUUID
()
for
partition
,
state
in
self
.
_
iterAssignedCells
():
for
partition
,
state
in
self
.
iterAssignedCells
():
if
state
>
tid
:
cell_list
.
append
((
partition
,
my_nid
,
tid
))
self
.
_deleteRange
(
partition
,
tid
)
...
...
neo/storage/database/mysqldb.py
View file @
8ba42463
...
...
@@ -117,9 +117,11 @@ class MySQLDatabaseManager(DatabaseManager):
return
super
(
MySQLDatabaseManager
,
self
).
__getattr__
(
attr
)
def
_tryConnect
(
self
):
kwd
=
{
'db'
:
self
.
db
,
'user'
:
self
.
user
}
if
self
.
passwd
is
not
None
:
kwd
[
'passwd'
]
=
self
.
passwd
kwd
=
{
'db'
:
self
.
db
}
if
self
.
user
:
kwd
[
'user'
]
=
self
.
user
if
self
.
passwd
is
not
None
:
kwd
[
'passwd'
]
=
self
.
passwd
if
self
.
socket
:
kwd
[
'unix_socket'
]
=
os
.
path
.
expanduser
(
self
.
socket
)
logging
.
info
(
'connecting to MySQL on the database %s with user %s'
,
...
...
@@ -198,6 +200,7 @@ class MySQLDatabaseManager(DatabaseManager):
self
.
_connect
()
def
_commit
(
self
):
# XXX: Should we translate OperationalError into MysqlError ?
self
.
conn
.
commit
()
self
.
_active
=
0
...
...
@@ -270,6 +273,12 @@ class MySQLDatabaseManager(DatabaseManager):
" ELSE 1-state"
" END as tid"
)
# Let's wait for a more important change to clean up,
# so that users can still downgrade.
if
0
:
def
_migrate4
(
self
,
schema_dict
):
self
.
_setConfiguration
(
'partitions'
,
None
)
def
_setup
(
self
,
dedup
=
False
):
self
.
_config
.
clear
()
q
=
self
.
query
...
...
@@ -295,6 +304,12 @@ class MySQLDatabaseManager(DatabaseManager):
p
+=
""" PARTITION BY LIST (`partition`) (
PARTITION dummy VALUES IN (NULL))"""
if
engine
==
"RocksDB"
:
cf
=
lambda
name
,
rev
=
False
:
" COMMENT '%scf_neo_%s'"
%
(
'rev:'
if
rev
else
''
,
name
)
else
:
cf
=
lambda
*
_
:
''
# The table "trans" stores information on committed transactions.
schema_dict
[
'trans'
]
=
"""CREATE TABLE %s (
`partition` SMALLINT UNSIGNED NOT NULL,
...
...
@@ -305,8 +320,8 @@ class MySQLDatabaseManager(DatabaseManager):
description BLOB NOT NULL,
ext BLOB NOT NULL,
ttid BIGINT UNSIGNED NOT NULL,
PRIMARY KEY (`partition`, tid)
) ENGINE=
"""
+
p
PRIMARY KEY (`partition`, tid)
{}
) ENGINE=
{}"""
.
format
(
cf
(
'append_meta'
),
p
)
# The table "obj" stores committed object metadata.
schema_dict
[
'obj'
]
=
"""CREATE TABLE %s (
...
...
@@ -315,10 +330,11 @@ class MySQLDatabaseManager(DatabaseManager):
tid BIGINT UNSIGNED NOT NULL,
data_id BIGINT UNSIGNED NULL,
value_tid BIGINT UNSIGNED NULL,
PRIMARY KEY (`partition`, oid, tid),
KEY tid (`partition`, tid, oid),
KEY (data_id)
) ENGINE="""
+
p
PRIMARY KEY (`partition`, oid, tid){},
KEY tid (`partition`, tid, oid){},
KEY (data_id){}
) ENGINE={}"""
.
format
(
cf
(
'obj_pk'
,
True
),
cf
(
'append_meta'
),
cf
(
'append_meta'
),
p
)
if
engine
==
"TokuDB"
:
engine
+=
" compression='tokudb_uncompressed'"
...
...
@@ -326,18 +342,21 @@ class MySQLDatabaseManager(DatabaseManager):
# The table "data" stores object data.
# We'd like to have partial index on 'hash' column (e.g. hash(4))
# but 'UNIQUE' constraint would not work as expected.
schema_dict
[
'data'
]
=
"""CREATE TABLE %
%
s (
id BIGINT UNSIGNED NOT NULL
PRIMARY KEY
,
schema_dict
[
'data'
]
=
"""CREATE TABLE %s (
id BIGINT UNSIGNED NOT NULL,
hash BINARY(20) NOT NULL,
compression TINYINT UNSIGNED NULL,
value MEDIUMBLOB NOT NULL%s
) ENGINE=%s"""
%
(
""",
UNIQUE (hash, compression)"""
if
dedup
else
""
,
engine
)
value MEDIUMBLOB NOT NULL,
PRIMARY KEY (id){}{}
) ENGINE={}"""
.
format
(
cf
(
'append'
),
""",
UNIQUE (hash, compression)"""
+
cf
(
'no_comp'
)
if
dedup
else
""
,
engine
)
schema_dict
[
'bigdata'
]
=
"""CREATE TABLE %s (
id INT UNSIGNED NOT NULL AUTO_INCREMENT PRIMARY KEY,
value MEDIUMBLOB NOT NULL
) ENGINE="""
+
engine
id INT UNSIGNED NOT NULL AUTO_INCREMENT,
value MEDIUMBLOB NOT NULL,
PRIMARY KEY (id){}
) ENGINE={}"""
.
format
(
cf
(
'append'
),
p
)
# The table "ttrans" stores information on uncommitted transactions.
schema_dict
[
'ttrans'
]
=
"""CREATE TABLE %s (
...
...
@@ -348,8 +367,9 @@ class MySQLDatabaseManager(DatabaseManager):
user BLOB NOT NULL,
description BLOB NOT NULL,
ext BLOB NOT NULL,
ttid BIGINT UNSIGNED NOT NULL
) ENGINE="""
+
engine
ttid BIGINT UNSIGNED NOT NULL,
PRIMARY KEY (ttid){}
) ENGINE={}"""
.
format
(
cf
(
'no_comp'
),
p
)
# The table "tobj" stores uncommitted object metadata.
schema_dict
[
'tobj'
]
=
"""CREATE TABLE %s (
...
...
@@ -358,8 +378,8 @@ class MySQLDatabaseManager(DatabaseManager):
tid BIGINT UNSIGNED NOT NULL,
data_id BIGINT UNSIGNED NULL,
value_tid BIGINT UNSIGNED NULL,
PRIMARY KEY (tid, oid)
) ENGINE=
"""
+
engine
PRIMARY KEY (tid, oid)
{}
) ENGINE=
{}"""
.
format
(
cf
(
'no_comp'
),
p
)
if
self
.
nonempty
(
'config'
)
is
None
:
q
(
schema_dict
.
pop
(
'config'
)
%
'config'
)
...
...
@@ -407,6 +427,9 @@ class MySQLDatabaseManager(DatabaseManager):
q
(
"ALTER TABLE config MODIFY value VARBINARY(%s) NULL"
%
len
(
value
))
q
(
sql
)
def
_getMaxPartition
(
self
):
return
self
.
query
(
"SELECT MAX(`partition`) FROM pt"
)[
0
][
0
]
def
_getPartitionTable
(
self
):
return
self
.
query
(
"SELECT * FROM pt"
)
...
...
@@ -965,7 +988,7 @@ class MySQLDatabaseManager(DatabaseManager):
cmd
+=
self
.
_cmdline
()
return
subprocess
.
check_output
(
cmd
)
def
restore
(
self
,
sql
):
def
_
restore
(
self
,
sql
):
import
subprocess
cmd
=
[
'mysql'
]
cmd
+=
self
.
_cmdline
()
...
...
neo/storage/database/sqlite.py
View file @
8ba42463
...
...
@@ -79,6 +79,7 @@ class SQLiteDatabaseManager(DatabaseManager):
def
_connect
(
self
):
logging
.
info
(
'connecting to SQLite database %r'
,
self
.
db
)
self
.
conn
=
sqlite3
.
connect
(
self
.
db
,
check_same_thread
=
False
)
self
.
conn
.
text_factory
=
str
self
.
lock
(
self
.
db
)
if
self
.
UNSAFE
:
q
=
self
.
query
...
...
@@ -144,6 +145,12 @@ class SQLiteDatabaseManager(DatabaseManager):
" WHEN 2 THEN -2"
# FEEDING
" ELSE 1-state END"
)
# Let's wait for a more important change to clean up,
# so that users can still downgrade.
if
0
:
def
_migrate4
(
self
,
schema_dict
,
index_dict
):
self
.
_setConfiguration
(
'partitions'
,
None
)
def
_setup
(
self
,
dedup
=
False
):
# BBB: SQLite has transactional DDL but before Python 3.6,
# the binding automatically commits between such statements.
...
...
@@ -265,6 +272,9 @@ class SQLiteDatabaseManager(DatabaseManager):
else
:
q
(
"REPLACE INTO config VALUES (?,?)"
,
(
key
,
str
(
value
)))
def
_getMaxPartition
(
self
):
return
self
.
query
(
"SELECT MAX(`partition`) FROM pt"
).
next
()[
0
]
def
_getPartitionTable
(
self
):
return
self
.
query
(
"SELECT * FROM pt"
)
...
...
@@ -451,8 +461,12 @@ class SQLiteDatabaseManager(DatabaseManager):
return
r
def
loadData
(
self
,
data_id
):
return
self
.
query
(
"SELECT compression, hash, value"
" FROM data WHERE id=?"
,
(
data_id
,)).
fetchone
()
compression
,
checksum
,
data
=
self
.
query
(
"SELECT compression, hash, value FROM data WHERE id=?"
,
(
data_id
,)).
fetchone
()
if
checksum
:
return
compression
,
str
(
checksum
),
str
(
data
)
return
compression
,
checksum
,
data
def
_getDataTID
(
self
,
oid
,
tid
=
None
,
before_tid
=
None
):
partition
=
self
.
_getReadablePartition
(
oid
)
...
...
@@ -712,5 +726,5 @@ class SQLiteDatabaseManager(DatabaseManager):
main
[
-
1
:
-
1
]
=
data
return
'
\
n
'
.
join
(
main
)
+
'
\
n
'
def
restore
(
self
,
sql
):
def
_
restore
(
self
,
sql
):
self
.
conn
.
executescript
(
sql
)
neo/storage/handlers/__init__.py
View file @
8ba42463
...
...
@@ -65,14 +65,14 @@ class BaseMasterHandler(BaseHandler):
# See comment in ClientOperationHandler.connectionClosed
self
.
app
.
tm
.
abortFor
(
uuid
,
even_if_voted
=
True
)
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
cell_list
):
def
notifyPartitionChanges
(
self
,
conn
,
ptid
,
num_replicas
,
cell_list
):
"""This is very similar to Send Partition Table, except that
the information is only about changes from the previous."""
app
=
self
.
app
if
ptid
!=
1
+
app
.
pt
.
getID
():
raise
ProtocolError
(
'wrong partition table id'
)
app
.
pt
.
update
(
ptid
,
cell_list
,
app
.
nm
)
app
.
dm
.
changePartitionTable
(
ptid
,
cell_list
)
app
.
pt
.
update
(
ptid
,
num_replicas
,
cell_list
,
app
.
nm
)
app
.
dm
.
changePartitionTable
(
ptid
,
num_replicas
,
cell_list
)
if
app
.
operational
:
app
.
replicator
.
notifyPartitionChanges
(
cell_list
)
app
.
dm
.
commit
()
...
...
neo/storage/handlers/client.py
View file @
8ba42463
...
...
@@ -53,7 +53,7 @@ class ClientOperationHandler(BaseHandler):
p
=
Errors
.
TidNotFound
(
'%s does not exist'
%
dump
(
tid
))
else
:
p
=
Packets
.
AnswerTransactionInformation
(
tid
,
t
[
1
],
t
[
2
],
t
[
3
],
t
[
4
]
,
t
[
0
])
bool
(
t
[
4
])
,
t
[
0
])
conn
.
answer
(
p
)
def
getEventQueue
(
self
):
...
...
@@ -183,12 +183,13 @@ class ClientOperationHandler(BaseHandler):
getObjectFromTransaction
=
app
.
tm
.
getObjectFromTransaction
object_tid_dict
=
{}
for
oid
in
oid_list
:
current_serial
,
undo_serial
,
is_current
=
findUndoTID
(
oid
,
ttid
,
r
=
findUndoTID
(
oid
,
ttid
,
ltid
,
undone_tid
,
getObjectFromTransaction
(
ttid
,
oid
))
if
current_serial
is
None
:
p
=
Errors
.
OidNotFound
(
dump
(
oid
))
break
object_tid_dict
[
oid
]
=
(
current_serial
,
undo_serial
,
is_current
)
if
r
:
if
not
r
[
0
]:
p
=
Errors
.
OidNotFound
(
dump
(
oid
))
break
object_tid_dict
[
oid
]
=
r
else
:
p
=
Packets
.
AnswerObjectUndoSerial
(
object_tid_dict
)
conn
.
answer
(
p
)
...
...
neo/storage/handlers/identification.py
View file @
8ba42463
...
...
@@ -32,7 +32,7 @@ class IdentificationHandler(EventHandler):
return
self
.
app
.
nm
def
requestIdentification
(
self
,
conn
,
node_type
,
uuid
,
address
,
name
,
devpath
,
id_timestamp
):
id_timestamp
,
devpath
,
new_nid
):
self
.
checkClusterName
(
name
)
app
=
self
.
app
# reject any incoming connections if not ready
...
...
@@ -65,6 +65,5 @@ class IdentificationHandler(EventHandler):
conn
.
setHandler
(
handler
)
node
.
setConnection
(
conn
,
force
)
# accept the identification and trigger an event
conn
.
answer
(
Packets
.
AcceptIdentification
(
NodeTypes
.
STORAGE
,
uuid
and
app
.
uuid
,
app
.
pt
.
getPartitions
(),
app
.
pt
.
getReplicas
(),
uuid
))
handler
.
connectionCompleted
(
conn
)
conn
.
answer
(
Packets
.
AcceptIdentification
(
NodeTypes
.
STORAGE
,
uuid
and
app
.
uuid
,
uuid
))
neo/storage/handlers/initialization.py
View file @
8ba42463
...
...
@@ -20,10 +20,10 @@ from neo.lib.protocol import Packets, ProtocolError, ZERO_TID
class
InitializationHandler
(
BaseMasterHandler
):
def
sendPartitionTable
(
self
,
conn
,
ptid
,
row_list
):
def
sendPartitionTable
(
self
,
conn
,
ptid
,
num_replicas
,
row_list
):
app
=
self
.
app
pt
=
app
.
pt
pt
.
load
(
ptid
,
row_list
,
app
.
nm
)
pt
.
load
(
ptid
,
num_replicas
,
row_list
,
app
.
nm
)
if
not
pt
.
filled
():
raise
ProtocolError
(
'Partial partition table received'
)
# Install the partition table into the database for persistence.
...
...
@@ -44,7 +44,7 @@ class InitializationHandler(BaseMasterHandler):
logging
.
debug
(
'drop data for partitions %r'
,
unassigned
)
dm
.
dropPartitions
(
unassigned
)
dm
.
changePartitionTable
(
ptid
,
cell_list
,
reset
=
True
)
dm
.
changePartitionTable
(
ptid
,
num_replicas
,
cell_list
,
reset
=
True
)
dm
.
commit
()
def
truncate
(
self
,
conn
,
tid
):
...
...
@@ -68,7 +68,8 @@ class InitializationHandler(BaseMasterHandler):
def
askPartitionTable
(
self
,
conn
):
pt
=
self
.
app
.
pt
conn
.
answer
(
Packets
.
AnswerPartitionTable
(
pt
.
getID
(),
pt
.
getRowList
()))
conn
.
answer
(
Packets
.
AnswerPartitionTable
(
pt
.
getID
(),
pt
.
getReplicas
(),
pt
.
getRowList
()))
def
askLockedTransactions
(
self
,
conn
):
conn
.
answer
(
Packets
.
AnswerLockedTransactions
(
...
...
neo/storage/handlers/storage.py
View file @
8ba42463
...
...
@@ -212,7 +212,7 @@ class StorageOperationHandler(EventHandler):
# Sending such packet does not mark the connection
# for writing if there's too little data in the buffer.
conn
.
send
(
Packets
.
AddTransaction
(
tid
,
user
,
desc
,
ext
,
packed
,
ttid
,
oid_list
),
msg_id
)
desc
,
ext
,
bool
(
packed
)
,
ttid
,
oid_list
),
msg_id
)
# To avoid delaying several connections simultaneously,
# and also prevent the backend from scanning different
# parts of the DB at the same time, we ask the
...
...
@@ -248,7 +248,7 @@ class StorageOperationHandler(EventHandler):
for
serial
,
oid
in
object_list
:
oid_set
=
object_dict
.
get
(
serial
)
if
oid_set
:
if
type
(
oid_set
)
is
list
:
if
type
(
oid_set
)
is
tuple
:
object_dict
[
serial
]
=
oid_set
=
set
(
oid_set
)
if
oid
in
oid_set
:
oid_set
.
remove
(
oid
)
...
...
neo/storage/replicator.py
View file @
8ba42463
...
...
@@ -350,7 +350,7 @@ class Replicator(object):
try
:
conn
.
ask
(
Packets
.
RequestIdentification
(
NodeTypes
.
STORAGE
,
None
if
name
else
app
.
uuid
,
app
.
server
,
name
or
app
.
name
,
(),
app
.
id_timestamp
))
app
.
id_timestamp
,
(),
()
))
except
ConnectionClosed
:
if
previous_node
is
self
.
current_node
:
return
...
...
neo/storage/transactions.py
View file @
8ba42463
...
...
@@ -98,9 +98,12 @@ class TransactionManager(EventQueue):
self
.
_load_lock_dict
=
{}
self
.
_replicated
=
{}
self
.
_replicating
=
set
()
def
getPartition
(
self
,
oid
):
from
neo.lib.util
import
u64
np
=
app
.
pt
.
getPartitions
()
np
=
self
.
_
app
.
pt
.
getPartitions
()
self
.
getPartition
=
lambda
oid
:
u64
(
oid
)
%
np
return
self
.
getPartition
(
oid
)
def
discarded
(
self
,
offset_list
):
self
.
_replicating
.
difference_update
(
offset_list
)
...
...
neo/tests/__init__.py
View file @
8ba42463
...
...
@@ -21,6 +21,7 @@ import gc
import
os
import
random
import
socket
import
subprocess
import
sys
import
tempfile
import
unittest
...
...
@@ -41,7 +42,7 @@ from .mock import Mock
from
neo.lib
import
debug
,
logging
,
protocol
from
neo.lib.protocol
import
NodeTypes
,
Packets
,
UUID_NAMESPACES
from
neo.lib.util
import
cached_property
from
time
import
time
from
time
import
time
,
sleep
from
struct
import
pack
,
unpack
from
unittest.case
import
_ExpectedFailure
,
_UnexpectedSuccess
try
:
...
...
@@ -72,6 +73,9 @@ DB_ADMIN = os.getenv('NEO_DB_ADMIN', 'root')
DB_PASSWD
=
os
.
getenv
(
'NEO_DB_PASSWD'
,
''
)
DB_USER
=
os
.
getenv
(
'NEO_DB_USER'
,
'test'
)
DB_SOCKET
=
os
.
getenv
(
'NEO_DB_SOCKET'
,
''
)
DB_INSTALL
=
os
.
getenv
(
'NEO_DB_INSTALL'
,
'mysql_install_db'
)
DB_MYSQLD
=
os
.
getenv
(
'NEO_DB_MYSQLD'
,
'/usr/sbin/mysqld'
)
DB_MYCNF
=
os
.
getenv
(
'NEO_DB_MYCNF'
)
IP_VERSION_FORMAT_DICT
=
{
socket
.
AF_INET
:
'127.0.0.1'
,
...
...
@@ -134,8 +138,12 @@ def getTempDirectory():
print
'Using temp directory %r.'
%
temp_dir
return
temp_dir
def
setupMySQLdb
(
db_list
,
user
=
DB_USER
,
password
=
''
,
clear_databases
=
True
):
def
setupMySQLdb
(
db_list
,
clear_databases
=
True
):
if
mysql_pool
:
return
mysql_pool
.
setup
(
db_list
,
clear_databases
)
from
MySQLdb.constants.ER
import
BAD_DB_ERROR
user
=
DB_USER
password
=
''
kw
=
{
'unix_socket'
:
os
.
path
.
expanduser
(
DB_SOCKET
)}
if
DB_SOCKET
else
{}
conn
=
MySQLdb
.
connect
(
user
=
DB_ADMIN
,
passwd
=
DB_PASSWD
,
**
kw
)
cursor
=
conn
.
cursor
()
...
...
@@ -154,6 +162,88 @@ def setupMySQLdb(db_list, user=DB_USER, password='', clear_databases=True):
cursor
.
close
()
conn
.
commit
()
conn
.
close
()
return
'{}:{}@%s{}'
.
format
(
user
,
password
,
DB_SOCKET
).
__mod__
class
MySQLPool
(
object
):
def
__init__
(
self
,
pool_dir
=
None
):
self
.
_args
=
{}
self
.
_mysqld_dict
=
{}
if
not
pool_dir
:
pool_dir
=
getTempDirectory
()
self
.
_base
=
pool_dir
+
os
.
sep
self
.
_sock_template
=
os
.
path
.
join
(
pool_dir
,
'%s'
,
'mysql.sock'
)
def
__del__
(
self
):
self
.
kill
(
*
self
.
_mysqld_dict
)
def
setup
(
self
,
db_list
,
clear_databases
):
start_list
=
set
(
db_list
).
difference
(
self
.
_mysqld_dict
)
if
start_list
:
start_list
=
sorted
(
start_list
)
x
=
[]
with
open
(
os
.
devnull
,
'wb'
)
as
f
:
for
db
in
start_list
:
base
=
self
.
_base
+
db
datadir
=
os
.
path
.
join
(
base
,
'datadir'
)
sock
=
self
.
_sock_template
%
db
tmpdir
=
os
.
path
.
join
(
base
,
'tmp'
)
args
=
[
DB_INSTALL
,
'--defaults-file='
+
DB_MYCNF
,
'--datadir='
+
datadir
,
'--socket='
+
sock
,
'--tmpdir='
+
tmpdir
,
'--log_error='
+
os
.
path
.
join
(
base
,
'error.log'
)]
if
os
.
path
.
exists
(
datadir
):
try
:
os
.
remove
(
sock
)
except
OSError
,
e
:
if
e
.
errno
!=
errno
.
ENOENT
:
raise
else
:
os
.
makedirs
(
tmpdir
)
x
.
append
(
subprocess
.
Popen
(
args
,
stdout
=
f
,
stderr
=
subprocess
.
STDOUT
))
args
[
0
]
=
DB_MYSQLD
self
.
_args
[
db
]
=
args
for
x
in
x
:
x
=
x
.
wait
()
if
x
:
raise
subprocess
.
CalledProcessError
(
x
,
DB_INSTALL
)
self
.
start
(
*
start_list
)
for
db
in
start_list
:
sock
=
self
.
_sock_template
%
db
p
=
self
.
_mysqld_dict
[
db
]
while
not
os
.
path
.
exists
(
sock
):
sleep
(
1
)
x
=
p
.
poll
()
if
x
is
not
None
:
raise
subprocess
.
CalledProcessError
(
x
,
DB_MYSQLD
)
for
db
in
db_list
:
db
=
MySQLdb
.
connect
(
unix_socket
=
self
.
_sock_template
%
db
,
user
=
'root'
)
if
clear_databases
:
db
.
query
(
'DROP DATABASE IF EXISTS neo'
)
db
.
query
(
'CREATE DATABASE IF NOT EXISTS neo'
)
db
.
close
()
return
(
'root@neo'
+
self
.
_sock_template
).
__mod__
def
start
(
self
,
*
db
,
**
kw
):
assert
set
(
db
).
isdisjoint
(
self
.
_mysqld_dict
)
for
db
in
db
:
self
.
_mysqld_dict
[
db
]
=
subprocess
.
Popen
(
self
.
_args
[
db
],
**
kw
)
def
kill
(
self
,
*
db
):
processes
=
[]
for
db
in
db
:
p
=
self
.
_mysqld_dict
.
pop
(
db
)
processes
.
append
(
p
)
p
.
kill
()
for
p
in
processes
:
p
.
wait
()
mysql_pool
=
MySQLPool
()
if
DB_MYCNF
else
None
def
ImporterConfigParser
(
adapter
,
zodb
,
**
kw
):
cfg
=
SafeConfigParser
()
...
...
@@ -244,13 +334,15 @@ class NeoUnitTestBase(NeoTestBase):
""" create empty databases """
adapter
=
os
.
getenv
(
'NEO_TESTS_ADAPTER'
,
'MySQL'
)
if
adapter
==
'MySQL'
:
setupMySQLdb
([
prefix
+
str
(
i
)
for
i
in
xrange
(
number
)])
db_template
=
setupMySQLdb
(
[
prefix
+
str
(
i
)
for
i
in
xrange
(
number
)])
self
.
db_template
=
lambda
i
:
db_template
(
prefix
+
str
(
i
))
elif
adapter
==
'SQLite'
:
temp_dir
=
getTempDirectory
()
self
.
db_template
=
os
.
path
.
join
(
getTempDirectory
(),
prefix
+
'%s.sqlite'
).
__mod__
for
i
in
xrange
(
number
):
try
:
os
.
remove
(
os
.
path
.
join
(
temp_dir
,
'%s%s.sqlite'
%
(
prefix
,
i
)))
os
.
remove
(
self
.
db_template
(
i
))
except
OSError
,
e
:
if
e
.
errno
!=
errno
.
ENOENT
:
raise
...
...
@@ -274,21 +366,14 @@ class NeoUnitTestBase(NeoTestBase):
def
getStorageConfiguration
(
self
,
cluster
=
'main'
,
master_number
=
2
,
index
=
0
,
prefix
=
DB_PREFIX
,
uuid
=
None
):
assert
master_number
>=
1
and
master_number
<=
10
assert
index
>=
0
and
index
<=
9
masters
=
[(
buildUrlFromString
(
self
.
local_ip
),
10010
+
i
)
for
i
in
xrange
(
master_number
)]
adapter
=
os
.
getenv
(
'NEO_TESTS_ADAPTER'
,
'MySQL'
)
if
adapter
==
'MySQL'
:
db
=
'%s@%s%s%s'
%
(
DB_USER
,
prefix
,
index
,
DB_SOCKET
)
elif
adapter
==
'SQLite'
:
db
=
os
.
path
.
join
(
getTempDirectory
(),
'test_neo%s.sqlite'
%
index
)
else
:
assert
False
,
adapter
return
{
'cluster'
:
cluster
,
'bind'
:
(
masters
[
0
],
10020
+
index
),
'masters'
:
masters
,
'database'
:
db
,
'database'
:
self
.
db_template
(
index
)
,
'uuid'
:
uuid
,
'adapter'
:
adapter
,
'wait'
:
0
,
...
...
neo/tests/functional/__init__.py
View file @
8ba42463
...
...
@@ -36,7 +36,7 @@ from neo.lib import logging
from
neo.lib.protocol
import
ClusterStates
,
NodeTypes
,
CellStates
,
NodeStates
,
\
UUID_NAMESPACES
from
neo.lib.util
import
dump
,
setproctitle
from
..
import
(
ADDRESS_TYPE
,
DB_SOCKET
,
DB_USER
,
IP_VERSION_FORMAT_DICT
,
SSL
,
from
..
import
(
ADDRESS_TYPE
,
IP_VERSION_FORMAT_DICT
,
SSL
,
buildUrlFromString
,
cluster
,
getTempDirectory
,
setupMySQLdb
,
ImporterConfigParser
,
NeoTestBase
,
Patch
)
from
neo.client.Storage
import
Storage
...
...
@@ -282,7 +282,7 @@ class NEOProcess(Process):
def
_args
(
self
):
args
=
super
(
NEOProcess
,
self
).
_args
()
if
self
.
uuid
:
args
[:
0
]
=
'--
uu
id'
,
str
(
self
.
uuid
)
args
[:
0
]
=
'--
n
id'
,
str
(
self
.
uuid
)
return
args
def
run
(
self
):
...
...
@@ -306,11 +306,11 @@ class NEOCluster(object):
SSL
=
None
def
__init__
(
self
,
db_list
,
master_count
=
1
,
partitions
=
1
,
replicas
=
0
,
db_user
=
DB_USER
,
db_password
=
''
,
name
=
None
,
name
=
None
,
cleanup_on_delete
=
False
,
temp_dir
=
None
,
clear_databases
=
True
,
adapter
=
os
.
getenv
(
'NEO_TESTS_ADAPTER'
),
address_type
=
ADDRESS_TYPE
,
bind_ip
=
None
,
logger
=
True
,
importer
=
None
):
importer
=
None
,
storage_kw
=
{}
):
if
not
adapter
:
adapter
=
'MySQL'
self
.
adapter
=
adapter
...
...
@@ -322,20 +322,28 @@ class NEOCluster(object):
temp_dir
=
tempfile
.
mkdtemp
(
prefix
=
'neo_'
)
print
'Using temp directory '
+
temp_dir
if
adapter
==
'MySQL'
:
self
.
db_user
=
db_user
self
.
db_password
=
db_password
self
.
db_template
=
(
'%s:%s@%%s%s'
%
(
db_user
,
db_password
,
DB_SOCKET
)).
__mod__
self
.
db_template
=
setupMySQLdb
(
db_list
,
clear_databases
)
elif
adapter
==
'SQLite'
:
self
.
db_template
=
(
lambda
t
:
lambda
db
:
':memory:'
if
db
is
None
else
db
if
os
.
sep
in
db
else
t
%
db
)(
os
.
path
.
join
(
temp_dir
,
'%s.sqlite'
))
if
clear_databases
:
for
db
in
self
.
db_list
:
if
db
is
None
:
continue
db
=
self
.
db_template
(
db
)
try
:
os
.
remove
(
db
)
except
OSError
,
e
:
if
e
.
errno
!=
errno
.
ENOENT
:
raise
else
:
logging
.
debug
(
'%r deleted'
,
db
)
else
:
assert
False
,
adapter
self
.
address_type
=
address_type
self
.
local_ip
=
local_ip
=
bind_ip
or
\
IP_VERSION_FORMAT_DICT
[
self
.
address_type
]
self
.
setupDB
(
clear_databases
)
if
importer
:
cfg
=
ImporterConfigParser
(
adapter
,
**
importer
)
cfg
.
set
(
"neo"
,
"database"
,
self
.
db_template
(
*
db_list
))
...
...
@@ -364,7 +372,8 @@ class NEOCluster(object):
# create storage nodes
for
i
,
db
in
enumerate
(
db_list
):
self
.
_newProcess
(
NodeTypes
.
STORAGE
,
logger
and
'storage_%u'
%
i
,
0
,
adapter
=
adapter
,
database
=
self
.
db_template
(
db
))
0
,
adapter
=
adapter
,
database
=
self
.
db_template
(
db
),
**
storage_kw
)
# create neoctl
self
.
neoctl
=
NeoCTL
((
self
.
local_ip
,
admin_port
),
ssl
=
self
.
SSL
)
...
...
@@ -382,23 +391,10 @@ class NEOCluster(object):
self
.
process_dict
.
setdefault
(
node_type
,
[]).
append
(
NEOProcess
(
command_dict
[
node_type
],
uuid
=
uuid
,
**
kw
))
def
setupDB
(
self
,
clear_databases
=
True
):
if
self
.
adapter
==
'MySQL'
:
setupMySQLdb
(
self
.
db_list
,
self
.
db_user
,
self
.
db_password
,
clear_databases
)
elif
self
.
adapter
==
'SQLite'
:
if
clear_databases
:
for
db
in
self
.
db_list
:
if
db
is
None
:
continue
db
=
self
.
db_template
(
db
)
try
:
os
.
remove
(
db
)
except
OSError
,
e
:
if
e
.
errno
!=
errno
.
ENOENT
:
raise
else
:
logging
.
debug
(
'%r deleted'
,
db
)
def
resetDB
(
self
):
for
db
in
self
.
db_list
:
dm
=
buildDatabaseManager
(
self
.
adapter
,
(
self
.
db_template
(
db
),))
dm
.
setup
(
True
)
def
run
(
self
,
except_storages
=
()):
""" Start cluster processes except some storage nodes """
...
...
@@ -437,7 +433,7 @@ class NEOCluster(object):
pending_count
+=
1
if
pending_count
==
target
[
0
]:
neoctl
.
startCluster
()
except
(
NotReadyException
,
RuntimeError
):
except
(
NotReadyException
,
SystemExit
):
pass
if
not
pdb
.
wait
(
test
,
MAX_START_TIME
):
raise
AssertionError
(
'Timeout when starting cluster'
)
...
...
@@ -449,7 +445,7 @@ class NEOCluster(object):
def
start
(
last_try
):
try
:
self
.
neoctl
.
startCluster
()
except
(
NotReadyException
,
RuntimeError
),
e
:
except
(
NotReadyException
,
SystemExit
),
e
:
return
False
,
e
return
True
,
None
self
.
expectCondition
(
start
)
...
...
@@ -653,10 +649,10 @@ class NEOCluster(object):
def
expectOudatedCells
(
self
,
number
,
*
args
,
**
kw
):
def
callback
(
last_try
):
row_list
=
self
.
neoctl
.
getPartitionRowList
()[
1
]
row_list
=
self
.
neoctl
.
getPartitionRowList
()[
2
]
number_of_outdated
=
0
for
row
in
row_list
:
for
cell
in
row
[
1
]
:
for
cell
in
row
:
if
cell
[
1
]
==
CellStates
.
OUT_OF_DATE
:
number_of_outdated
+=
1
return
number_of_outdated
==
number
,
number_of_outdated
...
...
@@ -664,10 +660,10 @@ class NEOCluster(object):
def
expectAssignedCells
(
self
,
process
,
number
,
*
args
,
**
kw
):
def
callback
(
last_try
):
row_list
=
self
.
neoctl
.
getPartitionRowList
()[
1
]
row_list
=
self
.
neoctl
.
getPartitionRowList
()[
2
]
assigned_cells_number
=
0
for
row
in
row_list
:
for
cell
in
row
[
1
]
:
for
cell
in
row
:
if
cell
[
0
]
==
process
.
getUUID
():
assigned_cells_number
+=
1
return
assigned_cells_number
==
number
,
assigned_cells_number
...
...
neo/tests/functional/testClient.py
View file @
8ba42463
...
...
@@ -62,8 +62,6 @@ class ClientTests(NEOFunctionalTest):
NEOFunctionalTest
.
_tearDown
(
self
,
success
)
def
__setup
(
self
):
# start cluster
self
.
neo
.
setupDB
()
self
.
neo
.
start
()
self
.
neo
.
expectClusterRunning
()
self
.
db
=
ZODB
.
DB
(
self
.
neo
.
getZODBStorage
())
...
...
neo/tests/functional/testCluster.py
View file @
8ba42463
...
...
@@ -71,7 +71,6 @@ class ClusterTests(NEOFunctionalTest):
def
testClusterBreaks
(
self
):
self
.
neo
=
NEOCluster
([
'test_neo1'
],
master_count
=
1
,
temp_dir
=
self
.
getTempDirectory
())
self
.
neo
.
setupDB
()
self
.
neo
.
start
()
self
.
neo
.
expectClusterRunning
()
self
.
neo
.
expectOudatedCells
(
number
=
0
)
...
...
@@ -82,7 +81,6 @@ class ClusterTests(NEOFunctionalTest):
self
.
neo
=
NEOCluster
([
'test_neo1'
,
'test_neo2'
],
partitions
=
2
,
master_count
=
1
,
replicas
=
0
,
temp_dir
=
self
.
getTempDirectory
())
self
.
neo
.
setupDB
()
self
.
neo
.
start
()
self
.
neo
.
expectClusterRunning
()
self
.
neo
.
expectOudatedCells
(
number
=
0
)
...
...
@@ -93,7 +91,6 @@ class ClusterTests(NEOFunctionalTest):
self
.
neo
=
NEOCluster
([
'test_neo1'
,
'test_neo2'
],
partitions
=
2
,
replicas
=
1
,
master_count
=
1
,
temp_dir
=
self
.
getTempDirectory
())
self
.
neo
.
setupDB
()
self
.
neo
.
start
()
self
.
neo
.
expectClusterRunning
()
self
.
neo
.
expectOudatedCells
(
number
=
0
)
...
...
neo/tests/functional/testMaster.py
View file @
8ba42463
...
...
@@ -47,7 +47,7 @@ class MasterTests(NEOFunctionalTest):
break
neoctl
.
killNode
(
uuid
)
self
.
neo
.
expectDead
(
master
)
self
.
assertRaises
(
RuntimeError
,
neoctl
.
killNode
,
primary_uuid
)
self
.
assertRaises
(
SystemExit
,
neoctl
.
killNode
,
primary_uuid
)
def
testStoppingPrimaryWithTwoSecondaries
(
self
):
# Wait for masters to stabilize
...
...
neo/tests/functional/testStorage.py
View file @
8ba42463
...
...
@@ -172,7 +172,7 @@ class StorageTests(NEOFunctionalTest):
self
.
neo
.
expectOudatedCells
(
2
)
self
.
neo
.
expectClusterRunning
()
self
.
assertRaises
(
RuntimeError
,
self
.
neo
.
neoctl
.
killNode
,
self
.
assertRaises
(
SystemExit
,
self
.
neo
.
neoctl
.
killNode
,
started
[
1
].
getUUID
())
started
[
1
].
stop
()
# Cluster not operational anymore. Only cells of second storage that
...
...
@@ -323,7 +323,7 @@ class StorageTests(NEOFunctionalTest):
self
.
neo
.
expectStorageUnknown
(
started
[
0
])
self
.
neo
.
expectAssignedCells
(
started
[
0
],
0
)
self
.
neo
.
expectAssignedCells
(
started
[
1
],
10
)
self
.
assertRaises
(
RuntimeError
,
self
.
neo
.
neoctl
.
dropNode
,
self
.
assertRaises
(
SystemExit
,
self
.
neo
.
neoctl
.
dropNode
,
started
[
1
].
getUUID
())
self
.
neo
.
expectClusterRunning
()
...
...
neo/tests/master/testClientHandler.py
View file @
8ba42463
...
...
@@ -30,8 +30,6 @@ class MasterClientHandlerTests(NeoUnitTestBase):
config
=
self
.
getMasterConfiguration
(
master_number
=
1
,
replicas
=
1
)
self
.
app
=
Application
(
config
)
self
.
app
.
em
.
close
()
self
.
app
.
pt
.
clear
()
self
.
app
.
pt
.
setID
(
1
)
self
.
app
.
em
=
Mock
()
self
.
app
.
loid
=
'
\
0
'
*
8
self
.
app
.
tm
.
setLastTID
(
'
\
0
'
*
8
)
...
...
@@ -73,7 +71,7 @@ class MasterClientHandlerTests(NeoUnitTestBase):
self
.
app
.
nm
.
getByUUID
(
storage_uuid
).
setConnection
(
storage_conn
)
self
.
service
.
askPack
(
conn
,
tid
)
self
.
checkNoPacketSent
(
conn
)
ptid
=
self
.
checkAskPacket
(
storage_conn
,
Packets
.
AskPack
).
decode
()
[
0
]
ptid
=
self
.
checkAskPacket
(
storage_conn
,
Packets
.
AskPack
).
_args
[
0
]
self
.
assertEqual
(
ptid
,
tid
)
self
.
assertTrue
(
self
.
app
.
packing
[
0
]
is
conn
)
self
.
assertEqual
(
self
.
app
.
packing
[
1
],
peer_id
)
...
...
@@ -85,7 +83,7 @@ class MasterClientHandlerTests(NeoUnitTestBase):
self
.
app
.
nm
.
getByUUID
(
storage_uuid
).
setConnection
(
storage_conn
)
self
.
service
.
askPack
(
conn
,
tid
)
self
.
checkNoPacketSent
(
storage_conn
)
status
=
self
.
checkAnswerPacket
(
conn
,
Packets
.
AnswerPack
).
decode
()
[
0
]
status
=
self
.
checkAnswerPacket
(
conn
,
Packets
.
AnswerPack
).
_args
[
0
]
self
.
assertFalse
(
status
)
if
__name__
==
'__main__'
:
...
...
neo/tests/master/testMasterApp.py
View file @
8ba42463
...
...
@@ -26,7 +26,6 @@ class MasterAppTests(NeoUnitTestBase):
# create an application object
config
=
self
.
getMasterConfiguration
()
self
.
app
=
Application
(
config
)
self
.
app
.
pt
.
clear
()
def
_tearDown
(
self
,
success
):
self
.
app
.
close
()
...
...
neo/tests/master/testMasterPT.py
View file @
8ba42463
...
...
@@ -289,7 +289,9 @@ class MasterPartitionTableTests(NeoUnitTestBase):
pt
.
addNodeList
(
sn
[
1
:
3
])
self
.
assertPartitionTable
(
pt
,
'U..|U..|U..|U..|U..|U..|U..'
)
self
.
update
(
pt
,
self
.
tweak
(
pt
,
sn
[:
1
]))
self
.
assertPartitionTable
(
pt
,
'.U.|..U|.U.|..U|.U.|..U|.U.'
)
# See note in PartitionTable.tweak() about drop_list.
#self.assertPartitionTable(pt,'.U.|..U|.U.|..U|.U.|..U|.U.')
self
.
assertPartitionTable
(
pt
,
'UU.|U.U|UU.|U.U|UU.|U.U|UU.'
)
def
test_18_tweakBigPT
(
self
):
seed
=
repr
(
time
.
time
())
...
...
neo/tests/master/testRecovery.py
deleted
100644 → 0
View file @
a33c624c
#
# Copyright (C) 2009-2019 Nexedi SA
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; either version 2
# of the License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import
unittest
from
..
import
NeoUnitTestBase
from
neo.lib.protocol
import
NodeTypes
,
NodeStates
,
CellStates
from
neo.master.recovery
import
RecoveryManager
from
neo.master.app
import
Application
class
MasterRecoveryTests
(
NeoUnitTestBase
):
def
setUp
(
self
):
NeoUnitTestBase
.
setUp
(
self
)
# create an application object
config
=
self
.
getMasterConfiguration
()
self
.
app
=
Application
(
config
)
self
.
app
.
pt
.
clear
()
self
.
recovery
=
RecoveryManager
(
self
.
app
)
self
.
app
.
unconnected_master_node_set
=
set
()
self
.
app
.
negotiating_master_node_set
=
set
()
for
node
in
self
.
app
.
nm
.
getMasterList
():
self
.
app
.
unconnected_master_node_set
.
add
(
node
.
getAddress
())
node
.
setState
(
NodeStates
.
RUNNING
)
# define some variable to simulate client and storage node
self
.
storage_port
=
10021
self
.
master_port
=
10011
def
_tearDown
(
self
,
success
):
self
.
app
.
close
()
NeoUnitTestBase
.
_tearDown
(
self
,
success
)
# Common methods
def
identifyToMasterNode
(
self
,
node_type
=
NodeTypes
.
STORAGE
,
ip
=
"127.0.0.1"
,
port
=
10021
):
"""Do first step of identification to MN
"""
address
=
(
ip
,
port
)
uuid
=
self
.
getNewUUID
(
node_type
)
self
.
app
.
nm
.
createFromNodeType
(
node_type
,
address
=
address
,
uuid
=
uuid
,
state
=
NodeStates
.
RUNNING
)
return
uuid
# Tests
def
test_10_answerPartitionTable
(
self
):
# XXX: This test does much less that it seems, because all 'for' loops
# iterate over empty lists. Currently, only testRecovery covers
# some paths in NodeManager._createNode: apart from that, we could
# delete it entirely.
recovery
=
self
.
recovery
uuid
=
self
.
identifyToMasterNode
(
NodeTypes
.
MASTER
,
port
=
self
.
master_port
)
# not from target node, ignore
uuid
=
self
.
identifyToMasterNode
(
NodeTypes
.
STORAGE
,
port
=
self
.
storage_port
)
conn
=
self
.
getFakeConnection
(
uuid
,
self
.
storage_port
)
node
=
self
.
app
.
nm
.
getByUUID
(
conn
.
getUUID
())
offset
=
1
cell_list
=
[(
offset
,
uuid
,
CellStates
.
UP_TO_DATE
)]
cells
=
self
.
app
.
pt
.
getRow
(
offset
)
for
cell
,
state
in
cells
:
self
.
assertEqual
(
state
,
CellStates
.
OUT_OF_DATE
)
recovery
.
target_ptid
=
2
node
.
setPending
()
recovery
.
answerPartitionTable
(
conn
,
1
,
cell_list
)
cells
=
self
.
app
.
pt
.
getRow
(
offset
)
for
cell
,
state
in
cells
:
self
.
assertEqual
(
state
,
CellStates
.
OUT_OF_DATE
)
# from target node, taken into account
conn
=
self
.
getFakeConnection
(
uuid
,
self
.
storage_port
)
offset
=
1
cell_list
=
[(
offset
,
((
uuid
,
CellStates
.
UP_TO_DATE
,),),)]
cells
=
self
.
app
.
pt
.
getRow
(
offset
)
for
cell
,
state
in
cells
:
self
.
assertEqual
(
state
,
CellStates
.
OUT_OF_DATE
)
node
.
setPending
()
recovery
.
answerPartitionTable
(
conn
,
None
,
cell_list
)
cells
=
self
.
app
.
pt
.
getRow
(
offset
)
for
cell
,
state
in
cells
:
self
.
assertEqual
(
state
,
CellStates
.
UP_TO_DATE
)
# give a bad offset, must send error
self
.
recovery
.
target_uuid
=
uuid
conn
=
self
.
getFakeConnection
(
uuid
,
self
.
storage_port
)
offset
=
1000000
self
.
assertFalse
(
self
.
app
.
pt
.
hasOffset
(
offset
))
cell_list
=
[(
offset
,
((
uuid
,
NodeStates
.
UNKNOWN
,),),)]
node
.
setPending
()
self
.
checkProtocolErrorRaised
(
recovery
.
answerPartitionTable
,
conn
,
2
,
cell_list
)
if
__name__
==
'__main__'
:
unittest
.
main
()
neo/tests/master/testStorageHandler.py
View file @
8ba42463
...
...
@@ -18,8 +18,8 @@ import unittest
from
..mock
import
Mock
from
..
import
NeoUnitTestBase
from
neo.lib.protocol
import
NodeTypes
,
Packets
from
neo.master.handlers.storage
import
StorageServiceHandler
from
neo.master.app
import
Application
from
neo.master.handlers.storage
import
StorageServiceHandler
class
MasterStorageHandlerTests
(
NeoUnitTestBase
):
...
...
@@ -29,7 +29,6 @@ class MasterStorageHandlerTests(NeoUnitTestBase):
config
=
self
.
getMasterConfiguration
(
master_number
=
1
,
replicas
=
1
)
self
.
app
=
Application
(
config
)
self
.
app
.
em
.
close
()
self
.
app
.
pt
.
clear
()
self
.
app
.
em
=
Mock
()
self
.
service
=
StorageServiceHandler
(
self
.
app
)
...
...
@@ -73,7 +72,7 @@ class MasterStorageHandlerTests(NeoUnitTestBase):
self
.
service
.
answerPack
(
conn2
,
False
)
packet
=
self
.
checkNotifyPacket
(
client_conn
,
Packets
.
AnswerPack
)
# TODO: verify packet peer id
self
.
assertTrue
(
packet
.
decode
()
[
0
])
self
.
assertTrue
(
packet
.
_args
[
0
])
self
.
assertEqual
(
self
.
app
.
packing
,
None
)
if
__name__
==
'__main__'
:
...
...
neo/tests/protocol
0 → 100644
View file @
8ba42463
# generated by running the whole test suite with -p
AbortTransaction(p64,[int])
AcceptIdentification(NodeTypes,?int,?int)
AddObject(p64,p64,int,bin,bin,?p64)
AddPendingNodes([int])
AddTransaction(p64,bin,bin,bin,bool,p64,[p64])
AnswerBeginTransaction(p64)
AnswerCheckCurrentSerial(?p64)
AnswerCheckSerialRange(int,bin,p64,bin,p64)
AnswerCheckTIDRange(int,bin,p64)
AnswerClusterState(?ClusterStates)
AnswerFetchObjects(?,?p64,?p64,{:})
AnswerFetchTransactions(?,?p64,[])
AnswerFinalTID(p64)
AnswerInformationLocked(p64)
AnswerLastIDs(?p64,?p64)
AnswerLastTransaction(p64)
AnswerLockedTransactions({p64:?p64})
AnswerNewOIDs([p64])
AnswerNodeList([(NodeTypes,?(bin,int),?int,NodeStates,?float)])
AnswerObject(p64,p64,?p64,?int,bin,bin,?p64)
AnswerObjectHistory(p64,[(p64,int)])
AnswerObjectUndoSerial({p64:(p64,?p64,bool)})
AnswerPack(bool)
AnswerPartitionList(int,int,[[(int,CellStates)]])
AnswerPartitionTable(int,int,[[(int,CellStates)]])
AnswerPrimary(int)
AnswerRebaseObject(?(p64,p64,?(int,bin,bin)))
AnswerRebaseTransaction([p64])
AnswerRecovery(?int,?p64,?p64)
AnswerStoreObject(?p64)
AnswerStoreTransaction()
AnswerTIDs([p64])
AnswerTIDsFrom([p64])
AnswerTransactionFinished(p64,p64)
AnswerTransactionInformation(p64,bin,bin,bin,bool,[p64])
AnswerTweakPartitionTable(bool,[[(int,CellStates)]])
AnswerUnfinishedTransactions(p64,[p64])
AnswerVoteTransaction()
AskBeginTransaction(?p64)
AskCheckCurrentSerial(p64,p64,p64)
AskCheckSerialRange(int,int,p64,p64,p64)
AskCheckTIDRange(int,int,p64,p64)
AskClusterState()
AskFetchObjects(int,int,p64,p64,p64,{p64:[p64]})
AskFetchTransactions(int,int,p64,p64,[p64])
AskFinalTID(p64)
AskFinishTransaction(p64,[p64],[p64])
AskLastIDs()
AskLastTransaction()
AskLockInformation(p64,p64)
AskLockedTransactions()
AskNewOIDs(int)
AskNodeList(NodeTypes)
AskObject(p64,?p64,?p64)
AskObjectHistory(p64,int,int)
AskObjectUndoSerial(p64,p64,p64,[p64])
AskPack(p64)
AskPartitionList(int,int,?)
AskPartitionTable()
AskPrimary()
AskRebaseObject(p64,p64)
AskRebaseTransaction(p64,p64)
AskRecovery()
AskStoreObject(p64,p64,int,bin,bin,?p64,?p64)
AskStoreTransaction(p64,bin,bin,bin,[p64])
AskTIDs(int,int,int)
AskTIDsFrom(p64,p64,int,int)
AskTransactionInformation(p64)
AskUnfinishedTransactions([int])
AskVoteTransaction(p64)
CheckPartition(int,(bin,?(bin,int)),p64,p64)
CheckReplicas({int:?int},p64,?)
Error(int,bin)
FailedVote(p64,[int])
InvalidateObjects(p64,[p64])
NotPrimaryMaster(?int,[(bin,int)])
NotifyClusterInformation(ClusterStates)
NotifyDeadlock(p64,p64)
NotifyNodeInformation(float,[(NodeTypes,?(bin,int),?int,NodeStates,?float)])
NotifyPartitionChanges(int,int,[(int,int,CellStates)])
NotifyPartitionCorrupted(int,[int])
NotifyReady()
NotifyRepair(bool)
NotifyReplicationDone(int,p64)
NotifyTransactionFinished(p64,p64)
NotifyUnlockInformation(p64)
Ping()
Pong()
Repair([int],bool)
Replicate(p64,bin,{int:?(bin,int)})
RequestIdentification(NodeTypes,?int,?(bin,int),bin,?float,any,[int])
SendPartitionTable(?int,int,[[(int,CellStates)]])
SetClusterState(ClusterStates)
SetNodeState(int,NodeStates)
SetNumReplicas(int)
StartOperation(bool)
StopOperation()
Truncate(p64)
TweakPartitionTable(bool,[int])
ValidateTransaction(p64,p64)
neo/tests/protocol_checker.py
0 → 100644
View file @
8ba42463
# The use of ast is convoluted, and the result quite verbose,
# but that remains simpler than writing a parser from scratch.
import
ast
,
os
from
contextlib
import
contextmanager
from
neo.lib.protocol
import
Packet
,
Enum
array
=
list
,
set
,
tuple
item
=
Enum
.
Item
class
_ast
(
object
):
def
__getattr__
(
self
,
k
):
v
=
lambda
*
args
:
getattr
(
ast
,
k
)(
lineno
=
0
,
col_offset
=
0
,
*
args
)
setattr
(
self
,
k
,
v
)
return
v
_ast
=
_ast
()
class
parseArgument
(
ast
.
NodeTransformer
):
def
visit_UnaryOp
(
self
,
node
):
assert
isinstance
(
node
.
op
,
ast
.
USub
)
return
_ast
.
Call
(
_ast
.
Name
(
'option'
,
ast
.
Load
()),
[
self
.
visit
(
node
.
operand
)],
[],
None
,
None
)
def
visit_Name
(
self
,
node
):
return
_ast
.
Str
(
node
.
id
.
replace
(
'_'
,
'?'
))
parseArgument
=
parseArgument
().
visit
class
Argument
(
object
):
merge
=
True
type
=
''
option
=
False
@
classmethod
def
load
(
cls
,
arg
):
arg
=
ast
.
parse
(
arg
.
rstrip
()
.
replace
(
'?('
,
'-('
).
replace
(
'?['
,
'-['
).
replace
(
'?{'
,
'-{'
)
.
replace
(
'?'
,
'_'
).
replace
(
'[]'
,
'[""]'
)
.
replace
(
'{:'
,
'{"":'
).
replace
(
':}'
,
':""}'
),
mode
=
"eval"
)
x
=
arg
.
body
name
=
x
.
func
.
id
arg
.
body
=
parseArgument
(
_ast
.
Tuple
(
x
.
args
,
ast
.
Load
()))
return
name
,
cls
.
_load
(
eval
(
compile
(
arg
,
''
,
mode
=
"eval"
),
{
'option'
:
cls
.
_option
}))
@
classmethod
def
_load
(
cls
,
arg
):
t
=
type
(
arg
)
if
t
is
cls
:
return
arg
x
=
object
.
__new__
(
cls
)
if
t
is
tuple
:
x
.
type
=
map
(
cls
.
_load
,
arg
)
elif
t
is
list
:
x
.
type
=
cls
.
_load
(
*
arg
),
elif
t
is
dict
:
(
k
,
v
),
=
arg
.
iteritems
()
x
.
type
=
cls
.
_load
(
k
),
cls
.
_load
(
v
)
else
:
if
arg
.
startswith
(
'?'
):
arg
=
arg
[
1
:]
x
.
option
=
True
x
.
type
=
arg
return
x
@
classmethod
def
_option
(
cls
,
arg
):
arg
=
cls
.
_load
(
arg
)
arg
.
option
=
True
return
arg
@
classmethod
def
_merge
(
cls
,
args
):
if
args
:
x
,
=
{
cls
(
x
)
for
x
in
args
}
return
x
return
object
.
__new__
(
cls
)
def
__init__
(
self
,
arg
,
root
=
False
):
if
arg
is
None
:
self
.
option
=
True
elif
isinstance
(
arg
,
tuple
)
and
(
root
or
len
(
arg
)
>
1
):
self
.
type
=
map
(
self
.
__class__
,
arg
)
elif
isinstance
(
arg
,
array
):
self
.
type
=
self
.
_merge
(
arg
),
elif
isinstance
(
arg
,
dict
):
self
.
type
=
self
.
_merge
(
arg
),
self
.
_merge
(
arg
.
values
())
else
:
self
.
type
=
((
'p64'
if
len
(
arg
)
==
8
else
'bin'
)
if
isinstance
(
arg
,
bytes
)
else
arg
.
_enum
.
_name
if
isinstance
(
arg
,
item
)
else
'str'
if
isinstance
(
arg
,
unicode
)
else
'int'
if
isinstance
(
arg
,
long
)
else
type
(
arg
).
__name__
)
def
__repr__
(
self
):
x
=
self
.
type
if
type
(
x
)
is
tuple
:
x
=
(
'[%s]'
if
len
(
x
)
==
1
else
'{%s:%s}'
)
%
x
elif
type
(
x
)
is
list
:
x
=
'(%s)'
%
','
.
join
(
map
(
repr
,
x
))
return
'?'
+
x
if
self
.
option
else
x
def
__hash__
(
self
):
return
0
def
__eq__
(
self
,
other
):
x
=
self
.
type
y
=
other
.
type
if
x
and
y
and
x
!=
'any'
:
# Since we don't know whether an array is fixed-size record of
# heterogeneous values or a collection of homogeneous values,
# we end up with the following complicated heuristic.
t
=
type
(
x
)
if
t
is
tuple
:
if
len
(
x
)
==
1
and
type
(
y
)
is
list
:
z
=
set
(
x
)
z
.
update
(
y
)
if
len
(
z
)
==
1
:
x
=
y
=
tuple
(
z
)
if
self
.
merge
:
self
.
type
=
x
elif
t
is
list
:
if
type
(
y
)
is
tuple
and
len
(
y
)
==
1
:
z
=
set
(
y
)
z
.
update
(
x
)
if
len
(
z
)
==
1
:
x
=
y
=
tuple
(
z
)
if
self
.
merge
:
self
.
type
=
x
t
=
tuple
elif
t
is
str
is
type
(
y
)
and
{
x
,
y
}.
issuperset
((
'bin'
,
'p64'
)):
x
=
y
=
'bin'
if
self
.
merge
:
self
.
type
=
x
if
not
(
t
is
type
(
y
)
and
(
t
is
not
tuple
or
len
(
x
)
==
len
(
y
))
and
x
==
y
):
if
not
self
.
merge
:
return
False
self
.
type
=
'any'
if
self
.
merge
:
if
not
x
:
self
.
type
=
y
if
not
self
.
option
:
self
.
option
=
other
.
option
elif
y
and
not
x
or
other
.
option
and
not
self
.
option
:
return
False
return
True
class
FrozenArgument
(
Argument
):
merge
=
False
@
contextmanager
def
protocolChecker
(
dump
):
x
=
'Packet(p64,?[(bin,{int:})],{:?(?,[])},?{?:float})'
assert
x
==
'%s%r'
%
Argument
.
load
(
x
)
assert
not
(
FrozenArgument
([])
==
Argument
([
0
]))
path
=
os
.
path
.
join
(
os
.
path
.
dirname
(
__file__
),
'protocol'
)
if
dump
:
import
threading
from
multiprocessing
import
Lock
lock
=
Lock
()
schema
=
{}
pid
=
os
.
getpid
()
r
,
w
=
os
.
pipe
()
def
_check
(
name
,
arg
):
try
:
schema
[
name
]
==
arg
except
KeyError
:
schema
[
name
]
=
arg
def
check
(
name
,
args
):
arg
=
Argument
(
args
,
True
)
if
pid
==
os
.
getpid
():
_check
(
name
,
arg
)
else
:
with
lock
:
os
.
write
(
w
,
'%s%r
\
n
'
%
(
name
,
arg
))
def
check_thread
(
r
):
for
x
in
os
.
fdopen
(
r
):
_check
(
*
Argument
.
load
(
x
))
check_thread
=
threading
.
Thread
(
target
=
check_thread
,
args
=
(
r
,))
check_thread
.
daemon
=
True
check_thread
.
start
()
else
:
with
open
(
path
)
as
p
:
x
=
p
.
readline
()
assert
x
[
0
]
==
'#'
,
x
schema
=
dict
(
map
(
FrozenArgument
.
load
,
p
))
def
check
(
name
,
args
):
arg
=
Argument
(
args
,
True
)
if
not
(
None
is
not
schema
.
get
(
name
)
==
arg
):
raise
Exception
(
'invalid packet: %s%r'
%
(
name
,
arg
))
w
=
None
Packet_encode
=
Packet
.
__dict__
[
'encode'
]
def
encode
(
packet
):
check
(
type
(
packet
).
__name__
,
packet
.
_args
)
return
Packet_encode
(
packet
)
Packet
.
encode
=
encode
try
:
yield
finally
:
Packet
.
encode
=
Packet_encode
if
w
:
os
.
close
(
w
)
check_thread
.
join
()
if
dump
:
with
open
(
path
,
'w'
)
as
p
:
p
.
write
(
'# generated by running the whole test suite with -p
\
n
'
)
for
x
in
sorted
(
schema
.
iteritems
()):
p
.
write
(
'%s%r
\
n
'
%
x
)
neo/tests/storage/testMasterHandler.py
View file @
8ba42463
...
...
@@ -56,7 +56,7 @@ class StorageMasterHandlerTests(NeoUnitTestBase):
self
.
app
.
pt
=
Mock
({
'getID'
:
1
})
count
=
len
(
self
.
app
.
nm
.
getList
())
self
.
assertRaises
(
ProtocolError
,
self
.
operation
.
notifyPartitionChanges
,
conn
,
0
,
())
conn
,
0
,
0
,
())
self
.
assertEqual
(
self
.
app
.
pt
.
getID
(),
1
)
self
.
assertEqual
(
len
(
self
.
app
.
nm
.
getList
()),
count
)
calls
=
self
.
app
.
replicator
.
mockGetNamedCalls
(
'removePartition'
)
...
...
@@ -84,13 +84,13 @@ class StorageMasterHandlerTests(NeoUnitTestBase):
ptid
=
2
app
.
dm
=
Mock
({
})
app
.
replicator
=
Mock
({})
self
.
operation
.
notifyPartitionChanges
(
conn
,
ptid
,
cells
)
self
.
operation
.
notifyPartitionChanges
(
conn
,
ptid
,
1
,
cells
)
# ptid set
self
.
assertEqual
(
app
.
pt
.
getID
(),
ptid
)
# dm call
calls
=
self
.
app
.
dm
.
mockGetNamedCalls
(
'changePartitionTable'
)
self
.
assertEqual
(
len
(
calls
),
1
)
calls
[
0
].
checkArgs
(
ptid
,
cells
)
calls
[
0
].
checkArgs
(
ptid
,
1
,
cells
)
if
__name__
==
"__main__"
:
unittest
.
main
()
neo/tests/storage/testStorageApp.py
deleted
100644 → 0
View file @
a33c624c
#
# Copyright (C) 2009-2019 Nexedi SA
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; either version 2
# of the License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import
unittest
from
..mock
import
Mock
from
..
import
NeoUnitTestBase
from
neo.storage.app
import
Application
from
neo.lib.protocol
import
CellStates
from
neo.lib.pt
import
PartitionTable
class
StorageAppTests
(
NeoUnitTestBase
):
def
setUp
(
self
):
NeoUnitTestBase
.
setUp
(
self
)
self
.
prepareDatabase
(
number
=
1
)
# create an application object
config
=
self
.
getStorageConfiguration
(
master_number
=
1
)
self
.
app
=
Application
(
config
)
def
_tearDown
(
self
,
success
):
self
.
app
.
close
()
del
self
.
app
super
(
StorageAppTests
,
self
).
_tearDown
(
success
)
def
test_01_loadPartitionTable
(
self
):
self
.
app
.
dm
=
Mock
({
'getPartitionTable'
:
[],
})
self
.
assertEqual
(
self
.
app
.
pt
,
None
)
num_partitions
=
3
num_replicas
=
2
self
.
app
.
pt
=
PartitionTable
(
num_partitions
,
num_replicas
)
self
.
assertFalse
(
self
.
app
.
pt
.
getNodeSet
())
self
.
assertFalse
(
self
.
app
.
pt
.
filled
())
for
x
in
xrange
(
num_partitions
):
self
.
assertFalse
(
self
.
app
.
pt
.
hasOffset
(
x
))
# load an empty table
self
.
app
.
loadPartitionTable
()
self
.
assertFalse
(
self
.
app
.
pt
.
getNodeSet
())
self
.
assertFalse
(
self
.
app
.
pt
.
filled
())
for
x
in
xrange
(
num_partitions
):
self
.
assertFalse
(
self
.
app
.
pt
.
hasOffset
(
x
))
# add some node, will be remove when loading table
master_uuid
=
self
.
getMasterUUID
()
master
=
self
.
app
.
nm
.
createMaster
(
uuid
=
master_uuid
)
storage_uuid
=
self
.
getStorageUUID
()
storage
=
self
.
app
.
nm
.
createStorage
(
uuid
=
storage_uuid
)
client_uuid
=
self
.
getClientUUID
()
self
.
app
.
pt
.
_setCell
(
0
,
master
,
CellStates
.
UP_TO_DATE
)
self
.
app
.
pt
.
_setCell
(
0
,
storage
,
CellStates
.
UP_TO_DATE
)
self
.
assertEqual
(
len
(
self
.
app
.
pt
.
getNodeSet
()),
2
)
self
.
assertFalse
(
self
.
app
.
pt
.
filled
())
for
x
in
xrange
(
num_partitions
):
if
x
==
0
:
self
.
assertTrue
(
self
.
app
.
pt
.
hasOffset
(
x
))
else
:
self
.
assertFalse
(
self
.
app
.
pt
.
hasOffset
(
x
))
# load an empty table, everything removed
self
.
app
.
loadPartitionTable
()
self
.
assertFalse
(
self
.
app
.
pt
.
getNodeSet
())
self
.
assertFalse
(
self
.
app
.
pt
.
filled
())
for
x
in
xrange
(
num_partitions
):
self
.
assertFalse
(
self
.
app
.
pt
.
hasOffset
(
x
))
# add some node
self
.
app
.
pt
.
_setCell
(
0
,
master
,
CellStates
.
UP_TO_DATE
)
self
.
app
.
pt
.
_setCell
(
0
,
storage
,
CellStates
.
UP_TO_DATE
)
self
.
assertEqual
(
len
(
self
.
app
.
pt
.
getNodeSet
()),
2
)
self
.
assertFalse
(
self
.
app
.
pt
.
filled
())
for
x
in
xrange
(
num_partitions
):
if
x
==
0
:
self
.
assertTrue
(
self
.
app
.
pt
.
hasOffset
(
x
))
else
:
self
.
assertFalse
(
self
.
app
.
pt
.
hasOffset
(
x
))
# fill partition table
self
.
app
.
dm
=
Mock
({
'getPartitionTable'
:
[
(
0
,
client_uuid
,
CellStates
.
UP_TO_DATE
),
(
1
,
client_uuid
,
CellStates
.
UP_TO_DATE
),
(
1
,
storage_uuid
,
CellStates
.
UP_TO_DATE
),
(
2
,
storage_uuid
,
CellStates
.
UP_TO_DATE
),
(
2
,
master_uuid
,
CellStates
.
UP_TO_DATE
),
],
'getPTID'
:
1
,
})
self
.
app
.
pt
.
clear
()
self
.
app
.
loadPartitionTable
()
self
.
assertTrue
(
self
.
app
.
pt
.
filled
())
for
x
in
xrange
(
num_partitions
):
self
.
assertTrue
(
self
.
app
.
pt
.
hasOffset
(
x
))
# check each row
cell_list
=
self
.
app
.
pt
.
getCellList
(
0
)
self
.
assertEqual
(
len
(
cell_list
),
1
)
self
.
assertEqual
(
cell_list
[
0
].
getUUID
(),
client_uuid
)
cell_list
=
self
.
app
.
pt
.
getCellList
(
1
)
self
.
assertEqual
(
len
(
cell_list
),
2
)
self
.
assertTrue
(
cell_list
[
0
].
getUUID
()
in
(
client_uuid
,
storage_uuid
))
self
.
assertTrue
(
cell_list
[
1
].
getUUID
()
in
(
client_uuid
,
storage_uuid
))
cell_list
=
self
.
app
.
pt
.
getCellList
(
2
)
self
.
assertEqual
(
len
(
cell_list
),
2
)
self
.
assertTrue
(
cell_list
[
0
].
getUUID
()
in
(
master_uuid
,
storage_uuid
))
self
.
assertTrue
(
cell_list
[
1
].
getUUID
()
in
(
master_uuid
,
storage_uuid
))
if
__name__
==
'__main__'
:
unittest
.
main
()
neo/tests/storage/testStorageDBTests.py
View file @
8ba42463
...
...
@@ -48,30 +48,15 @@ class StorageDBTests(NeoUnitTestBase):
raise
NotImplementedError
def
setNumPartitions
(
self
,
num_partitions
,
reset
=
0
):
try
:
db
=
self
.
_db
except
AttributeError
:
self
.
_db
=
db
=
self
.
getDB
(
reset
)
else
:
if
reset
:
db
.
setup
(
reset
)
else
:
try
:
n
=
db
.
getNumPartitions
()
except
KeyError
:
n
=
0
if
num_partitions
==
n
:
return
if
num_partitions
<
n
:
db
.
dropPartitions
(
n
)
db
.
setNumPartitions
(
num_partitions
)
self
.
assertEqual
(
num_partitions
,
db
.
getNumPartitions
())
assert
not
hasattr
(
self
,
'_db'
)
self
.
_db
=
db
=
self
.
getDB
(
reset
)
uuid
=
self
.
getStorageUUID
()
db
.
setUUID
(
uuid
)
self
.
assertEqual
(
uuid
,
db
.
getUUID
())
db
.
changePartitionTable
(
1
,
db
.
changePartitionTable
(
1
,
0
,
[(
i
,
uuid
,
CellStates
.
UP_TO_DATE
)
for
i
in
xrange
(
num_partitions
)],
reset
=
True
)
self
.
assertEqual
(
num_partitions
,
1
+
db
.
_getMaxPartition
())
db
.
commit
()
def
checkConfigEntry
(
self
,
get_call
,
set_call
,
value
):
...
...
@@ -102,16 +87,6 @@ class StorageDBTests(NeoUnitTestBase):
db
=
self
.
getDB
()
self
.
checkConfigEntry
(
db
.
getName
,
db
.
setName
,
'TEST_NAME'
)
def
test_getPartitionTable
(
self
):
db
=
self
.
getDB
()
db
.
setNumPartitions
(
3
)
uuid1
,
uuid2
=
self
.
getStorageUUID
(),
self
.
getStorageUUID
()
cell1
=
(
0
,
uuid1
,
CellStates
.
OUT_OF_DATE
)
cell2
=
(
1
,
uuid1
,
CellStates
.
UP_TO_DATE
)
db
.
changePartitionTable
(
1
,
[
cell1
,
cell2
],
1
)
result
=
db
.
getPartitionTable
()
self
.
assertEqual
(
set
(
result
),
{
cell1
,
cell2
})
def
getOIDs
(
self
,
count
):
return
map
(
p64
,
xrange
(
count
))
...
...
@@ -202,52 +177,6 @@ class StorageDBTests(NeoUnitTestBase):
self
.
assertEqual
(
self
.
db
.
getObject
(
oid1
,
before_tid
=
tid2
),
OBJECT_T1_NEXT
)
def
test_setPartitionTable
(
self
):
db
=
self
.
getDB
()
db
.
setNumPartitions
(
3
)
ptid
=
1
uuid
=
self
.
getStorageUUID
()
cell1
=
0
,
uuid
,
CellStates
.
OUT_OF_DATE
cell2
=
1
,
uuid
,
CellStates
.
UP_TO_DATE
cell3
=
1
,
uuid
,
CellStates
.
DISCARDED
# no partition table
self
.
assertEqual
(
list
(
db
.
getPartitionTable
()),
[])
# set one
db
.
changePartitionTable
(
ptid
,
[
cell1
],
1
)
result
=
db
.
getPartitionTable
()
self
.
assertEqual
(
list
(
result
),
[
cell1
])
# then another
db
.
changePartitionTable
(
ptid
,
[
cell2
],
1
)
result
=
db
.
getPartitionTable
()
self
.
assertEqual
(
list
(
result
),
[
cell2
])
# drop discarded cells
db
.
changePartitionTable
(
ptid
,
[
cell2
,
cell3
],
1
)
result
=
db
.
getPartitionTable
()
self
.
assertEqual
(
list
(
result
),
[])
def
test_changePartitionTable
(
self
):
db
=
self
.
getDB
()
db
.
setNumPartitions
(
3
)
ptid
=
1
uuid
=
self
.
getStorageUUID
()
cell1
=
0
,
uuid
,
CellStates
.
OUT_OF_DATE
cell2
=
1
,
uuid
,
CellStates
.
UP_TO_DATE
cell3
=
1
,
uuid
,
CellStates
.
DISCARDED
# no partition table
self
.
assertEqual
(
list
(
db
.
getPartitionTable
()),
[])
# set one
db
.
changePartitionTable
(
ptid
,
[
cell1
])
result
=
db
.
getPartitionTable
()
self
.
assertEqual
(
list
(
result
),
[
cell1
])
# add more entries
db
.
changePartitionTable
(
ptid
,
[
cell2
])
result
=
db
.
getPartitionTable
()
self
.
assertEqual
(
set
(
result
),
{
cell1
,
cell2
})
# drop discarded cells
db
.
changePartitionTable
(
ptid
,
[
cell2
,
cell3
])
result
=
db
.
getPartitionTable
()
self
.
assertEqual
(
list
(
result
),
[
cell1
])
def
test_commitTransaction
(
self
):
oid1
,
oid2
=
self
.
getOIDs
(
2
)
tid1
,
tid2
=
self
.
getTIDs
(
2
)
...
...
neo/tests/storage/testStorageMySQL.py
View file @
8ba42463
...
...
@@ -22,7 +22,7 @@ from MySQLdb.constants.ER import UNKNOWN_STORAGE_ENGINE
from
..mock
import
Mock
from
neo.lib.protocol
import
ZERO_OID
from
neo.lib.util
import
p64
from
..
import
DB_PREFIX
,
DB_
SOCKET
,
DB_USER
,
Patch
from
..
import
DB_PREFIX
,
DB_
USER
,
Patch
,
setupMySQLdb
from
.testStorageDBTests
import
StorageDBTests
from
neo.storage.database
import
DatabaseFailure
from
neo.storage.database.mysqldb
import
MySQLDatabaseManager
...
...
@@ -46,8 +46,8 @@ class StorageMySQLdbTests(StorageDBTests):
engine
=
None
def
_test_lockDatabase_open
(
self
):
self
.
prepareDatabase
(
number
=
1
,
prefix
=
DB_PREFIX
)
database
=
'%s@%s0%s'
%
(
DB_USER
,
DB_PREFIX
,
DB_SOCKET
)
self
.
prepareDatabase
(
1
)
database
=
self
.
db_template
(
0
)
return
MySQLDatabaseManager
(
database
,
self
.
engine
)
def
getDB
(
self
,
reset
=
0
):
...
...
neo/tests/stress.py
View file @
8ba42463
...
...
@@ -19,12 +19,9 @@ class Handler(MasterEventHandler):
super
(
Handler
,
self
).
answerClusterState
(
conn
,
state
)
self
.
app
.
refresh
(
'state'
)
def
answerPartitionTable
(
self
,
*
args
):
super
(
Handler
,
self
).
answerPartitionTable
(
*
args
)
self
.
app
.
refresh
(
'pt'
)
def
sendPartitionTable
(
self
,
*
args
):
raise
AssertionError
super
(
Handler
,
self
).
sendPartitionTable
(
*
args
)
self
.
app
.
refresh
(
'pt'
)
def
notifyPartitionChanges
(
self
,
*
args
):
super
(
Handler
,
self
).
notifyPartitionChanges
(
*
args
)
...
...
@@ -50,6 +47,7 @@ class StressApplication(AdminApplication):
cluster_state
=
server
=
uuid
=
None
listening_conn
=
True
fault_probability
=
1
restart_ratio
=
float
(
'inf'
)
# no firewall support
_stress
=
False
...
...
@@ -191,7 +189,7 @@ class StressApplication(AdminApplication):
self
.
loid
=
loid
self
.
ltid
=
ltid
self
.
em
.
setTimeout
(
int
(
time
.
time
()
+
1
),
self
.
askLastIDs
)
if
self
.
_stress
:
if
self
.
_stress
and
random
.
random
()
<
self
.
fault_probability
:
node_list
=
self
.
nm
.
getStorageList
()
random
.
shuffle
(
node_list
)
fw
=
[]
...
...
neo/tests/testHandler.py
View file @
8ba42463
...
...
@@ -33,9 +33,9 @@ class HandlerTests(NeoUnitTestBase):
def
getFakePacket
(
self
):
p
=
Mock
({
'decode'
:
(),
'__repr__'
:
'Fake Packet'
,
})
p
.
_args
=
()
p
.
handler_method_name
=
'fake_method'
return
p
...
...
@@ -53,13 +53,6 @@ class HandlerTests(NeoUnitTestBase):
self
.
handler
.
dispatch
(
conn
,
packet
)
self
.
checkErrorPacket
(
conn
)
self
.
checkAborted
(
conn
)
# raise PacketMalformedError
conn
.
mockCalledMethods
=
{}
def
fake
(
c
):
raise
PacketMalformedError
(
'message'
)
self
.
setFakeMethod
(
fake
)
self
.
handler
.
dispatch
(
conn
,
packet
)
self
.
checkClosed
(
conn
)
# raise NotReadyError
conn
.
mockCalledMethods
=
{}
def
fake
(
c
):
...
...
neo/tests/testUtil.py
View file @
8ba42463
...
...
@@ -17,7 +17,7 @@
import
unittest
import
socket
from
.
import
NeoUnitTestBase
from
neo.lib.util
import
ReadBuffer
,
parseNodeAddress
from
neo.lib.util
import
parseNodeAddress
class
UtilTests
(
NeoUnitTestBase
):
...
...
@@ -40,24 +40,6 @@ class UtilTests(NeoUnitTestBase):
self
.
assertIn
(
parseNodeAddress
(
'localhost'
),
local_address
(
0
))
self
.
assertIn
(
parseNodeAddress
(
'localhost:10'
),
local_address
(
10
))
def
testReadBufferRead
(
self
):
""" Append some chunk then consume the data """
buf
=
ReadBuffer
()
self
.
assertEqual
(
len
(
buf
),
0
)
buf
.
append
(
'abc'
)
self
.
assertEqual
(
len
(
buf
),
3
)
# no enough data
self
.
assertEqual
(
buf
.
read
(
4
),
None
)
self
.
assertEqual
(
len
(
buf
),
3
)
buf
.
append
(
'def'
)
# consume a part
self
.
assertEqual
(
len
(
buf
),
6
)
self
.
assertEqual
(
buf
.
read
(
4
),
'abcd'
)
self
.
assertEqual
(
len
(
buf
),
2
)
# consume the rest
self
.
assertEqual
(
buf
.
read
(
3
),
None
)
self
.
assertEqual
(
buf
.
read
(
2
),
'ef'
)
if
__name__
==
"__main__"
:
unittest
.
main
()
neo/tests/threaded/__init__.py
View file @
8ba42463
...
...
@@ -40,7 +40,7 @@ from neo.lib.util import cached_property, parseMasterList, p64
from
neo.master.recovery
import
RecoveryManager
from
..
import
(
getTempDirectory
,
setupMySQLdb
,
ImporterConfigParser
,
NeoTestBase
,
Patch
,
ADDRESS_TYPE
,
IP_VERSION_FORMAT_DICT
,
DB_PREFIX
,
DB_SOCKET
,
DB_USER
)
ADDRESS_TYPE
,
IP_VERSION_FORMAT_DICT
,
DB_PREFIX
)
BIND
=
IP_VERSION_FORMAT_DICT
[
ADDRESS_TYPE
],
0
LOCAL_IP
=
socket
.
inet_pton
(
ADDRESS_TYPE
,
IP_VERSION_FORMAT_DICT
[
ADDRESS_TYPE
])
...
...
@@ -304,7 +304,13 @@ class TestSerialized(Serialized):
class
Node
(
object
):
def
getConnectionList
(
self
,
*
peers
):
addr
=
lambda
c
:
c
and
(
c
.
addr
if
c
.
is_server
else
c
.
getAddress
())
def
addr
(
c
):
# Do not identify only by source address because 2 TCP connections
# can have same source host:port to different destinations.
if
c
:
a
=
c
.
addr
b
=
c
.
getAddress
()
return
(
b
,
a
)
if
c
.
is_server
else
(
ServerNode
.
resolv
(
a
),
b
)
addr_set
=
{
addr
(
c
.
connector
)
for
peer
in
peers
for
c
in
peer
.
em
.
connection_dict
.
itervalues
()
if
isinstance
(
c
,
Connection
)}
...
...
@@ -377,7 +383,10 @@ class ServerNode(Node):
assert
not
self
.
is_alive
()
init_args
=
self
.
_init_args
init_args
[
'reset'
]
=
False
assert
set
(
kw
).
issubset
(
init_args
),
(
kw
,
init_args
)
if
__debug__
:
x
=
set
(
kw
).
difference
(
init_args
)
assert
not
x
or
x
.
issubset
(
self
.
option_parser
.
getOptionDict
()),
(
kw
,
init_args
)
init_args
.
update
(
kw
)
self
.
close
()
self
.
__init__
(
**
init_args
)
...
...
@@ -708,7 +717,7 @@ class NEOCluster(object):
def
__init__
(
self
,
master_count
=
1
,
partitions
=
1
,
replicas
=
0
,
upstream
=
None
,
adapter
=
os
.
getenv
(
'NEO_TESTS_ADAPTER'
,
'SQLite'
),
storage_count
=
None
,
db_list
=
None
,
clear_databases
=
True
,
db_user
=
DB_USER
,
db_password
=
''
,
compress
=
True
,
compress
=
True
,
importer
=
None
,
autostart
=
None
,
dedup
=
False
,
name
=
None
):
self
.
name
=
name
or
'neo_%s'
%
self
.
_allocate
(
'name'
,
lambda
:
random
.
randint
(
0
,
100
))
...
...
@@ -735,21 +744,20 @@ class NEOCluster(object):
db_list
=
[
'%s%u'
%
(
DB_PREFIX
,
self
.
_allocate
(
'db'
,
index
))
for
_
in
xrange
(
storage_count
)]
if
adapter
==
'MySQL'
:
setupMySQLdb
(
db_list
,
db_user
,
db_password
,
clear_databases
)
db
=
'%s:%s@%%s%s'
%
(
db_user
,
db_password
,
DB_SOCKET
)
db
=
setupMySQLdb
(
db_list
,
clear_databases
)
elif
adapter
==
'SQLite'
:
db
=
os
.
path
.
join
(
getTempDirectory
(),
'%s.sqlite'
)
db
=
os
.
path
.
join
(
getTempDirectory
(),
'%s.sqlite'
)
.
__mod__
else
:
assert
False
,
adapter
if
importer
:
cfg
=
ImporterConfigParser
(
adapter
,
**
importer
)
cfg
.
set
(
"neo"
,
"database"
,
db
%
tuple
(
db_list
))
db
=
os
.
path
.
join
(
getTempDirectory
(),
'%s.conf'
)
with
open
(
db
%
tuple
(
db_list
),
"w"
)
as
f
:
cfg
.
set
(
"neo"
,
"database"
,
db
(
*
db_list
))
db
=
os
.
path
.
join
(
getTempDirectory
(),
'%s.conf'
)
.
__mod__
with
open
(
db
(
*
db_list
),
"w"
)
as
f
:
cfg
.
write
(
f
)
kw
[
"adapter"
]
=
"Importer"
kw
[
'wait'
]
=
0
self
.
storage_list
=
[
StorageApplication
(
database
=
db
%
x
,
**
kw
)
self
.
storage_list
=
[
StorageApplication
(
database
=
db
(
x
)
,
**
kw
)
for
x
in
db_list
]
self
.
admin_list
=
[
AdminApplication
(
**
kw
)]
...
...
@@ -805,7 +813,7 @@ class NEOCluster(object):
master_list
=
self
.
master_list
if
storage_list
is
None
:
storage_list
=
self
.
storage_list
def
answer
PartitionTable
(
release
,
orig
,
*
args
):
def
send
PartitionTable
(
release
,
orig
,
*
args
):
orig
(
*
args
)
release
()
def
dispatch
(
release
,
orig
,
handler
,
*
args
):
...
...
@@ -821,7 +829,7 @@ class NEOCluster(object):
if
state
in
expected_state
:
release
()
with
Serialized
.
until
(
MasterEventHandler
,
answerPartitionTable
=
answer
PartitionTable
)
as
tic1
,
\
sendPartitionTable
=
send
PartitionTable
)
as
tic1
,
\
Serialized
.
until
(
RecoveryManager
,
dispatch
=
dispatch
)
as
tic2
,
\
Serialized
.
until
(
MasterEventHandler
,
notifyClusterInformation
=
notifyClusterInformation
)
as
tic3
:
...
...
@@ -846,9 +854,13 @@ class NEOCluster(object):
expected_state
=
(
NodeStates
.
PENDING
if
state
==
ClusterStates
.
RECOVERING
else
NodeStates
.
RUNNING
)
for
node
in
self
.
storage_list
if
storage_list
is
None
else
storage_list
:
for
node
,
expected_state
in
(
storage_list
if
isinstance
(
storage_list
,
dict
)
else
dict
.
fromkeys
(
self
.
storage_list
if
storage_list
is
None
else
storage_list
,
expected_state
)
).
iteritems
():
state
=
self
.
getNodeState
(
node
)
assert
state
==
expected_state
,
(
repr
(
node
),
state
)
assert
state
==
expected_state
,
(
repr
(
node
),
state
,
expected_state
)
def
stop
(
self
,
clear_database
=
False
,
__print_exc
=
traceback
.
print_exc
,
**
kw
):
if
self
.
started
:
...
...
@@ -922,7 +934,7 @@ class NEOCluster(object):
def
startCluster
(
self
):
try
:
self
.
neoctl
.
startCluster
()
except
RuntimeError
:
except
SystemExit
:
Serialized
.
tic
()
if
self
.
neoctl
.
getClusterState
()
not
in
(
ClusterStates
.
BACKINGUP
,
...
...
@@ -1001,18 +1013,18 @@ class NEOCluster(object):
"""Sort storages so that storage_list[i] has partition i for all i"""
pt
=
[{
x
.
getUUID
()
for
x
in
x
}
for
x
in
self
.
primary_master
.
pt
.
partition_list
]
n
=
len
(
self
.
storage_list
)
r
=
[]
x
=
[
iter
(
pt
[
0
])]
try
:
while
1
:
try
:
r
.
append
(
next
(
x
[
-
1
]))
except
StopIteration
:
del
r
[
-
1
],
x
[
-
1
]
else
:
x
.
append
(
iter
(
pt
[
len
(
r
)].
difference
(
r
)))
except
IndexError
:
assert
len
(
r
)
==
len
(
self
.
storage_list
)
while
1
:
try
:
r
.
append
(
next
(
x
[
-
1
]))
except
StopIteration
:
del
r
[
-
1
],
x
[
-
1
]
else
:
if
len
(
r
)
==
n
:
break
x
.
append
(
iter
(
pt
[
len
(
r
)].
difference
(
r
)))
x
=
{
x
.
uuid
:
x
for
x
in
self
.
storage_list
}
self
.
storage_list
[:]
=
(
x
[
r
]
for
r
in
r
)
return
self
.
storage_list
...
...
neo/tests/threaded/test.py
View file @
8ba42463
...
...
@@ -42,6 +42,7 @@ from neo.lib.util import add64, makeChecksum, p64, u64
from
neo.client.exception
import
NEOPrimaryMasterLost
,
NEOStorageError
from
neo.client.transactions
import
Transaction
from
neo.master.handlers.client
import
ClientServiceHandler
from
neo.master.pt
import
PartitionTable
from
neo.storage.database
import
DatabaseFailure
from
neo.storage.handlers.client
import
ClientOperationHandler
from
neo.storage.handlers.identification
import
IdentificationHandler
...
...
@@ -148,6 +149,7 @@ class Test(NEOThreadedTest):
c
.
root
()[
0
]
=
ob
=
PCounterWithResolution
()
t
.
commit
()
tids
=
[]
c
.
readCurrent
(
c
.
root
())
for
x
in
inc
:
ob
.
value
+=
x
t
.
commit
()
...
...
@@ -424,6 +426,42 @@ class Test(NEOThreadedTest):
self
.
assertEqual
([
x
[
'tid'
]
for
x
in
c1
.
db
().
history
(
oid
,
size
=
10
)],
[
tid3
,
tid2
,
tid1
,
tid0
])
@
with_cluster
()
def
testSlowConflictResolution
(
self
,
cluster
):
"""
Check that a slow conflict resolution does not always result in a new
conflict because a concurrent client keeps modifying the same object
quickly.
An idea to fix it is to take the lock before the second attempt to
resolve.
"""
t1
,
c1
=
cluster
.
getTransaction
()
c1
.
root
()[
''
]
=
ob
=
PCounterWithResolution
()
t1
.
commit
()
l1
=
threading
.
Lock
();
l1
.
acquire
()
l2
=
threading
.
Lock
();
l2
.
acquire
()
conflicts
=
[]
def
_p_resolveConflict
(
orig
,
*
args
):
conflicts
.
append
(
get_ident
())
l1
.
release
();
l2
.
acquire
()
return
orig
(
*
args
)
with
cluster
.
newClient
(
1
)
as
db
,
Patch
(
PCounterWithResolution
,
_p_resolveConflict
=
_p_resolveConflict
):
t2
,
c2
=
cluster
.
getTransaction
(
db
)
c2
.
root
()[
''
].
value
+=
1
for
i
in
xrange
(
10
):
ob
.
value
+=
1
t1
.
commit
()
if
i
:
l2
.
release
()
else
:
t
=
self
.
newThread
(
t2
.
commit
)
l1
.
acquire
()
l2
.
release
()
t
.
join
()
with
self
.
expectedFailure
():
\
self
.
assertIn
(
get_ident
(),
conflicts
)
@
with_cluster
()
def
testDelayedLoad
(
self
,
cluster
):
"""
...
...
@@ -471,6 +509,7 @@ class Test(NEOThreadedTest):
self
.
assertFalse
(
conn
.
isClosed
())
getCellSortKey
=
cluster
.
client
.
getCellSortKey
self
.
assertEqual
(
getCellSortKey
(
s0
,
good
),
0
)
cluster
.
neoctl
.
killNode
(
s0
.
getUUID
())
cluster
.
neoctl
.
dropNode
(
s0
.
getUUID
())
self
.
assertEqual
([
s1
],
cluster
.
client
.
nm
.
getStorageList
())
self
.
assertTrue
(
conn
.
isClosed
())
...
...
@@ -776,6 +815,7 @@ class Test(NEOThreadedTest):
checkNodeState
(
NodeStates
.
RUNNING
)
self
.
assertEqual
([],
cluster
.
getOutdatedCells
())
# drop one
cluster
.
neoctl
.
killNode
(
s1
.
uuid
)
cluster
.
neoctl
.
dropNode
(
s1
.
uuid
)
checkNodeState
(
None
)
self
.
tic
()
# Let node state update reach remaining storage
...
...
@@ -1123,6 +1163,10 @@ class Test(NEOThreadedTest):
# Check that the storage hasn't answered to the store,
# which means that a lock is still taken for r['x'] by t2.
self
.
tic
()
try
:
txn
=
txn
.
data
(
c1
)
except
(
AttributeError
,
KeyError
):
# BBB: ZODB < 5
pass
txn_context
=
cluster
.
client
.
_txn_container
.
get
(
txn
)
empty
=
txn_context
.
queue
.
empty
()
ll
()
...
...
@@ -1202,7 +1246,7 @@ class Test(NEOThreadedTest):
# Also check that the master reset the last oid to a correct value.
t
.
begin
()
self
.
assertEqual
(
1
,
u64
(
c
.
root
()[
'x'
].
_p_oid
))
self
.
assertFalse
(
cluster
.
client
.
new_oid
_list
)
self
.
assertFalse
(
cluster
.
client
.
new_oid
s
)
self
.
assertEqual
(
2
,
u64
(
cluster
.
client
.
new_oid
()))
@
with_cluster
()
...
...
@@ -1371,7 +1415,7 @@ class Test(NEOThreadedTest):
del
conn
.
_queue
[:]
# XXX
conn
.
close
()
if
1
:
with
Patch
(
cluster
.
master
.
pt
,
make
=
make
),
\
with
Patch
(
PartitionTable
,
make
=
make
),
\
Patch
(
InitializationHandler
,
askPartitionTable
=
askPartitionTable
)
as
p
:
cluster
.
start
()
...
...
@@ -1562,7 +1606,7 @@ class Test(NEOThreadedTest):
bad
.
append
(
s
.
getDataLockInfo
())
s
.
dm
.
commit
()
def
check
(
dry_run
,
expected
):
cluster
.
neoctl
.
repair
(
node_list
,
dry_run
)
cluster
.
neoctl
.
repair
(
node_list
,
bool
(
dry_run
)
)
for
e
,
s
in
zip
(
expected
,
cluster
.
storage_list
):
while
1
:
self
.
tic
()
...
...
@@ -1902,18 +1946,7 @@ class Test(NEOThreadedTest):
x
.
value
+=
1
c2
.
root
()[
'x'
].
value
+=
2
TransactionalResource
(
t1
,
1
,
tpc_begin
=
begin1
)
# BUG: Very rarely, getConnectionList returns more that 1
# connection ("too many values to unpack"), which is
# a mystery and impossible to reproduce:
# - 1st time: v1.8.1 on a test machine (no SSL)
# - last: current revision on my laptop (SSL),
# at the first iteration of this loop
_sm
=
list
(
s1
.
getConnectionList
(
cluster
.
master
))
try
:
s1m
,
=
_sm
except
ValueError
:
self
.
fail
((
_sm
,
list
(
s1
.
getConnectionList
(
cluster
.
master
))))
s1m
,
=
s1
.
getConnectionList
(
cluster
.
master
)
try
:
s1
.
em
.
removeReader
(
s1m
)
with
ConnectionFilter
()
as
f
,
\
...
...
@@ -1979,7 +2012,7 @@ class Test(NEOThreadedTest):
except
threading
.
ThreadError
:
l
[
j
].
acquire
()
threads
[
j
-
1
].
start
()
if
x
!=
'StoreTransaction'
:
if
x
!=
'
Ask
StoreTransaction'
:
try
:
l
[
i
].
acquire
()
except
IndexError
:
...
...
@@ -2056,15 +2089,16 @@ class Test(NEOThreadedTest):
x
=
self
.
_testComplexDeadlockAvoidanceWithOneStorage
(
changes
,
(
1
,
1
,
0
,
1
,
2
,
2
,
2
,
2
,
0
,
1
,
2
,
1
,
0
,
0
,
1
,
0
,
0
,
1
),
(
'tpc_begin'
,
'tpc_begin'
,
1
,
2
,
3
,
'tpc_begin'
,
1
,
2
,
4
,
3
,
4
,
'StoreTransaction'
,
'RebaseTransaction'
,
'RebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'RebaseTransaction'
,
'AnswerRebaseTransaction'
),
'AskStoreTransaction'
,
'AskRebaseTransaction'
,
'AskRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AskRebaseTransaction'
,
'AnswerRebaseTransaction'
),
[
4
,
6
,
2
,
6
])
try
:
x
[
1
].
remove
(
1
)
except
ValueError
:
pass
self
.
assertEqual
(
x
,
{
0
:
[
2
,
'StoreTransaction'
],
1
:
[
'tpc_abort'
]})
self
.
assertEqual
(
x
,
{
0
:
[
2
,
'
Ask
StoreTransaction'
],
1
:
[
'tpc_abort'
]})
def
testCascadedDeadlockAvoidanceWithOneStorage2
(
self
):
def
changes
(
r1
,
r2
,
r3
):
...
...
@@ -2087,8 +2121,8 @@ class Test(NEOThreadedTest):
(
0
,
1
,
1
,
0
,
1
,
2
,
2
,
2
,
2
,
0
,
1
,
2
,
1
,
0
,
0
,
0
,
0
,
0
,
0
,
1
,
1
,
0
,
1
,
1
,
1
),
(
'tpc_begin'
,
1
,
'tpc_begin'
,
1
,
2
,
3
,
'tpc_begin'
,
2
,
3
,
4
,
3
,
4
,
'
StoreTransaction'
,
'
RebaseTransaction'
,
'RebaseTransaction'
,
'AnswerRebaseTransaction'
),
2
,
3
,
4
,
3
,
4
,
'
AskStoreTransaction'
,
'Ask
RebaseTransaction'
,
'
Ask
RebaseTransaction'
,
'AnswerRebaseTransaction'
),
[
1
,
7
,
9
,
0
])
x
[
0
].
sort
(
key
=
str
)
try
:
...
...
@@ -2097,8 +2131,8 @@ class Test(NEOThreadedTest):
pass
self
.
assertEqual
(
x
,
{
0
:
[
2
,
3
,
'AnswerRebaseTransaction'
,
'
RebaseTransaction'
,
'
StoreTransaction'
],
1
:
[
'AnswerRebaseTransaction'
,
'RebaseTransaction'
,
'
AskRebaseTransaction'
,
'Ask
StoreTransaction'
],
1
:
[
'AnswerRebaseTransaction'
,
'
Ask
RebaseTransaction'
,
'AnswerRebaseTransaction'
,
'tpc_abort'
],
})
...
...
@@ -2131,7 +2165,7 @@ class Test(NEOThreadedTest):
end
=
self
.
_testComplexDeadlockAvoidanceWithOneStorage
(
changes
,
(
0
,
1
,
1
,
0
,
1
,
1
,
0
,
0
,
2
,
2
,
2
,
2
,
1
,
vote_t2
,
tic_t1
),
(
'tpc_begin'
,
1
)
*
2
,
[
3
,
0
,
0
,
0
],
None
)
self
.
assertLessEqual
(
2
,
end
[
0
].
count
(
'RebaseTransaction'
))
self
.
assertLessEqual
(
2
,
end
[
0
].
count
(
'
Ask
RebaseTransaction'
))
def
testFailedConflictOnBigValueDuringDeadlockAvoidance
(
self
):
def
changes
(
r1
,
r2
,
r3
):
...
...
@@ -2147,10 +2181,10 @@ class Test(NEOThreadedTest):
x
=
self
.
_testComplexDeadlockAvoidanceWithOneStorage
(
changes
,
(
1
,
1
,
1
,
2
,
2
,
2
,
1
,
2
,
2
,
0
,
0
,
1
,
1
,
1
,
0
),
(
'tpc_begin'
,
'tpc_begin'
,
1
,
2
,
'tpc_begin'
,
1
,
3
,
3
,
4
,
'
StoreTransaction'
,
2
,
4
,
'
RebaseTransaction'
,
'
AskStoreTransaction'
,
2
,
4
,
'Ask
RebaseTransaction'
,
'AnswerRebaseTransaction'
,
'tpc_abort'
),
[
5
,
1
,
0
,
2
],
POSException
.
ConflictError
)
self
.
assertEqual
(
x
,
{
0
:
[
'StoreTransaction'
]})
self
.
assertEqual
(
x
,
{
0
:
[
'
Ask
StoreTransaction'
]})
@
with_cluster
(
replicas
=
1
,
partitions
=
4
)
def
testNotifyReplicated
(
self
,
cluster
):
...
...
@@ -2237,7 +2271,7 @@ class Test(NEOThreadedTest):
def
delayConflict
(
conn
,
packet
):
app
=
self
.
getConnectionApp
(
conn
)
if
(
isinstance
(
packet
,
Packets
.
AnswerStoreObject
)
and
packet
.
decode
()
[
0
]):
and
packet
.
_args
[
0
]):
conn
,
=
cluster
.
client
.
getConnectionList
(
app
)
kw
=
conn
.
_handlers
.
_pending
[
0
][
0
][
packet
.
_id
][
1
]
return
1
==
u64
(
kw
[
'oid'
])
and
delay_conflict
[
app
.
uuid
].
pop
()
...
...
@@ -2255,8 +2289,9 @@ class Test(NEOThreadedTest):
self
.
thread_switcher
(
threads
,
(
1
,
2
,
3
,
0
,
1
,
0
,
2
,
t3_c
,
1
,
3
,
2
,
t3_resolve
,
0
,
0
,
0
,
t1_rebase
,
2
,
t3_b
,
3
,
t4_d
,
0
,
2
,
2
),
(
'tpc_begin'
,
'tpc_begin'
,
'tpc_begin'
,
'tpc_begin'
,
2
,
1
,
1
,
3
,
3
,
4
,
4
,
3
,
1
,
'RebaseTransaction'
,
'RebaseTransaction'
,
(
'tpc_begin'
,
'tpc_begin'
,
'tpc_begin'
,
'tpc_begin'
,
2
,
1
,
1
,
3
,
3
,
4
,
4
,
3
,
1
,
'AskRebaseTransaction'
,
'AskRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AnswerRebaseTransaction'
,
2
))
as
end
:
delay
=
f
.
delayAskFetchTransactions
()
...
...
@@ -2268,11 +2303,11 @@ class Test(NEOThreadedTest):
t4
.
begin
()
self
.
assertEqual
([
15
,
11
,
13
,
16
],
[
r
[
x
].
value
for
x
in
'abcd'
])
self
.
assertEqual
([
2
,
2
],
map
(
end
.
pop
(
2
).
count
,
[
'RebaseTransaction'
,
'AnswerRebaseTransaction'
]))
[
'
Ask
RebaseTransaction'
,
'AnswerRebaseTransaction'
]))
self
.
assertEqual
(
end
,
{
0
:
[
1
,
'StoreTransaction'
],
1
:
[
'StoreTransaction'
],
3
:
[
4
,
'StoreTransaction'
],
0
:
[
1
,
'
Ask
StoreTransaction'
],
1
:
[
'
Ask
StoreTransaction'
],
3
:
[
4
,
'
Ask
StoreTransaction'
],
})
self
.
assertFalse
(
s1
.
dm
.
getOrphanList
())
...
...
@@ -2308,7 +2343,8 @@ class Test(NEOThreadedTest):
self
.
thread_switcher
((
thread
,),
(
1
,
0
,
1
,
1
,
t2_b
,
0
,
0
,
1
,
t2_vote
,
0
,
0
),
(
'tpc_begin'
,
'tpc_begin'
,
1
,
1
,
2
,
2
,
'RebaseTransaction'
,
'RebaseTransaction'
,
'StoreTransaction'
,
'AskRebaseTransaction'
,
'AskRebaseTransaction'
,
'AskStoreTransaction'
,
'AnswerRebaseTransaction'
,
'AnswerRebaseTransaction'
,
))
as
end
:
delay
=
f
.
delayAskFetchTransactions
()
...
...
@@ -2361,15 +2397,20 @@ class Test(NEOThreadedTest):
# Check that the storage hasn't answered to the store,
# which means that a lock is still taken for r[''] by t1.
self
.
tic
()
try
:
txn
=
txn
.
data
(
c3
)
except
(
AttributeError
,
KeyError
):
# BBB: ZODB < 5
pass
txn_context
=
db
.
storage
.
app
.
_txn_container
.
get
(
txn
)
raise
Abort
(
txn_context
.
queue
.
empty
())
TransactionalResource
(
t3
,
1
,
commit
=
t3_commit
)
with
self
.
thread_switcher
((
commit23
,),
(
1
,
1
,
0
,
0
,
t1_rebase
,
0
,
0
,
0
,
1
,
1
,
1
,
1
,
0
),
(
'tpc_begin'
,
'tpc_begin'
,
0
,
1
,
0
,
'RebaseTransaction'
,
'RebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'StoreTransaction'
,
'tpc_begin'
,
1
,
'tpc_abort'
))
as
end
:
'AskRebaseTransaction'
,
'AskRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AskStoreTransaction'
,
'tpc_begin'
,
1
,
'tpc_abort'
,
))
as
end
:
self
.
assertRaises
(
POSException
.
ConflictError
,
t1
.
commit
)
commit23
.
join
()
self
.
assertEqual
(
end
,
{
0
:
[
'tpc_abort'
]})
...
...
@@ -2407,8 +2448,8 @@ class Test(NEOThreadedTest):
for
x
in
'ab'
:
r
[
x
]
=
PCounterWithResolution
()
t1
.
commit
()
cluster
.
stop
(
replicas
=
1
)
cluster
.
start
()
cluster
.
neoctl
.
setNumReplicas
(
1
)
self
.
tic
()
s0
,
s1
=
cluster
.
sortStorageList
()
t1
,
c1
=
cluster
.
getTransaction
()
r
=
c1
.
root
()
...
...
@@ -2456,9 +2497,9 @@ class Test(NEOThreadedTest):
self
.
thread_switcher
((
commit2
,),
(
1
,
1
,
0
,
0
,
t1_b
,
t1_resolve
,
0
,
0
,
0
,
0
,
1
,
t2_vote
,
t1_end
),
(
'tpc_begin'
,
'tpc_begin'
,
2
,
1
,
2
,
1
,
1
,
'
RebaseTransaction'
,
'
RebaseTransaction'
,
'
AskRebaseTransaction'
,
'Ask
RebaseTransaction'
,
'AnswerRebaseTransaction'
,
'AnswerRebaseTransaction'
,
'StoreTransaction'
))
as
end
:
'
Ask
StoreTransaction'
))
as
end
:
t1
.
commit
()
commit2
.
join
()
t1
.
begin
()
...
...
@@ -2466,7 +2507,7 @@ class Test(NEOThreadedTest):
self
.
assertEqual
(
r
[
'a'
].
value
,
9
)
self
.
assertEqual
(
r
[
'b'
].
value
,
6
)
t1
=
end
.
pop
(
0
)
self
.
assertEqual
(
t1
.
pop
(),
'StoreTransaction'
)
self
.
assertEqual
(
t1
.
pop
(),
'
Ask
StoreTransaction'
)
self
.
assertEqual
(
sorted
(
t1
),
[
1
,
2
])
self
.
assertFalse
(
end
)
self
.
assertPartitionTable
(
cluster
,
'UU|UU'
)
...
...
@@ -2568,9 +2609,9 @@ class Test(NEOThreadedTest):
with
Patch
(
cluster
.
client
,
_loadFromStorage
=
load
)
as
p
,
\
self
.
thread_switcher
((
commit2
,),
(
1
,
0
,
tic1
,
0
,
t1_resolve
,
1
,
t2_begin
,
0
,
1
,
1
,
0
),
(
'tpc_begin'
,
'tpc_begin'
,
1
,
1
,
1
,
'StoreTransaction'
,
'tpc_begin'
,
'
RebaseTransaction'
,
'RebaseTransaction'
,
1
,
'
StoreTransaction'
))
as
end
:
(
'tpc_begin'
,
'tpc_begin'
,
1
,
1
,
1
,
'
Ask
StoreTransaction'
,
'tpc_begin'
,
'
AskRebaseTransaction'
,
'AskRebaseTransaction'
,
1
,
'Ask
StoreTransaction'
))
as
end
:
self
.
assertRaisesRegexp
(
NEOStorageError
,
'^partition 0 not fully write-locked$'
,
t1
.
commit
)
...
...
@@ -2592,8 +2633,8 @@ class Test(NEOThreadedTest):
for
x
in
'ab'
:
r
[
x
]
=
PCounterWithResolution
()
t1
.
commit
()
cluster
.
stop
(
replicas
=
1
)
cluster
.
start
()
cluster
.
neoctl
.
setNumReplicas
(
1
)
self
.
tic
()
s0
,
s1
=
cluster
.
sortStorageList
()
t1
,
c1
=
cluster
.
getTransaction
()
r
=
c1
.
root
()
...
...
@@ -2623,13 +2664,14 @@ class Test(NEOThreadedTest):
f
.
remove
(
delayFinish
)
with
self
.
thread_switcher
((
commit2
,),
(
1
,
0
,
0
,
1
,
t2_b
,
0
,
t1_resolve
),
(
'tpc_begin'
,
'tpc_begin'
,
0
,
2
,
2
,
'StoreTransaction'
))
as
end
:
(
'tpc_begin'
,
'tpc_begin'
,
0
,
2
,
2
,
'AskStoreTransaction'
)
)
as
end
:
t1
.
commit
()
commit2
.
join
()
t1
.
begin
()
self
.
assertEqual
(
c1
.
root
()[
'b'
].
value
,
6
)
self
.
assertPartitionTable
(
cluster
,
'UU|UU'
)
self
.
assertEqual
(
end
,
{
0
:
[
2
,
2
,
'StoreTransaction'
]})
self
.
assertEqual
(
end
,
{
0
:
[
2
,
2
,
'
Ask
StoreTransaction'
]})
self
.
assertFalse
(
s1
.
dm
.
getOrphanList
())
@
with_cluster
(
storage_count
=
2
,
partitions
=
2
)
...
...
@@ -2652,19 +2694,19 @@ class Test(NEOThreadedTest):
yield
1
self
.
tic
()
with
self
.
thread_switcher
((
t
,),
(
1
,
0
,
1
,
0
,
t1_b
,
0
,
0
,
0
,
1
),
(
'tpc_begin'
,
'tpc_begin'
,
1
,
3
,
3
,
1
,
'RebaseTransaction'
,
(
'tpc_begin'
,
'tpc_begin'
,
1
,
3
,
3
,
1
,
'
Ask
RebaseTransaction'
,
2
,
'AnswerRebaseTransaction'
))
as
end
:
t1
.
commit
()
t
.
join
()
t2
.
begin
()
self
.
assertEqual
([
6
,
9
,
6
],
[
r
[
x
].
value
for
x
in
'abc'
])
self
.
assertEqual
([
2
,
2
],
map
(
end
.
pop
(
1
).
count
,
[
'RebaseTransaction'
,
'AnswerRebaseTransaction'
]))
[
'
Ask
RebaseTransaction'
,
'AnswerRebaseTransaction'
]))
# Rarely, there's an extra deadlock for t1:
# 0: ['AnswerRebaseTransaction', 'RebaseTransaction',
# 'RebaseTransaction', 'AnswerRebaseTransaction',
# 0: ['AnswerRebaseTransaction', '
Ask
RebaseTransaction',
# '
Ask
RebaseTransaction', 'AnswerRebaseTransaction',
# 'AnswerRebaseTransaction', 2, 3, 1,
# 'StoreTransaction', 'VoteTransaction']
# '
Ask
StoreTransaction', 'VoteTransaction']
self
.
assertEqual
(
end
.
pop
(
0
)[
0
],
'AnswerRebaseTransaction'
)
self
.
assertFalse
(
end
)
...
...
@@ -2694,13 +2736,13 @@ class Test(NEOThreadedTest):
threads
=
map
(
self
.
newPausedThread
,
(
t2
.
commit
,
t3
.
commit
))
with
self
.
thread_switcher
(
threads
,
(
1
,
2
,
0
,
1
,
2
,
1
,
0
,
2
,
0
,
1
,
2
),
(
'tpc_begin'
,
'tpc_begin'
,
'tpc_begin'
,
1
,
2
,
3
,
4
,
4
,
4
,
'
RebaseTransaction'
,
'
StoreTransaction'
))
as
end
:
'
AskRebaseTransaction'
,
'Ask
StoreTransaction'
))
as
end
:
t1
.
commit
()
for
t
in
threads
:
t
.
join
()
self
.
assertEqual
(
end
,
{
0
:
[
'AnswerRebaseTransaction'
,
'StoreTransaction'
],
2
:
[
'StoreTransaction'
]})
0
:
[
'AnswerRebaseTransaction'
,
'
Ask
StoreTransaction'
],
2
:
[
'
Ask
StoreTransaction'
]})
@
with_cluster
(
replicas
=
1
)
def
testConflictAfterDeadlockWithSlowReplica1
(
self
,
cluster
,
...
...
@@ -2743,16 +2785,16 @@ class Test(NEOThreadedTest):
order
[
-
1
]
=
t1_resolve
delay
=
f
.
delayAskStoreObject
()
with
self
.
thread_switcher
((
t
,),
order
,
(
'tpc_begin'
,
'tpc_begin'
,
1
,
1
,
2
,
2
,
'RebaseTransaction'
,
'RebaseTransaction'
,
'AnswerRebaseTransaction'
,
'StoreTransaction'
))
as
end
:
(
'tpc_begin'
,
'tpc_begin'
,
1
,
1
,
2
,
2
,
'
Ask
RebaseTransaction'
,
'
Ask
RebaseTransaction'
,
'AnswerRebaseTransaction'
,
'
Ask
StoreTransaction'
))
as
end
:
t1
.
commit
()
t
.
join
()
self
.
assertNotIn
(
delay
,
f
)
t2
.
begin
()
end
[
0
].
sort
(
key
=
str
)
self
.
assertEqual
(
end
,
{
0
:
[
1
,
'AnswerRebaseTransaction'
,
'StoreTransaction'
]})
'
Ask
StoreTransaction'
]})
self
.
assertEqual
([
4
,
2
],
[
r
[
x
].
value
for
x
in
'ab'
])
def
testConflictAfterDeadlockWithSlowReplica2
(
self
):
...
...
@@ -2803,7 +2845,7 @@ class Test(NEOThreadedTest):
with
ConnectionFilter
()
as
f
:
f
.
add
(
lambda
conn
,
packet
:
isinstance
(
packet
,
Packets
.
RequestIdentification
)
and
packet
.
decode
()
[
0
]
==
NodeTypes
.
STORAGE
)
and
packet
.
_args
[
0
]
==
NodeTypes
.
STORAGE
)
self
.
tic
()
m2
.
start
()
self
.
tic
()
...
...
@@ -2843,7 +2885,7 @@ class Test(NEOThreadedTest):
with
ConnectionFilter
()
as
f
:
f
.
add
(
lambda
conn
,
packet
:
isinstance
(
packet
,
Packets
.
RequestIdentification
)
and
packet
.
decode
()
[
0
]
==
NodeTypes
.
MASTER
)
and
packet
.
_args
[
0
]
==
NodeTypes
.
MASTER
)
cluster
.
start
(
recovering
=
True
)
neoctl
=
cluster
.
neoctl
getClusterState
=
neoctl
.
getClusterState
...
...
@@ -2894,9 +2936,9 @@ class Test(NEOThreadedTest):
dm
=
s
.
dm
dm
.
commit
()
dump_dict
[
s
.
uuid
]
=
dm
.
dump
()
dm
.
erase
()
with
open
(
path
%
(
s
.
getAdapter
(),
s
.
uuid
))
as
f
:
dm
.
restore
(
f
.
read
())
dm
.
setConfiguration
(
'partitions'
,
None
)
# XXX: see dm._migrate4
with
NEOCluster
(
storage_count
=
3
,
partitions
=
3
,
replicas
=
1
,
name
=
self
.
_testMethodName
)
as
cluster
:
s1
,
s2
,
s3
=
cluster
.
storage_list
...
...
neo/tests/threaded/testImporter.py
View file @
8ba42463
...
...
@@ -17,16 +17,19 @@
from
cPickle
import
Pickler
,
Unpickler
from
cStringIO
import
StringIO
from
itertools
import
izip_longest
import
os
,
random
,
shutil
,
time
,
unittest
import
os
,
random
,
shutil
,
t
hreading
,
t
ime
,
unittest
import
transaction
,
ZODB
from
neo.client.exception
import
NEOPrimaryMasterLost
from
neo.lib
import
logging
from
neo.lib.util
import
u64
from
neo.lib.util
import
cached_property
,
p64
,
u64
from
neo.master.transactions
import
TransactionManager
from
neo.storage.database
import
getAdapterKlass
,
importer
,
manager
from
neo.storage.database.importer
import
Repickler
,
TransactionRecord
from
neo.storage.database.importer
import
\
Repickler
,
TransactionRecord
,
WriteBack
from
..
import
expectedFailure
,
getTempDirectory
,
random_tree
,
Patch
from
.
import
NEOCluster
,
NEOThreadedTest
from
ZODB
import
serialize
from
ZODB.DB
import
TransactionalUndo
from
ZODB.FileStorage
import
FileStorage
class
Equal
:
...
...
@@ -128,31 +131,51 @@ class ImporterTests(NEOThreadedTest):
self
.
assertIs
(
Obj
,
load
())
self
.
assertDictEqual
(
state
,
load
())
def
_importFromFileStorage
(
self
,
multi
=
(),
root_filter
=
None
,
sub_filter
=
None
):
import_hash
=
'1d4ff03730fe6bcbf235e3739fbe5f5b'
@
cached_property
def
getFS
(
self
):
fs_dir
=
os
.
path
.
join
(
getTempDirectory
(),
self
.
id
())
shutil
.
rmtree
(
fs_dir
,
1
)
# for --loop
os
.
mkdir
(
fs_dir
)
def
getFS
(
db
=
'root'
):
path
=
os
.
path
.
join
(
fs_dir
,
'%s.fs'
%
db
)
return
path
,
{
"storage"
:
"<filestorage>
\
n
path %s
\
n
</filestorage>"
%
path
}
return
getFS
def
getData
(
self
,
tree
=
random_tree
.
generateTree
(
random
.
Random
(
0
))):
txn_size
=
10
tree
=
random_tree
.
generateTree
(
random
.
Random
(
0
))
i
=
len
(
tree
)
//
3
assert
i
>
txn_size
before_tree
=
tree
[:
i
]
after_tree
=
tree
[
i
:]
fs_dir
=
os
.
path
.
join
(
getTempDirectory
(),
self
.
id
())
shutil
.
rmtree
(
fs_dir
,
1
)
# for --loop
os
.
mkdir
(
fs_dir
)
def
beforeCheck
(
h
,
count
=
52
):
self
.
assertEqual
(
count
,
h
())
self
.
assertEqual
(
'1d4ff03730fe6bcbf235e3739fbe5f5b'
,
h
.
hexdigest
())
def
finalCheck
(
r
):
h
=
random_tree
.
hashTree
(
r
)
self
.
assertEqual
(
93
,
h
())
self
.
assertEqual
(
'6bf0f0cb2d6c1aae9e52c412ef0e25b6'
,
h
.
hexdigest
())
return
(
beforeCheck
,
lambda
r
,
*
f
:
random_tree
.
importTree
(
r
,
before_tree
,
txn_size
,
*
f
),
finalCheck
,
lambda
r
:
random_tree
.
importTree
(
r
,
after_tree
,
txn_size
),
)
def
_importFromFileStorage
(
self
,
multi
=
(),
root_filter
=
None
,
sub_filter
=
None
):
beforeCheck
,
before
,
finalCheck
,
after
=
self
.
getData
()
iter_list
=
[]
db_list
=
[]
# Setup several FileStorage databases.
for
i
,
db
in
enumerate
((
'root'
,)
+
multi
):
fs_path
=
os
.
path
.
join
(
fs_dir
,
'%s.fs'
%
db
)
fs_path
,
cfg
=
self
.
getFS
(
db
)
c
=
ZODB
.
DB
(
FileStorage
(
fs_path
)).
open
()
r
=
c
.
root
()[
'tree'
]
=
random_tree
.
Node
()
transaction
.
commit
()
iter_list
.
append
(
random_tree
.
importTree
(
r
,
before_tree
,
txn_size
,
sub_filter
(
db
)
if
i
else
root_filter
))
db_list
.
append
((
db
,
r
,
{
"storage"
:
"<filestorage>
\
n
path %s
\
n
</filestorage>"
%
fs_path
}))
iter_list
.
append
(
before
(
r
,
sub_filter
(
db
)
if
i
else
root_filter
))
db_list
.
append
((
db
,
r
,
cfg
))
# Populate FileStorage databases.
for
i
,
iter_list
in
enumerate
(
izip_longest
(
*
iter_list
)):
for
r
in
iter_list
:
...
...
@@ -167,9 +190,7 @@ class ImporterTests(NEOThreadedTest):
for
x
in
multi
:
cfg
[
'_%s'
%
x
]
=
str
(
u64
(
r
[
x
].
_p_oid
))
else
:
h
=
random_tree
.
hashTree
(
r
)
h
()
self
.
assertEqual
(
import_hash
,
h
.
hexdigest
())
beforeCheck
(
random_tree
.
hashTree
(
r
))
importer
[
'writeback'
]
=
'true'
else
:
cfg
[
"oid"
]
=
str
(
u64
(
r
[
db
].
_p_oid
))
...
...
@@ -179,7 +200,7 @@ class ImporterTests(NEOThreadedTest):
del
db_list
,
iter_list
#del zodb[0][1][zodb.pop()[0]]
# Start NEO cluster with transparent import.
with
NEOCluster
(
importer
=
importer
)
as
cluster
:
with
NEOCluster
(
importer
=
importer
,
partitions
=
2
)
as
cluster
:
# Suspend import for a while, so that import
# is finished in the middle of the below 'for' loop.
# Use a slightly different main loop for storage so that it
...
...
@@ -214,25 +235,26 @@ class ImporterTests(NEOThreadedTest):
logging
.
info
(
"start migration"
)
dm
.
doOperation
(
cluster
.
storage
)
# Adjust if needed. Must remain > 0.
self
.
assertEqual
(
22
,
h
())
self
.
assertEqual
(
import_hash
,
h
.
hexdigest
())
beforeCheck
(
h
,
22
)
# New writes after the switch to NEO.
last_import
=
-
1
for
i
,
r
in
enumerate
(
random_tree
.
importTree
(
r
,
after_tree
,
txn_size
)):
for
i
,
r
in
enumerate
(
after
(
r
)):
c
.
readCurrent
(
r
)
t
.
commit
()
if
cluster
.
storage
.
dm
.
_import
:
last_import
=
i
for
x
in
0
,
1
:
undo
=
TransactionalUndo
(
c
.
db
(),
[
storage
.
lastTransaction
()])
txn
=
transaction
.
Transaction
()
undo
.
tpc_begin
(
txn
)
undo
.
commit
(
txn
)
undo
.
tpc_vote
(
txn
)
undo
.
tpc_finish
(
txn
)
self
.
tic
()
# Same as above. We want last_import smaller enough compared to i
assert
i
<
last_import
*
3
<
2
*
i
,
(
last_import
,
i
)
self
.
assertFalse
(
cluster
.
storage
.
dm
.
_import
)
storage
.
_cache
.
clear
()
def
finalCheck
(
r
):
h
=
random_tree
.
hashTree
(
r
)
self
.
assertEqual
(
93
,
h
())
self
.
assertEqual
(
'6bf0f0cb2d6c1aae9e52c412ef0e25b6'
,
h
.
hexdigest
())
finalCheck
(
r
)
if
dm
.
_writeback
:
dm
.
commit
()
...
...
@@ -243,10 +265,10 @@ class ImporterTests(NEOThreadedTest):
db
.
close
()
@
unittest
.
skipUnless
(
importer
.
FORK
,
'no os.fork'
)
def
test
1
(
self
):
def
test
MultiProcessWriteBack
(
self
):
self
.
_importFromFileStorage
()
def
testThreadedWriteback
(
self
):
def
testThreadedWriteback
AndDBReconnection
(
self
):
# Also check reconnection to the underlying DB for relevant backends.
tid_list
=
[]
def
__init__
(
orig
,
tr
,
db
,
tid
):
...
...
@@ -272,7 +294,25 @@ class ImporterTests(NEOThreadedTest):
Patch
(
time
,
sleep
=
sleep
)
as
p
:
self
.
_importFromFileStorage
()
self
.
assertFalse
(
p
.
applied
)
self
.
assertEqual
(
len
(
tid_list
),
11
)
self
.
assertEqual
(
len
(
tid_list
),
13
)
def
testThreadedWritebackWithUnbalancedPartitions
(
self
):
N
=
7
nonlocal_
=
[
0
]
def
committed
(
orig
,
self
):
if
nonlocal_
[
0
]
>
N
:
orig
(
self
)
def
_nextTID
(
orig
,
self
,
*
args
):
if
args
:
return
orig
(
self
,
*
args
)
nonlocal_
[
0
]
+=
1
return
orig
(
self
,
p64
(
nonlocal_
[
0
]
==
N
),
2
)
with
Patch
(
importer
,
FORK
=
False
),
\
Patch
(
TransactionManager
,
_nextTID
=
_nextTID
),
\
Patch
(
WriteBack
,
chunk_size
=
N
-
2
),
\
Patch
(
WriteBack
,
committed
=
committed
):
self
.
_importFromFileStorage
()
self
.
assertEqual
(
nonlocal_
[
0
],
12
)
def
testMerge
(
self
):
multi
=
1
,
2
,
3
...
...
@@ -285,5 +325,52 @@ class ImporterTests(NEOThreadedTest):
# merge several DB.
testMerge
=
expectedFailure
(
NEOPrimaryMasterLost
)(
testMerge
)
def
testIncremental
(
self
):
"""
This reproduces an undocumented way to speed up the import of a single
ZODB by doing most of the work before switching to NEO.
"""
beforeCheck
,
before
,
finalCheck
,
after
=
self
.
getData
()
fs_path
,
cfg
=
self
.
getFS
()
c
=
ZODB
.
DB
(
FileStorage
(
fs_path
)).
open
()
r
=
c
.
root
()[
'tree'
]
=
random_tree
.
Node
()
transaction
.
commit
()
for
_
in
before
(
r
):
transaction
.
commit
()
c
.
db
().
close
()
importer
=
{
'zodb'
:
[(
'root'
,
cfg
)]}
# Start NEO cluster with transparent import.
with
NEOCluster
(
importer
=
importer
,
partitions
=
2
)
as
cluster
:
s
=
cluster
.
storage
l
=
threading
.
Lock
()
l
.
acquire
()
def
_finished
(
orig
):
orig
()
l
.
release
()
with
Patch
(
s
.
dm
,
_finished
=
_finished
):
cluster
.
start
()
l
.
acquire
()
t
,
c
=
cluster
.
getTransaction
()
r
=
c
.
root
()[
'tree'
]
beforeCheck
(
random_tree
.
hashTree
(
r
))
c
=
ZODB
.
DB
(
FileStorage
(
fs_path
)).
open
()
for
_
in
after
(
c
.
root
()[
'tree'
]):
transaction
.
commit
()
c
.
db
().
close
()
# TODO: Add a storage option that only does this and exits.
# Such command would also check that there's no data after
# what's already imported.
s
.
dm
.
setConfiguration
(
'zodb'
,
None
)
s
.
stop
()
cluster
.
join
((
s
,))
s
.
resetNode
()
with
Patch
(
s
.
dm
,
_finished
=
_finished
):
s
.
start
()
self
.
tic
()
l
.
acquire
()
t
.
begin
()
finalCheck
(
r
)
if
__name__
==
"__main__"
:
unittest
.
main
()
neo/tests/threaded/testReplication.py
View file @
8ba42463
...
...
@@ -29,7 +29,7 @@ from neo.storage.database.manager import DatabaseManager
from
neo.storage
import
replicator
from
neo.lib.connector
import
SocketConnector
from
neo.lib.connection
import
ClientConnection
from
neo.lib.protocol
import
CellStates
,
ClusterStates
,
Packets
,
\
from
neo.lib.protocol
import
CellStates
,
ClusterStates
,
NodeStates
,
Packets
,
\
ZERO_OID
,
ZERO_TID
,
MAX_TID
,
uuid_str
from
neo.lib.util
import
add64
,
p64
,
u64
from
..
import
Patch
,
TransactionalResource
...
...
@@ -74,6 +74,8 @@ class ReplicationTests(NEOThreadedTest):
source_dict
=
{
x
.
uuid
:
x
for
x
in
cluster
.
upstream
.
storage_list
}
for
storage
in
cluster
.
storage_list
:
self
.
assertFalse
(
storage
.
dm
.
_uncommitted_data
)
if
storage
.
pt
is
None
:
storage
.
loadPartitionTable
()
self
.
assertEqual
(
np
,
storage
.
pt
.
getPartitions
())
for
partition
in
pt
.
getAssignedPartitionList
(
storage
.
uuid
):
cell_list
=
upstream_pt
.
getCellList
(
partition
,
readable
=
True
)
...
...
@@ -89,6 +91,7 @@ class ReplicationTests(NEOThreadedTest):
checksum_list
=
[
self
.
checksumPartition
(
storage_dict
[
x
.
getUUID
()],
offset
)
for
x
in
pt
.
getCellList
(
offset
)]
self
.
assertLess
(
1
,
len
(
checksum_list
))
self
.
assertEqual
(
1
,
len
(
set
(
checksum_list
)),
(
offset
,
checksum_list
))
...
...
@@ -103,7 +106,7 @@ class ReplicationTests(NEOThreadedTest):
importZODB
(
3
)
def
delaySecondary
(
conn
,
packet
):
if
isinstance
(
packet
,
Packets
.
Replicate
):
tid
,
upstream_name
,
source_dict
=
packet
.
decode
()
tid
,
upstream_name
,
source_dict
=
packet
.
_args
return
not
upstream_name
and
all
(
source_dict
.
itervalues
())
with
NEOCluster
(
partitions
=
np
,
replicas
=
nr
-
1
,
storage_count
=
5
,
upstream
=
upstream
)
as
backup
:
...
...
@@ -346,6 +349,22 @@ class ReplicationTests(NEOThreadedTest):
self
.
tic
()
self
.
assertTrue
(
backup
.
master
.
is_alive
())
@
with_cluster
(
master_count
=
2
)
def
testBackupFromUpstreamWithSecondaryMaster
(
self
,
upstream
):
"""
Check that the backup master reacts correctly when connecting first
to a secondary master of the upstream cluster.
"""
with
NEOCluster
(
upstream
=
upstream
)
as
backup
:
primary
=
upstream
.
primary_master
m
,
=
(
m
for
m
in
upstream
.
master_list
if
m
is
not
primary
)
backup
.
master
.
resetNode
(
upstream_masters
=
[
m
.
server
])
backup
.
start
()
backup
.
neoctl
.
setClusterState
(
ClusterStates
.
STARTING_BACKUP
)
self
.
tic
()
self
.
assertEqual
(
backup
.
neoctl
.
getClusterState
(),
ClusterStates
.
BACKINGUP
)
@
backup_test
()
def
testCreationUndone
(
self
,
backup
):
"""
...
...
@@ -443,15 +462,15 @@ class ReplicationTests(NEOThreadedTest):
"""
def
delayAskFetch
(
conn
,
packet
):
return
isinstance
(
packet
,
delayed
)
and
\
packet
.
decode
()
[
0
]
==
offset
and
\
packet
.
_args
[
0
]
==
offset
and
\
conn
in
s1
.
getConnectionList
(
s0
)
def
changePartitionTable
(
orig
,
ptid
,
cell_list
):
def
changePartitionTable
(
orig
,
ptid
,
num_replicas
,
cell_list
):
if
(
offset
,
s0
.
uuid
,
CellStates
.
DISCARDED
)
in
cell_list
:
connection_filter
.
remove
(
delayAskFetch
)
# XXX: this is currently not done by
# default for performance reason
orig
.
im_self
.
dropPartitions
((
offset
,))
return
orig
(
ptid
,
cell_list
)
return
orig
(
ptid
,
num_replicas
,
cell_list
)
np
=
cluster
.
num_partitions
s0
,
s1
,
s2
=
cluster
.
storage_list
for
delayed
in
Packets
.
AskFetchTransactions
,
Packets
.
AskFetchObjects
:
...
...
@@ -511,7 +530,9 @@ class ReplicationTests(NEOThreadedTest):
for
x
in
'ab'
:
r
[
x
]
=
PCounter
()
t
.
commit
()
cluster
.
stop
(
replicas
=
1
)
cluster
.
neoctl
.
setNumReplicas
(
1
)
self
.
tic
()
cluster
.
stop
()
cluster
.
start
((
s1
,
s2
))
with
ConnectionFilter
()
as
f
:
f
.
delayAddObject
()
...
...
@@ -596,8 +617,9 @@ class ReplicationTests(NEOThreadedTest):
tweak
()
t
.
commit
()
t2
.
join
()
cluster
.
neoctl
.
dropNode
(
S
[
2
].
uuid
)
cluster
.
neoctl
.
dropNode
(
S
[
3
].
uuid
)
for
s
in
S
[
2
:]:
cluster
.
neoctl
.
killNode
(
s
.
uuid
)
cluster
.
neoctl
.
dropNode
(
s
.
uuid
)
cluster
.
neoctl
.
tweakPartitionTable
()
if
done
:
f
.
remove
(
delay
)
...
...
@@ -695,7 +717,7 @@ class ReplicationTests(NEOThreadedTest):
def
logReplication
(
conn
,
packet
):
if
isinstance
(
packet
,
(
Packets
.
AskFetchTransactions
,
Packets
.
AskFetchObjects
)):
ask
.
append
(
packet
.
decode
()
[
2
:])
ask
.
append
(
packet
.
_args
[
2
:])
def
getTIDList
():
return
[
t
.
tid
for
t
in
c
.
db
().
storage
.
iterator
()]
s0
,
s1
=
cluster
.
storage_list
...
...
@@ -796,7 +818,7 @@ class ReplicationTests(NEOThreadedTest):
return
True
elif
not
isinstance
(
packet
,
Packets
.
AskFetchTransactions
):
return
ask
.
append
(
packet
.
decode
()
)
ask
.
append
(
packet
.
_args
)
conn
,
=
upstream
.
master
.
getConnectionList
(
backup
.
master
)
with
ConnectionFilter
()
as
f
,
Patch
(
replicator
.
Replicator
,
_nextPartitionSortKey
=
lambda
orig
,
self
,
offset
:
offset
):
...
...
@@ -857,11 +879,11 @@ class ReplicationTests(NEOThreadedTest):
@
f
.
add
def
delayReplicate
(
conn
,
packet
):
if
isinstance
(
packet
,
Packets
.
AskFetchTransactions
):
trans
.
append
(
packet
.
decode
()
[
2
])
trans
.
append
(
packet
.
_args
[
2
])
elif
isinstance
(
packet
,
Packets
.
AskFetchObjects
):
if
obj
:
return
True
obj
.
append
(
packet
.
decode
()
[
2
])
obj
.
append
(
packet
.
_args
[
2
])
s2
.
start
()
self
.
tic
()
cluster
.
neoctl
.
enableStorageList
([
s2
.
uuid
])
...
...
@@ -928,6 +950,74 @@ class ReplicationTests(NEOThreadedTest):
def
testReplicationBlockedByUnfinished2
(
self
):
self
.
testReplicationBlockedByUnfinished1
(
True
)
@
with_cluster
(
partitions
=
6
,
storage_count
=
5
,
start_cluster
=
0
)
def
testSplitAndMakeResilientUsingClone
(
self
,
cluster
):
"""
Test cloning of storage nodes using --new-nid instead NEO replication.
"""
s0
=
cluster
.
storage_list
[
0
]
s12
=
cluster
.
storage_list
[
1
:
3
]
s34
=
cluster
.
storage_list
[
3
:]
cluster
.
start
(
storage_list
=
(
s0
,))
cluster
.
importZODB
()(
6
)
for
s
in
s12
:
s
.
start
()
self
.
tic
()
drop_list
=
[
s0
.
uuid
]
self
.
assertRaises
(
SystemExit
,
cluster
.
neoctl
.
tweakPartitionTable
,
drop_list
)
cluster
.
enableStorageList
(
s12
)
def
expected
(
changed
):
s0
=
1
,
CellStates
.
UP_TO_DATE
s
=
CellStates
.
OUT_OF_DATE
if
changed
else
CellStates
.
UP_TO_DATE
return
changed
,
3
*
((
s0
,
(
2
,
s
)),
(
s0
,
(
3
,
s
)))
for
dry_run
in
True
,
False
:
self
.
assertEqual
(
expected
(
True
),
cluster
.
neoctl
.
tweakPartitionTable
(
drop_list
,
dry_run
))
self
.
tic
()
self
.
assertEqual
(
expected
(
False
),
cluster
.
neoctl
.
tweakPartitionTable
(
drop_list
))
for
s
,
d
in
zip
(
s12
,
s34
):
s
.
stop
()
cluster
.
join
((
s
,))
s
.
resetNode
()
d
.
dm
.
restore
(
s
.
dm
.
dump
())
d
.
resetNode
(
new_nid
=
True
)
s
.
start
()
d
.
start
()
self
.
tic
()
self
.
assertEqual
(
cluster
.
getNodeState
(
s
),
NodeStates
.
RUNNING
)
self
.
assertEqual
(
cluster
.
getNodeState
(
d
),
NodeStates
.
DOWN
)
cluster
.
join
((
d
,))
d
.
resetNode
(
new_nid
=
False
)
d
.
start
()
self
.
tic
()
self
.
checkReplicas
(
cluster
)
expected
=
'|'
.
join
([
'UU.U.|U.U.U'
]
*
3
)
self
.
assertPartitionTable
(
cluster
,
expected
)
cluster
.
neoctl
.
setNumReplicas
(
1
)
cluster
.
neoctl
.
tweakPartitionTable
(
drop_list
)
self
.
tic
()
self
.
assertPartitionTable
(
cluster
,
expected
)
s0
.
stop
()
cluster
.
join
((
s0
,))
cluster
.
neoctl
.
dropNode
(
s0
.
uuid
)
expected
=
'|'
.
join
([
'U.U.|.U.U'
]
*
3
)
self
.
assertPartitionTable
(
cluster
,
expected
)
@
with_cluster
(
partitions
=
3
,
replicas
=
1
,
storage_count
=
3
)
def
testAdminOnerousOperationCondition
(
self
,
cluster
):
s
=
cluster
.
storage_list
[
2
]
cluster
.
neoctl
.
killNode
(
s
.
uuid
)
tweak
=
cluster
.
neoctl
.
tweakPartitionTable
self
.
assertRaises
(
SystemExit
,
tweak
)
self
.
assertRaises
(
SystemExit
,
tweak
,
dry_run
=
True
)
self
.
assertTrue
(
tweak
((
s
.
uuid
,))[
0
])
self
.
tic
()
cluster
.
neoctl
.
dropNode
(
s
.
uuid
)
s
=
cluster
.
storage_list
[
1
]
self
.
assertRaises
(
SystemExit
,
cluster
.
neoctl
.
dropNode
,
s
.
uuid
)
@
with_cluster
(
partitions
=
5
,
replicas
=
2
,
storage_count
=
3
)
def
testCheckReplicas
(
self
,
cluster
):
from
neo.storage
import
checker
...
...
@@ -940,8 +1030,8 @@ class ReplicationTests(NEOThreadedTest):
return
s0
.
uuid
def
check
(
expected_state
,
expected_count
):
self
.
assertEqual
(
expected_count
,
len
([
None
for
row
in
cluster
.
neoctl
.
getPartitionRowList
()[
1
]
for
cell
in
row
[
1
]
for
row
in
cluster
.
neoctl
.
getPartitionRowList
()[
2
]
for
cell
in
row
if
cell
[
1
]
==
CellStates
.
CORRUPTED
]))
self
.
assertEqual
(
expected_state
,
cluster
.
neoctl
.
getClusterState
())
np
=
cluster
.
num_partitions
...
...
neo/tests/zodb/testRecovery.py
View file @
8ba42463
...
...
@@ -33,8 +33,6 @@ class RecoveryTests(ZODBTestCase, StorageTestBase, RecoveryStorage):
os
.
makedirs
(
dst_temp_dir
)
self
.
neo_dst
=
NEOCluster
([
'test_neo1-dst'
],
partitions
=
1
,
replicas
=
0
,
master_count
=
1
,
temp_dir
=
dst_temp_dir
)
self
.
neo_dst
.
stop
()
self
.
neo_dst
.
setupDB
()
self
.
neo_dst
.
start
()
self
.
_dst
=
self
.
neo
.
getZODBStorage
()
self
.
_dst_db
=
ZODB
.
DB
(
self
.
_dst
)
...
...
setup.py
View file @
8ba42463
...
...
@@ -53,7 +53,7 @@ extras_require = {
'master'
:
[],
'storage-sqlite'
:
[],
'storage-mysqldb'
:
[
'mysqlclient'
],
'storage-importer'
:
zodb_require
+
[
'
msgpack>=0.5.6'
,
'
setproctitle'
],
'storage-importer'
:
zodb_require
+
[
'setproctitle'
],
}
extras_require
[
'tests'
]
=
[
'coverage'
,
'zope.testing'
,
'psutil>=2'
,
'neoppod[%s]'
%
', '
.
join
(
extras_require
)]
...
...
@@ -78,7 +78,7 @@ else:
setup
(
name
=
'neoppod'
,
version
=
'1.1
1
'
,
version
=
'1.1
2.0
'
,
description
=
__doc__
.
strip
(),
author
=
'Nexedi SA'
,
author_email
=
'neo-dev@erp5.org'
,
...
...
@@ -108,6 +108,7 @@ setup(
],
},
install_requires
=
[
'msgpack>=0.5.6'
,
'python-dateutil'
,
# neolog --from
],
extras_require
=
extras_require
,
...
...
tools/matrix
View file @
8ba42463
...
...
@@ -129,7 +129,7 @@ class MatrixImportBenchmark(BenchmarkRunner):
finally
:
zodb
.
stop
()
# Clear DB if no error happened.
zodb
.
setup
DB
()
zodb
.
reset
DB
()
return
end
-
start
except
:
traceback
.
print_exc
()
...
...
tools/perfs
View file @
8ba42463
...
...
@@ -53,7 +53,7 @@ class ImportBenchmark(BenchmarkRunner):
finally
:
neo
.
stop
()
# Clear DB if no error happened.
neo
.
setup
DB
()
neo
.
reset
DB
()
return
result
except
:
return
'Perf: import failed'
,
''
.
join
(
traceback
.
format_exc
())
...
...
tools/stress
View file @
8ba42463
...
...
@@ -7,6 +7,7 @@ from contextlib import contextmanager
from
datetime
import
datetime
from
functools
import
partial
from
multiprocessing
import
Lock
,
RawArray
from
multiprocessing.queues
import
SimpleQueue
from
struct
import
Struct
from
netfilterqueue
import
NetfilterQueue
import
gevent.socket
# preload for subprocesses
...
...
@@ -19,7 +20,7 @@ from neo.lib.protocol import NodeTypes
from
neo.lib.util
import
timeStringFromTID
,
p64
,
u64
from
neo.storage.app
import
DATABASE_MANAGER_DICT
,
\
Application
as
StorageApplication
from
neo.tests
import
getTempDirectory
from
neo.tests
import
getTempDirectory
,
mysql_pool
from
neo.tests.ConflictFree
import
ConflictFreeLog
from
neo.tests.functional
import
AlreadyStopped
,
NEOCluster
,
Process
from
neo.tests.stress
import
StressApplication
...
...
@@ -312,13 +313,15 @@ class NEOCluster(NEOCluster):
class
Application
(
StressApplication
):
_blocking
=
None
_blocking
=
_kill_mysqld
=
None
def
__init__
(
self
,
client_count
,
thread_count
,
restart_ratio
,
logrotate
,
*
args
,
**
kw
):
def
__init__
(
self
,
client_count
,
thread_count
,
fault_probability
,
restart_ratio
,
kill_mysqld
,
logrotate
,
*
args
,
**
kw
):
self
.
client_count
=
client_count
self
.
thread_count
=
thread_count
self
.
logrotate
=
logrotate
self
.
fault_probability
=
fault_probability
self
.
restart_ratio
=
restart_ratio
self
.
cluster
=
cluster
=
NEOCluster
(
*
args
,
**
kw
)
# Make the firewall also affect connections between storage nodes.
...
...
@@ -326,7 +329,24 @@ class Application(StressApplication):
def
__init__
(
self
,
config
):
dscpPatch
(
1
)
StorageApplication__init__
(
self
,
config
)
StorageApplication
.
__init__
=
__init__
if
kill_mysqld
:
from
neo.scripts
import
neostorage
from
neo.storage.database
import
mysqldb
neostorage_main
=
neostorage
.
main
self
.
_kill_mysqld
=
kill_mysqld
=
SimpleQueue
()
def
main
():
pid
=
os
.
getpid
()
try
:
neostorage_main
()
except
mysqldb
.
OperationalError
as
e
:
code
=
e
.
args
[
0
]
except
mysqldb
.
MysqlError
as
e
:
code
=
e
.
code
if
mysqldb
.
SERVER_LOST
!=
code
!=
mysqldb
.
SERVER_GONE_ERROR
:
raise
kill_mysqld
.
put
(
pid
)
neostorage
.
main
=
main
super
(
Application
,
self
).
__init__
(
cluster
.
SSL
,
util
.
parseMasterList
(
cluster
.
master_nodes
))
...
...
@@ -398,6 +418,10 @@ class Application(StressApplication):
t
=
threading
.
Thread
(
target
=
self
.
_logrotate_thread
)
t
.
daemon
=
1
t
.
start
()
if
self
.
_kill_mysqld
:
t
=
threading
.
Thread
(
target
=
self
.
_watch_storage_thread
)
t
.
daemon
=
1
t
.
start
()
def
stopCluster
(
self
,
wait
=
None
):
self
.
restart_lock
.
acquire
()
...
...
@@ -471,13 +495,30 @@ class Application(StressApplication):
except
ValueError
:
pass
def
_watch_storage_thread
(
self
):
get
=
self
.
_kill_mysqld
.
get
storage_list
=
self
.
cluster
.
getStorageProcessList
()
while
1
:
pid
=
get
()
p
,
=
(
p
for
p
in
storage_list
if
p
.
pid
==
pid
)
p
.
wait
()
p
.
start
()
def
restartStorages
(
self
,
nids
):
processes
=
[
p
for
p
in
self
.
cluster
.
getStorageProcessList
()
if
p
.
uuid
in
nids
]
for
p
in
processes
:
p
.
kill
(
signal
.
SIGKILL
)
time
.
sleep
(
1
)
for
p
in
processes
:
p
.
wait
()
for
p
in
processes
:
p
.
start
()
storage_list
=
self
.
cluster
.
getStorageProcessList
()
if
self
.
_kill_mysqld
:
db_list
=
[
db
for
db
,
p
in
zip
(
self
.
cluster
.
db_list
,
storage_list
)
if
p
.
uuid
in
nids
]
mysql_pool
.
kill
(
*
db_list
)
time
.
sleep
(
1
)
with
open
(
os
.
devnull
,
"wb"
)
as
f
:
mysql_pool
.
start
(
*
db_list
,
stderr
=
f
)
else
:
processes
=
[
p
for
p
in
storage_list
if
p
.
uuid
in
nids
]
for
p
in
processes
:
p
.
kill
(
signal
.
SIGKILL
)
time
.
sleep
(
1
)
for
p
in
processes
:
p
.
wait
()
for
p
in
processes
:
p
.
start
()
def
_cleanFirewall
(
self
):
with
open
(
os
.
devnull
,
"wb"
)
as
f
:
...
...
@@ -548,6 +589,7 @@ def main():
default
=
socket
.
AF_INET
,
const
=
socket
.
AF_INET6
,
help
=
'(default: IPv4)'
)
_
(
'-a'
,
'--adapter'
,
choices
=
adapters
,
default
=
default_adapter
)
_
(
'-d'
,
'--datadir'
,
help
=
"(default: same as unit tests)"
)
_
(
'-e'
,
'--engine'
,
help
=
"database engine (MySQL only)"
)
_
(
'-l'
,
'--logdir'
,
help
=
"(default: same as --datadir)"
)
_
(
'-m'
,
'--masters'
,
type
=
int
,
default
=
1
)
_
(
'-s'
,
'--storages'
,
type
=
int
,
default
=
8
)
...
...
@@ -571,9 +613,14 @@ def main():
help
=
'number of client processes'
)
_
(
'-t'
,
'--threads'
,
type
=
int
,
default
=
1
,
help
=
'number of thread workers per client process'
)
_
(
'-f'
,
'--fault-probability'
,
type
=
ratio
,
default
=
1
,
metavar
=
'P'
,
help
=
'probability to cause faults every second'
)
_
(
'-r'
,
'--restart-ratio'
,
type
=
ratio
,
default
=
.
5
,
metavar
=
'RATIO'
,
help
=
'probability to kill/restart a storage node, rather than just'
' RSTing a TCP connection with this node'
)
_
(
'--kill-mysqld'
,
action
=
'store_true'
,
help
=
'if r != 0 and if NEO_DB_MYCNF is set,'
' kill mysqld rather than storage node'
)
_
(
'-C'
,
'--console'
,
type
=
int
,
default
=
0
,
help
=
'console port (localhost) (default: any)'
)
_
(
'-D'
,
'--delay'
,
type
=
float
,
default
=
.
01
,
...
...
@@ -594,18 +641,31 @@ def main():
db_list
=
[
'stress_neo%s'
%
x
for
x
in
xrange
(
args
.
storages
)]
if
args
.
datadir
:
if
args
.
adapter
!=
'SQLite'
:
parser
.
error
(
'--datadir is only for SQLite adapter'
)
db_list
=
[
os
.
path
.
join
(
args
.
datadir
,
x
+
'.sqlite'
)
for
x
in
db_list
]
if
args
.
adapter
==
'SQLite'
:
db_list
=
[
os
.
path
.
join
(
args
.
datadir
,
x
+
'.sqlite'
)
for
x
in
db_list
]
elif
mysql_pool
:
mysql_pool
.
__init__
(
args
.
datadir
)
else
:
parser
.
error
(
'--datadir: meaningless when using an existing MySQL server'
)
kw
=
dict
(
db_list
=
db_list
,
name
=
'stress'
,
partitions
=
args
.
partitions
,
replicas
=
args
.
replicas
,
adapter
=
args
.
adapter
,
address_type
=
args
.
address_type
,
temp_dir
=
args
.
logdir
or
args
.
datadir
or
getTempDirectory
())
temp_dir
=
args
.
logdir
or
args
.
datadir
or
getTempDirectory
(),
storage_kw
=
{
'engine'
:
args
.
engine
,
'wait'
:
-
1
})
if
args
.
command
==
'run'
:
NFQueue
.
delay
=
args
.
delay
app
=
Application
(
args
.
clients
,
args
.
threads
,
args
.
restart_ratio
,
error
=
args
.
kill_mysqld
and
(
'invalid adapter'
if
args
.
adapter
!=
'MySQL'
else
None
if
mysql_pool
else
'NEO_DB_MYCNF not set'
)
if
error
:
parser
.
error
(
'--kill-mysqld: '
+
error
)
app
=
Application
(
args
.
clients
,
args
.
threads
,
args
.
fault_probability
,
args
.
restart_ratio
,
args
.
kill_mysqld
,
int
(
round
(
args
.
logrotate
*
3600
,
0
)),
**
kw
)
t
=
threading
.
Thread
(
target
=
console
,
args
=
(
args
.
console
,
app
))
t
.
daemon
=
1
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment