Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
N
neoppod
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
1
Issues
1
List
Boards
Labels
Milestones
Merge Requests
2
Merge Requests
2
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
nexedi
neoppod
Commits
80038d2f
Commit
80038d2f
authored
Jan 10, 2012
by
Julien Muchembled
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Fix update of partition when cluster is not operational due a storage lost
parent
6aa372d9
Changes
5
Show whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
36 additions
and
24 deletions
+36
-24
neo/master/app.py
neo/master/app.py
+0
-4
neo/master/handlers/storage.py
neo/master/handlers/storage.py
+5
-7
neo/master/pt.py
neo/master/pt.py
+18
-6
neo/tests/functional/testStorage.py
neo/tests/functional/testStorage.py
+13
-6
neo/tests/master/testStorageHandler.py
neo/tests/master/testStorageHandler.py
+0
-1
No files found.
neo/master/app.py
View file @
80038d2f
...
@@ -258,10 +258,6 @@ class Application(object):
...
@@ -258,10 +258,6 @@ class Application(object):
if
selector
(
node
):
if
selector
(
node
):
node
.
notify
(
packet
)
node
.
notify
(
packet
)
def
outdateAndBroadcastPartition
(
self
):
""" Outdate cell of non-working nodes and broadcast changes """
self
.
broadcastPartitionChanges
(
self
.
pt
.
outdate
())
def
broadcastLastOID
(
self
):
def
broadcastLastOID
(
self
):
oid
=
self
.
tm
.
getLastOID
()
oid
=
self
.
tm
.
getLastOID
()
neo
.
lib
.
logging
.
debug
(
neo
.
lib
.
logging
.
debug
(
...
...
neo/master/handlers/storage.py
View file @
80038d2f
...
@@ -40,14 +40,12 @@ class StorageServiceHandler(BaseServiceHandler):
...
@@ -40,14 +40,12 @@ class StorageServiceHandler(BaseServiceHandler):
def
nodeLost
(
self
,
conn
,
node
):
def
nodeLost
(
self
,
conn
,
node
):
neo
.
lib
.
logging
.
info
(
'storage node lost'
)
neo
.
lib
.
logging
.
info
(
'storage node lost'
)
assert
not
node
.
isRunning
(),
node
.
getState
()
assert
not
node
.
isRunning
(),
node
.
getState
()
app
=
self
.
app
if
not
self
.
app
.
pt
.
operational
():
app
.
broadcastPartitionChanges
(
app
.
pt
.
outdate
(
node
))
if
not
app
.
pt
.
operational
():
raise
OperationFailure
,
'cannot continue operation'
raise
OperationFailure
,
'cannot continue operation'
# this is intentionaly placed after the raise because the last cell in a
app
.
tm
.
forget
(
conn
.
getUUID
())
# partition must not oudated to allows a cluster restart.
if
app
.
packing
is
not
None
:
self
.
app
.
outdateAndBroadcastPartition
()
self
.
app
.
tm
.
forget
(
conn
.
getUUID
())
if
self
.
app
.
packing
is
not
None
:
self
.
answerPack
(
conn
,
False
)
self
.
answerPack
(
conn
,
False
)
def
askLastIDs
(
self
,
conn
):
def
askLastIDs
(
self
,
conn
):
...
...
neo/master/pt.py
View file @
80038d2f
...
@@ -291,16 +291,28 @@ class PartitionTable(PartitionTable):
...
@@ -291,16 +291,28 @@ class PartitionTable(PartitionTable):
self
.
log
()
self
.
log
()
return
changed_cell_list
return
changed_cell_list
def
outdate
(
self
):
def
outdate
(
self
,
lost_node
=
None
):
"""Outdate all non-working nodes."""
"""Outdate all non-working nodes
cell_list
=
[]
Do not outdate cells of 'lost_node' for partitions it was the last node
to serve. This allows a cluster restart.
"""
change_list
=
[]
for
offset
,
row
in
enumerate
(
self
.
partition_list
):
for
offset
,
row
in
enumerate
(
self
.
partition_list
):
lost
=
lost_node
cell_list
=
[]
for
cell
in
row
:
for
cell
in
row
:
if
not
cell
.
getNode
().
isRunning
()
and
not
cell
.
isOutOfDate
():
if
cell
.
isUpToDate
()
or
cell
.
isFeeding
():
if
cell
.
getNode
().
isRunning
():
lost
=
None
else
:
cell_list
.
append
(
cell
)
for
cell
in
cell_list
:
if
cell
.
getNode
()
is
not
lost
:
cell
.
setState
(
CellStates
.
OUT_OF_DATE
)
cell
.
setState
(
CellStates
.
OUT_OF_DATE
)
c
ell
_list
.
append
((
offset
,
cell
.
getUUID
(),
c
hange
_list
.
append
((
offset
,
cell
.
getUUID
(),
CellStates
.
OUT_OF_DATE
))
CellStates
.
OUT_OF_DATE
))
return
c
ell
_list
return
c
hange
_list
def
getUpToDateCellNodeSet
(
self
):
def
getUpToDateCellNodeSet
(
self
):
"""
"""
...
...
neo/tests/functional/testStorage.py
View file @
80038d2f
...
@@ -189,19 +189,26 @@ class StorageTests(NEOFunctionalTest):
...
@@ -189,19 +189,26 @@ class StorageTests(NEOFunctionalTest):
down, the cluster remains up since there is a replica """
down, the cluster remains up since there is a replica """
# populate the two storages
# populate the two storages
(
started
,
_
)
=
self
.
__setup
(
storage_number
=
2
,
replicas
=
1
)
started
,
_
=
self
.
__setup
(
partitions
=
3
,
replicas
=
1
,
storage_number
=
3
)
self
.
neo
.
expectRunning
(
started
[
0
])
self
.
neo
.
expectRunning
(
started
[
0
])
self
.
neo
.
expectRunning
(
started
[
1
])
self
.
neo
.
expectRunning
(
started
[
1
])
self
.
neo
.
expectRunning
(
started
[
2
])
self
.
neo
.
expectOudatedCells
(
number
=
0
)
self
.
neo
.
expectOudatedCells
(
number
=
0
)
self
.
__populate
()
self
.
__checkReplicationDone
()
self
.
neo
.
expectClusterRunning
()
# stop one storage and check outdated cells
started
[
0
].
stop
()
started
[
0
].
stop
()
self
.
neo
.
expectOudatedCells
(
number
=
10
)
# Cluster still operational. All cells of first storage should be
# outdated.
self
.
neo
.
expectUnavailable
(
started
[
0
])
self
.
neo
.
expectOudatedCells
(
2
)
self
.
neo
.
expectClusterRunning
()
self
.
neo
.
expectClusterRunning
()
started
[
1
].
stop
()
# Cluster not operational anymore. Only cells of second storage that
# were shared with the third one should become outdated.
self
.
neo
.
expectUnavailable
(
started
[
1
])
self
.
neo
.
expectClusterVerifying
()
self
.
neo
.
expectOudatedCells
(
3
)
def
testVerificationTriggered
(
self
):
def
testVerificationTriggered
(
self
):
""" Check that the verification stage is executed when a storage node
""" Check that the verification stage is executed when a storage node
required to be operationnal is lost, and the cluster come back in
required to be operationnal is lost, and the cluster come back in
...
...
neo/tests/master/testStorageHandler.py
View file @
80038d2f
...
@@ -192,7 +192,6 @@ class MasterStorageHandlerTests(NeoUnitTestBase):
...
@@ -192,7 +192,6 @@ class MasterStorageHandlerTests(NeoUnitTestBase):
# Some shortcuts to simplify test code
# Some shortcuts to simplify test code
self
.
app
.
pt
=
Mock
({
'operational'
:
True
})
self
.
app
.
pt
=
Mock
({
'operational'
:
True
})
self
.
app
.
outdateAndBroadcastPartition
=
lambda
:
None
# Register some transactions
# Register some transactions
tm
=
self
.
app
.
tm
tm
=
self
.
app
.
tm
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment