Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
M
mariadb
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
mariadb
Commits
61f87f5d
Commit
61f87f5d
authored
Oct 25, 2006
by
jonas@perch.ndb.mysql.com
Browse files
Options
Browse Files
Download
Plain Diff
Merge perch.ndb.mysql.com:/home/jonas/src/50-work
into perch.ndb.mysql.com:/home/jonas/src/mysql-5.0-ndb
parents
6e4a0ec0
822e638d
Changes
4
Show whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
12 additions
and
5 deletions
+12
-5
ndb/include/kernel/ndb_limits.h
ndb/include/kernel/ndb_limits.h
+2
-0
ndb/src/kernel/blocks/backup/Backup.cpp
ndb/src/kernel/blocks/backup/Backup.cpp
+2
-2
ndb/src/kernel/blocks/backup/Backup.hpp
ndb/src/kernel/blocks/backup/Backup.hpp
+4
-2
ndb/src/kernel/blocks/dbdih/DbdihMain.cpp
ndb/src/kernel/blocks/dbdih/DbdihMain.cpp
+4
-1
No files found.
ndb/include/kernel/ndb_limits.h
View file @
61f87f5d
...
@@ -64,6 +64,8 @@
...
@@ -64,6 +64,8 @@
#define MAX_NULL_BITS 4096
#define MAX_NULL_BITS 4096
#define MAX_FRAGMENT_DATA_BYTES (4+(2 * 8 * MAX_REPLICAS * MAX_NDB_NODES))
#define MAX_FRAGMENT_DATA_BYTES (4+(2 * 8 * MAX_REPLICAS * MAX_NDB_NODES))
#define MAX_WORDS_META_FILE 24576
#define MIN_ATTRBUF ((MAX_ATTRIBUTES_IN_TABLE/24) + 1)
#define MIN_ATTRBUF ((MAX_ATTRIBUTES_IN_TABLE/24) + 1)
/*
/*
* Max Number of Records to fetch per SCAN_NEXTREQ in a scan in LQH. The
* Max Number of Records to fetch per SCAN_NEXTREQ in a scan in LQH. The
...
...
ndb/src/kernel/blocks/backup/Backup.cpp
View file @
61f87f5d
...
@@ -2547,8 +2547,8 @@ Backup::execDEFINE_BACKUP_REQ(Signal* signal)
...
@@ -2547,8 +2547,8 @@ Backup::execDEFINE_BACKUP_REQ(Signal* signal)
0
// 3M
0
// 3M
};
};
const
Uint32
maxInsert
[]
=
{
const
Uint32
maxInsert
[]
=
{
2048
,
// Temporarily to solve TR515
MAX_WORDS_META_FILE
,
4096
,
//
4
k
4096
,
//
16
k
16
*
3000
,
// Max 16 tuples
16
*
3000
,
// Max 16 tuples
};
};
Uint32
minWrite
[]
=
{
Uint32
minWrite
[]
=
{
...
...
ndb/src/kernel/blocks/backup/Backup.hpp
View file @
61f87f5d
...
@@ -518,7 +518,9 @@ public:
...
@@ -518,7 +518,9 @@ public:
Config
c_defaults
;
Config
c_defaults
;
Uint32
m_diskless
;
Uint32
m_diskless
;
STATIC_CONST
(
NO_OF_PAGES_META_FILE
=
2
);
STATIC_CONST
(
NO_OF_PAGES_META_FILE
=
(
MAX_WORDS_META_FILE
+
BACKUP_WORDS_PER_PAGE
-
1
)
/
BACKUP_WORDS_PER_PAGE
);
/**
/**
* Pools
* Pools
...
...
ndb/src/kernel/blocks/dbdih/DbdihMain.cpp
View file @
61f87f5d
...
@@ -8481,14 +8481,17 @@ Dbdih::resetReplicaSr(TabRecordPtr tabPtr){
...
@@ -8481,14 +8481,17 @@ Dbdih::resetReplicaSr(TabRecordPtr tabPtr){
ConstPtr
<
ReplicaRecord
>
constReplicaPtr
;
ConstPtr
<
ReplicaRecord
>
constReplicaPtr
;
constReplicaPtr
.
i
=
replicaPtr
.
i
;
constReplicaPtr
.
i
=
replicaPtr
.
i
;
constReplicaPtr
.
p
=
replicaPtr
.
p
;
constReplicaPtr
.
p
=
replicaPtr
.
p
;
if
(
setup_create_replica
(
fragPtr
,
if
(
tabPtr
.
p
->
storedTable
==
0
||
setup_create_replica
(
fragPtr
,
&
createReplica
,
constReplicaPtr
))
&
createReplica
,
constReplicaPtr
))
{
{
jam
();
removeOldStoredReplica
(
fragPtr
,
replicaPtr
);
removeOldStoredReplica
(
fragPtr
,
replicaPtr
);
linkStoredReplica
(
fragPtr
,
replicaPtr
);
linkStoredReplica
(
fragPtr
,
replicaPtr
);
}
}
else
else
{
{
jam
();
infoEvent
(
"Forcing take-over of node %d due to unsufficient REDO"
infoEvent
(
"Forcing take-over of node %d due to unsufficient REDO"
" for table %d fragment: %d"
,
" for table %d fragment: %d"
,
nodePtr
.
i
,
tabPtr
.
i
,
i
);
nodePtr
.
i
,
tabPtr
.
i
,
i
);
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment