Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
M
mariadb
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
mariadb
Commits
e39ba639
Commit
e39ba639
authored
Apr 13, 2005
by
mskold@mysql.com
Browse files
Options
Browse Files
Download
Plain Diff
Merge
parents
d5646eb6
f5855f2d
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
22 additions
and
20 deletions
+22
-20
ndb/src/ndbapi/Ndb.cpp
ndb/src/ndbapi/Ndb.cpp
+7
-7
sql/ha_ndbcluster.cc
sql/ha_ndbcluster.cc
+14
-12
sql/mysqld.cc
sql/mysqld.cc
+1
-1
No files found.
ndb/src/ndbapi/Ndb.cpp
View file @
e39ba639
...
...
@@ -727,7 +727,7 @@ Ndb::getAutoIncrementValue(const char* aTableName, Uint32 cacheSize)
Ndb_local_table_info
*
info
=
theDictionary
->
get_local_table_info
(
internalTableName
,
false
);
if
(
info
==
0
)
DBUG_RETURN
(
~
0
);
DBUG_RETURN
(
~
(
Uint64
)
0
);
const
NdbTableImpl
*
table
=
info
->
m_table_impl
;
Uint64
tupleId
=
getTupleIdFromNdb
(
table
->
m_tableId
,
cacheSize
);
DBUG_PRINT
(
"info"
,
(
"value %u"
,
tupleId
));
...
...
@@ -739,7 +739,7 @@ Ndb::getAutoIncrementValue(const NdbDictionary::Table * aTable, Uint32 cacheSize
{
DBUG_ENTER
(
"getAutoIncrementValue"
);
if
(
aTable
==
0
)
DBUG_RETURN
(
~
0
);
DBUG_RETURN
(
~
(
Uint64
)
0
);
const
NdbTableImpl
*
table
=
&
NdbTableImpl
::
getImpl
(
*
aTable
);
Uint64
tupleId
=
getTupleIdFromNdb
(
table
->
m_tableId
,
cacheSize
);
DBUG_PRINT
(
"info"
,
(
"value %u"
,
tupleId
));
...
...
@@ -751,7 +751,7 @@ Ndb::getTupleIdFromNdb(const char* aTableName, Uint32 cacheSize)
{
const
NdbTableImpl
*
table
=
theDictionary
->
getTable
(
aTableName
);
if
(
table
==
0
)
return
~
0
;
return
~
(
Uint64
)
0
;
return
getTupleIdFromNdb
(
table
->
m_tableId
,
cacheSize
);
}
...
...
@@ -780,7 +780,7 @@ Ndb::readAutoIncrementValue(const char* aTableName)
const
NdbTableImpl
*
table
=
theDictionary
->
getTable
(
aTableName
);
if
(
table
==
0
)
{
theError
=
theDictionary
->
getNdbError
();
DBUG_RETURN
(
~
0
);
DBUG_RETURN
(
~
(
Uint64
)
0
);
}
Uint64
tupleId
=
readTupleIdFromNdb
(
table
->
m_tableId
);
DBUG_PRINT
(
"info"
,
(
"value %u"
,
tupleId
));
...
...
@@ -792,7 +792,7 @@ Ndb::readAutoIncrementValue(const NdbDictionary::Table * aTable)
{
DBUG_ENTER
(
"readtAutoIncrementValue"
);
if
(
aTable
==
0
)
DBUG_RETURN
(
~
0
);
DBUG_RETURN
(
~
(
Uint64
)
0
);
const
NdbTableImpl
*
table
=
&
NdbTableImpl
::
getImpl
(
*
aTable
);
Uint64
tupleId
=
readTupleIdFromNdb
(
table
->
m_tableId
);
DBUG_PRINT
(
"info"
,
(
"value %u"
,
tupleId
));
...
...
@@ -829,7 +829,7 @@ Ndb::setAutoIncrementValue(const NdbDictionary::Table * aTable, Uint64 val, bool
{
DEBUG_TRACE
(
"setAutoIncrementValue "
<<
val
);
if
(
aTable
==
0
)
return
~
0
;
return
~
(
Uint64
)
0
;
const
NdbTableImpl
*
table
=
&
NdbTableImpl
::
getImpl
(
*
aTable
);
return
setTupleIdInNdb
(
table
->
m_tableId
,
val
,
increase
);
}
...
...
@@ -979,7 +979,7 @@ Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op)
setDatabaseName
(
currentDb
.
c_str
());
setDatabaseSchemaName
(
currentSchema
.
c_str
());
return
~
0
;
return
~
(
Uint64
)
0
;
}
Uint32
...
...
sql/ha_ndbcluster.cc
View file @
e39ba639
...
...
@@ -1880,7 +1880,7 @@ int ha_ndbcluster::write_row(byte *record)
m_rows_inserted
++
;
no_uncommitted_rows_update
(
1
);
m_bulk_insert_not_flushed
=
TRUE
;
if
((
m_rows_to_insert
==
1
)
||
if
((
m_rows_to_insert
==
(
ha_rows
)
1
)
||
((
m_rows_inserted
%
m_bulk_insert_rows
)
==
0
)
||
set_blob_value
)
{
...
...
@@ -2945,8 +2945,8 @@ void ha_ndbcluster::start_bulk_insert(ha_rows rows)
DBUG_ENTER
(
"start_bulk_insert"
);
DBUG_PRINT
(
"enter"
,
(
"rows: %d"
,
(
int
)
rows
));
m_rows_inserted
=
0
;
if
(
rows
==
0
)
m_rows_inserted
=
(
ha_rows
)
0
;
if
(
rows
==
(
ha_rows
)
0
)
/* We don't know how many will be inserted, guess */
m_rows_to_insert
=
m_autoincrement_prefetch
;
else
...
...
@@ -2984,7 +2984,7 @@ int ha_ndbcluster::end_bulk_insert()
// Send rows to NDB
DBUG_PRINT
(
"info"
,
(
"Sending inserts to NDB, "
\
"rows_inserted:%d, bulk_insert_rows: %d"
,
m_rows_inserted
,
m_bulk_insert_rows
));
(
int
)
m_rows_inserted
,
(
int
)
m_bulk_insert_rows
));
m_bulk_insert_not_flushed
=
FALSE
;
if
(
execute_no_commit
(
this
,
trans
)
!=
0
)
{
no_uncommitted_rows_execute_failure
();
...
...
@@ -2992,8 +2992,8 @@ int ha_ndbcluster::end_bulk_insert()
}
}
m_rows_inserted
=
0
;
m_rows_to_insert
=
1
;
m_rows_inserted
=
(
ha_rows
)
0
;
m_rows_to_insert
=
(
ha_rows
)
1
;
DBUG_RETURN
(
error
);
}
...
...
@@ -3182,7 +3182,8 @@ int ha_ndbcluster::external_lock(THD *thd, int lock_type)
// store thread specific data first to set the right context
m_force_send
=
thd
->
variables
.
ndb_force_send
;
m_ha_not_exact_count
=
!
thd
->
variables
.
ndb_use_exact_count
;
m_autoincrement_prefetch
=
thd
->
variables
.
ndb_autoincrement_prefetch_sz
;
m_autoincrement_prefetch
=
(
ha_rows
)
thd
->
variables
.
ndb_autoincrement_prefetch_sz
;
if
(
!
thd
->
transaction
.
on
)
m_transaction_on
=
FALSE
;
else
...
...
@@ -3596,7 +3597,7 @@ static int create_ndb_column(NDBCOL &col,
static
void
ndb_set_fragmentation
(
NDBTAB
&
tab
,
TABLE
*
form
,
uint
pk_length
)
{
if
(
form
->
max_rows
==
0
)
/* default setting, don't set fragmentation */
if
(
form
->
max_rows
==
(
ha_rows
)
0
)
/* default setting, don't set fragmentation */
return
;
/**
* get the number of fragments right
...
...
@@ -3964,6 +3965,7 @@ longlong ha_ndbcluster::get_auto_increment()
/* We guessed too low */
m_rows_to_insert
+=
m_autoincrement_prefetch
;
int
cache_size
=
(
int
)
(
m_rows_to_insert
-
m_rows_inserted
<
m_autoincrement_prefetch
)
?
m_rows_to_insert
-
m_rows_inserted
:
(
m_rows_to_insert
>
m_autoincrement_prefetch
)
?
...
...
@@ -3998,9 +4000,9 @@ ha_ndbcluster::ha_ndbcluster(TABLE *table_arg):
m_primary_key_update
(
FALSE
),
m_retrieve_all_fields
(
FALSE
),
m_retrieve_primary_key
(
FALSE
),
m_rows_to_insert
(
1
),
m_rows_inserted
(
0
),
m_bulk_insert_rows
(
1024
),
m_rows_to_insert
(
(
ha_rows
)
1
),
m_rows_inserted
(
(
ha_rows
)
0
),
m_bulk_insert_rows
(
(
ha_rows
)
1024
),
m_bulk_insert_not_flushed
(
FALSE
),
m_ops_pending
(
0
),
m_skip_auto_increment
(
TRUE
),
...
...
@@ -4010,7 +4012,7 @@ ha_ndbcluster::ha_ndbcluster(TABLE *table_arg):
m_dupkey
((
uint
)
-
1
),
m_ha_not_exact_count
(
FALSE
),
m_force_send
(
TRUE
),
m_autoincrement_prefetch
(
32
),
m_autoincrement_prefetch
(
(
ha_rows
)
32
),
m_transaction_on
(
TRUE
),
m_use_local_query_cache
(
FALSE
)
{
...
...
sql/mysqld.cc
View file @
e39ba639
...
...
@@ -4551,7 +4551,7 @@ Disable with --skip-ndbcluster (will save memory).",
"Specify number of autoincrement values that are prefetched."
,
(
gptr
*
)
&
global_system_variables
.
ndb_autoincrement_prefetch_sz
,
(
gptr
*
)
&
global_system_variables
.
ndb_autoincrement_prefetch_sz
,
0
,
GET_
INT
,
REQUIRED_ARG
,
32
,
1
,
256
,
0
,
0
,
0
},
0
,
GET_
ULONG
,
REQUIRED_ARG
,
32
,
1
,
256
,
0
,
0
,
0
},
{
"ndb-force-send"
,
OPT_NDB_FORCE_SEND
,
"Force send of buffers to ndb immediately without waiting for "
"other threads."
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment