Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
M
MariaDB
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
nexedi
MariaDB
Commits
5cefbe59
Commit
5cefbe59
authored
May 03, 2006
by
unknown
Browse files
Options
Browse Files
Download
Plain Diff
Merge svojtovich@bk-internal.mysql.com:/home/bk/mysql-5.0
into production.mysql.com:/usersnfs/svojtovich/mysql-5.0
parents
7f1cc138
b5732e7c
Changes
9
Hide whitespace changes
Inline
Side-by-side
Showing
9 changed files
with
86 additions
and
43 deletions
+86
-43
innobase/include/dict0dict.ic
innobase/include/dict0dict.ic
+0
-2
mysql-test/r/ndb_blob.result
mysql-test/r/ndb_blob.result
+13
-5
mysql-test/t/ndb_blob.test
mysql-test/t/ndb_blob.test
+22
-3
ndb/include/kernel/signaldata/TcKeyReq.hpp
ndb/include/kernel/signaldata/TcKeyReq.hpp
+1
-0
ndb/include/ndbapi/NdbBlob.hpp
ndb/include/ndbapi/NdbBlob.hpp
+1
-0
ndb/src/ndbapi/NdbBlob.cpp
ndb/src/ndbapi/NdbBlob.cpp
+23
-1
ndb/test/ndbapi/testBlobs.cpp
ndb/test/ndbapi/testBlobs.cpp
+9
-14
ndb/tools/delete_all.cpp
ndb/tools/delete_all.cpp
+14
-17
sql/sql_table.cc
sql/sql_table.cc
+3
-1
No files found.
innobase/include/dict0dict.ic
View file @
5cefbe59
...
@@ -92,7 +92,6 @@ dict_table_get_n_user_cols(
...
@@ -92,7 +92,6 @@ dict_table_get_n_user_cols(
{
{
ut_ad(table);
ut_ad(table);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
ut_ad(table->cached);
return(table->n_cols - DATA_N_SYS_COLS);
return(table->n_cols - DATA_N_SYS_COLS);
}
}
...
@@ -126,7 +125,6 @@ dict_table_get_n_cols(
...
@@ -126,7 +125,6 @@ dict_table_get_n_cols(
{
{
ut_ad(table);
ut_ad(table);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
ut_ad(table->cached);
return(table->n_cols);
return(table->n_cols);
}
}
...
...
mysql-test/r/ndb_blob.result
View file @
5cefbe59
...
@@ -481,14 +481,22 @@ msg text NOT NULL
...
@@ -481,14 +481,22 @@ msg text NOT NULL
insert into t1 (msg) values(
insert into t1 (msg) values(
'Tries to validate (8 byte length + inline bytes) as UTF8 :(
'Tries to validate (8 byte length + inline bytes) as UTF8 :(
Fast fix: removed validation for Text. It is not yet indexable
Fast fix: removed validation for Text. It is not yet indexable
so bad data will not crash kernel.
so bad data will not crash kernel.');
Proper fix: Set inline bytes to multiple of mbmaxlen and
validate it (after the 8 byte length).');
select * from t1;
select * from t1;
id msg
id msg
1 Tries to validate (8 byte length + inline bytes) as UTF8 :(
1 Tries to validate (8 byte length + inline bytes) as UTF8 :(
Fast fix: removed validation for Text. It is not yet indexable
Fast fix: removed validation for Text. It is not yet indexable
so bad data will not crash kernel.
so bad data will not crash kernel.
Proper fix: Set inline bytes to multiple of mbmaxlen and
drop table t1;
validate it (after the 8 byte length).
create table t1 (
a int primary key not null auto_increment,
b text
) engine=ndbcluster;
select count(*) from t1;
count(*)
500
truncate t1;
select count(*) from t1;
count(*)
0
drop table t1;
drop table t1;
mysql-test/t/ndb_blob.test
View file @
5cefbe59
...
@@ -403,10 +403,29 @@ create table t1 (
...
@@ -403,10 +403,29 @@ create table t1 (
insert
into
t1
(
msg
)
values
(
insert
into
t1
(
msg
)
values
(
'Tries to validate (8 byte length + inline bytes) as UTF8 :(
'Tries to validate (8 byte length + inline bytes) as UTF8 :(
Fast fix: removed validation for Text. It is not yet indexable
Fast fix: removed validation for Text. It is not yet indexable
so bad data will not crash kernel.
so bad data will not crash kernel.'
);
Proper fix: Set inline bytes to multiple of mbmaxlen and
validate it (after the 8 byte length).'
);
select
*
from
t1
;
select
*
from
t1
;
drop
table
t1
;
drop
table
t1
;
# -- bug #19201
create
table
t1
(
a
int
primary
key
not
null
auto_increment
,
b
text
)
engine
=
ndbcluster
;
--
disable_query_log
set
autocommit
=
1
;
# more rows than batch size (64)
# for this bug no blob parts would be necessary
let
$
1
=
500
;
while
(
$
1
)
{
insert
into
t1
(
b
)
values
(
repeat
(
'x'
,
4000
));
dec
$
1
;
}
--
enable_query_log
select
count
(
*
)
from
t1
;
truncate
t1
;
select
count
(
*
)
from
t1
;
drop
table
t1
;
# End of 4.1 tests
# End of 4.1 tests
ndb/include/kernel/signaldata/TcKeyReq.hpp
View file @
5cefbe59
...
@@ -39,6 +39,7 @@ class TcKeyReq {
...
@@ -39,6 +39,7 @@ class TcKeyReq {
friend
class
NdbOperation
;
friend
class
NdbOperation
;
friend
class
NdbIndexOperation
;
friend
class
NdbIndexOperation
;
friend
class
NdbScanOperation
;
friend
class
NdbScanOperation
;
friend
class
NdbBlob
;
friend
class
DbUtil
;
friend
class
DbUtil
;
/**
/**
...
...
ndb/include/ndbapi/NdbBlob.hpp
View file @
5cefbe59
...
@@ -290,6 +290,7 @@ private:
...
@@ -290,6 +290,7 @@ private:
bool
isWriteOp
();
bool
isWriteOp
();
bool
isDeleteOp
();
bool
isDeleteOp
();
bool
isScanOp
();
bool
isScanOp
();
bool
isTakeOverOp
();
// computations
// computations
Uint32
getPartNumber
(
Uint64
pos
);
Uint32
getPartNumber
(
Uint64
pos
);
Uint32
getPartCount
();
Uint32
getPartCount
();
...
...
ndb/src/ndbapi/NdbBlob.cpp
View file @
5cefbe59
...
@@ -23,6 +23,7 @@
...
@@ -23,6 +23,7 @@
#include <NdbBlob.hpp>
#include <NdbBlob.hpp>
#include "NdbBlobImpl.hpp"
#include "NdbBlobImpl.hpp"
#include <NdbScanOperation.hpp>
#include <NdbScanOperation.hpp>
#include <signaldata/TcKeyReq.hpp>
/*
/*
* Reading index table directly (as a table) is faster but there are
* Reading index table directly (as a table) is faster but there are
...
@@ -264,6 +265,13 @@ NdbBlob::isScanOp()
...
@@ -264,6 +265,13 @@ NdbBlob::isScanOp()
theNdbOp
->
theOperationType
==
NdbOperation
::
OpenRangeScanRequest
;
theNdbOp
->
theOperationType
==
NdbOperation
::
OpenRangeScanRequest
;
}
}
inline
bool
NdbBlob
::
isTakeOverOp
()
{
return
TcKeyReq
::
getTakeOverScanFlag
(
theNdbOp
->
theScanInfo
);
}
// computations (inline)
// computations (inline)
inline
Uint32
inline
Uint32
...
@@ -1203,8 +1211,22 @@ NdbBlob::preExecute(NdbTransaction::ExecType anExecType, bool& batch)
...
@@ -1203,8 +1211,22 @@ NdbBlob::preExecute(NdbTransaction::ExecType anExecType, bool& batch)
if
(
isUpdateOp
()
||
isWriteOp
()
||
isDeleteOp
())
{
if
(
isUpdateOp
()
||
isWriteOp
()
||
isDeleteOp
())
{
// add operation before this one to read head+inline
// add operation before this one to read head+inline
NdbOperation
*
tOp
=
theNdbCon
->
getNdbOperation
(
theTable
,
theNdbOp
);
NdbOperation
*
tOp
=
theNdbCon
->
getNdbOperation
(
theTable
,
theNdbOp
);
/*
* If main op is from take over scan lock, the added read is done
* as committed read:
*
* In normal transactional case, the row is locked by us and
* committed read returns same as normal read.
*
* In current TRUNCATE TABLE, the deleting trans is committed in
* batches and then restarted with new trans id. A normal read
* would hang on the scan delete lock and then fail.
*/
NdbOperation
::
LockMode
lockMode
=
!
isTakeOverOp
()
?
NdbOperation
::
LM_Read
:
NdbOperation
::
LM_CommittedRead
;
if
(
tOp
==
NULL
||
if
(
tOp
==
NULL
||
tOp
->
readTuple
()
==
-
1
||
tOp
->
readTuple
(
lockMode
)
==
-
1
||
setTableKeyValue
(
tOp
)
==
-
1
||
setTableKeyValue
(
tOp
)
==
-
1
||
getHeadInlineValue
(
tOp
)
==
-
1
)
{
getHeadInlineValue
(
tOp
)
==
-
1
)
{
setErrorCode
(
tOp
);
setErrorCode
(
tOp
);
...
...
ndb/test/ndbapi/testBlobs.cpp
View file @
5cefbe59
...
@@ -44,6 +44,7 @@ struct Opt {
...
@@ -44,6 +44,7 @@ struct Opt {
bool
m_dbg
;
bool
m_dbg
;
bool
m_dbgall
;
bool
m_dbgall
;
const
char
*
m_dbug
;
const
char
*
m_dbug
;
bool
m_fac
;
bool
m_full
;
bool
m_full
;
unsigned
m_loop
;
unsigned
m_loop
;
unsigned
m_parts
;
unsigned
m_parts
;
...
@@ -72,6 +73,7 @@ struct Opt {
...
@@ -72,6 +73,7 @@ struct Opt {
m_dbg
(
false
),
m_dbg
(
false
),
m_dbgall
(
false
),
m_dbgall
(
false
),
m_dbug
(
0
),
m_dbug
(
0
),
m_fac
(
false
),
m_full
(
false
),
m_full
(
false
),
m_loop
(
1
),
m_loop
(
1
),
m_parts
(
10
),
m_parts
(
10
),
...
@@ -110,6 +112,7 @@ printusage()
...
@@ -110,6 +112,7 @@ printusage()
<<
" -dbg print debug"
<<
endl
<<
" -dbg print debug"
<<
endl
<<
" -dbgall print also NDB API debug (if compiled in)"
<<
endl
<<
" -dbgall print also NDB API debug (if compiled in)"
<<
endl
<<
" -dbug opt dbug options"
<<
endl
<<
" -dbug opt dbug options"
<<
endl
<<
" -fac fetch across commit in scan delete ["
<<
d
.
m_fac
<<
"]"
<<
endl
<<
" -full read/write only full blob values"
<<
endl
<<
" -full read/write only full blob values"
<<
endl
<<
" -loop N loop N times 0=forever ["
<<
d
.
m_loop
<<
"]"
<<
endl
<<
" -loop N loop N times 0=forever ["
<<
d
.
m_loop
<<
"]"
<<
endl
<<
" -parts N max parts in blob value ["
<<
d
.
m_parts
<<
"]"
<<
endl
<<
" -parts N max parts in blob value ["
<<
d
.
m_parts
<<
"]"
<<
endl
...
@@ -1255,23 +1258,11 @@ deleteScan(bool idx)
...
@@ -1255,23 +1258,11 @@ deleteScan(bool idx)
CHK
((
ret
=
g_ops
->
nextResult
(
false
))
==
0
||
ret
==
1
||
ret
==
2
);
CHK
((
ret
=
g_ops
->
nextResult
(
false
))
==
0
||
ret
==
1
||
ret
==
2
);
if
(
++
n
==
g_opt
.
m_batch
||
ret
==
2
)
{
if
(
++
n
==
g_opt
.
m_batch
||
ret
==
2
)
{
DBG
(
"execute batch: n="
<<
n
<<
" ret="
<<
ret
);
DBG
(
"execute batch: n="
<<
n
<<
" ret="
<<
ret
);
switch
(
0
)
{
if
(
!
g_opt
.
m_fac
)
{
case
0
:
// works normally
CHK
(
g_con
->
execute
(
NoCommit
)
==
0
);
CHK
(
g_con
->
execute
(
NoCommit
)
==
0
);
CHK
(
true
||
g_con
->
restart
()
==
0
);
}
else
{
break
;
case
1
:
// nonsense - g_con is invalid for 2nd batch
CHK
(
g_con
->
execute
(
Commit
)
==
0
);
CHK
(
true
||
g_con
->
restart
()
==
0
);
break
;
case
2
:
// DBTC sendSignalErrorRefuseLab
CHK
(
g_con
->
execute
(
NoCommit
)
==
0
);
CHK
(
g_con
->
restart
()
==
0
);
break
;
case
3
:
// 266 time-out
CHK
(
g_con
->
execute
(
Commit
)
==
0
);
CHK
(
g_con
->
execute
(
Commit
)
==
0
);
CHK
(
g_con
->
restart
()
==
0
);
CHK
(
g_con
->
restart
()
==
0
);
break
;
}
}
n
=
0
;
n
=
0
;
}
}
...
@@ -1817,6 +1808,10 @@ NDB_COMMAND(testOdbcDriver, "testBlobs", "testBlobs", "testBlobs", 65535)
...
@@ -1817,6 +1808,10 @@ NDB_COMMAND(testOdbcDriver, "testBlobs", "testBlobs", "testBlobs", 65535)
continue
;
continue
;
}
}
}
}
if
(
strcmp
(
arg
,
"-fac"
)
==
0
)
{
g_opt
.
m_fac
=
true
;
continue
;
}
if
(
strcmp
(
arg
,
"-full"
)
==
0
)
{
if
(
strcmp
(
arg
,
"-full"
)
==
0
)
{
g_opt
.
m_full
=
true
;
g_opt
.
m_full
=
true
;
continue
;
continue
;
...
...
ndb/tools/delete_all.cpp
View file @
5cefbe59
...
@@ -23,17 +23,21 @@
...
@@ -23,17 +23,21 @@
#include <NDBT.hpp>
#include <NDBT.hpp>
static
int
clear_table
(
Ndb
*
pNdb
,
const
NdbDictionary
::
Table
*
pTab
,
static
int
clear_table
(
Ndb
*
pNdb
,
const
NdbDictionary
::
Table
*
pTab
,
bool
commit_across_open_cursor
,
int
parallelism
=
240
);
bool
fetch_across_commit
,
int
parallelism
=
240
);
NDB_STD_OPTS_VARS
;
NDB_STD_OPTS_VARS
;
static
const
char
*
_dbname
=
"TEST_DB"
;
static
const
char
*
_dbname
=
"TEST_DB"
;
static
my_bool
_transactional
=
false
;
static
struct
my_option
my_long_options
[]
=
static
struct
my_option
my_long_options
[]
=
{
{
NDB_STD_OPTS
(
"ndb_desc"
),
NDB_STD_OPTS
(
"ndb_desc"
),
{
"database"
,
'd'
,
"Name of database table is in"
,
{
"database"
,
'd'
,
"Name of database table is in"
,
(
gptr
*
)
&
_dbname
,
(
gptr
*
)
&
_dbname
,
0
,
(
gptr
*
)
&
_dbname
,
(
gptr
*
)
&
_dbname
,
0
,
GET_STR
,
REQUIRED_ARG
,
0
,
0
,
0
,
0
,
0
,
0
},
GET_STR
,
REQUIRED_ARG
,
0
,
0
,
0
,
0
,
0
,
0
},
{
"transactional"
,
't'
,
"Single transaction (may run out of operations)"
,
(
gptr
*
)
&
_transactional
,
(
gptr
*
)
&
_transactional
,
0
,
GET_BOOL
,
NO_ARG
,
0
,
0
,
0
,
0
,
0
,
0
},
{
0
,
0
,
0
,
0
,
0
,
0
,
GET_NO_ARG
,
NO_ARG
,
0
,
0
,
0
,
0
,
0
,
0
}
{
0
,
0
,
0
,
0
,
0
,
0
,
GET_NO_ARG
,
NO_ARG
,
0
,
0
,
0
,
0
,
0
,
0
}
};
};
static
void
usage
()
static
void
usage
()
...
@@ -84,18 +88,11 @@ int main(int argc, char** argv){
...
@@ -84,18 +88,11 @@ int main(int argc, char** argv){
ndbout
<<
" Table "
<<
argv
[
i
]
<<
" does not exist!"
<<
endl
;
ndbout
<<
" Table "
<<
argv
[
i
]
<<
" does not exist!"
<<
endl
;
return
NDBT_ProgramExit
(
NDBT_WRONGARGS
);
return
NDBT_ProgramExit
(
NDBT_WRONGARGS
);
}
}
// Check if we have any blobs
ndbout
<<
"Deleting all from "
<<
argv
[
i
];
bool
commit_across_open_cursor
=
true
;
if
(
!
_transactional
)
for
(
int
j
=
0
;
j
<
pTab
->
getNoOfColumns
();
j
++
)
{
ndbout
<<
" (non-transactional)"
;
NdbDictionary
::
Column
::
Type
t
=
pTab
->
getColumn
(
j
)
->
getType
();
ndbout
<<
" ..."
;
if
(
t
==
NdbDictionary
::
Column
::
Blob
||
if
(
clear_table
(
&
MyNdb
,
pTab
,
!
_transactional
)
==
NDBT_FAILED
){
t
==
NdbDictionary
::
Column
::
Text
)
{
commit_across_open_cursor
=
false
;
break
;
}
}
ndbout
<<
"Deleting all from "
<<
argv
[
i
]
<<
"..."
;
if
(
clear_table
(
&
MyNdb
,
pTab
,
commit_across_open_cursor
)
==
NDBT_FAILED
){
res
=
NDBT_FAILED
;
res
=
NDBT_FAILED
;
ndbout
<<
"FAILED"
<<
endl
;
ndbout
<<
"FAILED"
<<
endl
;
}
}
...
@@ -105,7 +102,7 @@ int main(int argc, char** argv){
...
@@ -105,7 +102,7 @@ int main(int argc, char** argv){
int
clear_table
(
Ndb
*
pNdb
,
const
NdbDictionary
::
Table
*
pTab
,
int
clear_table
(
Ndb
*
pNdb
,
const
NdbDictionary
::
Table
*
pTab
,
bool
commit_across_open_cursor
,
int
parallelism
)
bool
fetch_across_commit
,
int
parallelism
)
{
{
// Scan all records exclusive and delete
// Scan all records exclusive and delete
// them one by one
// them one by one
...
@@ -136,7 +133,7 @@ int clear_table(Ndb* pNdb, const NdbDictionary::Table* pTab,
...
@@ -136,7 +133,7 @@ int clear_table(Ndb* pNdb, const NdbDictionary::Table* pTab,
}
}
goto
failed
;
goto
failed
;
}
}
pOp
=
pTrans
->
getNdbScanOperation
(
pTab
->
getName
());
pOp
=
pTrans
->
getNdbScanOperation
(
pTab
->
getName
());
if
(
pOp
==
NULL
)
{
if
(
pOp
==
NULL
)
{
goto
failed
;
goto
failed
;
...
@@ -166,7 +163,7 @@ int clear_table(Ndb* pNdb, const NdbDictionary::Table* pTab,
...
@@ -166,7 +163,7 @@ int clear_table(Ndb* pNdb, const NdbDictionary::Table* pTab,
}
while
((
check
=
pOp
->
nextResult
(
false
))
==
0
);
}
while
((
check
=
pOp
->
nextResult
(
false
))
==
0
);
if
(
check
!=
-
1
){
if
(
check
!=
-
1
){
if
(
commit_across_open_cursor
)
{
if
(
fetch_across_commit
)
{
check
=
pTrans
->
execute
(
NdbTransaction
::
Commit
);
check
=
pTrans
->
execute
(
NdbTransaction
::
Commit
);
pTrans
->
restart
();
// new tx id
pTrans
->
restart
();
// new tx id
}
else
{
}
else
{
...
@@ -197,7 +194,7 @@ int clear_table(Ndb* pNdb, const NdbDictionary::Table* pTab,
...
@@ -197,7 +194,7 @@ int clear_table(Ndb* pNdb, const NdbDictionary::Table* pTab,
}
}
goto
failed
;
goto
failed
;
}
}
if
(
!
commit_across_open_cursor
&&
if
(
!
fetch_across_commit
&&
pTrans
->
execute
(
NdbTransaction
::
Commit
)
!=
0
)
{
pTrans
->
execute
(
NdbTransaction
::
Commit
)
!=
0
)
{
err
=
pTrans
->
getNdbError
();
err
=
pTrans
->
getNdbError
();
goto
failed
;
goto
failed
;
...
...
sql/sql_table.cc
View file @
5cefbe59
...
@@ -1941,7 +1941,9 @@ mysql_rename_table(enum db_type base,
...
@@ -1941,7 +1941,9 @@ mysql_rename_table(enum db_type base,
}
}
}
}
delete
file
;
delete
file
;
if
(
error
)
if
(
error
==
HA_ERR_WRONG_COMMAND
)
my_error
(
ER_NOT_SUPPORTED_YET
,
MYF
(
0
),
"ALTER TABLE"
);
else
if
(
error
)
my_error
(
ER_ERROR_ON_RENAME
,
MYF
(
0
),
from
,
to
,
error
);
my_error
(
ER_ERROR_ON_RENAME
,
MYF
(
0
),
from
,
to
,
error
);
DBUG_RETURN
(
error
!=
0
);
DBUG_RETURN
(
error
!=
0
);
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment