Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
M
MariaDB
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
nexedi
MariaDB
Commits
a880ec94
Commit
a880ec94
authored
Jan 24, 2002
by
monty@hundin.mysql.fi
Browse files
Options
Browse Files
Download
Plain Diff
Merge
parents
1d86021c
4609e627
Changes
10
Show whitespace changes
Inline
Side-by-side
Showing
10 changed files
with
416 additions
and
261 deletions
+416
-261
Docs/manual.texi
Docs/manual.texi
+78
-33
client/mysql.cc
client/mysql.cc
+2
-2
client/mysqldump.c
client/mysqldump.c
+2
-2
dbug/dbug.c
dbug/dbug.c
+1
-1
mysql-test/t/innodb.test
mysql-test/t/innodb.test
+48
-0
sql/handler.cc
sql/handler.cc
+30
-10
sql/handler.h
sql/handler.h
+1
-1
sql/opt_range.cc
sql/opt_range.cc
+3
-0
sql/sql_select.cc
sql/sql_select.cc
+250
-211
sql/sql_select.h
sql/sql_select.h
+1
-1
No files found.
Docs/manual.texi
View file @
a880ec94
...
@@ -818,14 +818,14 @@ size is pushed up to 8 million terabytes (2 ^ 63 bytes).
...
@@ -818,14 +818,14 @@ size is pushed up to 8 million terabytes (2 ^ 63 bytes).
Note, however, that operating systems have their own file size
Note, however, that operating systems have their own file size
limits. Here are some examples:
limits. Here are some examples:
@multitable @columnfractions .
5 .5
@multitable @columnfractions .
6 .4
@item @strong{Operating System} @tab @strong{File Size Limit}
@item @strong{Operating System} @tab @strong{File Size Limit}
@item Linux-Intel 32 bit @tab 2G, 4G or more, depends on Linux version
@item Linux-Intel 32 bit @tab 2G, 4G or more, depends on Linux version
@item Linux-Alpha @tab 8T (?)
@item Linux-Alpha @tab 8T (?)
@item Solaris 2.5.1 @tab 2G (possible 4G with patch)
@item Solaris 2.5.1 @tab 2G (possible 4G with patch)
@item Solaris 2.6 @tab 4G
@item Solaris 2.6 @tab 4G
@item Solaris 2.7 Intel @tab 4G
@item Solaris 2.7 Intel @tab 4G
@item Solaris 2.7 ULTRA-SPARC
@tab 8T (?)
@item Solaris 2.7 ULTRA-SPARC
@tab 512G
@end multitable
@end multitable
On Linux 2.2 you can get bigger tables than 2G by using the LFS patch for
On Linux 2.2 you can get bigger tables than 2G by using the LFS patch for
...
@@ -5507,7 +5507,6 @@ pipes.
...
@@ -5507,7 +5507,6 @@ pipes.
All of the above binaries are optimised for the Pentium Pro
All of the above binaries are optimised for the Pentium Pro
processor but should work on any Intel processor >= i386.
processor but should work on any Intel processor >= i386.
In the following circumstances you will need to use the MySQL
In the following circumstances you will need to use the MySQL
configuration file:
configuration file:
@itemize @bullet
@itemize @bullet
...
@@ -5525,17 +5524,23 @@ If you want to use one of the these servers:
...
@@ -5525,17 +5524,23 @@ If you want to use one of the these servers:
If you need to tune the server settings.
If you need to tune the server settings.
@end itemize
@end itemize
There are two configuration files with the same function:
Normally you can use the @code{WinMySQLAdmin} tool to edit the
@file{my.cnf} and @file{my.ini} file, however please note that
configuration file @code{my.ini}. In this case you don't have to worry
only of one these should can used. Both files are plain text.
about the following section.
The @file{my.cnf} file should be created in the root directory
of drive C and the @file{my.ini} file on the WinDir directory
There are two configuration files with the same function: @file{my.cnf}
e.g: @file{C:\WINDOWS} or @file{C:\WINNT}.
and @file{my.ini} file, however please note that only of one these
If your PC uses a boot loader where the C drive isn't the boot
should be used to avoid confusion. Both files are plain text. The
drive, then your only option is to use the @file{my.ini} file.
@file{my.cnf} file , if used, should be created in the root directory of
Also note that if you use the @code{WinMySQLAdmin} tool, only
drive C and the @file{my.ini} file on the WinDir directory e.g:
the @file{my.ini} file is used. The @file{\mysql\bin} directory
@file{C:\WINDOWS} or @file{C:\WINNT}. MySQL will first read the
contains a help file with instructions for using this tool.
@code{my.ini} file, followed by the @code{my.cnf} file.
If your PC uses a boot loader where the C drive isn't the boot drive,
then your only option is to use the @file{my.ini} file. Also note that
if you use the @code{WinMySQLAdmin} tool, only the @file{my.ini} file is
used by this tool. The @file{\mysql\bin} directory contains a help file
with instructions for using this tool.
Using @code{notepad.exe}, create the configuration file and
Using @code{notepad.exe}, create the configuration file and
edit the base section and keys:
edit the base section and keys:
...
@@ -5808,11 +5813,12 @@ release or the last stable release:
...
@@ -5808,11 +5813,12 @@ release or the last stable release:
@itemize @bullet
@itemize @bullet
@item
@item
Normally, if you are beginning to use MySQL for the first time
Normally, if you are beginning to use MySQL for the first time or trying
or trying to port it to some system for which there is no binary
to port it to some system for which there is no binary distribution, we
distribution, we recommend going with the stable release (currently
recommend going with the stable release (currently Version 3.23. Note
Version 3.23. Note that all MySQL releases are checked with the
that all MySQL releases are checked with the MySQL benchmarks and an
MySQL benchmarks and an extensive test suite before each release.
extensive test suite before each release (even the development
releases).
@item
@item
Otherwise, if you are running an old system and want to upgrade, but
Otherwise, if you are running an old system and want to upgrade, but
...
@@ -6073,7 +6079,8 @@ Stable tested releases are meant to appear about 1-2 times a year, but
...
@@ -6073,7 +6079,8 @@ Stable tested releases are meant to appear about 1-2 times a year, but
if small bugs are found, a release with only bug fixes will be released.
if small bugs are found, a release with only bug fixes will be released.
@item
@item
Working releases are meant to appear about every 1-8 weeks.
Working releases/bug fixes to old releases are meant to appear about
every 1-8 weeks.
@item
@item
Binary distributions for some platforms will be made by us for major releases.
Binary distributions for some platforms will be made by us for major releases.
...
@@ -6082,12 +6089,12 @@ less frequently.
...
@@ -6082,12 +6089,12 @@ less frequently.
@item
@item
We usually make patches available as soon as we have located and fixed
We usually make patches available as soon as we have located and fixed
small bugs.
small bugs. They are posted to @email{bugs@@lists.mysql.com} and will
be added to the next release.
@item
@item
For non-critical but annoying bugs, we will make patches available if they
For non-critical but annoying bugs, we will add them the MySQL source
are sent to us. Otherwise we will combine many of them into a larger
repository and they will be fixed in the next release.
patch.
@item
@item
If there is, by any chance, a fatal bug in a release we will make a new
If there is, by any chance, a fatal bug in a release we will make a new
...
@@ -6102,6 +6109,10 @@ and things that ``must be done.'' ``Somewhat frozen'' means that we may
...
@@ -6102,6 +6109,10 @@ and things that ``must be done.'' ``Somewhat frozen'' means that we may
add small things that ``almost surely will not affect anything that's
add small things that ``almost surely will not affect anything that's
already working.''
already working.''
MySQL uses a little different naming scheme that most other products.
In general it's relatively safe to use any version that has been out for
a couple of weeks without being replaced with a new version.
@xref{Which version}.
@node MySQL binaries, Installing binary, Many versions, General Installation Issues
@node MySQL binaries, Installing binary, Many versions, General Installation Issues
@subsection MySQL Binaries Compiled by MySQL AB
@subsection MySQL Binaries Compiled by MySQL AB
...
@@ -19514,7 +19525,7 @@ bytes when needed. This value by default is small, to catch big (possibly
...
@@ -19514,7 +19525,7 @@ bytes when needed. This value by default is small, to catch big (possibly
wrong) packets. You must increase this value if you are using big
wrong) packets. You must increase this value if you are using big
@code{BLOB} columns. It should be as big as the biggest @code{BLOB} you want
@code{BLOB} columns. It should be as big as the biggest @code{BLOB} you want
to use. The protocol limits for @code{max_allowed_packet} is 16M in MySQL
to use. The protocol limits for @code{max_allowed_packet} is 16M in MySQL
3.23 and
4
G in MySQL 4.0.
3.23 and
2
G in MySQL 4.0.
@item @code{max_binlog_cache_size}
@item @code{max_binlog_cache_size}
If a multi-statement transaction requires more than this amount of memory,
If a multi-statement transaction requires more than this amount of memory,
...
@@ -36773,6 +36784,11 @@ The following are the known problems with @code{MERGE} tables:
...
@@ -36773,6 +36784,11 @@ The following are the known problems with @code{MERGE} tables:
@itemize @bullet
@itemize @bullet
@item
@item
@code{MERGE} table cannot maintain UNIQUE constraints over the whole
table. When you do INSERT, the data goes into the first or last table
(according to @code{INSERT_METHOD=xxx}) and this MyISAM table ensures
that the data are unique, but it knows nothing about the first MyISAM table.
@item
@code{DELETE FROM merge_table} used without a @code{WHERE}
@code{DELETE FROM merge_table} used without a @code{WHERE}
will only clear the mapping for the table, not delete everything in the
will only clear the mapping for the table, not delete everything in the
mapped tables.
mapped tables.
...
@@ -38738,6 +38754,7 @@ Finland
...
@@ -38738,6 +38754,7 @@ Finland
* BDB characteristics:: Characteristics of @code{BDB} tables:
* BDB characteristics:: Characteristics of @code{BDB} tables:
* BDB TODO:: Things we need to fix for BDB in the near future:
* BDB TODO:: Things we need to fix for BDB in the near future:
* BDB portability:: Operating systems supported by @strong{BDB}
* BDB portability:: Operating systems supported by @strong{BDB}
* BDB restrictions::
* BDB errors:: Errors That May Occur When Using BDB Tables
* BDB errors:: Errors That May Occur When Using BDB Tables
@end menu
@end menu
...
@@ -38946,7 +38963,7 @@ Change to not use page locks at all when we are scanning tables.
...
@@ -38946,7 +38963,7 @@ Change to not use page locks at all when we are scanning tables.
@end itemize
@end itemize
@node BDB portability, BDB
error
s, BDB TODO, BDB
@node BDB portability, BDB
restriction
s, BDB TODO, BDB
@subsection Operating systems supported by @strong{BDB}
@subsection Operating systems supported by @strong{BDB}
If you after having built MySQL with support for BDB tables get
If you after having built MySQL with support for BDB tables get
...
@@ -38986,8 +39003,28 @@ Linux 2.x Alpha
...
@@ -38986,8 +39003,28 @@ Linux 2.x Alpha
Max OS X
Max OS X
@end itemize
@end itemize
@node BDB restrictions, BDB errors, BDB portability, BDB
@subsection Restrictions on BDB Tables
Here follows the restrictions you have when using BDB tables:
@node BDB errors, , BDB portability, BDB
@itemize @bullet
@item
BDB tables store in the .db file the path to the file as it was created
(My guess is that this is to be able to detect locks in a multi-user
environment that supports symlinks).
The effect of this is that BDB tables are not movable between directories!
@item
When taking backups of BDB tables, you have to either use
@code{mysqldump} or take a backup of all @code{table_name.db} files and
the BDB log files. The BDB log files are the files in the base data
directory named @code{log.XXXXXX} (6 digits);
The BDB table handler stores unfinished transactions in the log files
and requires these to be present when @code{mysqld} starts.
@end itemize
@node BDB errors, , BDB restrictions, BDB
@subsection Errors That May Occur When Using BDB Tables
@subsection Errors That May Occur When Using BDB Tables
@itemize @bullet
@itemize @bullet
...
@@ -45376,6 +45413,11 @@ When a MySQL client or the @code{mysqld} server gets a packet bigger
...
@@ -45376,6 +45413,11 @@ When a MySQL client or the @code{mysqld} server gets a packet bigger
than @code{max_allowed_packet} bytes, it issues a @code{Packet too large}
than @code{max_allowed_packet} bytes, it issues a @code{Packet too large}
error and closes the connection.
error and closes the connection.
In MySQL 3.23 the biggest possible packet is 16M (due to limits in the
client/server protocol). In MySQL 4.0.1 and up, this is only limited by
the amount on memory you have on your server (up to a theoretical
maximum of 2G).
A communication packet is a single SQL statement sent to the MySQL server
A communication packet is a single SQL statement sent to the MySQL server
or a single row that is sent to the client.
or a single row that is sent to the client.
...
@@ -47185,11 +47227,6 @@ New version of netadmin. See above for details.
...
@@ -47185,11 +47227,6 @@ New version of netadmin. See above for details.
Atronic's MySQL client for Windows 2.0.3.0.
Atronic's MySQL client for Windows 2.0.3.0.
The home page for this can be found at @uref{http://www.artronic.hr/}.
The home page for this can be found at @uref{http://www.artronic.hr/}.
@item @uref{http://www.mysql.com/Downloads/Win32/mysqlfront.zip}
MySQL Front, home page: @uref{http://www.mysqlfront.de/}.
Win32-Client for accessing and managing dbs, tables, table-data, indexes,
import-/export-files. (Freeware). By Ansgar Becker.
@item @uref{http://www.mysql.com/Downloads/Win32/W9xstop.zip}
@item @uref{http://www.mysql.com/Downloads/Win32/W9xstop.zip}
Utility from Artronic to stop MySQL on win9x.
Utility from Artronic to stop MySQL on win9x.
...
@@ -48230,7 +48267,15 @@ Our TODO section contains what we plan to have in 4.0. @xref{TODO MySQL 4.0}.
...
@@ -48230,7 +48267,15 @@ Our TODO section contains what we plan to have in 4.0. @xref{TODO MySQL 4.0}.
@itemize @bullet
@itemize @bullet
@item
@item
Made same changes in multi table updates as in multi table delete
Fixed optimization problem where a MySQL was a long time in a
"preparing" state when selecting from an empty table which had contained
a lot of rows.
@item
Fixed bug in complicated join with @code{const} tables. This fix also
improves performance a bit when referring to another table from a
@code{const} table.
@item
First pre-version of multi table updates.
@item
@item
Fixed bug in multi table delete.
Fixed bug in multi table delete.
@item
@item
client/mysql.cc
View file @
a880ec94
...
@@ -477,8 +477,8 @@ static struct option long_options[] =
...
@@ -477,8 +477,8 @@ static struct option long_options[] =
CHANGEABLE_VAR
changeable_vars
[]
=
{
CHANGEABLE_VAR
changeable_vars
[]
=
{
{
"connect_timeout"
,
(
long
*
)
&
opt_connect_timeout
,
0
,
0
,
3600
*
12
,
0
,
1
},
{
"connect_timeout"
,
(
long
*
)
&
opt_connect_timeout
,
0
,
0
,
3600
*
12
,
0
,
1
},
{
"max_allowed_packet"
,
(
long
*
)
&
max_allowed_packet
,
16
*
1024L
*
1024L
,
4096
,
{
"max_allowed_packet"
,
(
long
*
)
&
max_allowed_packet
,
16
*
1024L
*
1024L
,
4096
,
24
*
1024L
*
1024L
,
MALLOC_OVERHEAD
,
1024
},
512
*
1024L
*
1024L
,
MALLOC_OVERHEAD
,
1024
},
{
"net_buffer_length"
,(
long
*
)
&
net_buffer_length
,
16384
,
1024
,
24
*
1024
*
1024L
,
{
"net_buffer_length"
,(
long
*
)
&
net_buffer_length
,
16384
,
1024
,
512
*
1024
*
1024L
,
MALLOC_OVERHEAD
,
1024
},
MALLOC_OVERHEAD
,
1024
},
{
"select_limit"
,
(
long
*
)
&
select_limit
,
1000L
,
1
,
~
0L
,
0
,
1
},
{
"select_limit"
,
(
long
*
)
&
select_limit
,
1000L
,
1
,
~
0L
,
0
,
1
},
{
"max_join_size"
,
(
long
*
)
&
max_join_size
,
1000000L
,
1
,
~
0L
,
0
,
1
},
{
"max_join_size"
,
(
long
*
)
&
max_join_size
,
1000000L
,
1
,
~
0L
,
0
,
1
},
...
...
client/mysqldump.c
View file @
a880ec94
...
@@ -147,9 +147,9 @@ static const char *load_default_groups[]= { "mysqldump","client",0 };
...
@@ -147,9 +147,9 @@ static const char *load_default_groups[]= { "mysqldump","client",0 };
CHANGEABLE_VAR
md_changeable_vars
[]
=
{
CHANGEABLE_VAR
md_changeable_vars
[]
=
{
{
"max_allowed_packet"
,
(
long
*
)
&
max_allowed_packet
,
24
*
1024
*
1024
,
4096
,
{
"max_allowed_packet"
,
(
long
*
)
&
max_allowed_packet
,
24
*
1024
*
1024
,
4096
,
24
*
1024L
*
1024L
,
MALLOC_OVERHEAD
,
1024
},
512
*
1024L
*
1024L
,
MALLOC_OVERHEAD
,
1024
},
{
"net_buffer_length"
,
(
long
*
)
&
net_buffer_length
,
1024
*
1024L
-
1025
,
4096
,
{
"net_buffer_length"
,
(
long
*
)
&
net_buffer_length
,
1024
*
1024L
-
1025
,
4096
,
24
*
1024L
*
1024L
,
MALLOC_OVERHEAD
,
1024
},
512
*
1024L
*
1024L
,
MALLOC_OVERHEAD
,
1024
},
{
0
,
0
,
0
,
0
,
0
,
0
,
0
}
{
0
,
0
,
0
,
0
,
0
,
0
,
0
}
};
};
...
...
dbug/dbug.c
View file @
a880ec94
...
@@ -1928,7 +1928,7 @@ static void dbug_flush(CODE_STATE *state)
...
@@ -1928,7 +1928,7 @@ static void dbug_flush(CODE_STATE *state)
{
{
if
(
!
(
freopen
(
stack
->
name
,
"a"
,
_db_fp_
)))
if
(
!
(
freopen
(
stack
->
name
,
"a"
,
_db_fp_
)))
{
{
(
void
)
fprintf
(
stderr
,
ERR_OPEN
,
_db_process_
);
(
void
)
fprintf
(
stderr
,
ERR_OPEN
,
_db_process_
,
stack
->
name
);
fflush
(
stderr
);
fflush
(
stderr
);
_db_fp_
=
stdout
;
_db_fp_
=
stdout
;
stack
->
out_file
=
_db_fp_
;
stack
->
out_file
=
_db_fp_
;
...
...
mysql-test/t/innodb.test
View file @
a880ec94
...
@@ -577,3 +577,51 @@ handler t1 read t first;
...
@@ -577,3 +577,51 @@ handler t1 read t first;
--
error
1109
--
error
1109
handler
t1
close
;
handler
t1
close
;
drop
table
t1
;
drop
table
t1
;
#
# Test of multi-table-delete
#
CREATE
TABLE
t1
(
number
bigint
(
20
)
NOT
NULL
default
'0'
,
cname
char
(
15
)
NOT
NULL
default
''
,
carrier_id
smallint
(
6
)
NOT
NULL
default
'0'
,
privacy
tinyint
(
4
)
NOT
NULL
default
'0'
,
last_mod_date
timestamp
(
14
)
NOT
NULL
,
last_mod_id
smallint
(
6
)
NOT
NULL
default
'0'
,
last_app_date
timestamp
(
14
)
NOT
NULL
,
last_app_id
smallint
(
6
)
default
'-1'
,
version
smallint
(
6
)
NOT
NULL
default
'0'
,
assigned_scps
int
(
11
)
default
'0'
,
status
tinyint
(
4
)
default
'0'
)
TYPE
=
InnoDB
;
INSERT
INTO
t1
VALUES
(
4077711111
,
'SeanWheeler'
,
90
,
2
,
20020111112846
,
500
,
00000000000000
,
-
1
,
2
,
3
,
1
);
INSERT
INTO
t1
VALUES
(
9197722223
,
'berry'
,
90
,
3
,
20020111112809
,
500
,
20020102114532
,
501
,
4
,
10
,
0
);
INSERT
INTO
t1
VALUES
(
650
,
'San Francisco'
,
0
,
0
,
20011227111336
,
342
,
00000000000000
,
-
1
,
1
,
24
,
1
);
INSERT
INTO
t1
VALUES
(
302467
,
'Sue\'s Subshop'
,
90
,
3
,
20020109113241
,
500
,
20020102115111
,
501
,
7
,
24
,
0
);
INSERT
INTO
t1
VALUES
(
6014911113
,
'SudzCarwash'
,
520
,
1
,
20020102115234
,
500
,
20020102115259
,
501
,
33
,
32768
,
0
);
INSERT
INTO
t1
VALUES
(
333
,
'tubs'
,
99
,
2
,
20020109113440
,
501
,
20020109113440
,
500
,
3
,
10
,
0
);
CREATE
TABLE
t2
(
number
bigint
(
20
)
NOT
NULL
default
'0'
,
cname
char
(
15
)
NOT
NULL
default
''
,
carrier_id
smallint
(
6
)
NOT
NULL
default
'0'
,
privacy
tinyint
(
4
)
NOT
NULL
default
'0'
,
last_mod_date
timestamp
(
14
)
NOT
NULL
,
last_mod_id
smallint
(
6
)
NOT
NULL
default
'0'
,
last_app_date
timestamp
(
14
)
NOT
NULL
,
last_app_id
smallint
(
6
)
default
'-1'
,
version
smallint
(
6
)
NOT
NULL
default
'0'
,
assigned_scps
int
(
11
)
default
'0'
,
status
tinyint
(
4
)
default
'0'
)
TYPE
=
InnoDB
;
INSERT
INTO
t2
VALUES
(
4077711111
,
'SeanWheeler'
,
0
,
2
,
20020111112853
,
500
,
00000000000000
,
-
1
,
2
,
3
,
1
);
INSERT
INTO
t2
VALUES
(
9197722223
,
'berry'
,
90
,
3
,
20020111112818
,
500
,
20020102114532
,
501
,
4
,
10
,
0
);
INSERT
INTO
t2
VALUES
(
650
,
'San Francisco'
,
90
,
0
,
20020109113158
,
342
,
00000000000000
,
-
1
,
1
,
24
,
1
);
INSERT
INTO
t2
VALUES
(
333
,
'tubs'
,
99
,
2
,
20020109113453
,
501
,
20020109113453
,
500
,
3
,
10
,
0
);
select
*
from
t1
;
select
*
from
t2
;
delete
t1
,
t2
from
t1
left
join
t2
on
t1
.
number
=
t2
.
number
where
(
t1
.
carrier_id
=
90
and
t1
.
number
=
t2
.
number
)
or
(
t2
.
carrier_id
=
90
and
t1
.
number
=
t2
.
number
)
or
(
t1
.
carrier_id
=
90
and
t2
.
number
is
null
);
select
*
from
t1
;
select
*
from
t2
;
select
*
from
t2
;
drop
table
t1
,
t2
;
sql/handler.cc
View file @
a880ec94
...
@@ -472,17 +472,36 @@ int handler::analyze(THD* thd, HA_CHECK_OPT* check_opt)
...
@@ -472,17 +472,36 @@ int handler::analyze(THD* thd, HA_CHECK_OPT* check_opt)
return
HA_ADMIN_NOT_IMPLEMENTED
;
return
HA_ADMIN_NOT_IMPLEMENTED
;
}
}
/* Read first row from a table */
/*
Read first row (only) from a table
This is never called for InnoDB or BDB tables, as these table types
has the HA_NOT_EXACT_COUNT set.
*/
int
handler
::
r
nd_first
(
byte
*
buf
)
int
handler
::
r
ead_first_row
(
byte
*
buf
,
uint
primary_key
)
{
{
register
int
error
;
register
int
error
;
DBUG_ENTER
(
"handler::r
nd_first
"
);
DBUG_ENTER
(
"handler::r
ead_first_row
"
);
statistic_increment
(
ha_read_first_count
,
&
LOCK_status
);
statistic_increment
(
ha_read_first_count
,
&
LOCK_status
);
/*
If there is very few deleted rows in the table, find the first row by
scanning the table.
*/
if
(
deleted
<
10
||
primary_key
>=
MAX_KEY
)
{
(
void
)
rnd_init
();
(
void
)
rnd_init
();
while
((
error
=
rnd_next
(
buf
))
==
HA_ERR_RECORD_DELETED
)
;
while
((
error
=
rnd_next
(
buf
))
==
HA_ERR_RECORD_DELETED
)
;
(
void
)
rnd_end
();
(
void
)
rnd_end
();
}
else
{
/* Find the first row through the primary key */
(
void
)
index_init
(
primary_key
);
error
=
index_first
(
buf
);
(
void
)
index_end
();
}
DBUG_RETURN
(
error
);
DBUG_RETURN
(
error
);
}
}
...
@@ -498,7 +517,7 @@ int handler::restart_rnd_next(byte *buf, byte *pos)
...
@@ -498,7 +517,7 @@ int handler::restart_rnd_next(byte *buf, byte *pos)
}
}
/* Set a timestamp in record */
/* Set a timestamp in record */
void
handler
::
update_timestamp
(
byte
*
record
)
void
handler
::
update_timestamp
(
byte
*
record
)
{
{
...
@@ -514,9 +533,10 @@ void handler::update_timestamp(byte *record)
...
@@ -514,9 +533,10 @@ void handler::update_timestamp(byte *record)
return
;
return
;
}
}
/* Updates field with field_type NEXT_NUMBER according to following:
/*
** if field = 0 change field to the next free key in database.
Updates field with field_type NEXT_NUMBER according to following:
*/
if field = 0 change field to the next free key in database.
*/
void
handler
::
update_auto_increment
()
void
handler
::
update_auto_increment
()
{
{
...
...
sql/handler.h
View file @
a880ec94
...
@@ -269,7 +269,7 @@ class handler :public Sql_alloc
...
@@ -269,7 +269,7 @@ class handler :public Sql_alloc
virtual
int
rnd_end
()
{
return
0
;
}
virtual
int
rnd_end
()
{
return
0
;
}
virtual
int
rnd_next
(
byte
*
buf
)
=
0
;
virtual
int
rnd_next
(
byte
*
buf
)
=
0
;
virtual
int
rnd_pos
(
byte
*
buf
,
byte
*
pos
)
=
0
;
virtual
int
rnd_pos
(
byte
*
buf
,
byte
*
pos
)
=
0
;
virtual
int
r
nd_first
(
byte
*
buf
);
virtual
int
r
ead_first_row
(
byte
*
buf
,
uint
primary_key
);
virtual
int
restart_rnd_next
(
byte
*
buf
,
byte
*
pos
);
virtual
int
restart_rnd_next
(
byte
*
buf
,
byte
*
pos
);
virtual
ha_rows
records_in_range
(
int
inx
,
virtual
ha_rows
records_in_range
(
int
inx
,
const
byte
*
start_key
,
uint
start_key_len
,
const
byte
*
start_key
,
uint
start_key_len
,
...
...
sql/opt_range.cc
View file @
a880ec94
...
@@ -585,6 +585,9 @@ int SQL_SELECT::test_quick_select(key_map keys_to_use, table_map prev_tables,
...
@@ -585,6 +585,9 @@ int SQL_SELECT::test_quick_select(key_map keys_to_use, table_map prev_tables,
uint
idx
;
uint
idx
;
double
scan_time
;
double
scan_time
;
DBUG_ENTER
(
"test_quick_select"
);
DBUG_ENTER
(
"test_quick_select"
);
DBUG_PRINT
(
"enter"
,(
"keys_to_use: %lu prev_tables: %lu const_tables: %lu"
,
(
ulong
)
keys_to_use
,
(
ulong
)
prev_tables
,
(
ulong
)
const_tables
));
delete
quick
;
delete
quick
;
quick
=
0
;
quick
=
0
;
...
...
sql/sql_select.cc
View file @
a880ec94
...
@@ -41,6 +41,8 @@ static bool update_ref_and_keys(THD *thd, DYNAMIC_ARRAY *keyuse,
...
@@ -41,6 +41,8 @@ static bool update_ref_and_keys(THD *thd, DYNAMIC_ARRAY *keyuse,
uint
tables
,
COND
*
conds
,
table_map
table_map
);
uint
tables
,
COND
*
conds
,
table_map
table_map
);
static
int
sort_keyuse
(
KEYUSE
*
a
,
KEYUSE
*
b
);
static
int
sort_keyuse
(
KEYUSE
*
a
,
KEYUSE
*
b
);
static
void
set_position
(
JOIN
*
join
,
uint
index
,
JOIN_TAB
*
table
,
KEYUSE
*
key
);
static
void
set_position
(
JOIN
*
join
,
uint
index
,
JOIN_TAB
*
table
,
KEYUSE
*
key
);
static
bool
create_ref_for_key
(
JOIN
*
join
,
JOIN_TAB
*
j
,
KEYUSE
*
org_keyuse
,
table_map
used_tables
);
static
void
find_best_combination
(
JOIN
*
join
,
table_map
rest_tables
);
static
void
find_best_combination
(
JOIN
*
join
,
table_map
rest_tables
);
static
void
find_best
(
JOIN
*
join
,
table_map
rest_tables
,
uint
index
,
static
void
find_best
(
JOIN
*
join
,
table_map
rest_tables
,
uint
index
,
double
record_count
,
double
read_time
);
double
record_count
,
double
read_time
);
...
@@ -84,7 +86,7 @@ static int end_unique_update(JOIN *join,JOIN_TAB *join_tab,
...
@@ -84,7 +86,7 @@ static int end_unique_update(JOIN *join,JOIN_TAB *join_tab,
static
int
end_write_group
(
JOIN
*
join
,
JOIN_TAB
*
join_tab
,
static
int
end_write_group
(
JOIN
*
join
,
JOIN_TAB
*
join_tab
,
bool
end_of_records
);
bool
end_of_records
);
static
int
test_if_group_changed
(
List
<
Item_buff
>
&
list
);
static
int
test_if_group_changed
(
List
<
Item_buff
>
&
list
);
static
int
join_read_const_table
s
(
JOIN
*
join
);
static
int
join_read_const_table
(
JOIN_TAB
*
tab
,
POSITION
*
pos
);
static
int
join_read_system
(
JOIN_TAB
*
tab
);
static
int
join_read_system
(
JOIN_TAB
*
tab
);
static
int
join_read_const
(
JOIN_TAB
*
tab
);
static
int
join_read_const
(
JOIN_TAB
*
tab
);
static
int
join_read_key
(
JOIN_TAB
*
tab
);
static
int
join_read_key
(
JOIN_TAB
*
tab
);
...
@@ -413,9 +415,8 @@ mysql_select(THD *thd,TABLE_LIST *tables,List<Item> &fields,COND *conds,
...
@@ -413,9 +415,8 @@ mysql_select(THD *thd,TABLE_LIST *tables,List<Item> &fields,COND *conds,
goto
err
;
goto
err
;
thd
->
proc_info
=
"preparing"
;
thd
->
proc_info
=
"preparing"
;
result
->
initialize_tables
(
&
join
);
result
->
initialize_tables
(
&
join
);
if
((
tmp
=
join_read_const_tables
(
&
join
))
>
0
)
if
(
join
.
const_table_map
!=
join
.
found_const_table_map
&&
goto
err
;
!
(
select_options
&
SELECT_DESCRIBE
))
if
(
tmp
&&
!
(
select_options
&
SELECT_DESCRIBE
))
{
{
error
=
return_zero_rows
(
result
,
tables
,
fields
,
error
=
return_zero_rows
(
result
,
tables
,
fields
,
join
.
tmp_table_param
.
sum_func_count
!=
0
&&
join
.
tmp_table_param
.
sum_func_count
!=
0
&&
...
@@ -903,7 +904,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -903,7 +904,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
{
{
int
error
;
int
error
;
uint
i
,
table_count
,
const_count
,
found_ref
,
refs
,
key
,
const_ref
,
eq_part
;
uint
i
,
table_count
,
const_count
,
found_ref
,
refs
,
key
,
const_ref
,
eq_part
;
table_map
const_table_map
,
all_table_map
;
table_map
const_table_map
,
found_const_table_map
,
all_table_map
;
TABLE
**
table_vector
;
TABLE
**
table_vector
;
JOIN_TAB
*
stat
,
*
stat_end
,
*
s
,
**
stat_ref
;
JOIN_TAB
*
stat
,
*
stat_end
,
*
s
,
**
stat_ref
;
SQL_SELECT
*
select
;
SQL_SELECT
*
select
;
...
@@ -923,7 +924,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -923,7 +924,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
join
->
best_ref
=
stat_vector
;
join
->
best_ref
=
stat_vector
;
stat_end
=
stat
+
table_count
;
stat_end
=
stat
+
table_count
;
const_table_map
=
all_table_map
=
0
;
const_table_map
=
found_const_table_map
=
all_table_map
=
0
;
const_count
=
0
;
const_count
=
0
;
for
(
s
=
stat
,
i
=
0
;
tables
;
s
++
,
tables
=
tables
->
next
,
i
++
)
for
(
s
=
stat
,
i
=
0
;
tables
;
s
++
,
tables
=
tables
->
next
,
i
++
)
...
@@ -938,13 +939,13 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -938,13 +939,13 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
bzero
((
char
*
)
table
->
const_key_parts
,
sizeof
(
key_part_map
)
*
table
->
keys
);
bzero
((
char
*
)
table
->
const_key_parts
,
sizeof
(
key_part_map
)
*
table
->
keys
);
all_table_map
|=
table
->
map
;
all_table_map
|=
table
->
map
;
s
->
join
=
join
;
s
->
join
=
join
;
s
->
info
=
0
;
// For describe
if
((
s
->
on_expr
=
tables
->
on_expr
))
if
((
s
->
on_expr
=
tables
->
on_expr
))
{
{
/* Left join */
if
(
!
table
->
file
->
records
)
if
(
!
table
->
file
->
records
)
{
// Empty table
{
// Empty table
s
->
key_dependent
=
s
->
dependent
=
0
;
s
->
key_dependent
=
s
->
dependent
=
0
;
// Ignore LEFT JOIN depend.
s
->
type
=
JT_SYSTEM
;
const_table_map
|=
table
->
map
;
set_position
(
join
,
const_count
++
,
s
,(
KEYUSE
*
)
0
);
set_position
(
join
,
const_count
++
,
s
,(
KEYUSE
*
)
0
);
continue
;
continue
;
}
}
...
@@ -965,8 +966,6 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -965,8 +966,6 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
if
((
table
->
system
||
table
->
file
->
records
<=
1
)
&&
!
s
->
dependent
&&
if
((
table
->
system
||
table
->
file
->
records
<=
1
)
&&
!
s
->
dependent
&&
!
(
table
->
file
->
option_flag
()
&
HA_NOT_EXACT_COUNT
))
!
(
table
->
file
->
option_flag
()
&
HA_NOT_EXACT_COUNT
))
{
{
s
->
type
=
JT_SYSTEM
;
const_table_map
|=
table
->
map
;
set_position
(
join
,
const_count
++
,
s
,(
KEYUSE
*
)
0
);
set_position
(
join
,
const_count
++
,
s
,(
KEYUSE
*
)
0
);
}
}
}
}
...
@@ -974,10 +973,10 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -974,10 +973,10 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
join
->
outer_join
=
outer_join
;
join
->
outer_join
=
outer_join
;
/*
/*
**
If outer join: Re-arrange tables in stat_vector so that outer join
If outer join: Re-arrange tables in stat_vector so that outer join
**
tables are after all tables it is dependent of.
tables are after all tables it is dependent of.
**
For example: SELECT * from A LEFT JOIN B ON B.c=C.c, C WHERE A.C=C.C
For example: SELECT * from A LEFT JOIN B ON B.c=C.c, C WHERE A.C=C.C
**
Will shift table B after table C.
Will shift table B after table C.
*/
*/
if
(
outer_join
)
if
(
outer_join
)
{
{
...
@@ -1014,31 +1013,66 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -1014,31 +1013,66 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
conds
,
~
outer_join
))
conds
,
~
outer_join
))
DBUG_RETURN
(
1
);
DBUG_RETURN
(
1
);
/* Read tables with 0 or 1 rows (system tables) */
join
->
const_table_map
=
const_table_map
;
for
(
POSITION
*
p_pos
=
join
->
positions
,
*
p_end
=
p_pos
+
const_count
;
p_pos
<
p_end
;
p_pos
++
)
{
int
tmp
;
s
=
p_pos
->
table
;
s
->
type
=
JT_SYSTEM
;
join
->
const_table_map
|=
s
->
table
->
map
;
if
((
tmp
=
join_read_const_table
(
s
,
p_pos
)))
{
if
(
tmp
>
0
)
DBUG_RETURN
(
1
);
// Fatal error
}
else
found_const_table_map
|=
s
->
table
->
map
;
}
/* loop until no more const tables are found */
/* loop until no more const tables are found */
int
ref_changed
;
int
ref_changed
;
do
do
{
{
ref_changed
=
0
;
ref_changed
=
0
;
found_ref
=
0
;
found_ref
=
0
;
for
(
JOIN_TAB
**
pos
=
stat_vector
+
const_count
;
(
s
=
*
pos
)
;
pos
++
)
/*
We only have to loop from stat_vector + const_count as
set_position() will move all const_tables first in stat_vector
*/
for
(
JOIN_TAB
**
pos
=
stat_vector
+
const_count
;
(
s
=
*
pos
)
;
pos
++
)
{
{
TABLE
*
table
=
s
->
table
;
if
(
s
->
dependent
)
// If dependent on some table
if
(
s
->
dependent
)
// If dependent on some table
{
{
if
(
s
->
dependent
&
~
(
const_table_map
))
// All dep. must be constants
// All dep. must be constants
if
(
s
->
dependent
&
~
(
join
->
const_table_map
))
continue
;
continue
;
if
(
s
->
table
->
file
->
records
<=
1L
&&
if
(
table
->
file
->
records
<=
1L
&&
!
(
s
->
table
->
file
->
option_flag
()
&
HA_NOT_EXACT_COUNT
))
!
(
table
->
file
->
option_flag
()
&
HA_NOT_EXACT_COUNT
))
{
// system table
{
// system table
int
tmp
;
s
->
type
=
JT_SYSTEM
;
s
->
type
=
JT_SYSTEM
;
const_table_map
|=
s
->
table
->
map
;
join
->
const_table_map
|=
table
->
map
;
set_position
(
join
,
const_count
++
,
s
,(
KEYUSE
*
)
0
);
set_position
(
join
,
const_count
++
,
s
,(
KEYUSE
*
)
0
);
if
((
tmp
=
join_read_const_table
(
s
,
join
->
positions
+
const_count
-
1
)))
{
if
(
tmp
>
0
)
DBUG_RETURN
(
1
);
// Fatal error
}
else
found_const_table_map
|=
table
->
map
;
continue
;
continue
;
}
}
}
}
/* check if table can be read by key or table only uses const refs */
/* check if table can be read by key or table only uses const refs */
if
((
keyuse
=
s
->
keyuse
))
if
((
keyuse
=
s
->
keyuse
))
{
{
TABLE
*
table
=
s
->
table
;
s
->
type
=
JT_REF
;
s
->
type
=
JT_REF
;
while
(
keyuse
->
table
==
table
)
while
(
keyuse
->
table
==
table
)
{
{
...
@@ -1051,7 +1085,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -1051,7 +1085,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
{
{
if
(
keyuse
->
val
->
type
()
!=
Item
::
NULL_ITEM
)
if
(
keyuse
->
val
->
type
()
!=
Item
::
NULL_ITEM
)
{
{
if
(
!
((
~
const_table_map
)
&
keyuse
->
used_tables
))
if
(
!
((
~
join
->
const_table_map
)
&
keyuse
->
used_tables
))
const_ref
|=
(
key_map
)
1
<<
keyuse
->
keypart
;
const_ref
|=
(
key_map
)
1
<<
keyuse
->
keypart
;
else
else
refs
|=
keyuse
->
used_tables
;
refs
|=
keyuse
->
used_tables
;
...
@@ -1065,10 +1099,22 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -1065,10 +1099,22 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
{
{
if
(
const_ref
==
eq_part
)
if
(
const_ref
==
eq_part
)
{
// Found everything for ref.
{
// Found everything for ref.
int
tmp
;
ref_changed
=
1
;
s
->
type
=
JT_CONST
;
s
->
type
=
JT_CONST
;
const_table_map
|=
table
->
map
;
join
->
const_table_map
|=
table
->
map
;
set_position
(
join
,
const_count
++
,
s
,
start_keyuse
);
set_position
(
join
,
const_count
++
,
s
,
start_keyuse
);
ref_changed
=
1
;
if
(
create_ref_for_key
(
join
,
s
,
start_keyuse
,
join
->
const_table_map
))
DBUG_RETURN
(
1
);
if
((
tmp
=
join_read_const_table
(
s
,
join
->
positions
+
const_count
-
1
)))
{
if
(
tmp
>
0
)
DBUG_RETURN
(
1
);
// Fatal error
}
else
found_const_table_map
|=
table
->
map
;
break
;
break
;
}
}
else
else
...
@@ -1077,7 +1123,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -1077,7 +1123,7 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
}
}
}
}
}
}
}
while
(
const_table_map
&
found_ref
&&
ref_changed
);
}
while
(
join
->
const_table_map
&
found_ref
&&
ref_changed
);
/* Calc how many (possible) matched records in each table */
/* Calc how many (possible) matched records in each table */
...
@@ -1093,8 +1139,10 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -1093,8 +1139,10 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
s
->
found_records
=
s
->
records
=
s
->
table
->
file
->
records
;
s
->
found_records
=
s
->
records
=
s
->
table
->
file
->
records
;
s
->
read_time
=
(
ha_rows
)
s
->
table
->
file
->
scan_time
();
s
->
read_time
=
(
ha_rows
)
s
->
table
->
file
->
scan_time
();
/* Set a max range of how many seeks we can expect when using keys */
/*
/* This was (s->read_time*5), but this was too low with small rows */
Set a max range of how many seeks we can expect when using keys
This was (s->read_time*5), but this was too low with small rows
*/
s
->
worst_seeks
=
(
double
)
s
->
found_records
/
5
;
s
->
worst_seeks
=
(
double
)
s
->
found_records
/
5
;
if
(
s
->
worst_seeks
<
2.0
)
// Fix for small tables
if
(
s
->
worst_seeks
<
2.0
)
// Fix for small tables
s
->
worst_seeks
=
2.0
;
s
->
worst_seeks
=
2.0
;
...
@@ -1105,15 +1153,14 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -1105,15 +1153,14 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
{
{
ha_rows
records
;
ha_rows
records
;
if
(
!
select
)
if
(
!
select
)
select
=
make_select
(
s
->
table
,
const_table_map
,
select
=
make_select
(
s
->
table
,
join
->
const_table_map
,
0
,
join
->
const_table_map
,
and_conds
(
conds
,
s
->
on_expr
),
&
error
);
and_conds
(
conds
,
s
->
on_expr
),
&
error
);
records
=
get_quick_record_count
(
select
,
s
->
table
,
s
->
const_keys
,
records
=
get_quick_record_count
(
select
,
s
->
table
,
s
->
const_keys
,
join
->
row_limit
);
join
->
row_limit
);
s
->
quick
=
select
->
quick
;
s
->
quick
=
select
->
quick
;
s
->
needed_reg
=
select
->
needed_reg
;
s
->
needed_reg
=
select
->
needed_reg
;
select
->
quick
=
0
;
select
->
quick
=
0
;
select
->
read_tables
=
const_table_map
;
if
(
records
!=
HA_POS_ERROR
)
if
(
records
!=
HA_POS_ERROR
)
{
{
s
->
found_records
=
records
;
s
->
found_records
=
records
;
...
@@ -1128,10 +1175,10 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
...
@@ -1128,10 +1175,10 @@ make_join_statistics(JOIN *join,TABLE_LIST *tables,COND *conds,
join
->
map2table
=
stat_ref
;
join
->
map2table
=
stat_ref
;
join
->
table
=
join
->
all_tables
=
table_vector
;
join
->
table
=
join
->
all_tables
=
table_vector
;
join
->
const_tables
=
const_count
;
join
->
const_tables
=
const_count
;
join
->
const_table_map
=
const_table_map
;
join
->
found_const_table_map
=
found_
const_table_map
;
if
(
join
->
const_tables
!=
join
->
tables
)
if
(
join
->
const_tables
!=
join
->
tables
)
find_best_combination
(
join
,
all_table_map
&
~
const_table_map
);
find_best_combination
(
join
,
all_table_map
&
~
join
->
const_table_map
);
else
else
{
{
memcpy
((
gptr
)
join
->
best_positions
,(
gptr
)
join
->
positions
,
memcpy
((
gptr
)
join
->
best_positions
,(
gptr
)
join
->
positions
,
...
@@ -2040,12 +2087,10 @@ prev_record_reads(JOIN *join,table_map found_ref)
...
@@ -2040,12 +2087,10 @@ prev_record_reads(JOIN *join,table_map found_ref)
static
bool
static
bool
get_best_combination
(
JOIN
*
join
)
get_best_combination
(
JOIN
*
join
)
{
{
uint
i
,
key
,
tablenr
;
uint
i
,
tablenr
;
table_map
used_tables
;
table_map
used_tables
;
TABLE
*
table
;
JOIN_TAB
*
join_tab
,
*
j
;
JOIN_TAB
*
join_tab
,
*
j
;
KEYUSE
*
keyuse
;
KEYUSE
*
keyuse
;
KEY
*
keyinfo
;
uint
table_count
;
uint
table_count
;
THD
*
thd
=
join
->
thd
;
THD
*
thd
=
join
->
thd
;
...
@@ -2054,8 +2099,6 @@ get_best_combination(JOIN *join)
...
@@ -2054,8 +2099,6 @@ get_best_combination(JOIN *join)
(
JOIN_TAB
*
)
thd
->
alloc
(
sizeof
(
JOIN_TAB
)
*
table_count
)))
(
JOIN_TAB
*
)
thd
->
alloc
(
sizeof
(
JOIN_TAB
)
*
table_count
)))
return
TRUE
;
return
TRUE
;
join
->
const_tables
=
0
;
/* for checking */
join
->
const_table_map
=
0
;
join
->
full_join
=
0
;
join
->
full_join
=
0
;
used_tables
=
0
;
used_tables
=
0
;
...
@@ -2064,41 +2107,52 @@ get_best_combination(JOIN *join)
...
@@ -2064,41 +2107,52 @@ get_best_combination(JOIN *join)
TABLE
*
form
;
TABLE
*
form
;
*
j
=
*
join
->
best_positions
[
tablenr
].
table
;
*
j
=
*
join
->
best_positions
[
tablenr
].
table
;
form
=
join
->
table
[
tablenr
]
=
j
->
table
;
form
=
join
->
table
[
tablenr
]
=
j
->
table
;
j
->
ref
.
key
=
-
1
;
j
->
ref
.
key_parts
=
0
;
j
->
info
=
0
;
// For describe
used_tables
|=
form
->
map
;
used_tables
|=
form
->
map
;
form
->
reginfo
.
join_tab
=
j
;
form
->
reginfo
.
join_tab
=
j
;
if
(
!
j
->
on_expr
)
if
(
!
j
->
on_expr
)
form
->
reginfo
.
not_exists_optimize
=
0
;
// Only with LEFT JOIN
form
->
reginfo
.
not_exists_optimize
=
0
;
// Only with LEFT JOIN
if
(
j
->
type
==
JT_CONST
)
continue
;
// Handled in make_join_stat..
j
->
ref
.
key
=
-
1
;
j
->
ref
.
key_parts
=
0
;
if
(
j
->
type
==
JT_SYSTEM
)
if
(
j
->
type
==
JT_SYSTEM
)
{
j
->
table
->
const_table
=
1
;
if
(
join
->
const_tables
==
tablenr
)
{
join
->
const_tables
++
;
join
->
const_table_map
|=
form
->
map
;
}
continue
;
continue
;
}
if
(
!
j
->
keys
||
!
(
keyuse
=
join
->
best_positions
[
tablenr
].
key
))
if
(
!
j
->
keys
||
!
(
keyuse
=
join
->
best_positions
[
tablenr
].
key
))
{
{
j
->
type
=
JT_ALL
;
j
->
type
=
JT_ALL
;
if
(
tablenr
!=
join
->
const_tables
)
if
(
tablenr
!=
join
->
const_tables
)
join
->
full_join
=
1
;
join
->
full_join
=
1
;
}
}
else
else
if
(
create_ref_for_key
(
join
,
j
,
keyuse
,
used_tables
))
{
return
TRUE
;
// Something went wrong
uint
keyparts
,
length
;
}
for
(
i
=
0
;
i
<
table_count
;
i
++
)
join
->
map2table
[
join
->
join_tab
[
i
].
table
->
tablenr
]
=
join
->
join_tab
+
i
;
update_depend_map
(
join
);
return
0
;
}
static
bool
create_ref_for_key
(
JOIN
*
join
,
JOIN_TAB
*
j
,
KEYUSE
*
org_keyuse
,
table_map
used_tables
)
{
KEYUSE
*
keyuse
=
org_keyuse
;
bool
ftkey
=
(
keyuse
->
keypart
==
FT_KEYPART
);
bool
ftkey
=
(
keyuse
->
keypart
==
FT_KEYPART
);
THD
*
thd
=
join
->
thd
;
uint
keyparts
,
length
,
key
;
TABLE
*
table
;
KEY
*
keyinfo
;
/*
/*
** Use best key from find_best
** Use best key from find_best
*/
*/
table
=
j
->
table
;
table
=
j
->
table
;
key
=
keyuse
->
key
;
key
=
keyuse
->
key
;
keyinfo
=
table
->
key_info
+
key
;
keyinfo
=
table
->
key_info
+
key
;
if
(
ftkey
)
if
(
ftkey
)
{
{
Item_func_match
*
ifm
=
(
Item_func_match
*
)
keyuse
->
val
;
Item_func_match
*
ifm
=
(
Item_func_match
*
)
keyuse
->
val
;
...
@@ -2138,7 +2192,7 @@ get_best_combination(JOIN *join)
...
@@ -2138,7 +2192,7 @@ get_best_combination(JOIN *join)
}
}
j
->
ref
.
key_buff2
=
j
->
ref
.
key_buff
+
ALIGN_SIZE
(
length
);
j
->
ref
.
key_buff2
=
j
->
ref
.
key_buff
+
ALIGN_SIZE
(
length
);
j
->
ref
.
key_err
=
1
;
j
->
ref
.
key_err
=
1
;
keyuse
=
join
->
best_positions
[
tablenr
].
key
;
keyuse
=
org_keyuse
;
store_key
**
ref_key
=
j
->
ref
.
key_copy
;
store_key
**
ref_key
=
j
->
ref
.
key_copy
;
byte
*
key_buff
=
j
->
ref
.
key_buff
;
byte
*
key_buff
=
j
->
ref
.
key_buff
;
...
@@ -2152,6 +2206,7 @@ get_best_combination(JOIN *join)
...
@@ -2152,6 +2206,7 @@ get_best_combination(JOIN *join)
}
}
else
else
{
{
uint
i
;
for
(
i
=
0
;
i
<
keyparts
;
keyuse
++
,
i
++
)
for
(
i
=
0
;
i
<
keyparts
;
keyuse
++
,
i
++
)
{
{
while
(
keyuse
->
keypart
!=
i
||
while
(
keyuse
->
keypart
!=
i
||
...
@@ -2178,7 +2233,7 @@ get_best_combination(JOIN *join)
...
@@ -2178,7 +2233,7 @@ get_best_combination(JOIN *join)
tmp
->
copy
();
tmp
->
copy
();
}
}
else
else
*
ref_key
++=
get_store_key
(
join
->
thd
,
*
ref_key
++=
get_store_key
(
thd
,
keyuse
,
join
->
const_table_map
,
keyuse
,
join
->
const_table_map
,
&
keyinfo
->
key_part
[
i
],
&
keyinfo
->
key_part
[
i
],
(
char
*
)
key_buff
,
maybe_null
);
(
char
*
)
key_buff
,
maybe_null
);
...
@@ -2188,15 +2243,9 @@ get_best_combination(JOIN *join)
...
@@ -2188,15 +2243,9 @@ get_best_combination(JOIN *join)
*
ref_key
=
0
;
// end_marker
*
ref_key
=
0
;
// end_marker
if
(
j
->
type
==
JT_FT
)
/* no-op */
;
if
(
j
->
type
==
JT_FT
)
/* no-op */
;
else
if
(
j
->
type
==
JT_CONST
)
else
if
(
j
->
type
==
JT_CONST
)
{
j
->
table
->
const_table
=
1
;
j
->
table
->
const_table
=
1
;
if
(
join
->
const_tables
==
tablenr
)
else
if
(((
keyinfo
->
flags
&
(
HA_NOSAME
|
HA_NULL_PART_KEY
))
{
!=
HA_NOSAME
)
||
join
->
const_tables
++
;
join
->
const_table_map
|=
form
->
map
;
}
}
else
if
(((
keyinfo
->
flags
&
(
HA_NOSAME
|
HA_NULL_PART_KEY
))
!=
HA_NOSAME
)
||
keyparts
!=
keyinfo
->
key_parts
)
keyparts
!=
keyinfo
->
key_parts
)
j
->
type
=
JT_REF
;
/* Must read with repeat */
j
->
type
=
JT_REF
;
/* Must read with repeat */
else
if
(
ref_key
==
j
->
ref
.
key_copy
)
else
if
(
ref_key
==
j
->
ref
.
key_copy
)
...
@@ -2209,24 +2258,14 @@ get_best_combination(JOIN *join)
...
@@ -2209,24 +2258,14 @@ get_best_combination(JOIN *join)
have a 'normal' value or a NULL value.
have a 'normal' value or a NULL value.
*/
*/
j
->
type
=
JT_CONST
;
j
->
type
=
JT_CONST
;
if
(
join
->
const_tables
==
tablenr
)
{
join
->
const_tables
++
;
join
->
const_table_map
|=
form
->
map
;
}
}
}
else
else
j
->
type
=
JT_EQ_REF
;
j
->
type
=
JT_EQ_REF
;
}
}
for
(
i
=
0
;
i
<
table_count
;
i
++
)
join
->
map2table
[
join
->
join_tab
[
i
].
table
->
tablenr
]
=
join
->
join_tab
+
i
;
update_depend_map
(
join
);
return
0
;
return
0
;
}
}
static
store_key
*
static
store_key
*
get_store_key
(
THD
*
thd
,
KEYUSE
*
keyuse
,
table_map
used_tables
,
get_store_key
(
THD
*
thd
,
KEYUSE
*
keyuse
,
table_map
used_tables
,
KEY_PART_INFO
*
key_part
,
char
*
key_buff
,
uint
maybe_null
)
KEY_PART_INFO
*
key_part
,
char
*
key_buff
,
uint
maybe_null
)
...
@@ -3163,13 +3202,13 @@ remove_eq_conds(COND *cond,Item::cond_result *cond_value)
...
@@ -3163,13 +3202,13 @@ remove_eq_conds(COND *cond,Item::cond_result *cond_value)
((
Item_func
*
)
cond
)
->
functype
()
==
Item_func
::
ISNULL_FUNC
)
((
Item_func
*
)
cond
)
->
functype
()
==
Item_func
::
ISNULL_FUNC
)
{
{
/*
/*
**
Handles this special case for some ODBC applications:
Handles this special case for some ODBC applications:
**
The are requesting the row that was just updated with a auto_increment
The are requesting the row that was just updated with a auto_increment
**
value with this construct:
value with this construct:
**
**
SELECT * from table_name where auto_increment_column IS NULL
SELECT * from table_name where auto_increment_column IS NULL
**
This will be changed to:
This will be changed to:
**
SELECT * from table_name where auto_increment_column = LAST_INSERT_ID
SELECT * from table_name where auto_increment_column = LAST_INSERT_ID
*/
*/
Item_func_isnull
*
func
=
(
Item_func_isnull
*
)
cond
;
Item_func_isnull
*
func
=
(
Item_func_isnull
*
)
cond
;
...
@@ -4365,45 +4404,44 @@ flush_cached_records(JOIN *join,JOIN_TAB *join_tab,bool skipp_last)
...
@@ -4365,45 +4404,44 @@ flush_cached_records(JOIN *join,JOIN_TAB *join_tab,bool skipp_last)
*****************************************************************************/
*****************************************************************************/
static
int
static
int
join_read_const_table
s
(
JOIN
*
join
)
join_read_const_table
(
JOIN_TAB
*
tab
,
POSITION
*
pos
)
{
{
uint
i
;
int
error
;
int
error
;
DBUG_ENTER
(
"join_read_const_tables"
);
DBUG_ENTER
(
"join_read_const_table"
);
for
(
i
=
0
;
i
<
join
->
const_tables
;
i
++
)
TABLE
*
table
=
tab
->
table
;
{
table
->
const_table
=
1
;
TABLE
*
form
=
join
->
table
[
i
];
table
->
null_row
=
0
;
form
->
null_row
=
0
;
table
->
status
=
STATUS_NO_RECORD
;
form
->
status
=
STATUS_NO_RECORD
;
if
(
join
->
join_tab
[
i
].
type
==
JT_SYSTEM
)
if
(
tab
->
type
==
JT_SYSTEM
)
{
{
if
((
error
=
join_read_system
(
join
->
join_tab
+
i
)))
if
((
error
=
join_read_system
(
tab
)))
{
// Info for DESCRIBE
{
// Info for DESCRIBE
join
->
join_tab
[
i
].
info
=
"const row not found"
;
tab
->
info
=
"const row not found"
;
join
->
best_positions
[
i
].
records_read
=
0.0
;
/* Mark for EXPLAIN that the row was not found */
if
(
!
form
->
outer_join
||
error
>
0
)
pos
->
records_read
=
0.0
;
if
(
!
table
->
outer_join
||
error
>
0
)
DBUG_RETURN
(
error
);
DBUG_RETURN
(
error
);
}
}
}
}
else
else
{
{
if
((
error
=
join_read_const
(
join
->
join_tab
+
i
)))
if
((
error
=
join_read_const
(
tab
)))
{
{
join
->
join_tab
[
i
].
info
=
"unique row not found"
;
tab
->
info
=
"unique row not found"
;
join
->
best_positions
[
i
].
records_read
=
0.0
;
/* Mark for EXPLAIN that the row was not found */
if
(
!
form
->
outer_join
||
error
>
0
)
pos
->
records_read
=
0.0
;
if
(
!
table
->
outer_join
||
error
>
0
)
DBUG_RETURN
(
error
);
DBUG_RETURN
(
error
);
}
}
}
}
if
(
join
->
join_tab
[
i
].
on_expr
&&
!
form
->
null_row
)
if
(
tab
->
on_expr
&&
!
table
->
null_row
)
{
{
if
((
form
->
null_row
=
test
(
join
->
join_tab
[
i
].
on_expr
->
val_int
()
==
0
)))
if
((
table
->
null_row
=
test
(
tab
->
on_expr
->
val_int
()
==
0
)))
empty_record
(
form
);
empty_record
(
table
);
}
if
(
!
form
->
null_row
)
form
->
maybe_null
=
0
;
}
}
if
(
!
table
->
null_row
)
table
->
maybe_null
=
0
;
DBUG_RETURN
(
0
);
DBUG_RETURN
(
0
);
}
}
...
@@ -4415,7 +4453,8 @@ join_read_system(JOIN_TAB *tab)
...
@@ -4415,7 +4453,8 @@ join_read_system(JOIN_TAB *tab)
int
error
;
int
error
;
if
(
table
->
status
&
STATUS_GARBAGE
)
// If first read
if
(
table
->
status
&
STATUS_GARBAGE
)
// If first read
{
{
if
((
error
=
table
->
file
->
rnd_first
(
table
->
record
[
0
])))
if
((
error
=
table
->
file
->
read_first_row
(
table
->
record
[
0
],
table
->
primary_key
)))
{
{
if
(
error
!=
HA_ERR_END_OF_FILE
)
if
(
error
!=
HA_ERR_END_OF_FILE
)
{
{
...
...
sql/sql_select.h
View file @
a880ec94
...
@@ -158,7 +158,7 @@ class JOIN {
...
@@ -158,7 +158,7 @@ class JOIN {
uint
send_group_parts
;
uint
send_group_parts
;
bool
sort_and_group
,
first_record
,
full_join
,
group
,
no_field_update
;
bool
sort_and_group
,
first_record
,
full_join
,
group
,
no_field_update
;
bool
do_send_rows
;
bool
do_send_rows
;
table_map
const_table_map
,
outer_join
;
table_map
const_table_map
,
found_const_table_map
,
outer_join
;
ha_rows
send_records
,
found_records
,
examined_rows
,
row_limit
;
ha_rows
send_records
,
found_records
,
examined_rows
,
row_limit
;
POSITION
positions
[
MAX_TABLES
+
1
],
best_positions
[
MAX_TABLES
+
1
];
POSITION
positions
[
MAX_TABLES
+
1
],
best_positions
[
MAX_TABLES
+
1
];
double
best_read
;
double
best_read
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment