Many hyperlinks are disabled.
Use anonymous login
to enable hyperlinks.
Overview
Comment: | Added the speed1.test script (CVS 3522) |
---|---|
Downloads: | Tarball | ZIP archive |
Timelines: | family | ancestors | descendants | both | trunk |
Files: | files | file ages | folders |
SHA1: |
30355dfbd920f3b6a78110aaf370371f |
User & Date: | drh 2006-11-29 20:53:00.000 |
Context
2006-11-29
| ||
21:03 | Test that terms longer than interior nodes work correctly. A bug prior to fts2.c r1.10 meant that such large terms caused an eventual stack overflow. (CVS 3523) (check-in: 66581162da user: shess tags: trunk) | |
20:53 | Added the speed1.test script (CVS 3522) (check-in: 30355dfbd9 user: drh tags: trunk) | |
05:17 |
http://www.sqlite.org/cvstrac/tktview?tn=2046
The virtual table interface allows for a cursor to field multiple xFilter() calls. For instance, if a join is done with a virtual table, there could be a call for each row which potentially matches. Unfortunately, fulltextFilter() assumes that it has a fresh cursor, and overwrites a prepared statement and a malloc'ed pointer, resulting in unfinalized statements and a memory leak. This change hacks the code to manually clean up offending items in fulltextFilter(), emphasis on "hacks", since it's a fragile fix insofar as future additions to fulltext_cursor could continue to have the problem. (CVS 3521) (check-in: 18142fdb6d user: shess tags: trunk) | |
Changes
Added test/speed1.test.
> > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 | # 2006 November 23 # # The author disclaims copyright to this source code. In place of # a legal notice, here is a blessing: # # May you do good and not evil. # May you find forgiveness for yourself and forgive others. # May you share freely, never taking more than you give. # #************************************************************************* # This file implements regression tests for SQLite library. The # focus of this script is measuring executing speed. # # $Id: speed1.test,v 1.1 2006/11/29 20:53:00 drh Exp $ # set testdir [file dirname $argv0] source $testdir/tester.tcl # The number_name procedure below converts its argment (an integer) # into a string which is the English-language name for that number. # # Example: # # puts [number_name 123] -> "one hundred twenty three" # set ones {zero one two three four five six seven eight nine ten eleven twelve thirteen fourteen fifteen sixteen seventeen eighteen nineteen} set tens {{} ten twenty thirty forty fifty sixty seventy eighty ninety} proc number_name {n} { if {$n>=1000} { set txt "[number_name [expr {$n/1000}]] thousand" set n [expr {$n%1000}] } else { set txt {} } if {$n>=100} { append txt " [lindex $::ones [expr {$n/100}]] hundred" set n [expr {$n%100}] } if {$n>=20} { append txt " [lindex $::tens [expr {$n/10}]]" set n [expr {$n%10}] } if {$n>0} { append txt " [lindex $::ones $n]" } set txt [string trim $txt] if {$txt==""} {set txt zero} return $txt } # Create a database schema. # do_test speed1-1.0 { execsql { CREATE TABLE t1(a INTEGER, b INTEGER, c TEXT); CREATE TABLE t2(a INTEGER, b INTEGER, c TEXT); CREATE INDEX i2a ON t2(a); CREATE INDEX i2b ON t2(b); SELECT name FROM sqlite_master ORDER BY 1; } } {i2a i2b t1 t2} # 50000 INSERTs on an unindexed table # set sql {} for {set i 1} {$i<=50000} {incr i} { set r [expr {int(rand()*500000)}] append sql "INSERT INTO t1 VALUES($i,$r,'[number_name $r]');\n" } db eval BEGIN speed_trial speed1-insert1 50000 row $sql db eval COMMIT # 50000 INSERTs on an indexed table # set sql {} for {set i 1} {$i<=50000} {incr i} { set r [expr {int(rand()*500000)}] append sql "INSERT INTO t2 VALUES($i,$r,'[number_name $r]');\n" } db eval BEGIN speed_trial speed1-insert2 50000 row $sql db eval COMMIT # 50 SELECTs on an integer comparison. There is no index so # a full table scan is required. # set sql {} for {set i 0} {$i<50} {incr i} { set lwr [expr {$i*100}] set upr [expr {($i+10)*100}] append sql "SELECT count(*), avg(b) FROM t1 WHERE b>=$lwr AND b<$upr;" } db eval BEGIN speed_trial speed1-select1 [expr {50*50000}] row $sql db eval COMMIT # 50 SELECTs on an LIKE comparison. There is no index so a full # table scan is required. # set sql {} for {set i 0} {$i<50} {incr i} { append sql \ "SELECT count(*), avg(b) FROM t1 WHERE c LIKE '%[number_name $i]%';" } db eval BEGIN speed_trial speed1-select2 [expr {50*50000}] row $sql db eval COMMIT # Create indices # db eval BEGIN speed_trial speed1-createidx 150000 row { CREATE INDEX i1a ON t1(a); CREATE INDEX i1b ON t1(b); CREATE INDEX i1c ON t1(c); } db eval COMMIT # 5000 SELECTs on an integer comparison where the integer is # indexed. # set sql {} for {set i 0} {$i<5000} {incr i} { set lwr [expr {$i*100}] set upr [expr {($i+10)*100}] append sql "SELECT count(*), avg(b) FROM t1 WHERE b>=$lwr AND b<$upr;" } db eval BEGIN speed_trial speed1-select3 5000 stmt $sql db eval COMMIT # 20000 random SELECTs against rowid. # set sql {} for {set i 1} {$i<=20000} {incr i} { set id [expr {int(rand()*50000)+1}] append sql "SELECT c FROM t1 WHERE rowid=$id;" } db eval BEGIN speed_trial speed1-select4 20000 row $sql db eval COMMIT # 20000 random SELECTs against a unique indexed column. # set sql {} for {set i 1} {$i<=20000} {incr i} { set id [expr {int(rand()*50000)+1}] append sql "SELECT c FROM t1 WHERE a=$id;" } db eval BEGIN speed_trial speed1-select5 20000 row $sql db eval COMMIT # 20000 random SELECTs against an indexed column text column # set sql {} db eval {SELECT c FROM t1 ORDER BY random() LIMIT 20000} { append sql "SELECT c FROM t1 WHERE c='$c';" } db eval BEGIN speed_trial speed1-select6 20000 row $sql db eval COMMIT # Vacuum speed_trial speed1-vacuum 100000 row VACUUM # 5000 updates of ranges where the field being compared is indexed. # set sql {} for {set i 0} {$i<5000} {incr i} { set lwr [expr {$i*2}] set upr [expr {($i+1)*2}] append sql "UPDATE t1 SET b=b*2 WHERE a>=$lwr AND a<$upr;" } db eval BEGIN speed_trial speed1-update1 5000 stmt $sql db eval COMMIT # 50000 single-row updates. An index is used to find the row quickly. # set sql {} for {set i 0} {$i<50000} {incr i} { set r [expr {int(rand()*500000)}] append sql "UPDATE t1 SET b=$r WHERE a=$i;" } db eval BEGIN speed_trial speed1-update2 50000 row $sql db eval COMMIT # 1 big text update that touches every row in the table. # speed_trial speed1-update3 50000 row { UPDATE t1 SET c=a; } # Many individual text updates. Each row in the table is # touched through an index. # set sql {} for {set i 1} {$i<=50000} {incr i} { set r [expr {int(rand()*500000)}] append sql "UPDATE t1 SET c='[number_name $r]' WHERE a=$i;" } db eval BEGIN speed_trial speed1-update4 50000 row $sql db eval COMMIT # Delete all content in a table. # speed_trial speed1-delete1 50000 row {DELETE FROM t1} # Copy one table into another # speed_trial speed1-copy1 50000 row {INSERT INTO t1 SELECT * FROM t2} # Delete all content in a table, one row at a time. # speed_trial speed1-delete2 50000 row {DELETE FROM t1 WHERE 1} # Refill the table yet again # speed_trial speed1-copy2 50000 row {INSERT INTO t1 SELECT * FROM t2} # Drop the table and recreate it without its indices. # db eval BEGIN speed_trial speed1-drop1 50000 row { DROP TABLE t1; CREATE TABLE t1(a INTEGER, b INTEGER, c TEXT); } db eval COMMIT # Refill the table yet again. This copy should be faster because # there are no indices to deal with. # speed_trial speed1-copy3 50000 row {INSERT INTO t1 SELECT * FROM t2} # Select 20000 rows from the table at random. # speed_trial speed1-random1 50000 row { SELECT rowid FROM t1 ORDER BY random() LIMIT 20000 } # Delete 20000 random rows from the table. # speed_trial speed1-random-del1 20000 row { DELETE FROM t1 WHERE rowid IN (SELECT rowid FROM t1 ORDER BY random() LIMIT 20000) } do_test speed1-1.1 { db one {SELECT count(*) FROM t1} } 30000 # Delete 20000 more rows at random from the table. # speed_trial speed1-random-del2 20000 row { DELETE FROM t1 WHERE rowid IN (SELECT rowid FROM t1 ORDER BY random() LIMIT 20000) } do_test speed1-1.2 { db one {SELECT count(*) FROM t1} } 10000 finish_test |