diff options
author | Jesse Morgan <jesse@jesterpm.net> | 2016-12-17 21:28:53 -0800 |
---|---|---|
committer | Jesse Morgan <jesse@jesterpm.net> | 2016-12-17 21:28:53 -0800 |
commit | 54df2afaa61c6a03cbb4a33c9b90fa572b6d07b8 (patch) | |
tree | 18147b92b969d25ffbe61935fb63035cac820dd0 /db-4.8.30/test/test111.tcl |
Berkeley DB 4.8 with rust build script for linux.
Diffstat (limited to 'db-4.8.30/test/test111.tcl')
-rw-r--r-- | db-4.8.30/test/test111.tcl | 370 |
1 files changed, 370 insertions, 0 deletions
diff --git a/db-4.8.30/test/test111.tcl b/db-4.8.30/test/test111.tcl new file mode 100644 index 0000000..836f1fa --- /dev/null +++ b/db-4.8.30/test/test111.tcl @@ -0,0 +1,370 @@ +# See the file LICENSE for redistribution information. +# +# Copyright (c) 2005-2009 Oracle. All rights reserved. +# +# $Id$ +# +# TEST test111 +# TEST Test database compaction. +# TEST +# TEST Populate a database. Remove a high proportion of entries. +# TEST Dump and save contents. Compact the database, dump again, +# TEST and make sure we still have the same contents. +# TEST Add back some entries, delete more entries (this time by +# TEST cursor), dump, compact, and do the before/after check again. + +proc test111 { method {nentries 10000} {tnum "111"} args } { + + # Compaction is an option for btree and recno databases only. + if { [is_hash $method] == 1 || [is_queue $method] == 1 } { + puts "Skipping test$tnum for method $method." + return + } + + # If a page size was specified, find out what it is. Pages + # might not be freed in the case of really large pages (64K) + # but we still want to run this test just to make sure + # nothing funny happens. + set pagesize 0 + set pgindex [lsearch -exact $args "-pagesize"] + if { $pgindex != -1 } { + incr pgindex + set pagesize [lindex $args $pgindex] + } + + source ./include.tcl + global rand_init + error_check_good set_random_seed [berkdb srand $rand_init] 0 + set args [convert_args $method $args] + set omethod [convert_method $method] + if { [is_partition_callback $args] == 1 } { + set nodump 1 + } else { + set nodump 0 + } + + # If we are using an env, then testfile should just be the db name. + # Otherwise it is the test directory and the name. + set txnenv 0 + set eindex [lsearch -exact $args "-env"] + if { $eindex == -1 } { + set basename $testdir/test$tnum + set env NULL + } else { + set basename test$tnum + incr eindex + set env [lindex $args $eindex] + set rpcenv [is_rpcenv $env] + if { $rpcenv == 1 } { + puts "Test$tnum: skipping for RPC" + return + } + set txnenv [is_txnenv $env] + if { $txnenv == 1 } { + append args " -auto_commit " + } + set testdir [get_home $env] + } + puts "Test$tnum: ($method $args) Database compaction." + set t1 $testdir/t1 + set t2 $testdir/t2 + set splitopts { "" "-revsplitoff" } + set txn "" + + if { [is_record_based $method] == 1 } { + set checkfunc test001_recno.check + } else { + set checkfunc test001.check + } + + foreach splitopt $splitopts { + set testfile $basename.db + if { $splitopt == "-revsplitoff" } { + set testfile $basename.rev.db + if { [is_record_based $method] == 1 } { + puts "Skipping\ + -revsplitoff option for method $method." + continue + } + } + set did [open $dict] + if { $env != "NULL" } { + set testdir [get_home $env] + } + cleanup $testdir $env + + puts "\tTest$tnum.a: Create and populate database ($splitopt)." + set db [eval {berkdb_open -create \ + -mode 0644} $splitopt $args $omethod $testfile] + error_check_good dbopen [is_valid_db $db] TRUE + + set count 0 + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + while { [gets $did str] != -1 && $count < $nentries } { + global kvals + + if { [is_record_based $method] == 1 } { + set key [expr $count + 1] + set kvals($key) [pad_data $method $str] + } else { + set key $str + set str [reverse $str] + } + + set ret [eval \ + {$db put} $txn {$key [chop_data $method $str]}] + error_check_good put $ret 0 + incr count + + } + if { $txnenv == 1 } { + error_check_good txn_commit [$t commit] 0 + } + close $did + error_check_good db_sync [$db sync] 0 + + if { $env != "NULL" } { + set testdir [get_home $env] + set filename $testdir/$testfile + } else { + set filename $testfile + } + set size1 [file size $filename] + set free1 [stat_field $db stat "Pages on freelist"] + set leaf1 [stat_field $db stat "Leaf pages"] + set internal1 [stat_field $db stat "Internal pages"] + + # Delete between 1 and maxdelete items, then skip over between + # 1 and maxskip items. This is to make the data bunchy, + # so we sometimes follow the code path where merging is + # done record by record, and sometimes the path where + # the whole page is merged at once. + + puts "\tTest$tnum.b: Delete most entries from database." + set did [open $dict] + set count [expr $nentries - 1] + set maxskip 4 + set maxdelete 48 + + # Since rrecno and rbtree renumber, we delete starting at + # nentries and working down to 0. + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + while { [gets $did str] != -1 && $count > 0 } { + + # Delete a random number of successive items. + set ndeletes [berkdb random_int 1 $maxdelete] + set target [expr $count - $ndeletes] + while { [expr $count > $target] && $count > 0 } { + if { [is_record_based $method] == 1 } { + set key [expr $count + 1] + } else { + set key [gets $did] + } + + set ret [eval {$db del} $txn {$key}] + error_check_good del $ret 0 + incr count -1 + } + # Skip over a random smaller number of items. + set skip [berkdb random_int 1 [expr $maxskip]] + set target [expr $count - $skip] + while { [expr $count > $target] && $count > 0 } { + incr count -1 + } + } + if { $txnenv == 1 } { + error_check_good t_commit [$t commit] 0 + } + error_check_good db_sync [$db sync] 0 + + puts "\tTest$tnum.c: Do a dump_file on contents." + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + dump_file $db $txn $t1 + if { $txnenv == 1 } { + error_check_good txn_commit [$t commit] 0 + } + + puts "\tTest$tnum.d: Compact and verify database." + for {set commit 0} {$commit <= $txnenv} {incr commit} { + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + set ret [eval $db compact $txn -freespace] + if { $txnenv == 1 } { + if { $commit == 0 } { + puts "\tTest$tnum.d: Aborting." + error_check_good txn_abort [$t abort] 0 + } else { + puts "\tTest$tnum.d: Committing." + error_check_good txn_commit [$t commit] 0 + } + } + error_check_good db_sync [$db sync] 0 + error_check_good verify_dir \ + [verify_dir $testdir "" 0 0 $nodump ] 0 + } + + set size2 [file size $filename] + set free2 [stat_field $db stat "Pages on freelist"] + set leaf2 [stat_field $db stat "Leaf pages"] + set internal2 [stat_field $db stat "Internal pages"] + + # The sum of internal pages, leaf pages, and pages freed + # should decrease on compaction, indicating that pages + # have been freed to the file system. + set sum1 [expr $free1 + $leaf1 + $internal1] + set sum2 [expr $free2 + $leaf2 + $internal2] + error_check_good pages_freed [expr $sum1 > $sum2] 1 + + # The on-disk file size should be smaller. + #### We should look at the partitioned files ##### + if { [is_partitioned $args] == 0 } { + set reduction .96 + error_check_good \ + file_size [expr [expr $size1 * $reduction] > $size2] 1 + } + + puts "\tTest$tnum.e: Contents are the same after compaction." + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + dump_file $db $txn $t2 + if { $txnenv == 1 } { + error_check_good txn_commit [$t commit] 0 + } + + error_check_good filecmp [filecmp $t1 $t2] 0 + + puts "\tTest$tnum.f: Add more entries to database." + # Use integers as keys instead of strings, just to mix it up + # a little. + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + for { set i 1 } { $i < $nentries } { incr i } { + set key $i + set str $i + set ret [eval \ + {$db put} $txn {$key [chop_data $method $str]}] + error_check_good put $ret 0 + } + if { $txnenv == 1 } { + error_check_good t_commit [$t commit] 0 + } + error_check_good db_sync [$db sync] 0 + + set size3 [file size $filename] + set free3 [stat_field $db stat "Pages on freelist"] + set leaf3 [stat_field $db stat "Leaf pages"] + set internal3 [stat_field $db stat "Internal pages"] + + puts "\tTest$tnum.g: Remove more entries, this time by cursor." + set count 0 + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + set dbc [eval {$db cursor} $txn] + + # Delete all items except those evenly divisible by + # $maxdelete -- so the db is nearly empty. + for { set dbt [$dbc get -first] } { [llength $dbt] > 0 }\ + { set dbt [$dbc get -next] ; incr count } { + if { [expr $count % $maxdelete] != 0 } { + error_check_good dbc_del [$dbc del] 0 + } + } + + error_check_good cursor_close [$dbc close] 0 + if { $txnenv == 1 } { + error_check_good t_commit [$t commit] 0 + } + error_check_good db_sync [$db sync] 0 + + puts "\tTest$tnum.h: Save contents." + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + dump_file $db $txn $t1 + if { $txnenv == 1 } { + error_check_good t_commit [$t commit] 0 + } + + puts "\tTest$tnum.i: Compact and verify database again." + for {set commit 0} {$commit <= $txnenv} {incr commit} { + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + set ret [eval $db compact $txn -freespace] + if { $txnenv == 1 } { + if { $commit == 0 } { + puts "\tTest$tnum.d: Aborting." + error_check_good txn_abort [$t abort] 0 + } else { + puts "\tTest$tnum.d: Committing." + error_check_good txn_commit [$t commit] 0 + } + } + error_check_good db_sync [$db sync] 0 + error_check_good verify_dir \ + [verify_dir $testdir "" 0 0 $nodump ] 0 + } + + set size4 [file size $filename] + set free4 [stat_field $db stat "Pages on freelist"] + set leaf4 [stat_field $db stat "Leaf pages"] + set internal4 [stat_field $db stat "Internal pages"] + + # The sum of internal pages, leaf pages, and pages freed + # should decrease on compaction, indicating that pages + # have been freed to the file system. + set sum3 [expr $free3 + $leaf3 + $internal3] + set sum4 [expr $free4 + $leaf4 + $internal4] + error_check_good pages_freed [expr $sum3 > $sum4] 1 + + # File should be smaller as well. + #### We should look at the partitioned files ##### + if { [is_partitioned $args] == 0 } { + error_check_good \ + file_size [expr [expr $size3 * $reduction] > $size4] 1 + } + + puts "\tTest$tnum.j: Contents are the same after compaction." + if { $txnenv == 1 } { + set t [$env txn] + error_check_good txn [is_valid_txn $t $env] TRUE + set txn "-txn $t" + } + dump_file $db $txn $t2 + if { $txnenv == 1 } { + error_check_good t_commit [$t commit] 0 + } + error_check_good filecmp [filecmp $t1 $t2] 0 + + error_check_good db_close [$db close] 0 + close $did + } +} |