Index: third_party/sqlite/sqlite-src-3080704/tool/spaceanal.tcl |
diff --git a/third_party/sqlite/sqlite-src-3080704/tool/spaceanal.tcl b/third_party/sqlite/sqlite-src-3080704/tool/spaceanal.tcl |
deleted file mode 100644 |
index a227b8524359c00b9e75f052c092d5e65ab0575f..0000000000000000000000000000000000000000 |
--- a/third_party/sqlite/sqlite-src-3080704/tool/spaceanal.tcl |
+++ /dev/null |
@@ -1,781 +0,0 @@ |
-# Run this TCL script using "testfixture" in order get a report that shows |
-# how much disk space is used by a particular data to actually store data |
-# versus how much space is unused. |
-# |
- |
-if {[catch { |
-# Get the name of the database to analyze |
-# |
-proc usage {} { |
- set argv0 [file rootname [file tail [info nameofexecutable]]] |
- puts stderr "Usage: $argv0 database-name" |
- exit 1 |
-} |
-set file_to_analyze {} |
-set flags(-pageinfo) 0 |
-set flags(-stats) 0 |
-append argv {} |
-foreach arg $argv { |
- if {[regexp {^-+pageinfo$} $arg]} { |
- set flags(-pageinfo) 1 |
- } elseif {[regexp {^-+stats$} $arg]} { |
- set flags(-stats) 1 |
- } elseif {[regexp {^-} $arg]} { |
- puts stderr "Unknown option: $arg" |
- usage |
- } elseif {$file_to_analyze!=""} { |
- usage |
- } else { |
- set file_to_analyze $arg |
- } |
-} |
-if {$file_to_analyze==""} usage |
-set root_filename $file_to_analyze |
-regexp {^file:(//)?([^?]*)} $file_to_analyze all x1 root_filename |
-if {![file exists $root_filename]} { |
- puts stderr "No such file: $root_filename" |
- exit 1 |
-} |
-if {![file readable $root_filename]} { |
- puts stderr "File is not readable: $root_filename" |
- exit 1 |
-} |
-set true_file_size [file size $root_filename] |
-if {$true_file_size<512} { |
- puts stderr "Empty or malformed database: $root_filename" |
- exit 1 |
-} |
- |
-# Compute the total file size assuming test_multiplexor is being used. |
-# Assume that SQLITE_ENABLE_8_3_NAMES might be enabled |
-# |
-set extension [file extension $root_filename] |
-set pattern $root_filename |
-append pattern {[0-3][0-9][0-9]} |
-foreach f [glob -nocomplain $pattern] { |
- incr true_file_size [file size $f] |
- set extension {} |
-} |
-if {[string length $extension]>=2 && [string length $extension]<=4} { |
- set pattern [file rootname $root_filename] |
- append pattern {.[0-3][0-9][0-9]} |
- foreach f [glob -nocomplain $pattern] { |
- incr true_file_size [file size $f] |
- } |
-} |
- |
-# Open the database |
-# |
-if {[catch {sqlite3 db $file_to_analyze -uri 1} msg]} { |
- puts stderr "error trying to open $file_to_analyze: $msg" |
- exit 1 |
-} |
-register_dbstat_vtab db |
- |
-db eval {SELECT count(*) FROM sqlite_master} |
-set pageSize [expr {wide([db one {PRAGMA page_size}])}] |
- |
-if {$flags(-pageinfo)} { |
- db eval {CREATE VIRTUAL TABLE temp.stat USING dbstat} |
- db eval {SELECT name, path, pageno FROM temp.stat ORDER BY pageno} { |
- puts "$pageno $name $path" |
- } |
- exit 0 |
-} |
-if {$flags(-stats)} { |
- db eval {CREATE VIRTUAL TABLE temp.stat USING dbstat} |
- puts "BEGIN;" |
- puts "CREATE TABLE stats(" |
- puts " name STRING, /* Name of table or index */" |
- puts " path INTEGER, /* Path to page from root */" |
- puts " pageno INTEGER, /* Page number */" |
- puts " pagetype STRING, /* 'internal', 'leaf' or 'overflow' */" |
- puts " ncell INTEGER, /* Cells on page (0 for overflow) */" |
- puts " payload INTEGER, /* Bytes of payload on this page */" |
- puts " unused INTEGER, /* Bytes of unused space on this page */" |
- puts " mx_payload INTEGER, /* Largest payload size of all cells */" |
- puts " pgoffset INTEGER, /* Offset of page in file */" |
- puts " pgsize INTEGER /* Size of the page */" |
- puts ");" |
- db eval {SELECT quote(name) || ',' || |
- quote(path) || ',' || |
- quote(pageno) || ',' || |
- quote(pagetype) || ',' || |
- quote(ncell) || ',' || |
- quote(payload) || ',' || |
- quote(unused) || ',' || |
- quote(mx_payload) || ',' || |
- quote(pgoffset) || ',' || |
- quote(pgsize) AS x FROM stat} { |
- puts "INSERT INTO stats VALUES($x);" |
- } |
- puts "COMMIT;" |
- exit 0 |
-} |
- |
-# In-memory database for collecting statistics. This script loops through |
-# the tables and indices in the database being analyzed, adding a row for each |
-# to an in-memory database (for which the schema is shown below). It then |
-# queries the in-memory db to produce the space-analysis report. |
-# |
-sqlite3 mem :memory: |
-set tabledef {CREATE TABLE space_used( |
- name clob, -- Name of a table or index in the database file |
- tblname clob, -- Name of associated table |
- is_index boolean, -- TRUE if it is an index, false for a table |
- nentry int, -- Number of entries in the BTree |
- leaf_entries int, -- Number of leaf entries |
- payload int, -- Total amount of data stored in this table or index |
- ovfl_payload int, -- Total amount of data stored on overflow pages |
- ovfl_cnt int, -- Number of entries that use overflow |
- mx_payload int, -- Maximum payload size |
- int_pages int, -- Number of interior pages used |
- leaf_pages int, -- Number of leaf pages used |
- ovfl_pages int, -- Number of overflow pages used |
- int_unused int, -- Number of unused bytes on interior pages |
- leaf_unused int, -- Number of unused bytes on primary pages |
- ovfl_unused int, -- Number of unused bytes on overflow pages |
- gap_cnt int, -- Number of gaps in the page layout |
- compressed_size int -- Total bytes stored on disk |
-);} |
-mem eval $tabledef |
- |
-# Create a temporary "dbstat" virtual table. |
-# |
-db eval {CREATE VIRTUAL TABLE temp.stat USING dbstat} |
-db eval {CREATE TEMP TABLE dbstat AS SELECT * FROM temp.stat |
- ORDER BY name, path} |
-db eval {DROP TABLE temp.stat} |
- |
-proc isleaf {pagetype is_index} { |
- return [expr {$pagetype == "leaf" || ($pagetype == "internal" && $is_index)}] |
-} |
-proc isoverflow {pagetype is_index} { |
- return [expr {$pagetype == "overflow"}] |
-} |
-proc isinternal {pagetype is_index} { |
- return [expr {$pagetype == "internal" && $is_index==0}] |
-} |
- |
-db func isleaf isleaf |
-db func isinternal isinternal |
-db func isoverflow isoverflow |
- |
-set isCompressed 0 |
-set compressOverhead 0 |
-set sql { SELECT name, tbl_name FROM sqlite_master WHERE rootpage>0 } |
-foreach {name tblname} [concat sqlite_master sqlite_master [db eval $sql]] { |
- |
- set is_index [expr {$name!=$tblname}] |
- db eval { |
- SELECT |
- sum(ncell) AS nentry, |
- sum(isleaf(pagetype, $is_index) * ncell) AS leaf_entries, |
- sum(payload) AS payload, |
- sum(isoverflow(pagetype, $is_index) * payload) AS ovfl_payload, |
- sum(path LIKE '%+000000') AS ovfl_cnt, |
- max(mx_payload) AS mx_payload, |
- sum(isinternal(pagetype, $is_index)) AS int_pages, |
- sum(isleaf(pagetype, $is_index)) AS leaf_pages, |
- sum(isoverflow(pagetype, $is_index)) AS ovfl_pages, |
- sum(isinternal(pagetype, $is_index) * unused) AS int_unused, |
- sum(isleaf(pagetype, $is_index) * unused) AS leaf_unused, |
- sum(isoverflow(pagetype, $is_index) * unused) AS ovfl_unused, |
- sum(pgsize) AS compressed_size |
- FROM temp.dbstat WHERE name = $name |
- } break |
- |
- set total_pages [expr {$leaf_pages+$int_pages+$ovfl_pages}] |
- set storage [expr {$total_pages*$pageSize}] |
- if {!$isCompressed && $storage>$compressed_size} { |
- set isCompressed 1 |
- set compressOverhead 14 |
- } |
- |
- # Column 'gap_cnt' is set to the number of non-contiguous entries in the |
- # list of pages visited if the b-tree structure is traversed in a top-down |
- # fashion (each node visited before its child-tree is passed). Any overflow |
- # chains present are traversed from start to finish before any child-tree |
- # is. |
- # |
- set gap_cnt 0 |
- set prev 0 |
- db eval { |
- SELECT pageno, pagetype FROM temp.dbstat |
- WHERE name=$name |
- ORDER BY pageno |
- } { |
- if {$prev>0 && $pagetype=="leaf" && $pageno!=$prev+1} { |
- incr gap_cnt |
- } |
- set prev $pageno |
- } |
- mem eval { |
- INSERT INTO space_used VALUES( |
- $name, |
- $tblname, |
- $is_index, |
- $nentry, |
- $leaf_entries, |
- $payload, |
- $ovfl_payload, |
- $ovfl_cnt, |
- $mx_payload, |
- $int_pages, |
- $leaf_pages, |
- $ovfl_pages, |
- $int_unused, |
- $leaf_unused, |
- $ovfl_unused, |
- $gap_cnt, |
- $compressed_size |
- ); |
- } |
-} |
- |
-proc integerify {real} { |
- if {[string is double -strict $real]} { |
- return [expr {wide($real)}] |
- } else { |
- return 0 |
- } |
-} |
-mem function int integerify |
- |
-# Quote a string for use in an SQL query. Examples: |
-# |
-# [quote {hello world}] == {'hello world'} |
-# [quote {hello world's}] == {'hello world''s'} |
-# |
-proc quote {txt} { |
- return [string map {' ''} $txt] |
-} |
- |
-# Output a title line |
-# |
-proc titleline {title} { |
- if {$title==""} { |
- puts [string repeat * 79] |
- } else { |
- set len [string length $title] |
- set stars [string repeat * [expr 79-$len-5]] |
- puts "*** $title $stars" |
- } |
-} |
- |
-# Generate a single line of output in the statistics section of the |
-# report. |
-# |
-proc statline {title value {extra {}}} { |
- set len [string length $title] |
- set dots [string repeat . [expr 50-$len]] |
- set len [string length $value] |
- set sp2 [string range { } $len end] |
- if {$extra ne ""} { |
- set extra " $extra" |
- } |
- puts "$title$dots $value$sp2$extra" |
-} |
- |
-# Generate a formatted percentage value for $num/$denom |
-# |
-proc percent {num denom {of {}}} { |
- if {$denom==0.0} {return ""} |
- set v [expr {$num*100.0/$denom}] |
- set of {} |
- if {$v==100.0 || $v<0.001 || ($v>1.0 && $v<99.0)} { |
- return [format {%5.1f%% %s} $v $of] |
- } elseif {$v<0.1 || $v>99.9} { |
- return [format {%7.3f%% %s} $v $of] |
- } else { |
- return [format {%6.2f%% %s} $v $of] |
- } |
-} |
- |
-proc divide {num denom} { |
- if {$denom==0} {return 0.0} |
- return [format %.2f [expr double($num)/double($denom)]] |
-} |
- |
-# Generate a subreport that covers some subset of the database. |
-# the $where clause determines which subset to analyze. |
-# |
-proc subreport {title where showFrag} { |
- global pageSize file_pgcnt compressOverhead |
- |
- # Query the in-memory database for the sum of various statistics |
- # for the subset of tables/indices identified by the WHERE clause in |
- # $where. Note that even if the WHERE clause matches no rows, the |
- # following query returns exactly one row (because it is an aggregate). |
- # |
- # The results of the query are stored directly by SQLite into local |
- # variables (i.e. $nentry, $nleaf etc.). |
- # |
- mem eval " |
- SELECT |
- int(sum(nentry)) AS nentry, |
- int(sum(leaf_entries)) AS nleaf, |
- int(sum(payload)) AS payload, |
- int(sum(ovfl_payload)) AS ovfl_payload, |
- max(mx_payload) AS mx_payload, |
- int(sum(ovfl_cnt)) as ovfl_cnt, |
- int(sum(leaf_pages)) AS leaf_pages, |
- int(sum(int_pages)) AS int_pages, |
- int(sum(ovfl_pages)) AS ovfl_pages, |
- int(sum(leaf_unused)) AS leaf_unused, |
- int(sum(int_unused)) AS int_unused, |
- int(sum(ovfl_unused)) AS ovfl_unused, |
- int(sum(gap_cnt)) AS gap_cnt, |
- int(sum(compressed_size)) AS compressed_size |
- FROM space_used WHERE $where" {} {} |
- |
- # Output the sub-report title, nicely decorated with * characters. |
- # |
- puts "" |
- titleline $title |
- puts "" |
- |
- # Calculate statistics and store the results in TCL variables, as follows: |
- # |
- # total_pages: Database pages consumed. |
- # total_pages_percent: Pages consumed as a percentage of the file. |
- # storage: Bytes consumed. |
- # payload_percent: Payload bytes used as a percentage of $storage. |
- # total_unused: Unused bytes on pages. |
- # avg_payload: Average payload per btree entry. |
- # avg_fanout: Average fanout for internal pages. |
- # avg_unused: Average unused bytes per btree entry. |
- # ovfl_cnt_percent: Percentage of btree entries that use overflow pages. |
- # |
- set total_pages [expr {$leaf_pages+$int_pages+$ovfl_pages}] |
- set total_pages_percent [percent $total_pages $file_pgcnt] |
- set storage [expr {$total_pages*$pageSize}] |
- set payload_percent [percent $payload $storage {of storage consumed}] |
- set total_unused [expr {$ovfl_unused+$int_unused+$leaf_unused}] |
- set avg_payload [divide $payload $nleaf] |
- set avg_unused [divide $total_unused $nleaf] |
- if {$int_pages>0} { |
- # TODO: Is this formula correct? |
- set nTab [mem eval " |
- SELECT count(*) FROM ( |
- SELECT DISTINCT tblname FROM space_used WHERE $where AND is_index=0 |
- ) |
- "] |
- set avg_fanout [mem eval " |
- SELECT (sum(leaf_pages+int_pages)-$nTab)/sum(int_pages) FROM space_used |
- WHERE $where AND is_index = 0 |
- "] |
- set avg_fanout [format %.2f $avg_fanout] |
- } |
- set ovfl_cnt_percent [percent $ovfl_cnt $nleaf {of all entries}] |
- |
- # Print out the sub-report statistics. |
- # |
- statline {Percentage of total database} $total_pages_percent |
- statline {Number of entries} $nleaf |
- statline {Bytes of storage consumed} $storage |
- if {$compressed_size!=$storage} { |
- set compressed_size [expr {$compressed_size+$compressOverhead*$total_pages}] |
- set pct [expr {$compressed_size*100.0/$storage}] |
- set pct [format {%5.1f%%} $pct] |
- statline {Bytes used after compression} $compressed_size $pct |
- } |
- statline {Bytes of payload} $payload $payload_percent |
- statline {Average payload per entry} $avg_payload |
- statline {Average unused bytes per entry} $avg_unused |
- if {[info exists avg_fanout]} { |
- statline {Average fanout} $avg_fanout |
- } |
- if {$showFrag && $total_pages>1} { |
- set fragmentation [percent $gap_cnt [expr {$total_pages-1}]] |
- statline {Non-sequential pages} $gap_cnt $fragmentation |
- } |
- statline {Maximum payload per entry} $mx_payload |
- statline {Entries that use overflow} $ovfl_cnt $ovfl_cnt_percent |
- if {$int_pages>0} { |
- statline {Index pages used} $int_pages |
- } |
- statline {Primary pages used} $leaf_pages |
- statline {Overflow pages used} $ovfl_pages |
- statline {Total pages used} $total_pages |
- if {$int_unused>0} { |
- set int_unused_percent [ |
- percent $int_unused [expr {$int_pages*$pageSize}] {of index space}] |
- statline "Unused bytes on index pages" $int_unused $int_unused_percent |
- } |
- statline "Unused bytes on primary pages" $leaf_unused [ |
- percent $leaf_unused [expr {$leaf_pages*$pageSize}] {of primary space}] |
- statline "Unused bytes on overflow pages" $ovfl_unused [ |
- percent $ovfl_unused [expr {$ovfl_pages*$pageSize}] {of overflow space}] |
- statline "Unused bytes on all pages" $total_unused [ |
- percent $total_unused $storage {of all space}] |
- return 1 |
-} |
- |
-# Calculate the overhead in pages caused by auto-vacuum. |
-# |
-# This procedure calculates and returns the number of pages used by the |
-# auto-vacuum 'pointer-map'. If the database does not support auto-vacuum, |
-# then 0 is returned. The two arguments are the size of the database file in |
-# pages and the page size used by the database (in bytes). |
-proc autovacuum_overhead {filePages pageSize} { |
- |
- # Set $autovacuum to non-zero for databases that support auto-vacuum. |
- set autovacuum [db one {PRAGMA auto_vacuum}] |
- |
- # If the database is not an auto-vacuum database or the file consists |
- # of one page only then there is no overhead for auto-vacuum. Return zero. |
- if {0==$autovacuum || $filePages==1} { |
- return 0 |
- } |
- |
- # The number of entries on each pointer map page. The layout of the |
- # database file is one pointer-map page, followed by $ptrsPerPage other |
- # pages, followed by a pointer-map page etc. The first pointer-map page |
- # is the second page of the file overall. |
- set ptrsPerPage [expr double($pageSize/5)] |
- |
- # Return the number of pointer map pages in the database. |
- return [expr wide(ceil( ($filePages-1.0)/($ptrsPerPage+1.0) ))] |
-} |
- |
- |
-# Calculate the summary statistics for the database and store the results |
-# in TCL variables. They are output below. Variables are as follows: |
-# |
-# pageSize: Size of each page in bytes. |
-# file_bytes: File size in bytes. |
-# file_pgcnt: Number of pages in the file. |
-# file_pgcnt2: Number of pages in the file (calculated). |
-# av_pgcnt: Pages consumed by the auto-vacuum pointer-map. |
-# av_percent: Percentage of the file consumed by auto-vacuum pointer-map. |
-# inuse_pgcnt: Data pages in the file. |
-# inuse_percent: Percentage of pages used to store data. |
-# free_pgcnt: Free pages calculated as (<total pages> - <in-use pages>) |
-# free_pgcnt2: Free pages in the file according to the file header. |
-# free_percent: Percentage of file consumed by free pages (calculated). |
-# free_percent2: Percentage of file consumed by free pages (header). |
-# ntable: Number of tables in the db. |
-# nindex: Number of indices in the db. |
-# nautoindex: Number of indices created automatically. |
-# nmanindex: Number of indices created manually. |
-# user_payload: Number of bytes of payload in table btrees |
-# (not including sqlite_master) |
-# user_percent: $user_payload as a percentage of total file size. |
- |
-### The following, setting $file_bytes based on the actual size of the file |
-### on disk, causes this tool to choke on zipvfs databases. So set it based |
-### on the return of [PRAGMA page_count] instead. |
-if 0 { |
- set file_bytes [file size $file_to_analyze] |
- set file_pgcnt [expr {$file_bytes/$pageSize}] |
-} |
-set file_pgcnt [db one {PRAGMA page_count}] |
-set file_bytes [expr {$file_pgcnt * $pageSize}] |
- |
-set av_pgcnt [autovacuum_overhead $file_pgcnt $pageSize] |
-set av_percent [percent $av_pgcnt $file_pgcnt] |
- |
-set sql {SELECT sum(leaf_pages+int_pages+ovfl_pages) FROM space_used} |
-set inuse_pgcnt [expr wide([mem eval $sql])] |
-set inuse_percent [percent $inuse_pgcnt $file_pgcnt] |
- |
-set free_pgcnt [expr {$file_pgcnt-$inuse_pgcnt-$av_pgcnt}] |
-set free_percent [percent $free_pgcnt $file_pgcnt] |
-set free_pgcnt2 [db one {PRAGMA freelist_count}] |
-set free_percent2 [percent $free_pgcnt2 $file_pgcnt] |
- |
-set file_pgcnt2 [expr {$inuse_pgcnt+$free_pgcnt2+$av_pgcnt}] |
- |
-set ntable [db eval {SELECT count(*)+1 FROM sqlite_master WHERE type='table'}] |
-set nindex [db eval {SELECT count(*) FROM sqlite_master WHERE type='index'}] |
-set sql {SELECT count(*) FROM sqlite_master WHERE name LIKE 'sqlite_autoindex%'} |
-set nautoindex [db eval $sql] |
-set nmanindex [expr {$nindex-$nautoindex}] |
- |
-# set total_payload [mem eval "SELECT sum(payload) FROM space_used"] |
-set user_payload [mem one {SELECT int(sum(payload)) FROM space_used |
- WHERE NOT is_index AND name NOT LIKE 'sqlite_master'}] |
-set user_percent [percent $user_payload $file_bytes] |
- |
-# Output the summary statistics calculated above. |
-# |
-puts "/** Disk-Space Utilization Report For $root_filename" |
-puts "" |
-statline {Page size in bytes} $pageSize |
-statline {Pages in the whole file (measured)} $file_pgcnt |
-statline {Pages in the whole file (calculated)} $file_pgcnt2 |
-statline {Pages that store data} $inuse_pgcnt $inuse_percent |
-statline {Pages on the freelist (per header)} $free_pgcnt2 $free_percent2 |
-statline {Pages on the freelist (calculated)} $free_pgcnt $free_percent |
-statline {Pages of auto-vacuum overhead} $av_pgcnt $av_percent |
-statline {Number of tables in the database} $ntable |
-statline {Number of indices} $nindex |
-statline {Number of defined indices} $nmanindex |
-statline {Number of implied indices} $nautoindex |
-if {$isCompressed} { |
- statline {Size of uncompressed content in bytes} $file_bytes |
- set efficiency [percent $true_file_size $file_bytes] |
- statline {Size of compressed file on disk} $true_file_size $efficiency |
-} else { |
- statline {Size of the file in bytes} $file_bytes |
-} |
-statline {Bytes of user payload stored} $user_payload $user_percent |
- |
-# Output table rankings |
-# |
-puts "" |
-titleline "Page counts for all tables with their indices" |
-puts "" |
-mem eval {SELECT tblname, count(*) AS cnt, |
- int(sum(int_pages+leaf_pages+ovfl_pages)) AS size |
- FROM space_used GROUP BY tblname ORDER BY size+0 DESC, tblname} {} { |
- statline [string toupper $tblname] $size [percent $size $file_pgcnt] |
-} |
-puts "" |
-titleline "Page counts for all tables and indices separately" |
-puts "" |
-mem eval { |
- SELECT |
- upper(name) AS nm, |
- int(int_pages+leaf_pages+ovfl_pages) AS size |
- FROM space_used |
- ORDER BY size+0 DESC, name} {} { |
- statline $nm $size [percent $size $file_pgcnt] |
-} |
-if {$isCompressed} { |
- puts "" |
- titleline "Bytes of disk space used after compression" |
- puts "" |
- set csum 0 |
- mem eval {SELECT tblname, |
- int(sum(compressed_size)) + |
- $compressOverhead*sum(int_pages+leaf_pages+ovfl_pages) |
- AS csize |
- FROM space_used GROUP BY tblname ORDER BY csize+0 DESC, tblname} {} { |
- incr csum $csize |
- statline [string toupper $tblname] $csize [percent $csize $true_file_size] |
- } |
- set overhead [expr {$true_file_size - $csum}] |
- if {$overhead>0} { |
- statline {Header and free space} $overhead [percent $overhead $true_file_size] |
- } |
-} |
- |
-# Output subreports |
-# |
-if {$nindex>0} { |
- subreport {All tables and indices} 1 0 |
-} |
-subreport {All tables} {NOT is_index} 0 |
-if {$nindex>0} { |
- subreport {All indices} {is_index} 0 |
-} |
-foreach tbl [mem eval {SELECT DISTINCT tblname name FROM space_used |
- ORDER BY name}] { |
- set qn [quote $tbl] |
- set name [string toupper $tbl] |
- set n [mem eval {SELECT count(*) FROM space_used WHERE tblname=$tbl}] |
- if {$n>1} { |
- set idxlist [mem eval "SELECT name FROM space_used |
- WHERE tblname='$qn' AND is_index |
- ORDER BY 1"] |
- subreport "Table $name and all its indices" "tblname='$qn'" 0 |
- subreport "Table $name w/o any indices" "name='$qn'" 1 |
- if {[llength $idxlist]>1} { |
- subreport "Indices of table $name" "tblname='$qn' AND is_index" 0 |
- } |
- foreach idx $idxlist { |
- set qidx [quote $idx] |
- subreport "Index [string toupper $idx] of table $name" "name='$qidx'" 1 |
- } |
- } else { |
- subreport "Table $name" "name='$qn'" 1 |
- } |
-} |
- |
-# Output instructions on what the numbers above mean. |
-# |
-puts "" |
-titleline Definitions |
-puts { |
-Page size in bytes |
- |
- The number of bytes in a single page of the database file. |
- Usually 1024. |
- |
-Number of pages in the whole file |
-} |
-puts " The number of $pageSize-byte pages that go into forming the complete |
- database" |
-puts { |
-Pages that store data |
- |
- The number of pages that store data, either as primary B*Tree pages or |
- as overflow pages. The number at the right is the data pages divided by |
- the total number of pages in the file. |
- |
-Pages on the freelist |
- |
- The number of pages that are not currently in use but are reserved for |
- future use. The percentage at the right is the number of freelist pages |
- divided by the total number of pages in the file. |
- |
-Pages of auto-vacuum overhead |
- |
- The number of pages that store data used by the database to facilitate |
- auto-vacuum. This is zero for databases that do not support auto-vacuum. |
- |
-Number of tables in the database |
- |
- The number of tables in the database, including the SQLITE_MASTER table |
- used to store schema information. |
- |
-Number of indices |
- |
- The total number of indices in the database. |
- |
-Number of defined indices |
- |
- The number of indices created using an explicit CREATE INDEX statement. |
- |
-Number of implied indices |
- |
- The number of indices used to implement PRIMARY KEY or UNIQUE constraints |
- on tables. |
- |
-Size of the file in bytes |
- |
- The total amount of disk space used by the entire database files. |
- |
-Bytes of user payload stored |
- |
- The total number of bytes of user payload stored in the database. The |
- schema information in the SQLITE_MASTER table is not counted when |
- computing this number. The percentage at the right shows the payload |
- divided by the total file size. |
- |
-Percentage of total database |
- |
- The amount of the complete database file that is devoted to storing |
- information described by this category. |
- |
-Number of entries |
- |
- The total number of B-Tree key/value pairs stored under this category. |
- |
-Bytes of storage consumed |
- |
- The total amount of disk space required to store all B-Tree entries |
- under this category. The is the total number of pages used times |
- the pages size. |
- |
-Bytes of payload |
- |
- The amount of payload stored under this category. Payload is the data |
- part of table entries and the key part of index entries. The percentage |
- at the right is the bytes of payload divided by the bytes of storage |
- consumed. |
- |
-Average payload per entry |
- |
- The average amount of payload on each entry. This is just the bytes of |
- payload divided by the number of entries. |
- |
-Average unused bytes per entry |
- |
- The average amount of free space remaining on all pages under this |
- category on a per-entry basis. This is the number of unused bytes on |
- all pages divided by the number of entries. |
- |
-Non-sequential pages |
- |
- The number of pages in the table or index that are out of sequence. |
- Many filesystems are optimized for sequential file access so a small |
- number of non-sequential pages might result in faster queries, |
- especially for larger database files that do not fit in the disk cache. |
- Note that after running VACUUM, the root page of each table or index is |
- at the beginning of the database file and all other pages are in a |
- separate part of the database file, resulting in a single non- |
- sequential page. |
- |
-Maximum payload per entry |
- |
- The largest payload size of any entry. |
- |
-Entries that use overflow |
- |
- The number of entries that user one or more overflow pages. |
- |
-Total pages used |
- |
- This is the number of pages used to hold all information in the current |
- category. This is the sum of index, primary, and overflow pages. |
- |
-Index pages used |
- |
- This is the number of pages in a table B-tree that hold only key (rowid) |
- information and no data. |
- |
-Primary pages used |
- |
- This is the number of B-tree pages that hold both key and data. |
- |
-Overflow pages used |
- |
- The total number of overflow pages used for this category. |
- |
-Unused bytes on index pages |
- |
- The total number of bytes of unused space on all index pages. The |
- percentage at the right is the number of unused bytes divided by the |
- total number of bytes on index pages. |
- |
-Unused bytes on primary pages |
- |
- The total number of bytes of unused space on all primary pages. The |
- percentage at the right is the number of unused bytes divided by the |
- total number of bytes on primary pages. |
- |
-Unused bytes on overflow pages |
- |
- The total number of bytes of unused space on all overflow pages. The |
- percentage at the right is the number of unused bytes divided by the |
- total number of bytes on overflow pages. |
- |
-Unused bytes on all pages |
- |
- The total number of bytes of unused space on all primary and overflow |
- pages. The percentage at the right is the number of unused bytes |
- divided by the total number of bytes. |
-} |
- |
-# Output a dump of the in-memory database. This can be used for more |
-# complex offline analysis. |
-# |
-titleline {} |
-puts "The entire text of this report can be sourced into any SQL database" |
-puts "engine for further analysis. All of the text above is an SQL comment." |
-puts "The data used to generate this report follows:" |
-puts "*/" |
-puts "BEGIN;" |
-puts $tabledef |
-unset -nocomplain x |
-mem eval {SELECT * FROM space_used} x { |
- puts -nonewline "INSERT INTO space_used VALUES" |
- set sep ( |
- foreach col $x(*) { |
- set v $x($col) |
- if {$v=="" || ![string is double $v]} {set v '[quote $v]'} |
- puts -nonewline $sep$v |
- set sep , |
- } |
- puts ");" |
-} |
-puts "COMMIT;" |
- |
-} err]} { |
- puts "ERROR: $err" |
- puts $errorInfo |
- exit 1 |
-} |