250 likes | 417 Views
Compression Suresh Gandhi. Compression – Methods around. The data grows in relational databases (new business requirements, managing audit data etc), the cost associated with the disk systems to store that data and the resources required to manage them becomes vital. .
E N D
Compression – Methods around • The data grows in relational databases (new business requirements, managing audit data etc), the cost associated with the disk systems to store that data and the resources required to manage them becomes vital.
Data is compressed when it is written to block – Not true any more! • Decompressed when it should read from the block • Basically a arithmetic algorithm’s to shorten or simplify your data • Huffman’s • CLEAMCOM • Swarm • Requires • Less data storage to hold the compress the data • More cpu to compress and decompress data • Primary Functions on • By removing redundant data at page/block level What is compression for RDBMS
Oracle 8i: Index compression introduced (Key Factoring) • Oracle 9iR2: Table compression introduced (batch/Basic) only) • Oracle Database 10g • RMAN: Unused block compression • Data Pump (export/import): Metadata compression • Lob compression – utl_compress • Oracle Database 11g • Structured/relational data segment compression (DML/Batch) • (Heap/IOT/Index/External/Materialized views) • Unstructured data compression (Lobs) (SecureFiles) • RMAN: Binary Compression for backup data • Data Pump: Expanded export file compression capabilities • Data Guard: Redo Transport Services – Network compression • Hybrid Column Compression – (Exadata only) • Oracle Database 12c? • Archive Log Compression (Surprise ????) Oracle’s journey thus so far…
Symbol Table – Dictionary Table (IBM/SQL etc) • Static, Table wide dictionary approch • Due to global in nature can yield higher compression factors • Symbol Table (Oracle’s) – Local (Row level compression) • Compressed blocks contain a structure called a symbol table that maintains compression metadata. The symbol table is stored as another table in the block. Each column in a row in a block references back to an entry in the symbol table in the block • Duplicate values are eliminated by first adding a single copy of the duplicate value to the symbol table when a block is compressed. Each duplicate value is then replaced by a short reference to the appropriate entry in the symbol table. • This makes compressed data self-contained within the database block as the metadata used to translate compressed data into its original state is contained within the block. • Multi-Column Compression • If a sequence of columns are having common data, then instead of saving each individual column as separate entry, the combination of the columns will be saved in the symbol table as a whole. This approach is particularly useful in case of materialized views using grouping sets and cubes (where the data is ordered). • Cross-Column Compression As explain above, the symbol table holds the common data. This is true for the column values for different columns. So, if column values from the different columns are different, they all will share the same symbol table entry. • Compression factor : Thresholds or factor that yields the benefit then only starts Compression - Terminologies
Symbols’ table is tab0 Example Symbols’ Table – How its look like in our block block_row_dump: tab 0, row 0, @0x1f66 tl: 13 fb: --H-FL-- lb: 0x0 cc: 2 col 0: [ 3] 53 59 53 col 1: [ 5] 49 4e 44 45 58 bindmp: 00 d4 02 cb 53 59 53 cd 49 4e 44 45 58 tab 0, row 1, @0x1f73 tl: 13 fb: --H-FL-- lb: 0x0 cc: 2 col 0: [ 3] 53 59 53 col 1: [ 5] 54 41 42 4c 45 bindmp: 00 b6 02 cb 53 59 53 cd 54 41 42 4c 45 tab 1, row 0, @0x1f5b tl: 11 fb: --H-FL-- lb: 0x0 cc: 3 col 0: [ 3] 53 59 53 col 1: [ 5] 54 41 42 4c 45 col 2: [ 5] 49 43 4f 4c 24 bindmp: 2c 00 02 02 01 cd 49 43 4f 4c 24 tab 1, row 1, @0x1f4e tl: 13 fb: --H-FL-- lb: 0x0 cc: 3 col 0: [ 3] 53 59 53 col 1: [ 5] 49 4e 44 45 58 col 2: [ 7] 49 5f 55 53 45 52 31 bindmp: 2c 00 02 02 00 cf 49 5f 55 53 45 52 31 tab 1, row 2, @0x1f44 tl: 10 fb: --H-FL-- lb: 0x0 cc: 3 col 0: [ 3] 53 59 53 col 1: [ 5] 54 41 42 4c 45 col 2: [ 4] 43 4f 4e 24 bindmp: 2c 00 02 02 01 cc 43 4f 4e 24 SYS INDEX Actual data: tab1 Row1: SYS ICOL$ TABLE Row2: SYS I_USER1 INDEX Row3: SYS CON$ TABLE … Uncompressed data Actual binary dump of row 7
Basic compression (9i) • Insert /%APPEND%/ • Direct Load operations kicks in • OLTP Table Compression Syntax: • CREATE TABLE emp ( • emp_id NUMBER • , first_name VARCHAR2(128) • , last_name VARCHAR2(128) • ) COMPRESSFOR ALL OPERATIONS; • Direct Load Compression Syntax (default): • CREATE TABLE emp ( • emp_id NUMBER • , first_name VARCHAR2(128) • , last_name VARCHAR2(128) • ) COMPRESS[FOR DIRECT_LOAD OPERATIONS]; • Exadata? • Create table comp_table • Compress for QUERY [LOW|HIGH] • Compress for ARCHIVE [LOW|HIGH] • as select * from MY_UNCOMP_TABLE; Oracle’s Syntax for OLTP Table/Row Compression
Compression is applied at block level • Blocks will only be compressed if • data is sufficiently large to fill the block • rows have low enough cardinality • Columns will be reordered within each block to achieve optimal compression ratios • A segment may contain • compressed and uncompressed blocks • blocks compressed on different columns Block Level Compression
Empty Block Initially Uncompressed Block Compressed Block Partially Compressed Block Compressed Block Legend Header Data Uncompressed Data Free Space Compressed Data • Allows ‘normal INSERT’ into the table • Compression will not start until the block hits PCTFREE threshold • For an update, the row will be decompressed and compressed Block (Row) Compression – When it starts PCTFREE Reached – Kick Compression
ID FIRST_NAME LAST_NAME 1 John Doe 2 Jane Doe 3 John Smith 4 Jane Doe How its looks like? Initially Uncompressed Block Employee Table Header 1•John•Doe 2•Jane• Doe 3•John•Smith4• Jane • Doe Free Space INSERT INTO EMPLOYEE VALUES (5, ‘Jack’, ‘Smith’); COMMIT;
Compressed Block John=|Doe=|Jane|Smith= ID FIRST_NAME LAST_NAME 1•• 2•• 3•• 4 • • 5•Jack• Free Space 1•John•Doe 2•Jane• Doe 3•John•Smith4• Jane • Doe Free Space 1 John Doe 2 Jane Doe 3 John Smith 4 Jane Doe 5 Jack Smith How its look like Local Symbol Table Employee Table Header
Compressed Block John=|Doe=|Jane=|Smith= 1•• 2•• 3•• 4 • • 5•Jack• Free Space 1•John•Doe 2•Jane• Doe 3•John•Smith4• Jane • Doe Free Space How its looks like? Uncompressed Block Header Header 1•John•Doe 2•Jane• Doe 3•John•Smith4• Jane • Doe 5•Jack •Smith Free Space Local Symbol Table More Data Per Block
Disk Space saving:This is the obvious reason for implementing data compression. Although the space saving comes at a cost of decreased query performance (time required to decompress the data), when used along with other features, this can be minimized. Coexisting compressed and non-compressed data:The same table can have both compressed and non-compressed data. This is possible because the compression is at block level, one block might hold compressed data and the other block might hold normal uncompressed data. This is particularly useful with partitioning where you can compress the old partitions while keep the current partition as non-compressed. Less Buffer Size required:The data from a compressed block is read and cached in its compressed format and it is decompressed only at data access time. As the data is cached in compressed form, more data can be hold into the sameamount of buffer cache. Advantages of Block/Row/Table Level Compression
Not an free – License is required (except Basic 9i method) , see how much you are saving before to use, since price is high • Do not capacity it while comparing to disks – rather look into max IOPS • RAM Demands: Since many blocks now demand and want more time to be in buffer • Must know which is suitable under available options • Performance impacts • Resource availability ’
SecureFiles is a new 11g feature designed to break the performance barrier keeping file data out of databases • Next-generation LOB • Superset of LOB interfaces allows easy migration from LOBs • Transparent deduplication, compression, and encryption • Leverage the security, reliability, and scalability of database • Enables consolidation of file data with associated relational data • Single security model • Single view of data • Single management of data • Scalable to any level using SMP scale-up or grid scale-out Secure Files (Lobs) - Compression
SecureFiles Deduplication Syntax CREATE TABLE images ( image_id NUMBER, image BLOB) LOB(image) STORE AS SECUREFILE (TABLESPACE lob_tbs DEDUPLICATE); • SecureFiles Compression Syntax CREATE TABLE images ( image_id NUMBER, image BLOB) LOB(image)STORE AS SECUREFILE (TABLESPACE lob_tbs COMPRESS); Secure Files Syntax
Secure Hash SecureFilesDeduplication • Deduplication is the mechanism to eliminate redundant physical files • Enables storage of a single physical image for duplicate data • Significantly reduces space consumption • Dramatically improves writes and copy operations • No adverse impact on read operations • May actually improve read performance for cache data • Duplicate detection happens within a table, partition or sub-partition • Specially useful for content management, email applications and data archival applications
Data Pump Compression • Metadata compression available since Oracle Database 10g • Oracle Database 11g extends compression to table data during exports • No need to decompress before import • Single step compression of both data and metadata • Compressed data directly hits disk resulting in reduced disk space requirements • Internal tests reduced dump file size up to 75% • Compression factor comparable to GNU gzip utility • Application transparent • Complete Data Pump functionality available on compressed files COMPRESSION = ALL|METADATA_ONLY|NONE Caveats • Performance cost: ~10% overhead • Compression Ratio: comparable to gzip
Backup Compression • Compression features in the RMAN utility: • Since inception: Null compression – does not backup data blocks that have never been allocated • Since Oracle Database 10g Release 2: Unused block compression – RMAN skips blocks that currently do not contain data • Since Oracle Database 11g: Binary compression (or backup set compression) – applies a compression algorithm to the blocks as they are backed up, if configured/requested • All RMAN compression techniques are totally transparent during restore operations • Fast RMAN Compression:-Compresses the backup set contents before writing them to disk or tapeNo extra decompression steps are required during recovery when you use High performance, industry standard compression algorithm40% faster backup compression versus Oracle Database 10gSuitable for fast, incremental daily backupsReduces network usage
Backup Compression : Examples RMAN> CONFIGURE DEVICE TYPE DISK BACKUP TYPE TO COMPRESSED BACKUPSET; followed by .. RMAN> CONFIGURE COMPRESSION ALGORITHM ‘HIGH’;RMAN> CONFIGURE COMPRESSION ALGORITHM ‘MEDIUM’;RMAN> CONFIGURE COMPRESSION ALGORITHM ‘LOW’;RMAN> CONFIGURE COMPRESSION ALGORITHM ‘BASIC’; Or 11gR1:RMAN> CONFIGURE COMPRESSION ALGORITHM 'BZIP2';11gR2:RMAN> CONFIGURE COMPRESSION ALGORITHM 'BASIC';Or for both, reset to default:RMAN> CONFIGURE COMPRESSION ALGORITHM CLEAR; Watch out for Trade-offs for each type: However the following gives overview of those LOW – corresponds to LZO (11gR2) – smallest compression ratio, fastest MEDIUM – corresponds to ZLIB (11gR1) – good compression ratio, slower than LOW HIGH – corresponds to unmodified BZIP2 (11gR2) – highest compression ratio, slowest BASIC (which is free) – corresponds to BZIP2 (10g style compression) – compression ratio in the range of MEDIUM, but slower
Network Compression Oracle Data Guard Redo Transport Services Fast re-sync of standby database after network outages • Lower bandwidth networks (<100Mbps) • 15-35% less time required to transmit 1 GB of data • Bandwidth consumption reduced up to 35% • High bandwidth networks (>100 Mbps) • Compression will not reduce transmission time • But will reduce bandwidth consumption up to 35% • The COMPRESSION attribute is used to specify whether redo data is transmitted to a redo transport destination in compressed form or uncompressed form when resolving redo data gaps. • The following example shows the COMPRESSION attribute with the LOG_ARCHIVE_DEST_n parameter. • LOG_ARCHIVE_DEST_3='SERVICE=denver SYNC COMPRESSION=ENABLE' LOG_ARCHIVE_DEST_STATE_3=ENABLE
Heard (I have only heard) there is a feature in 10g partially developed and still partial in 11g , that archive’s can be compressed. • CAUTION: DO NOT SET IN BUSINESS SYSTEM IT WILL CRASH THE DB, though this works in 10g,11g, do not set especially below 11.2.0.2 • TO ENABLE THE COMPRESSION THE FOLLOWING COMMAND IS ISSUED • SQL> ALTER DATABASE ARCHIVELOG COMPRESS ENABLE; • TO DISABLE THE COMPRESSION THE FOLLOWING COMMAND IS ISSUED • SQL> ALTER DATABASE ARCHIVELOG COMPRESS DISABLE; • Interested? Search for an undocumented parameter _log_archive_compress_enable • Test’s shown no size reduction for archive logs(10/11g), assuming its still in development, Lets wait for other release, if we can get archive compression Archive compression, what? – Next release?