480 likes | 657 Views
Distributed DBMSs - Concepts and Design Transparencies. Concepts. Distributed Database A logically interrelated collection of shared data (and a description of this data), physically distributed over a computer network. Distributed DBMS
E N D
Distributed DBMSs - Concepts and Design Transparencies
Concepts Distributed Database A logically interrelated collection of shared data (and a description of this data), physically distributed over a computer network. Distributed DBMS Software system that permits the management of the distributed database and makes the distribution transparent to users.
Date’s Rules for a DDBMS Fundamental Principle: To the user, a distributed system should look exactly like a non-distributed system. • Local Autonomy • No Reliance on a Central Site • Location Independence • Fragmentation Independence • Replication Independence • Distributed Query Processing • Distributed Transaction Processing
Concepts • Collection of logically-related shared data. • Data split into fragments. • Fragments may be replicated. • Fragments/replicas allocated to sites. • Sites linked by a communications network. • Each DBMS participates in at least one global application.
Advantages of DDBMSs • Reflects organizational structure • Improved shareability and local autonomy • Improved availability • Improved reliability • Improved performance • Economics
Disadvantages of DDBMSs • Complexity • Cost • Security • Integrity control more difficult • Database design more complex
Types of DDBMS • Homogeneous DDBMS • Heterogeneous DDBMS • Sites may run different DBMS products, with possibly different underlying data models. • Occurs when sites have implemented their own databases and integration is considered later.
Functions of a DDBMS • DDBMS to have at least the functionality of a DBMS. • Also to have following functionality: • Distributed query processing. • Extended concurrency control. • Extended recovery services.
Distributed Database Design • Three key issues: Fragmentation Relation may be divided into a number of sub-relations, which are then distributed. Allocation Each fragment is stored at site with "optimal" distribution (see principles of distribution design). Replication Copy of fragment may be maintained at several sites.
Fragmentation • Definition and allocation of fragments carried out strategically to achieve: • Locality of Reference • Improved Reliability and Availability • Improved Performance • Balanced Storage Capacities and Costs • Minimal Communication Costs. • Involves analyzing most important applications, based on quantitative/qualitative information.
Fragmentation • Quantitative information (replication) used for may include: • frequency with which an application is run; • site from which an application is run; • performance criteria for transactions and applications. • Qualitative information (fragmentation) may include transactions that are executed by application: relations, attributes and tuples.
Data Allocation Centralized: Consists of single database and DBMS stored at one site with users distributed across the network. Partitioned: Database partitioned into fragments, each fragment assigned to one site. Complete Replication: Consists of maintaining complete copy of database at each site. Selective Replication: Combination of partitioning, replication, and centralization. © Pearson Education Limited 1995, 2005
Why Fragment? • Usage • Applications work with views rather than entire relations. • Efficiency • Data is stored close to where it is most frequently used. • Data that is not needed by local applications is not stored. • Security • Data not required by local applications is not stored and so not available to unauthorized users.
Correctness of Fragmentation • Three correctness rules: Completeness If relation R is decomposed into fragments R1, R2, ... Rn, each data item that can be found in R must appear in at least one fragment. Reconstruction • Must be possible to define a relational operation that will reconstruct R from the fragments. • Reconstruction for horizontal fragmentation is Union operation and Join for vertical . Disjointness • If data item di appears in fragment Ri, then it should not appear in any other fragment.; Exception: vertical fragmentation, where primary key attributes must be repeated to allow reconstruction. • For horizontal fragmentation, data item is a tuple • For vertical fragmentation, data item is an attribute.
Types of Fragmentation • Four types of fragmentation: • Horizontal • Vertical • Mixed
Horizontal Fragmentation • Consists of a subset of the tuples of a relation. • Defined using Selection operation of relational algebra: p(R) • For example: P1 = type='House'(PropertyForRent) P2 = type='Flat' (PropertyForRent) Result (PNo., St, City, postcode,type,room,rent,ownerno.,staffno., branchno.) • This strategy is determined by looking at predicates used by transactions. • Reconstruction involves using a union eg R = r1 U r2
Vertical Fragmentation • Consists of a subset of attributes of a relation. • Defined using Projection operation of relational algebra: a1, ... ,an(R) • For example: S1 = staffNo, position, sex, DOB, salary(Staff) S2 = staffNo, fName, lName, branchNo(Staff) • Determined by establishing affinity of one attribute to another. • For vertical fragements reconstruction involves the join operation; Each fragment is disjointed except for the primary key
a b c a1 bb1 bb2 bb3 x y Union Intersection Difference Relational Algebra Select Project Product x x x y y y a b c b c c Divide a1 bb3 a1 bb2 b1 cc1 a1 bb1 b1 bb2 Join a1 b1 c1 a2 b2 c2 a1 b1 a2 b2 b1 c1 b2 c2
Mixed Fragmentation • Consists of a horizontal fragment that is vertically fragmented, or a vertical fragment that is horizontally fragmented. • Defined using Selection and Projection operations of relational algebra: p(a1, ... ,an(R)) or a1, ... ,an(σp(R))
Example - Mixed Fragmentation S1 = staffNo, position, sex, DOB, salary(Staff) S2 = staffNo, fName, lName, branchNo(Staff) S21 = branchNo='B003'(S2) S22 = branchNo='B005'(S2) S23 = branchNo='B007'(S2) Explain and Illustrate the result of the above example
Transparencies in a DDBMS • Distribution Transparency • Fragmentation Transparency • Location Transparency • Replication Transparency • Transaction Transparency • Concurrency Transparency • Failure Transparency • Performance Transparency • DBMS Transparency
Distribution Transparency • Distribution transparency allows user to perceive database as single, logical entity. • If DDBMS exhibits distribution transparency, user does not need to know: • data is fragmented (fragmentation transparency), • location of data items (location transparency), • otherwise call this local mapping transparency. • With replication transparency, user is unaware of replication of fragments .
Transaction Transparency • Ensures that all distributed transactions maintain distributed database’s integrity and consistency. • Distributed transaction accesses data stored at more than one location. • Each transaction is divided into number of subtransactions, one for each site that has to be accessed. • DDBMS must ensure the indivisibility of both the global transaction and each subtransactions.
Concurrency Transparency • All transactions must execute independently and be logically consistent with results obtained if transactions executed one at a time, in some arbitrary serial order. • Same fundamental principles as for centralized DBMS. • DDBMS must ensure both global and local transactions do not interfere with each other. • Similarly, DDBMS must ensure consistency of all subtransactions of global transaction.
Concurrency Transparency (Replication) • Replication makes concurrency more complex. • If a copy of a replicated data item is updated, update must be propagated to all copies. • However, if one site holding copy is not reachable, then transaction is delayed until site is reachable.
Failure Transparency • DDBMS must ensure atomicity and durability of global transaction. • Means ensuring that subtransactions of global transaction either all commit or all abort. • Thus, DDBMS must synchronize global transaction to ensure that all subtransactions have completed successfully before recording a final COMMIT for global transaction. • Must do this in the presence of site and network failures.
Performance Transparency • DDBMS must perform as if it were a centralized DBMS. • DDBMS should not suffer any performance degradation due to distributed architecture. • DDBMS should determine most cost-effective strategy to execute a request. i.e. query optimisation (the order of selects and projects) applied to a distributed database
Performance Transparency • Must consider fragmentation, replication, and allocation schemas. • DQP has to decide e.g. : • which fragment to access; • which copy of a fragment to use; • which location to use.
Performance Transparency • DQP produces execution strategy optimized with respect to some cost function. • Typically, costs associated with a distributed request include: • I/O cost; • Communication cost.
Performance Transparency - Example Property(propNo, city) 10000 records in London Client(clientNo,maxPrice) 100000 records in Glasgow Viewing(propNo, clientNo) 1000000 records in London SELECT p.propNo FROM Property p INNER JOIN Client c INNER JOIN Viewing v ON c.clientNo = v.clientNo) ON p.propNo = v.propNo WHERE p.city=‘Aberdeen’ AND c.maxPrice > 200000; • This query selects properties that viewed in aberdeen that have a price greater than £200, 000.
Performance Transparency - Example Assume: • Each tuple in each relation is 100 characters long. • 10 renters with maximum price greater than £200,000. • 100 000 viewings for properties in Aberdeen. • In addition the data transmission rate is 10,000 characters per sec and there is a 1 sec access delay to send a message.
Performance Transparency - Example • Derive the following :
Distributed system architecture and parallel Data management.DT211
A distributed system is a collection of computers that communicate by means of some networked media. There are a number of key issues which must be considered when discussing distributed system architectures: • The principle of locality. This means that parts of a system which are associated with each other should be in close proximity. For example, programs which exchange large amounts of data should, ideally, be on the same computer or, less ideally, the same local area network. • The principle of sharing. This means that ideally resources (memory, file space, processor power) should be carefully shared in order to minimise the load on some of the elements of a distributed system. • The parallelism principle. This means that maximum use should be made of the main rationale behind distributed systems: the fact that a heavy degree ofscaling up can be achieved by means of the careful deployment of serverssharing processing load.
Principle of locality: This is the principle that entities which comprise a distributed system should belocated close together, ideally on the same server or client. These entities can befiles, programs, objects or hardware. • (1)Keeping data together: • Probably the best known example of the locality principle is that data that isrelated to each other should be grouped together. One example of this where two tables which are related by virtue of the fact that they are often accessed together are movedonto the same server. • (2)Keeping code together • The idea behind this is that if two programs communicate with each other in adistributed system then, ideally, they should be located on the same computer or,at worst, they should be located on the same local area network. The worst caseis where programs communicate by passing data over the slow communicationmedia used in wide area networks.
(3)Bringing users and data close together.There are two popular ways of implementing this principle. The first is the use of replicated data and the second is caching. • Replicated data is data that is duplicated at various locations in a distributedsystem. The rationale for replicating a file, database, part of a file or part of adatabase is simple. By replicating data which is stored on a wide area networkand placing it on a local area network major improvements in performance can beobtained as local area technology can be orders of magnitude faster than widearea technology in delivering data. However, replicating and synchronisation of the data can be counter productive • (4)Keeping programs and data together • A good precept to be used when designing a distributed system is that programsand the data that they operate on should be as close together as possible.
The principle of sharing • This principle is concerned with the sharing of resources - memory, software and processor. • (1)Sharing amongst servers: A major decision to be made about the design of a distributed system is how theservers in a system are going to have the work performed by the systempartitioned among them. The main rationale for sharing work amongst servers isto avoid bottlenecks where servers are overloaded with work which could bereallocated to other servers. • (2)Sharing data: A distributed system will have, as a given, the fact that data should beshared between users. The two main decisions are where to situate the tables that make upa relational application and what locking strategy to adopt.
Locking:Most of the decisions about locking will be made with respect to the databasemanagement system that is used. In general such systems allow locking at one ormore sizes: page locks, table locks, database locks and row lock. • Although many locking decisions are straightforward, some locking issues aremuch more subtle. For instance, the relationship between deadlock occurrence,the locking strategy adopted and performance can be a subtle one, wheretrade-offs have to be considered. For example, locking at the row level willenable more concurrency to take place at the cost of increased deadlocking whereresources have to be expended in order to monitor and release deadlocks;conversely locking at the table level can lead to a major reduction in deadlockoccurrence, but at the expense of efficient concurrent operations
The parallelism principle • A key idea behind the parallel principle is that of load balancing: • For example, one decision that the designer of a distributed system has to make is what to do about very large programs which could theoretically execute on the same processor. Should this program be split up into different threads which execute in parallel, either on a single server or on a number of distributed servers?
Another ideal which designers try and aim for is to partition a database into tables and files in order that the transactions are evenly spread around the file storage devices that are found on a distributed system. • Finally another area where parallelism can increase the performance of a distributed system is I/O parallelism. Usually design decisions which are made in this category involve the use of special software or hardware, for example the use of RAID devices or parallel database packages. • In conclusion employing Distributed design principles is always a compromise; a solution which is optimal for a particular system workload will often be not as efficient for another workload.
Parallel Data Management • A topic that’s closely linked to Data Warehousing is that of Parallel Data Management. • The argument goes: • if your main problem is that your queries run too slowly, use more than one machine at a time to make them run faster (Parallel Processing). • Oracle uses this strategy in its warehousing products. • There are two types of parallel processing - Symmetric Multiprocessing (SMP), and Massively Parallel Processing (MPP)
SMP – All the processors share the same memory and the O.S. runs and schedules tasks on more than one processor without distinction. • in other words, all processors are treated equally in an effort to get the list of jobs done. • However, SMP can suffer from bottleneck problems when all the CPUs attempt to access the same memory at once. • MPP - more varied in its design, but essentially consists of multiple processors, each running their own program on their own memory i.e. memory is not shared between processors. • the problem with MPP is to harness all these processors to solve a single problem. • But they do not suffer from bottleneck problems
Regardless of the architecture used, there are still alternatives regarding the use of the parallel processing capabilities. • There are two possible solutions dividing up the data: Static and Dynamic Partitioning. • InStatic Partitioningyou break up the data into a number of sections. Each section is placed on a different processor with its own data storage and memory. The query is then run on each of the processors, and the results combined at the end to give the entire picture. This is like joining a queue in a supermarket. You stay with it until you reach the check-out. • The main problem with Static Partitioning is that you can’t tell how much processing the various sections need. If most of the relevant data is processed by one processor you could end up waiting almost as long as if you didn’t use parallel processing at all. • In Dynamic Partitioningthe data is stored in one place, and the data server takes care of splitting the query into multiple tasks, which are allocated to processors as they become available. This is like the single queue in a bank. As a counter position becomes free the person at the head of the queue takes that position • With Dynamic Partitioning the performance improvement can be dramatic, but the partitioning is out of the users hands.
Sample type question • Fragmentation, replication and allocation are the three important characteristics discuss their importance in relation to distributed databases. • Discuss the major principles of distributed systems design and how they can be utilised in geographically distributed database.