1 / 58

Query Processing and Query Optmization

CS157B Lecture 19. Query Processing and Query Optmization. Prof. Sin-Min Lee Department of Computer Science San Jose State University. Transaction Management. Until now, our concept of database has been one in which programs accessing the database are run one

Download Presentation

Query Processing and Query Optmization

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CS157B Lecture 19 Query Processing and Query Optmization Prof. Sin-Min Lee Department of Computer Science San Jose State University

  2. TransactionManagement Until now, our concept of database has been one in which programs accessing the database are run one at a time (serially). Often this is indeed the case. However, there are also numerous applications in which more than one program, or different executions of the same program, run simultaneously (concurrently). This is where TRANSACTION MANAGEMENT comes in handy

  3. SAFETY Example #1 An airline reservation system, where at one time, several agents may be selling tickets, and therefore, changing lists of passengers and counts of available seats. The problem here is that if we are not careful when we allow two or more processes to access the database, we could sell the same seat twice. Here 2 processes that read and change the value of the same object must not be allowed to run concurrently, because they might interact in undesirable ways.

  4. SPEED Example #2 Statistical database, such as census data, where many people may be querying the database at once. Here, as long as no one is changing the data, we do not really care in what order the processes read data. We can let the operating system schedule simultaneous read requests. Here we want to allow maximum concurrent operations, so time can be saved

  5. Important Terms Serializable schedule -- is a linear arrangement of the database calls from several transactions with the property: the final database state obtained by executing the calls in schedule order is the same as the obtained by running the transactions in some unspecified serial order

  6. Important Terms (cont…) Lock -- is an access privilege on a database object, which the DBMS grants to a particular transaction. To allow the privileged transaction to complete its work without undue interference, the lock restricts the access of competing transactions to the database object.

  7. Important Terms (cont…) Two-phase locking protocol -- is a discipline that transactions can use to ensure serializability. The first phase is characterized by the monotonic acquisition of new locks or the strengthening of existing locks. The second phase involves the monotonic downgrade or release of existing locks. This lock is the most important technique for managing concurrency.

  8. Important Terms (cont…) Strict two-phase locking protocol -- each transaction holds all its locks until it commits or rolls back. When the commit or rollback is secure, it releases all the locks. In other words, the shrinking phase occurs all at once, immediately after the transaction terminates.

  9. Important Terms (cont…) Timestamping -- is another method for securing concurrency in the face of conflicting transactions. Timestamp -- is a time of origin of each transaction given by transaction manager portion of DBMS Items -- units of data to which access is controlled

  10. Important Terms (cont…) Transaction -- is simply a single execution of a program. This program may be a simple query expressed in one of the query languages or an elaborate host language program with embedded calls to a query language. Several independent executions of the same program may be in progress simultaneously; each is a different transaction

  11. Main Idea To a large extent, transaction management can be seen as an attempt to make complex operations appear atomic. That is, they either occur in their entirety or do not occur at all, and if they occur, nothing else apparently went on during the time of their occurrence. The normal approach to ensuring atomicity of transactions is ‘serialization’, which forces transactions to run concurrently in a way that makes it appear that they ran one-at-a-time (serially).

  12. Main Idea In reality, transaction are sequences of more elementary steps, such as reading or writing of single items from the database, and performing simple arithmetic steps in the workspace. When concurrency control is provided, other primitive steps are also needed, steps which set and release locks, commit (complete) transactions, and others

  13. Items To manage concurrency, the database must be partitioned into items, which are the units of data to which access is controlled. The nature and size of items are for the system designer to choose. In the relational model of data, for example, we could choose large items, like relations, or small items like individual tuples or even components of tuples.

  14. Locks The most common way in which access to items is controlled is by “locks.” Lock manager is the part of a DBMS that records, for each item I, whether one or more transactions are reading or writing any part of I. If so, the manager will forbid another transaction from gaining access to I, provided the type of access (read or write) could cause a conflict, such as the duplicate selling of an airline seat.

  15. Locks As it is typical for only a small subset of the items to have locks on them at any one time, the lock manager can store the current locks in a lock table which consists of records (<item>,<lock type>,<transaction> The meaning of record (I,L,T) is that transaction T has a lock of type L on item I.

  16. Example of locks Lets consider two transaction T1 and T2. Each accesses an item A, which we assume has an integer value, and adds one to A. Read A;A:=A+1;Write A; ----------------------------------------------------------- T1: Read A A:=A+1 Write A T2: Read A A:=A+1 Write A -----------------------------------------------------------

  17. Example of locks (cont…) The most common solution to this problem is to provide a lock on A. Before reading A, a transaction T must lock A, which prevents another transaction from accessing A until T is finished with A. Furthermore, the need for T to set a lock on A prevents T from accessing A if some other transaction is already using A. T must wait until the other transaction unlocks A, which it should do only after finishing with A.

  18. Introduction Goals: • get the right answer to a query • get it fast Possible query representations (after parsing) • relational algebra (not usually) • query graph (usually)

  19. Strategies for making relational algebra queries “better” • Push down selects • eg.: r1(r1no, a, b, c, d) r2(r2no, x, y, z, r1no) σ r1.a = 7 (r1 JOIN r2) (σ r1.a = 7 r1) JOIN r2 • push down projects • Πr1.a, r1.b(r1 JOIN r2) Πr1.a, r1.b( (Πr1no,a,b (r1)) JOIN r2)

  20. Strategies for making relational algebra queries “better”(cntd) • Eliminate  products • σr1.r1no=r2.r1no (r1  r2) r1 JOIN r1.r1no=r2.r1no r2 • replace • σp1 and p2 (e1) by σp1 (σp2 (e1))

  21. Strategies for making relational algebra queries “better”(cntd) • maybe there is an index to support p2 solution eg,: σ (sal > 10000) and (name = “elmo”) emp do name = “elmo” first with index test results for sal > 10000

  22. Strategies for making relational algebra queries “better”(cntd) • Join relations in increasing order of size to keep intermediate results small ( saves time) | e1| = 1 | e2| = 200 | e3| = 100000 eg.: e1 JOIN (e2 JOIN e3) (e1 JOIN e2) JOIN e3

  23. Query Processing Cost Estimation • Catalog Statistics • r = a relation • nr = |r| = # tuples in r • sr = size of record r in bytes • V(A,r) = # unique attribute A values in r Given these statistics, • |r  s| = nrns • size of a tuple of r  s = sr + ss

  24. Query Processing Cost Estimation If ‘a’ is a constant, then |σ r.A = a| = 1 * nr V(A,r) Note: histograms can be better for selectivity estimation than V(A,r) statistic.

  25. Selection Predicate Selectivity Estimates sel(r.A=a) 1 / V(A,r) sel(r.A<a) (a-min(A,r))/(max(A,r)-min(A,r)) sel(a1 <r.A< a2) (a2 - a1) / (max(A,r) - min(A,r)) sel(p(R.a)) 1/3 guess!!

  26. Join Predicate Selectivity • What is | r1 (R1) JOIN r2 (R2)| ? • If R1  R2 is a key of r1, then at most 1 tuple of r1 joins to each r2 tuple. Hence, | r1 JOIN r2|  | r2 | • if R1  R2 is a key of neither r1 nor r2, then, • assume R1  R2 = {A} • a tuple t  r1 produces approximately n r2 / V(A, r2) tuples when joined to r2. • Using the above, | r1 JOIN r2 |  (n r1n r2 )/V(A, r2)

  27. Join Predicate Selectivity(cntd) • If we had used t  r2 instead of r1 to compute the above estimate, we would have gotten | r1 JOIN r2 | (n r1 n r2 )/V(A, r1) If V(A,r1)  V(A,r2), there are likely to be some dangling tuples that don’t participate in the join. Hence, the smaller of the two (join size) estimates is probably better.

  28. Maintaining Statistics • Doing it after every update is too much overhead • Hence, most systems gather statistics periodically during slack time ( often this is initiated manually).

  29. 1-table selection subquery cost • Without index • must use a sequential scan • C I/O = cost of an I/O  11 msec • C CPU = cost of CPU to process 1 tuple .1msec cost(σ p (r)) = nr C CPU + pages ( r ) * C I/O

  30. 1-table selection subquery cost(cntd) with clustered index • B+ tree on r.A • p= ( r.A = a) cost (σ p (r)) = (1/ V(A,r)) * nr C CPU+ (height (r.A B+tree) +(1/V(A,r))*pages ( r )) C I/O

  31. 1-table selection subquery cost(cntd) With unclustered index • secondary B+ tree on r.A • p = (r.A = a) cost (σ p (r)) = (1/ V(A,r)) * nr C CPU+ (height (r.A B+tree) +(1/V(A,r)) * nr) C I/O

  32. Exercise • What is the cost of solving σ a1 < r.A < a2 (r) using a clustered B+ tree on R?

  33. Join Strategies Nested Loop Join ( Simple Iteration) NLJ to solve r1 JOINp r2 for each t1 r1 do begin for t2 r2 do begin if p(t1 , t2) = true then output(t1 , t2) as part of result end end NOTE: outer scan is over r1 inner scan is over r2 (this could have been reversed)

  34. QUIZ • Should outer scan be over biggest or smallest table? • Hint: Consider buffer pool size. • Variation of nested loop join: block oriented iteration • Scan inner table once per outer table block instead of outer table tuple.

  35. Sort Merge Join To do r1 JOIN r2r1.a = r2 . b • sort r1 on a • sort r2 on b • merge • scan r1 sequentially • iterate over only needed portion of r2 for each r1 tuple

  36. Sort Merge Join(cntd) r1 x a r2 b y a 1 1 p b 1 2 q c 2 2 r d 2 3 s e 3 cost = sort cost + constant * | r1 JOIN r2|

  37. Use of an Index • In nested loop join, if there is an index on JOIN attribute of inner table, you can use it. • In sort-merge join, if there is a B+tree index (preferably clustered) on the join attribute(s) you can avoid sorting that table (read it in sorted order).

  38. Conclusion • Query processing • Query processing operations • sequential scan • index scan • join • nested loop • sort merge • Query processing operation costs • Next time: Hash join

More Related