Chapter 1 Quiz: What Is IBM DB2?
Test your understanding of Chapter 1. Attempt all questions from memory before revealing the answers. The act of retrieval — even when you get it wrong — strengthens learning more than re-reading.
Scoring Guide: - 20-25 correct: Excellent — you have strong command of the material - 15-19 correct: Good — review the sections where you missed questions - 10-14 correct: Fair — re-read the chapter with the quiz questions in mind - Below 10: Re-read the chapter carefully, then retake this quiz in 48 hours
Multiple Choice Questions
Question 1
What year was the foundational paper on the relational model published?
A) 1965 B) 1970 C) 1973 D) 1983
Answer
**B) 1970** Ted Codd published "A Relational Model of Data for Large Shared Data Banks" in the *Communications of the ACM* in June 1970. This paper laid the theoretical foundation for all relational databases, including DB2.Question 2
Which IBM research project directly preceded DB2 as a commercial product?
A) IMS B) SEQUEL C) System R D) System/360
Answer
**C) System R** System R was the prototype relational database system built at IBM's San Jose Research Laboratory between 1973 and 1979. It produced both SQL and the cost-based optimizer. DB2 Version 1 (1983) was the commercial product that grew from System R's research.Question 3
DB2 Version 1 was initially positioned primarily as a:
A) High-performance OLTP system to replace IMS B) Decision-support tool for ad hoc queries C) Distributed database for client-server computing D) Object-oriented database for multimedia data
Answer
**B) Decision-support tool for ad hoc queries** DB2 V1 was slower than IMS for OLTP workloads and was initially marketed as a system for running ad hoc queries against data extracted from IMS. Its evolution into a mission-critical OLTP platform happened over subsequent versions.Question 4
What hardware component enables DB2 for z/OS data sharing in a Parallel Sysplex?
A) Cluster Caching Facility B) Coupling Facility C) Workload Manager D) RACF
Answer
**B) Coupling Facility** The Coupling Facility is a specialized hardware device that enables cross-system communication in a Parallel Sysplex. It allows multiple DB2 subsystems on separate mainframes to share a single database simultaneously. The Cluster Caching Facility (A) is the LUW equivalent used by pureScale.Question 5
Which of the following is NOT shared between DB2 for z/OS and DB2 for LUW?
A) SQL dialect (approximately 90% compatible) B) ACID guarantees C) The underlying codebase D) Support for JDBC/ODBC access
Answer
**C) The underlying codebase** DB2 for z/OS and DB2 for LUW were developed by different teams and do NOT share a codebase. They share a SQL dialect, ACID guarantees, core data types, and application access methods — but the internal implementations are substantially different.Question 6
"Five-nines availability" (99.999%) corresponds to approximately how much unplanned downtime per year?
A) 52 minutes B) 26 minutes C) 5 minutes and 15 seconds D) 31 seconds
Answer
**C) 5 minutes and 15 seconds** 99.999% uptime means 0.001% downtime. 365.25 days * 24 hours * 60 minutes * 0.00001 = approximately 5.26 minutes per year.Question 7
What was the significance of DB2 Version 9 (2007), codenamed "Viper"?
A) It introduced BLU Acceleration for columnar analytics B) It added native XML storage (pureXML) C) It was the first version to support Linux D) It introduced continuous delivery
Answer
**B) It added native XML storage (pureXML)** DB2 9, codenamed "Viper," introduced pureXML — native storage and querying of XML data alongside traditional relational data. BLU Acceleration came in DB2 10.5 LUW (2013), and continuous delivery came with DB2 12 for z/OS (2017).Question 8
What is "polyglot persistence"?
A) The ability to store data in multiple languages B) Using different database technologies for different workloads within a single organization C) Replicating data across multiple DB2 instances D) The practice of supporting multiple SQL dialects
Answer
**B) Using different database technologies for different workloads within a single organization** Polyglot persistence recognizes that no single database engine is optimal for all workloads. Organizations may use DB2 for OLTP, a document store for unstructured data, a graph database for relationship queries, and an in-memory cache for high-speed lookups — all in the same enterprise.Question 9
DB2 for z/OS licensing is primarily based on:
A) Number of named users B) Number of processor cores C) Mainframe capacity (MSUs/MIPS) D) Data volume stored
Answer
**C) Mainframe capacity (MSUs/MIPS)** DB2 for z/OS licensing costs are tied to the processing capacity of the mainframe partition, measured in Millions of Service Units (MSUs). This means software costs scale with CPU consumption, which has led to significant focus on CPU optimization in mainframe shops.Question 10
Which DB2 for LUW edition is free for both development and small production use?
A) Db2 Developer Edition B) Db2 Express-C C) Db2 Community Edition D) Db2 Standard Edition
Answer
**C) Db2 Community Edition** Db2 Community Edition is free for development and small production use, limited to 4 cores and 16 GB RAM. Db2 Developer Edition is free only for development and testing (no production use).Question 11
Who developed the cost-based query optimizer approach that all modern relational databases use?
A) Edgar F. Codd B) Donald Chamberlin C) Patricia Selinger D) Raymond Boyce
Answer
**C) Patricia Selinger** Patricia Selinger's 1979 paper, "Access Path Selection in a Relational Database Management System," described how a database could automatically choose the most efficient query execution plan by estimating costs. This approach is the foundation of query optimization in every major RDBMS today.Question 12
What is the primary advantage of DB2 12 for z/OS's continuous delivery model?
A) It eliminates the need for DBAs B) It delivers new capabilities through function levels without requiring a full version migration C) It makes DB2 free to use D) It removes the need for backup and recovery
Answer
**B) It delivers new capabilities through function levels without requiring a full version migration** Continuous delivery means IBM delivers new features through incremental function levels rather than major version releases. This reduces the disruption and risk associated with version upgrades while allowing organizations to adopt new capabilities at their own pace.Short Answer Questions
Question 13
Name the two revolutionary innovations that came out of the System R project.
Answer
1. **SQL** (Structured Query Language, originally SEQUEL) — a declarative query language that allows users to specify *what* data they want without specifying *how* to retrieve it. 2. **The cost-based query optimizer** — an approach where the database automatically estimates the cost of different execution strategies and chooses the most efficient one.Question 14
Explain why a large bank might choose NOT to migrate from DB2 for z/OS to PostgreSQL, even though PostgreSQL is free. Give at least three reasons.
Answer
Any three of the following (or similar valid reasons): 1. **Migration cost**: Migrating thousands of tables, stored procedures, and application code is a multi-year, multi-hundred-million-dollar effort with significant risk. 2. **Proven reliability at scale**: DB2 for z/OS has decades of proven performance at the transaction volumes and availability levels banks require. PostgreSQL has not been tested to the same degree in these specific environments. 3. **Regulatory familiarity**: Regulators have audited DB2 on z/OS thousands of times and understand its compliance capabilities. Introducing a new platform means re-education and re-certification. 4. **Availability**: DB2 for z/OS with Parallel Sysplex achieves five-nines availability through hardware and software integration that PostgreSQL cannot easily replicate. 5. **Operational maturity**: Decades of backup, recovery, monitoring, and performance-tuning procedures would need to be rebuilt. 6. **Total cost of ownership**: When migration costs, risk, retraining, and compliance are included, the TCO of migration may exceed the licensing savings.Question 15
Describe the Meridian National Bank's dual-platform DB2 architecture. What role does each platform play?
Answer
Meridian National Bank runs: - **DB2 12 for z/OS** on an IBM z15 mainframe — this is the core banking database handling accounts, transactions, and customer records. It serves the CICS-based online transaction processing and batch applications. - **Db2 11.5 for LUW** on Linux servers — this is the operational database for the digital banking platform, mobile app backend, and customer analytics. It supports Java/Spring Boot microservices. - **Db2 Warehouse on Cloud Pak for Data** — the enterprise data warehouse, receiving replicated data from both z/OS and LUW environments. Data flows between the mainframe and distributed environments via Apache Kafka event streaming.Question 16
What does the acronym ACID stand for? Define each property in one sentence.
Answer
- **Atomicity**: A transaction either completes entirely or has no effect at all — there are no partial transactions. - **Consistency**: A transaction brings the database from one valid state to another, satisfying all defined constraints and rules. - **Isolation**: Concurrent transactions do not interfere with each other; each transaction appears to execute in isolation. - **Durability**: Once a transaction is committed, its effects are permanent and survive any subsequent system failure.Question 17
What is the difference between DB2's "instance model" on LUW and its architecture on z/OS? Why does this difference exist?
Answer
**[LUW]** DB2 for LUW uses an instance-based architecture where a single installation can host multiple instances, each running as an independent set of operating-system processes. This is similar to PostgreSQL and Oracle on distributed platforms. **[z/OS]** DB2 for z/OS runs as a subsystem within z/OS, integrated with the operating system's resource management (Workload Manager), security (RACF), and recovery infrastructure. Multiple DB2 subsystems can share data through Parallel Sysplex and the Coupling Facility. The difference exists because the underlying platforms are fundamentally different. z/OS provides system-level services (resource management, security, recovery) that DB2 can leverage natively. On distributed platforms, DB2 must manage more of these capabilities itself within the instance model.Question 18
What was the "Universal Database" (UDB) initiative, and what capabilities did it add to DB2?
Answer
The "Universal Database" initiative, launched in 1997 with DB2 UDB V5, reflected IBM's ambition to make DB2 a database for *all* types of data, not just structured relational data. It added: - User-defined types (UDTs) - User-defined functions (UDFs) - Large object (LOB) support for binary and text data - Object-relational extensions While the broader object-relational revolution did not play out as IBM predicted, the LOB support and extensibility features remain important in modern DB2.Question 19
Name three features that are shared between DB2 for z/OS and DB2 for LUW, and two features that are specific to only one platform.
Answer
**Shared features** (any three): - SQL dialect (approximately 90% compatible) - ACID guarantees - Core data types (INTEGER, VARCHAR, DECIMAL, TIMESTAMP, etc.) - SQL PL stored procedures - JDBC/ODBC application access - Cost-based query optimization **z/OS-specific** (any one): - Data sharing with Parallel Sysplex and Coupling Facility - Workload Manager (WLM) integration - RACF security integration - STOGROUP-based storage management **LUW-specific** (any one): - Instance-based architecture - pureScale clustering with Cluster Caching Facility - Automatic storage management - db2top monitoring toolQuestion 20
Why is the aging workforce of DB2 for z/OS professionals both a risk and an opportunity?
Answer
**Risk**: Many experienced mainframe DB2 DBAs are approaching retirement, and fewer young professionals are entering the field. Organizations that depend on DB2 for z/OS face the possibility of losing critical institutional knowledge and operational expertise, with difficulty finding replacements. **Opportunity**: The combination of high demand and limited supply means that professionals who invest in learning DB2 for z/OS skills can command strong salaries ($120,000-$180,000+) and enjoy strong job security. For early-career professionals willing to learn mainframe skills, this represents a differentiated career path with less competition than more popular technology stacks.Application Questions
Question 21
A small credit union (50,000 members, 10 million transactions per year) is choosing a database for a new core banking system. They have no existing mainframe infrastructure and a small IT team of five people. Would you recommend DB2? If so, which platform? If not, what would you recommend instead? Justify your answer.
Answer
**Recommended approach**: DB2 for LUW (Community Edition or Standard Edition) or PostgreSQL would both be reasonable choices. **DB2 for LUW rationale**: The Db2 Community Edition is free and capable of handling this workload easily (10M transactions/year is roughly 0.3 TPS average). It provides enterprise features, ACID guarantees, and a growth path. If the credit union grows, they can scale up to a licensed edition without changing platforms. **Against z/OS**: A mainframe is not justified. The transaction volume is far below mainframe-scale needs, the IT team lacks mainframe skills, and the cost would be disproportionate. **PostgreSQL alternative**: For a team of five with no IBM infrastructure, PostgreSQL might be more practical due to the larger community, more available talent, and zero licensing cost. The decision depends on whether the credit union values IBM's enterprise support or prefers the broader PostgreSQL ecosystem. This is a case where DB2's strengths (extreme scale, mainframe integration, regulatory maturity) are less relevant, and its costs (licensing, specialized skills) are more significant.Question 22
Explain the concept of "data gravity" and how it relates to why organizations continue to use DB2.
Answer
**Data gravity** is the principle that large accumulations of data tend to attract applications, services, and additional data to them — much like physical gravity. The more data you have in a system, the harder it becomes to move. For DB2 installations, data gravity manifests in several ways: 1. **Volume**: A large bank might have petabytes of data in DB2 — transaction history, customer records, audit trails. Moving this volume is operationally expensive and time-consuming. 2. **Application dependencies**: Hundreds or thousands of applications are written to query, insert, and update this data. Each application embeds assumptions about the database platform. 3. **Operational infrastructure**: Backup procedures, monitoring tools, disaster recovery plans, and compliance documentation all reference the existing database. 4. **Institutional knowledge**: The organization's staff knows how to operate, optimize, and troubleshoot DB2 specifically. Data gravity helps explain why "just migrate to X" is rarely as simple as it sounds, and why DB2 installations measured in decades are common in enterprise environments.Question 23
A technology journalist writes: "The mainframe is dead. No organization should be investing in DB2 for z/OS in 2026." Write a response (3-4 sentences) that corrects this claim using evidence from the chapter.
Answer
The claim that the mainframe is dead is contradicted by observable reality. IBM continues to release new Z hardware — the z16 added an on-chip AI accelerator — and global mainframe MIPS consumption grows approximately 4-5% per year. DB2 for z/OS processes transactions for more than 90 of the world's 100 largest banks, and IBM's continuous delivery model means the platform receives ongoing feature enhancements. The organizations running these systems have evaluated alternatives and continue to invest in mainframe DB2 because no other platform matches its combination of transaction throughput, availability, and regulatory maturity at comparable scale.Question 24
If you were starting a career in database administration today, would you specialize in DB2, Oracle, PostgreSQL, or SQL Server? Explain your reasoning, considering market demand, salary potential, competition, and long-term viability.
Answer
*Note: This is an opinion question. A well-reasoned answer considering multiple factors is more important than the specific choice.* **One strong argument**: Specialize in DB2 for z/OS as a primary skill while maintaining PostgreSQL as a secondary skill. **Reasoning**: - **Market demand**: DB2 for z/OS has strong demand driven by the aging workforce. The supply-demand imbalance favors the DBA. - **Salary**: DB2 for z/OS DBAs command premium salaries ($120K-$180K+) due to scarcity. - **Competition**: Fewer people are entering the mainframe DB2 space, so there is less competition for positions. - **Long-term viability**: The systems running on DB2 for z/OS are deeply entrenched and will not be migrated away in the foreseeable future. - **PostgreSQL complement**: Adding PostgreSQL skills provides versatility and exposure to the fastest-growing database ecosystem, making you valuable in polyglot persistence environments. Other valid choices exist (Oracle for consulting opportunities, SQL Server for the Microsoft ecosystem, PostgreSQL for startups and cloud-native companies), and the best choice depends on individual interests, geographic market, and career goals.Question 25
Explain in your own words why Ted Codd's principle of "data independence" was revolutionary, and give one example of how it manifests in modern DB2.
Answer
**Why it was revolutionary**: Before the relational model, applications were tightly coupled to the physical storage structure of data. If you changed how data was stored on disk — reorganized the files, added a new index, moved data to a different device — the applications that accessed that data would break. They had to be rewritten. This made database administration extremely rigid and expensive. Codd's principle of data independence separated the *logical* structure of data (tables, columns, relationships — what applications see) from the *physical* structure (how data is stored on disk, what indexes exist, how files are organized). Applications interact with the logical layer, and the database manages the physical layer independently. **Modern DB2 example**: A DB2 DBA can reorganize a tablespace (changing the physical layout of data on disk), add or drop indexes, or even move data to different storage devices — all without changing a single line of application code. The SQL queries that applications use continue to work identically because they reference tables and columns (logical), not disk files and page layouts (physical). This is data independence in action, and it is so fundamental to modern databases that we take it for granted — which is precisely why it was revolutionary.Return to Chapter 1 | Continue to Case Study 1