Data Management Complete Study Guide

CDMP Certified Data Management Professional Data Management Examination Guide Data Exam Series Vol. 1 The Education &

Views 78 Downloads 0 File size 1MB

Report DMCA / Copyright

DOWNLOAD FILE

Recommend stories

Citation preview

CDMP Certified Data Management Professional

Data Management Examination Guide Data Exam Series Vol. 1

The Education & Research Affiliate of DAMA International

The Data Management Examination Guide

Acknowledgement is made for permission to use the full exam outline from the jointly developed DAMA International – Institute for Certification of Computing Professionals (ICCP) Data Management exam outline, copyright © 2006 ICCP. Written by Diane C. Johnson, PMP, for DAMA International & DAMA International Foundation. Published by DAMA International & DAMA International Foundation, Bellevue, WA, U.S.A. Data Management Examination Guide Data Exam Series, Vol. 1 To order copies, please contact

[email protected] PO Box 5786 Bellevue, WA 98006-5786 1-425-562-2636 www.dama.org

For exam administration questions, please contact

[email protected] or call 1.800.843.8227 2350 E. Devon Avenue, Suite 115, Des Plaines, IL 60018 USA www.iccp.org

Copyright © 2006 by DAMA International & DAMA International Foundation All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form, or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior consent of the publisher. ISBN 0-9676674-3-7 First Edition, April 2006

Page 2 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Table of Contents Introduction......................................................................................................................... 4 How Do You Obtain a CDMP? ...................................................................................... 5 CDMP Examination Criteria........................................................................................... 5 Additional CDMP Certification Criteria......................................................................... 6 Recommended Exams Based on Candidate’s Work Experience.................................... 8 Preparation for Taking Exams ........................................................................................ 9 Taking CDMP Exams ..................................................................................................... 9 Professional Development / Recertification ................................................................. 10 CDMP Contact Information.......................................................................................... 10 ICCP Data Management Specialty Examination Outline................................................. 12 How To Read The Data Management Subject Outline ................................................ 12 Data Management Exam Subject Outline..................................................................... 13 1.0 Data Management Function........................................................................................ 21 Overview....................................................................................................................... 21 Topics............................................................................................................................ 21 Questions....................................................................................................................... 22 Quick Answers.............................................................................................................. 33 Detailed Answers .......................................................................................................... 34 2.0 Data & Metadata Infrastructures Creation / Maintenance .......................................... 42 Overview....................................................................................................................... 42 Topics............................................................................................................................ 42 Questions....................................................................................................................... 43 Quick Answers.............................................................................................................. 51 Detailed Answers .......................................................................................................... 52 3.0 Data Analysis and Modeling....................................................................................... 57 Overview....................................................................................................................... 57 Topics............................................................................................................................ 57 Questions....................................................................................................................... 58 Quick Answers.............................................................................................................. 76 Detailed Answers .......................................................................................................... 77 4.0 Data / Metadata Infrastructure Management .............................................................. 88 Overview....................................................................................................................... 88 Topics............................................................................................................................ 88 Questions....................................................................................................................... 89 Quick Answers.............................................................................................................. 96 Detailed Answers .......................................................................................................... 97 5.0 Information Quality Management............................................................................. 102 Overview..................................................................................................................... 102 Topics.......................................................................................................................... 102 Questions..................................................................................................................... 103 Quick Answers............................................................................................................ 114 Detailed Answers ........................................................................................................ 115 Selected Bibliography..................................................................................................... 122

Page 3 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Introduction The Certified Data Management Professional (CDMP) credential validates knowledge and experience of Data Management Professionals. CDMP Credentials can be a doorway to opportunities either measuring your standing by demonstrating Mastery level or providing a starting point for new professions through a Practitioner designation. The Certified Data Management Professional (CDMP) credential is awarded to those who qualify based on a combination of criteria including education, experience and testbased examination of professional level knowledge. To maintain certified status and continued use of the credential, an annual recertification fee along with a 3-year cycle of continuing education and professional activity is required. The Data Management Association International (DAMA) authorizes the Certified Data Management Professional certification program and granting of the CDMP designation in partnership with the Institute for Certification of Computing Professionals (ICCP), which administers testing and recertification. The ICCP Data Management exam is meant to be an experience exam, meaning that it tests what you know at the time. This study guide is meant to be a refresher to test taking and the concepts behind data management. You can focus on the sections that you need to learn, or take the practise exam to see where your strengths lie. The study guide is broken down into the five major sections of the exam: 1.0 Data Management Function 2.0 Data & Metadata Infrastructures Creation / Maintenance 3.0 Data Analysis and Modeling 4.0.Data / Metadata Infrastructure Management 5.0 Information Quality Management The DAMA International Foundation welcomes feedback on this Study Guide, as revisions will occur in the future. We encourage you to let us know how you are using these materials and how they might be improved. Your comments can be sent to: Vice President of Education, [email protected]

Page 4 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Overview of the Certified Data Management Professional (CDMP) Certification Program How Do You Obtain a CDMP? The CDMP certification process is as follows: 1. 2. 3. 4. 5.

Fill out an application (from www.dama.org or www.iccp.org). Arrange to take the exam(s). Pass the IT Core exam (required). Take two specialty exams. One specialty exam must be taken from a. Data Management b. Data Warehousing c. Database Administration 6. Meet the experience and education qualifications. 7. Sign the ICCP code of ethics. There is a professional development / recertification aspect to keeping your certification current after you are certified. This recertification activity is handled through the ICCP office.

CDMP Examination Criteria Three ICCP exams must be passed with the following scores: Score Pass all exams at 50% or higher Pass all exams at 70% or higher

Credential Earned CDMP Practitioner Certificate CDMP Mastery Certificate

The CDMP Practitioner certification is awarded to professionals who scored above 50% on all three exams. These individuals can contribute as a team member on assigned tasks for they have a working knowledge of concepts, skills and techniques in a particular data specialization. The CDMP Mastery certification is awarded to professionals who scored 70% or higher on all three exams. These individuals have the ability to lead and mentor a team of professionals as they have mastered the concepts, skills and practices of their data specialization. Exams may be retaken to improve your score and go from the Practitioner to the Mastery certificate level. You may be able to substitute select vendor certifications for up to one specialty exam.

Page 5 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Additional CDMP Certification Criteria The following criteria must also be met in order to qualify for the CDMP: CDMP Criteria # Years Data Professional Work Experience Substitute Up to 2 Years –Bachelor or Master Degree in an appropriate discipline for Work Experience Recertification Required Continuing Professional Education/Activity Required ICCP Code of Ethics

CDMP Practitioner Certificate 2

CDMP Mastery Certificate 4+

2

2

Yes

Yes

120 hours every 3-year cycle

120 hours every 3-year cycle

Yes

Yes

Sample Qualifications for the CDMP Other qualifications may be accepted. Check with the DAMA contacts or ICCP office. Education Bachelor of Science Degree Major in: Computer Science Information Systems Management Information Systems Information and Communications Technology Major in another discipline with minor in any of the above Masters Degree Computer Science Information Systems Information Resource Management Information and Communications Technology MBA with concentration in one of the above

Page 6 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Work Experience Sample qualifying position/role titles: VP, Director, or Manager of Data Management Data Architect, Data Administrator, Data Analyst, Data Modeler Data Specialist, Database Administrator, Data Warehousing Analyst Systems Architect, Systems Analyst, Project Manager, Project Leader Business Analyst, Repository Analyst, Repository Architect

Professional Examinations The CDMP requires three ICCP exams: IT Core, one specified data oriented exam, and one other exam. If you already passed one or more ICCP exams, these exams can be used toward a CDMP if considered current by ICCP standards, and the exams are listed within your CDMP area of specialization. For information on your status, contact the ICCP. If you wish to demonstrate expertise in exam specialty areas specifically, the ICCP will issue Expert (Proficiency) Certificates for each specialty exam passed at 70% or higher. If you wish to know how these exams were developed, go to (http://www.iccp.org/iccpnew/iwg2.html). These exams are product and vendor neutral, and international in scope.

Page 7 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Suggested Exams Based on Candidate’s Work Experience The following table shows the Data Management areas by which ICCP exams are either suggested (S) or your choice (C) for a total of three exams. The IT Core exam is required for all candidates. Your work experience in the field will let you determine what exams you are best suited to pass. ICCP Exams

Mgmt

Architecture

Data Analysis & Design

DBA

Data Warehousing

Metadata / Repository Mgmt

Data / Information Quality (Future)

Req’d

Req’d

Req’d

Req’d

Req’d

Req’d

Req’d

S

S

S

C

C

S

S

Database Administration

C

C

S

C

C

C

Data Warehousing

C

C

C

S

C

C C

IT Core Specialty Exams Data Management

Integrated Project Mgmt

C

C

C

IT Management

C

C

C

Systems Development

C

Object Oriented Analysis & Design

C

C

C

C

Systems Security

C

Future ICCP Exams Business Intelligence & Analytics

S

Data & Information Quality Acceptable Exam Substitutes (Third Party)

C

C

C

C

C

C

S

C

C

C

C

C (future: e.g. MIT or Berkeley – DQ programs)

Page 8 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Accepted Vendor/Training Certifications Any of the following certifications may be substituted for one of the "candidate's choice" specialty exams required for the CDMP. Other certification programs may be accepted, but need to be evaluated. Check with the ICCP office or the DAMA contacts. IBM - IBM Certified Database Administrator - DB2 Universal Database - IBM Certified Advanced Database Administrator – DB2 Universal Database - IBM Certified Solutions Expert - DB2 Universal Database - IBM Certified Solutions Expert - DB2 Content Manager Information Engineering Services Pty Ltd - Certified Business Data Modeller Insurance Data Management Association (IDMA) - Certified Insurance Data Manager Microsoft - Microsoft Certified Database Administrator NCR (Teradata) - Teradata Certified Professional Oracle - Oracle (xx) Certified Professional - Oracle9i Database Administrator Certified Professional (for Practitioner Level CDMP) - Oracle9i Database Administrator Certified Master (for Mastery Level CDMP) Project Management Institute - Project Management Professional (PMP)

Preparation for Taking Exams There are various ways of learning the process of taking ICCP exams: • Sponsor ICCP Exam Review courses for your DAMA chapter membership • Refer to the exam subject outlines (at level 1 & 2) posted on http://www.iccp.org/iccpnew/outlines.html to become familiar with the subject coverage of each exam • Contact the ICCP for the CDMP Study Guide which covers all the exams in the CDMP program and has sample exams/questions for self-study • Contact DAMA International for the Data Management Exam Study Guide. Other individual data exam study guides are planned for the future. The ICCP exams are also offered at the DAMA International Symposiums.

Taking CDMP Exams ICCP Testing can be done anywhere in the world, with an approved ICCP Proctor to verify physical identity and supervise/invigilate the delivery of the examination.

Page 9 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

A DAMA chapter can set up exam sessions during their chapter meetings. What is needed is a volunteer proctor from the chapter. A proctor is an individual authorized by ICCP to oversee the writing of an exam by an ICCP exam taker. This person must meet specific guidelines (http://www.iccp.org/iccpnew/testing.html) and be willing to supervise the exam taker. The ICCP reserves the right to reject proposed proctors. Contact [email protected] or phone 847.299.4227 or 800.843.8227 if you require assistance in determining an appropriate proctor. The exams run off the USB drive of an individual’s laptop. There are 110 questions with 110 being scored and 10 are beta questions to complete in 90 minutes. You will not know which type of question you are answering. Questions and possible distracters (answers) are randomly listed in a different order for each exam taker. Therefore, although this guide contains sample questions that allow for “all or none of the above” type answers meant for study purposes, you will not find this type of answer to choose from on the actual exam. Computer based testing allows for the immediate scoring after the exam is taken. An ICCP Performance Profile is then available for downloading, and one will be sent later to the individual by the ICCP. This Profile shows your exam strengths and weaknesses.

Professional Development / Recertification To keep your CDMP current, you must earn 120 approved contact hours of continuing education over a 3-year period. Many educational activities count including DAMA Symposiums and chapter meetings. For further information, contact the ICCP ([email protected]) for an ICCP Recertification Guidelines Booklet or go to www.iccp.org/iccpnew/Recertification%20Guidelines2005.pdf. Recertification credits can be entered on an ICCP Educational Activity Form or through www.iccp.org/cgi-win/pdform.exe. Your DAMA chapter can also keep track of meeting attendance for the purpose of recertification and submit on a timely basis. There is an annual maintenance fee to ICCP for keeping track of your recertification credits. You will receive an annual transcript from the ICCP.

CDMP Contact Information For Questions on the CDMP Certification Program: Contact DAMA International ICCP Directors at: [email protected] or [email protected] To Order the DAMA Data Management Examination Guide: [email protected] P.O. Box 5786 Bellevue, WA 98006-5786 415-562-2636 Page 10 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

The DAMA website is www.dama.org for further information and an application. To Order the ICCP CDMP Study Guide or For Questions on CDMP Testing, Administration and Recertification: Contact the ICCP Office at: 847-299-4227 or 800-843-8227 (phone) 847-299-4280 (fax), or [email protected]. The ICCP website is www.iccp.org for further information and an application.

Page 11 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

ICCP Data Management Specialty Examination Outline How To Read The Data Management Subject Outline In the examination outline, a consistent set of syntax conventions has been used: • • •



Outline elements with numeric level leaders imply inclusivity. Concepts not within the numbered structure will not be tested. Outline elements with a “•” bullet leader are examples to clarify the content of a numbered element, and are not necessarily inclusive. Numbers in parentheses after an element name indicate the number of questions in the exam, which will be presented on the subject indicated by the element name and all subordinate elements. These allocations are guidelines established by the Test Management Council, and are followed as closely as possible in selecting questions for the exam. There are 100 multiple-choice questions on each exam version and this outline reflects this total. The characters “D#” after an element name indicate the target “depth” of questions to be posed on the subject indicated by the element name and all subordinate elements. The depths of knowledge are defined as follows: D1 D2 D3 D4 D5 D6

Recognition Knowing what a concept is called. Differentiation Knowing the external differences between a concept and a neighboring concept. Description Knowing the external characteristics of a concept. Usage Knowing how to use instances of the concept and why. Structure Knowing the internal structure of the concept — its components and the relationships among these components. Construction Knowing how to put together instances of the concept tailored to specific purposes.

Page 12 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Data Management Exam Subject Outline 1.0. Data Management Function

(18) Section 1 Total

1.1 Planning

(8)

1.1.1. • • • 1.1.2. • • • •

D4

Scope & Charter Data Management customer base Vision, goals, objectives Functions / services Data Management Plans Strategic data management plan (linked to business plan) Organizational structure plan and budgets Metadata management budgets, metrics, audits Data management oversight/control, e.g., data standards approval committee, technology change management committee, data management process change management committee, enterprise data management ‘board of directors’ • Enterprise data / information framework • Data portfolio management plan • Relationship management plan (vendor, customer, employee) • Data quality management plan • Data management process maturity improvement plan • Data and data management configuration management plan • Data and data management standards management plan 1.1.3. Policies / Standards / Processes / Procedures / Guidelines • Internal to data management organization • Customer data / metadata guidelines 1.2. Organization 1.2.1. • • 1.2.2. • • •

(2)

D3

(8)

D3

Types of Staff Training Orientation for new employees Continuing education for required skills, retraining Communication Marketing data services and benefits Customer education / training Publishing newsletters and web site news

1.3. Roles & Responsibilities 1.3.1. Data Administration • Data planning, policy development • Data architecture

Page 13 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

• • • • • 1.3.2. • • • • 1.3.3. • • • • • 1.3.4. • • • 1.3.5. • • • • • • • • 1.3.6 • • • •

Data requirements modeling: conceptual (entity types and their relationships), logical (ERA), and physical (database design) Data model management Data resource control and quality Standards management, setting, communication and enforcement Liaison with Database Administrators, Business Analysts, Management, Users Metadata Administration Metadata planning, policy development Metadata requirements gathering Metamodeling (metadata modeling) Metadata tool administration (metadata registries and repositories) Database Administration Definition and organization of physical database Protection and recovery of physical database Data archiving and deletion Optimization and documentation of physical databases Liaison with Data Administrators, Business Analysts, Management, Users Data Warehouse Administration Warehouse modeling, design, implementation, and operation Operational data store modeling, design, implementation, and operation Data access administration Information Stewardship Business information steward Managerial information steward Physical data trustee Originator of business rules Information producer Data quality accountability Metadata creation Information usage and knowledge worker stewardship Configuration Management Database Data models Data standards Metadata management tools

Page 14 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

2.0 Data & Metadata Infrastructures Creation / Maintenance (15) Section 2 Total 2.1 Planning for Data & Metadata 2.1.1 • • • • • • • • 2.1.2. • • • • •

D4

Architectures Enterprise Data Data Sourcing Data Distribution Data Integration Change Authorization Zachman Framework Data Processing Architectures (i.e. client-server, distributed data, etc.) Metadata Architectures Data Architecture Methods Information Engineering Enterprise Architecture Planning Data Life Cycle Data Reengineering Prototyping

2.2.Tools and Technology Types 2.2.1. • • • 2.2.2. • • • • • 2.2.3. • • •

(6)

(9)

Data Database Management Systems (DBMS & ODBMS) Data modeling tools Extract, transform, and load (ETL) tools Metadata & Descriptive Information Data dictionaries Data directories Data encyclopaedias Metadata registries (e.g. ISO/IEC 11179) Metadata repositories Data Issues Business intelligence technologies (OLAP, Data Mining, etc.) Data management and the Internet / Intranet Data management and unstructured data

Page 15 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

D3

The Data Management Examination Guide

3.0. Data Analysis and Modeling 3.1. Data / Metadata Analysis & Design 3.1.1. • • • • 3.1.2. • • • • • •

(6)

D5

Fact Finding Techniques Interviewing Surveys, questionnaires JAD sessions Legacy systems analysis Requirements Definition and Management Evaluation of current environment and documentation Future state Gap analysis Business rules (discovery, validation and documentation) Data / process matrices Requirements tracking and management to implementation

3.2. Data Model Components 3.2.1. • • • • • • • • • • 3.2.2. • • 3.2.3. • • • • 3.2.4. • • • •

(37) Section 3 Total

(21)

Logical Data Model Entity type Relationship type Attributes and their roles Definitions Key Cardinality Optionality Metadata type Rules: Business / data integrity Normalization Dimensional Warehouse Fact Dimension Object Oriented / UML Object Class type Attribute type Relationship type Data Representations in Process Models Business views / presentation level Trigger Stored procedure Object method

Page 16 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

D5

The Data Management Examination Guide

3.3. Data / Metadata Model Management 3.3.1. • • • • • • • • 3.3.2. • • • • 3.3.3. • • • • • • •

(10)

D5

Types of Data Models Conceptual Logical Physical Data warehouse Metamodels / meta-metamodels Universal / industry models Object class Data life cycle Scope of Model and Metadata Enterprise wide Business area Project oriented Subject area Data Model Support Creation Maintenance Version control Comparison Merging Importing / exporting Linkages and mappings between enterprise, logical, physical data models, and process models

Page 17 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

4.0.Data / Metadata Infrastructure Management 4.1. Standards, Policies, Procedures, Guidelines

(12) Section 4 Total (7)

D5

4.1.1. Standards Management Process • Awareness of external standards, e.g. ANSI and ISO/IEC data and data management related standards • Creation/identification • Approval • Enforcement • Maintenance 4.1.2. Data Models • Naming conventions for entities, relationships, attributes, etc. • Business and data integrity rules 4.1.3. Data Elements • Element types • Naming conventions • Metadata / definition principles • Legacy element linkages • Data element audit 4.2. Data Security and Privacy

(5)

4.2.1 Data Security Principles • Accountability • Authorization • Availability 4.2.2. Data Security Policy Types • Data stewardship and trustee responsibilities • Data and instance value access sensitivities (e.g. privacy, corporate confidentiality, data aggregation sensitivity issues) • Trans-border data flow • Data content

Page 18 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

D3

The Data Management Examination Guide

5.0. Information Quality Management 5.1. Information Quality Principles

(18) Section 5 Total (6)

D3

5.1.1. Definition • Data and Information • Information quality • Data definition as Information Product Specification • Data definition quality • Information architecture (data model) quality • Business drivers • Costs of nonquality information 5.1.2. Information Quality Characteristics • Conformance to definition • Completeness • Validity • Accuracy • Precision • Non duplication • Consistency of redundant data • Timeliness • Usefulness • Objectivity (of presentation) • Presentation clarity 5.1.3 Data Definition (or Information Product Specification) Quality Characteristics • Properly formed name, in accordance with approved naming convention standard • Standard, single enterprise abbreviations for new development • Name appropriate to knowledge workers • Correct, clear, and complete definition • Business term (used in data definition) defined in glossary • Correctly specified value domain and definition (of code values) • Properly defined data value type (not just alphanumeric, etc., but domain type (corresponding to class words, e.g., data, code, amount, ID, etc.) • Correct, complete, and useful business rule specification 5.2. Information Quality Assessment / Audit

(4)

5.2.1. Quality Assessment Characteristics • Data definition quality assessment process/techniques • Data model / requirements quality assessment process/techniques

Page 19 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

D4

The Data Management Examination Guide

5.2.2. Quality /Cost Measurement • Baseline data cost calculation (Measurement of cost of redundancy and interfaces) • Cost of non quality information • Value chain relationship between quality information and business drivers 5.3. Information Quality Improvement

(8)

5.3.1. Data Corrective Maintenance • Data correction of defective data • Redesign processes/systems producing poor quality data content or presentation 5.3.2. Data Movement Control • Mapping, transforming, data for data movement planning • Quality audit and control of data movement 5.3.3. Information Quality Process Improvement • Root Cause Analysis and Cause-and-Effect Diagrams • Shewhart Cycle (Plan-Do-Check-Act) process for improvement • Information defect prevention techniques 5.3.4. Information Quality Culture Transformation • Employee training in information quality techniques • Management accountability for information quality (managerial information stewardship • Information quality management maturity assessment • Gap analysis • Information quality performance measures

Page 20 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

D3

The Data Management Examination Guide

1.0 Data Management Function Overview The Data Management mission consists of goals and objectives that guide the creation, use, management and deletion of data across enterprise. Data Managements needs to be planned into overall Information Technology Strategy which is linked to the Business plan. To support the functions and services related to Data Management within the organization, a dedicated group of professionals with oversight committees need to be in place e.g., data standards approval committee, technology change management committee. An important aspect of data management oversight is to establish common Policies, Standards Procedures and Guidelines for data ontology. Support services of training and communication will further the knowledge and usage of Data Management within an organization. When new employees are hired, orientation sessions are required from the Data Management group to gain an understanding of the data environment and the use of data to do their job. For individuals involved in the dayto-day management of the data, the training should be comprehensive and potentially involve mentoring of the Data Policies, Standards, Procedures and Guidelines. Communication of data services and updates through the use of newsletters assists in keeping the organization up-to-date with the progress and accomplishments of the group. To support the Data Management Function, major roles and responsibilities for managing data need to be defined. There are no standards that define titles or team structure. Typically, there is a Manager role for the group that is responsible for planning, organizing and directing, plus may be a subject matter specialist. The job descriptions outlined in this section are based on standard guidelines of job titles and responsibilities based on experience.

Topics Data Management Planning Data Management Scope & Charter Data Management Plans Policies / Standards / Processes / Procedures / Guidelines Data Management Organization Types of Staff Training Communication Roles & Responsibilities Data Administration Metadata Administration Database Administration Data Warehouse Administration Information Stewardship Configuration Management

Page 21 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Questions 1.1. Planning 1.1.1

Scope & Charter

1. What is the primary mission/vision of Data Management? A. To facilitate the development, management, and use of the data resources as a vital asset. B. Committed to excellence. C. Database maintenance and enhancement for production application systems D. Data analysis and modeling for projects in planning or analysis 2. Which is not an objective of Data Management? A. Maintain the physical integrity and efficiency of data resources. B. Education about the benefits of and methods for enhancing data quality. C. Provide the architecture and guidelines for documenting and implementing data resources. D. Provide a cost effective and robust document and content management capabilities, workflow and business process management capabilities. 3. Which of the following is not a valid scope of Data Management Function? A. Requirements analysis and modeling B. Enterprise-wide data coordination, integration, stewardship and use C. Data security and quality D. Economies of scale in purchasing. 4. Which one of the following is not a typical type of service in a Data Management function? A. Support for warehouse initiatives. B. Database maintenance and enhancement for production application systems C. Database design for projects in development D. Data analysis and modeling for projects in planning or analysis

1.1.2 Data Management Plans 5. The goal of the Data Management Plan is to describe the resources and process used to ensure high quality data. The Data Management Plan is usually part of what overall strategy? A. Information Technology Strategy B. Information Technology Infrastructure Strategy C. Application Infrastructure Plan D. Information Technology Architecture Plan Page 22 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

6. A Data Management Plan is usually part of what overall strategy? A. Information Technology Strategy B. Information Technology Infrastructure Strategy C. Application Infrastructure Plan D. Information Technology Architecture Plan 7. Which one of the following is not a part of a Data Management Plan? A. Describe the roles and resources of program staff. B. Define future direction of data management activities in a work plan. C. Implement facilities and tools for managing metadata resources. D. Development of a quality management plan. 8. Which Committee is not an oversight committee regarding Data Management? A. Data Standards Approval Committee B. Data Management Process Change Management Committee C. Enterprise Data Management Board of Directors D. Project Change Committee 9. What is the purpose of conducting a metadata management audit? A. To ensure metadata management controls have achieved intended results. B. Provide an additional source of information for the budget. C. To determine that decisions made in a timely fashion with appropriate criteria/guidance that uses all necessary data/information from automated systems and, if applicable, users. D. To determine that all appropriate policies/procedures been developed, disseminated, kept up to date, and test checks made to ensure compliance. 10. Benchmarking will provide comparative information. Which of the following is not a result of benchmarking? A. Efficiency and how well resources and services (outputs) are delivered B. Quality of services and extent to which customers are satisfied (outcomes) C. Measure for evaluating outcomes or the results of program activity compared to its intended purpose and program objectives D. Best practices 11. Which role is not typically involved in a data management oversight committee? A. Program Director/Manager B. Users C. Database Administrator D. Data Analysts Page 23 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

12. Which one of the following is not a benefit of a strong data portfolio management program? A. Maximize value of IT data investments while minimizing the risk B. Improve communication and alignment between technology and business. C. Encourage reuse of data throughout the organization. D. Allow planners to schedule resources more efficiently. 13. Which one of the following is least desirable benefit of Enterprise data / information framework? A. Provides enterprise-wide definitions of concepts and data. B. Provides a scoping tool for new initiatives. C. Reduces data redundancy by providing transparency as to the meaning of data items D. Encourages re-use and consistent data structures across the enterprise 14. What party would not be considered when creating a relationship management plan? A. Vendor / Supplier B. Customer C. Employee D. President 15. A Relationship Management Plan when dealing with vendors / suppliers should be part of which overall strategy? A. Procurement Strategy B. Quality Management Strategy C. Enterprise Architecture Strategy D. IT Strategy 16. Which area does a Quality Management Plan does not address? A. Quality policies and procedures. B. Roles, responsibilities and authorities. C. Description of quality system. D. Meta-metamodel 17. Which one of the following is not true when describing Capability Maturity Model Integration (CMMI)? A. Model framework to assess data and process maturity. B. Model framework to determine priorities. C. Model framework to institute process and data improvement. D. Defines six levels of process maturity. Page 24 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

18. What level is organizations CMMI maturity, if the data management requirements are not being met? A. Level 0 B. Level 1 C. Level 2 D. Level 3 19. What level is organizations CMMI maturity, if the data management requirements are being met? A. Level 0 B. Level 1 C. Level 2 D. Level 3 20. What level is organizations CMMI maturity, if the data management requirements are being managed and tracked? A. Level 1 B. Level 2 C. Level 3 D. Level 4 21. What level is organizations CMMI maturity, if the data management requirements meet EIA Standard 859 Industry Standard for Data Management that includes nine high level Data Management Principles? A. Level 1 B. Level 2 C. Level 3 D. Level 4 22. Which one is not the purpose of the data management configuration management plan? A. Identify and describe the overall policies and methods for Configuration Management. B. Establish and provide the basis for a uniform and concise Configuration Management practice C. Manage the data for its entire lifecycle. D. Retain data commensurate with value.

Page 25 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

23. Who is responsible for developing and implementing Data-Management planning for projects, for ensuring that the activities are completed according to agreed standards and timelines and for coordinating ongoing data management to support the business? A. Data Manager B. Data Analyst C. Database Administrator D. Business Manager

1.1.3 Policies / Standards / Processes / Procedures / Guidelines 24. What is the following statement: Data archives must include easily accessible information about the data holdings, including quality assessments, supporting ancillary information, and guidance and aids for locating and obtaining the data? A. Policy B. Standard C. Procedure D. Guideline 25. What is the following statement: Contact Information offers data groupings that are used to describe a point of contact, address, and communication information? A. Policy B. Standard C. Procedure D. Guideline 26. What is the following statement: To keep the hard drives from getting full, please back-up your data. 1. Put CD data you want to back up in one folder. The name of the folder is the name of the CD. 2. Start the "Backup" program. 3. Click on "Data." 4. Move the data you want to back up in "Data window." 5. Click on "Done." 6. Put the CD in the CD-R and close the door. 7. Write the CD. A. Policy B. Standard C. Procedure D. Guideline

Page 26 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

27. What is the following statement: Aggregation of data values is appropriate for fields with a large numbers of values, such as dates, age, and geographic areas; it is the primary method used to collapse a dataset in order to create tables with no small numbers as denominators or numerators in cells? A. Policy B. Standard C. Procedure D. Guideline 28. What is the following statement: Data custodians are responsible for creating and maintaining metadata for their datasets? A. Policy B. Standard C. Procedure D. Guideline 29. Which of the following is the best answer for the definition of cost when following the metadata procedure, to “state what the concept is, not only what it is not”. A. Total spent for goods or services including money and time and labor. B. Cost is a price paid. C. Costs, which are not related to external costs. D. Direct cost to the business owner of those items, which will be sold to customers.

1.2. Organization 1.2.1. Types of Staff Training 30. Which one of the following is not appropriate for an orientation of the data environment for new employees? A. Acronym list. B. Customer Service Policy. C. Data Policy and Procedure. D. WWW Design and Programming. 31. When embarking on continuing education for required skills or retraining, which training method is least desirable? A. Mentoring with another employee. B. Workshops and seminars. C. Classroom or computer based courses. D. Booklets and information sheets.

Page 27 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

1.2.2. Communication 32. Which one of the following is the least beneficial when promoting data services? A. Communicating data services and benefits. B. Publishing newsletters and web site news. C. Customer education and training. D. Convening a meeting of the Organizations Management Team. 1.3. 1.3.1.

Roles & Responsibilities Data Administration

33. Who is responsible for identifying and analyzing information needs for the enterprise or business area, and develops and maintains data architecture? A. Data Administrator B. Manager, Data Administration C. Data Administration Consultant D. Database Administrator 34. Which one is not a responsibility of the Data Administrator? A. Identify and analyze customer information needs. B. Develop and maintain data architecture. C. Develop and maintain strategic data plan. D. Provide approval authority over metadata policies and design. 35. When hiring a Data Administrator which skill is the least preferred? A. Relational Database experience. B. Logical and Physical Data Modeling. C. Project Management experience. D. Strong written and oral communication skills. 36. Which role would a Data Administrator not typically interact? A. Business Analyst B. CEO C. Repository Administrator D. Management

Page 28 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

37. Who is responsible for planning, organizing, directing and controlling data definition, data use, and ensure data availability for the enterprise? A. Data Administrator B. Manager, Data Administration C. Metadata Administrator D. Database Administrator

1.3.2. Metadata Administration 38. Who is responsible for creating, administrating and enforcing of standards, guidelines and procedures for the use of metadata? A. Data Administrator B. Manager, Data Administration C. Metadata Administrator D. Database Administrator 39. Which is not a responsibility of the Metadata Administrator role? A. Establish and maintain the metadata architecture. B. Provide approval authority over metadata policies and design. C. Maintain repository security profiles. D. Provide final review and approval authority over data design for an application system. 40. In a company with a Metadata team, which role would collect the requirements and design the metadata solution? A. Metadata Administrator. B. Manager, Metadata Administration. C. Metadata Analyst. D. Metamodelers.

1.3.3. Database Administration 41. Which responsibility is not typically a responsibility of the Database Administrator? A. Establish and maintain sound backup and recovery policies and procedures. B. Implement and maintain database security (create and maintain users and roles, assign privileges). C. Perform database tuning and performance monitoring. D. Perform application tuning and performance monitoring.

Page 29 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

42. Which role would a Database Administrator not typically interact? A. Business Analyst B. CEO C. Data Administrator D. Management 43. When hiring a Database Administrator which skill is the least preferred? A. Relational Database, related utilities and tools experience. B. Physical Data Modeling. C. Ability to perform both Relational Database and Operating System performance tuning and monitoring. D. Network security administration. 44. Which role in an organization would develop the referential integrity constraint scripts? A. Data Administrator B. Manager, Data Administration C. Data Analyst D. Database Administrator 45. Who has the responsibility to recover the physical database in the event of a power disruption? A. Data Administrator B. Manager, Data Administration C. Data Analyst D. Database Administrator

1.3.4 Data Warehouse Administration 46. Which responsibility is not typically a responsibility of the Data Warehouse Administrator? A. Data Warehouse data modeling and design. B. Data Warehouse implementation and refresh. C. Data Access administration. D. Installing the Operating System on the Data Warehouse server.

Page 30 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

47. When hiring a Data Warehouse Administrator which skill is the least preferred? A. Relational Database, related utilities and tools experience. B. Ability to calculate Data Warehouse return on investment, costs and benefits. C. Expert in data structure including parallel data structure. D. Logical and Physical Data Modeling.

1.3.5. Information Stewardship 48. Which role would review and approve data definitions and domain value specifications for business data? A. Business Information Steward B. Managerial Information Steward C. Physical Data Trustee D. Information Producer 49. Who has authority to select and mandate Business Information Stewards? A. Business Information Steward B. Managerial Information Steward C. Physical Data Trustee D. Information Producer 50. Which role is not responsible for data quality? A. Business Information Steward or Managerial Information Steward B. Physical Data Trustee C. Information Producer D. Everyone is responsible for data quality. 51. Which one is not a responsibility of the Physical data trustee? A. Creation of data standards. B. Enforcement of physical security. C. Performance tuning of physical databases. D. Backup and Recovery of physical databases. 52. Who should be responsible for creating the data dictionary entries? A. Repository Administrator B. End User C. Data Modeler D. Data Librarian

Page 31 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

1.3.6

Configuration Management

53. A new data model is created and rolled into Production. Which process is responsible for registering the modification in the Configuration Management Database (CMDB)? A. Change Management B. Configuration Management C. Problem Management D. Release Management 54. Which of the following is a Configuration Item (CI)? A. Organization Structure B. Data Model C. An incident D. A process 55. Which one is not a discipline of Data Management Configuration? A. Status Accounting B. Collection C. Approval D. Distribution 56. Which item is not a responsibility of the Configuration and Data Management team? A. Management of all documentation and specifications. B. Configuration and data management of programs. C. Maintaining requirements of deliverables through the data change process. D. Providing storage, retrieval, distribution, and management of program data.

Page 32 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Quick Answers 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19.

A D D A A A C D A D C D B D A D D A B

20. 21. 22. 23. 24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38.

B C D A A B C D A A D D D A D C B B C

39. 40. 41. 42. 43. 44. 45. 46. 47. 48. 49. 50. 51. 52. 53. 54. 55. 56.

D C D B D D D D B A B D A B B B A A

Page 33 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Detailed Answers 1. Answer: A. To facilitate the development, management, and use of the data resources as a vital asset. The primary mission of Data Management is to facilitate the development, management, and use of the data resources as a vital asset. The services of Data Management comprise of the following: Database maintenance and enhancement for production application systems and Data analysis and modeling for projects in planning or analysis. 2. Answer: D. Provide a cost effective and robust document and content management capabilities, workflow and business process management capabilities. Objectives of Data Management are: Maintain the physical integrity and efficiency of data resources; education about the benefits of and methods for enhancing data quality; and provide the architecture and guidelines for documenting and implementing data resources. 3. Answer: D. Economies of scale in purchasing. The scope of Data Management function include: Requirements analysis and modeling; Enterprise-wide data coordination, integration, stewardship and use; and data security and quality. 4. Answer: A. Support for warehouse initiatives. Data Management services include data maintenance and enhancement for production application systems; Database design for projects in development; and Data analysis and modeling for projects in planning or analysis. 5. Answer: A. Information Technology Strategy. A Data Management Plan is usually part of the overall Information Technology Strategy. The Information Technology Strategy leads to Infrastructure Strategy, Information Technology Architecture Plan and Application Infrastructure Plan. 6. Answer: A. Information Technology Strategy. A Data Management Plan is usually part of the overall Information Technology Strategy. The Information Technology Strategy leads to Infrastructure Strategy, Information Technology Architecture Plan and Application Infrastructure Plan. 7. Answer: C. Implement facilities and tools for managing metadata resources. Data Management Plans are high level and describe the roles and resources of program staff, define future direction of data management activities in a work plan and the development of a quality management plan. 8. Answer: D. Project Change Committee. Data Management oversight committees have been called: Data Standards Approval Committee; Data Management Process Change Management Committee; and Enterprise Data Management Board of Directors. Project Change Committee refers to changes made to project scope, time, or cost. 9. Answer: A. To ensure metadata management controls have achieved intended results. The purpose of conducting a metadata management audit is to ensure metadata Page 34 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

management controls have achieved intended results as systematic and proactive measures. Conducting a Metadata Management Control Review assists managers in developing and implementing appropriate, cost-effective management controls for results-oriented metadata management; assessing the adequacy of metadata management controls in programs and operations; identify needed improvements; take corresponding corrective action; and report on the status of metadata management control improvements. 10. Answer: D. Best practices. The results of benchmarking will measure: efficiency and how well resources and services (outputs) are delivered; quality of services and extent to which customers are satisfied (outcomes); and measure for evaluating outcomes or the results of program activity compared to its intended purpose and program objectives. Only one benchmark will not provide comparative information. After measuring over time, will benchmarks will provide comparative information. 11. Answer: C. Database Administrator. Typically the data management oversight committee comprises of Program Director/Manager, Users, and Data Analysts. A Database Administrator is not typically part of the data management oversight committee. 12. Answer: D. Allow planners to schedule resources more efficiently. A strong data portfolio management program 13. Answer: B. Provides a scoping tool for new initiatives. The least desirable benefit of Enterprise data / information framework is provides a scoping tool for new initiatives. New initiatives are first scoped by the business needs to gain the requirements. Enterprise data / information framework: Provides an enterprise-wide definitions of concepts and data; Reduces data redundancy by providing transparency as to the meaning of data items; and Encourages re-use and consistent data structures across the enterprise. 14. Answer: D. President. When creating a relationship management plan, the vendor, customer and employee should be considered. A relationship management plan should be approved for each project and the goal is to improve consistency in the way we approach relationships. 15. Answer: A. Procurement Strategy. A Relationship Management Plan when dealing with vendors / suppliers should be part of the Procurement Strategy. The Procurement Strategy should address when and how potential suppliers are to be engaged in the development of Relationship Management Plans. 16. Answer: D. Meta-metamodel. The Quality Management Plan addresses: quality policies and procedures; roles, responsibilities and authorities; and description of quality system. 17. Answer: D. Defines six levels of process maturity. The Capability Maturity Model defines five levels of process maturity; Model framework to assess data and process Page 35 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

maturity; Model framework to determine priorities; and Model framework to institute process and data improvement. 18. Answer: A. Level 0. If the data management requirements are not being met, the organization is at CMMI maturity level 0. CMMI Levels: 0 – not performed, 1 – initial, 2 – managed (disciplined process); 3 Defined (Standard, consistent process); 4 – Quantitatively Managed (Measured Predictable Process); 5 – Optimizing (Continuously Improving Process). 19. Answer: B. Level 1. If the data management requirements are being met, the organization is at CMMI maturity level 1. CMMI Levels: 0 – not performed, 1 – initial, 2 – managed (disciplined process); 3 Defined (Standard, consistent process); 4 – Quantitatively Managed (Measured Predictable Process); 5 – Optimizing (Continuously Improving Process). 20. Answer: B. Level 2. If the data management requirements are being met, the organization is at CMMI maturity level 2. CMMI Levels: 0 – not performed, 1 – initial, 2 – managed (disciplined process); 3 Defined (Standard, consistent process); 4 – Quantitatively Managed (Measured Predictable Process); 5 – Optimizing (Continuously Improving Process). 21. Answer: C. Level 3. If the data management requirements meet EIA standard 859 Industry Standard for Data Management, the organization is at CMMI maturity level 3. CMMI Levels: 0 – not performed, 1 – initial, 2 – managed (disciplined process); 3 Defined (Standard, consistent process); 4 – Quantitatively Managed (Measured Predictable Process); 5 – Optimizing (Continuously Improving Process). EIA standard 859 includes nine high level Data Management Principles. The principles address functions of Data Management: 1. Define the organizationally relevant scope of Data Management. 2. Plan for, acquire, and provide data responsive to customer requirements. 3. Develop DM processes to fit the context and business environment in which they will be performed. 4. Identify data products and views so their requirements and attributes can be controlled. 5. Control data repositories, data products, data views, and metadata using approved change control process. 6. Establish and maintain an identification process for intellectual property, proprietary, and competition-sensitive data. 7. Retain data commensurate with value. 8. Continuously improve data management. 9. Effectively integrate data management and knowledge management. 22. Answer: D. Retain data commensurate with value. The purpose of the data management configuration management plan is to identify and describe the overall policies and methods for Configuration Management; Establish and provide the basis for

Page 36 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

a uniform and concise Configuration Management practice; and Manage the data for its entire lifecycle. 23. Answer: A. Data Manager. A data manager is responsible for developing and implementing Data-Management planning for projects, for ensuring that the activities are completed according to agreed standards and timelines and for coordinating ongoing data management to support the business, known as the data management standards management plan. 24. Answer: A. Policy. The sentence is a Policy statement: Data archives must include easily accessible information about the data holdings, including quality assessments, supporting ancillary information, and guidance and aids for locating and obtaining the data. A Policy is a prescribed or proscribed course of action or behavior, which is to be followed with respect to the acquisition, deployment, implementation or use of information technology resources. It is not a standard, as it does not outline a specific technical approach. It is not a procedure, as it does not offer a set of administrative instructions for implementation of a policy or standard. It is not guideline that should offer a detailed plan or explanation to guide you in setting standards or determining a course of action. 25. Answer: B. Standard. The sentence is a Standard statement: Contact Information offers data groupings that are used to describe a point of contact, address, and communication information. Standard(s) is a prescribed or proscribed specific technical approach, solution, methodology, product or protocol which must be adhered to in the design, development, implementation or upgrade of data architecture. Standards are intended to establish uniformity in data. Standards should be designated as either "preferred" or "mandatory". It is not a procedure, as it does not offer a set of administrative instructions for implementation of a policy or standard. It is not a guideline, which should offer a detailed plan or explanation to guide you in setting standards or determining a course of action. 26. Answer: C. Procedure. The sentence is a Procedure statement: To keep the hard drives from getting full, please back-up your data. Procedure is a set of administrative instructions for implementation of a policy or standard. It is not a guideline, which should offer a detailed plan or explanation to guide you in setting standards or determining a course of action. 27. Answer: D. Guideline. The sentence is a Guideline statement: Aggregation of data values is appropriate for fields with a large numbers of values, such as dates, age, and geographic areas; it is the primary method used to collapse a dataset in order to create tables with no small numbers as denominators or numerators in cells. A guideline offers a detailed plan or explanation to guide you in setting standards or determining a course of action. 28. Answer: A. Policy. The sentence is a Metadata Policy statement: Data custodians are responsible for creating and maintaining metadata for their datasets A Policy is a Page 37 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

prescribed or proscribed course of action or behavior, which is to be followed with respect to the acquisition, deployment, implementation or use of information technology resources. It is not a standard, as it does not outline a specific technical approach. It is not a procedure, as it does not offer a set of administrative instructions for implementation of a policy or standard. It is not guideline that should offer a detailed plan or explanation to guide you in setting standards or determining a course of action. 29. Answer: A. Total spent for goods or services including money and time and labor. When creating metadata definitions a guideline to put in place according to ISO/IEC 11179, is to “state what the concept is, not only what it is not”. 30. Answer: D. WWW Design and Programming. When creating an orientation of the data environment for new employees it is appropriate to have items like: Acronym list, Customer Service Policy, Data Policy and Procedure, Data Standards, Data Guidelines, IT Strategy, and potentially even Data Models where appropriate. WWW Design and Programming would be most beneficial to Analyst and Programmers. 31. Answer: D. Booklets and information sheets. When embarking on continuing education for required skills or retraining the least desirable method of training is booklets and information sheets due to the non-interactiveness of the information. 32. Answer: D. Convening a meeting of the Organizations Management Team. When promoting data services, the goal is to publicize the efforts to the Organizations data customers. Convening a meeting of the Organization Management Team only reaches a limited audience and they may not be suitable people. Using a wide variety of marketing and communication vehicles will assist in targeting the message like: Marketing data services and benefits; Customer education / training; and Publishing newsletters and web site news. 33. Answer: A. Data Administrator. The Data Administrator identifies and analyzes information needs for the enterprise or business area, and develops and maintains data architecture plus the strategic data plan. The Data Administrator provides project support for data processing applications like data modeling and designing physical database. The Metadata Administrator is responsible for creating, administrating and enforcing of standards, guidelines and procedures for the use of metadata plus metadata query and analysis tools. The Manager, Data Administration is responsible for planning, organizing, directing and controlling data definition, data use, and ensuring data availability for the enterprise. 34. Answer: D. Provide approval authority over metadata policies and design. The Data Administrator identifies and analyzes information needs for the enterprise or business area, and develops and maintains data architecture plus the strategic data plan. The Data Administrator provides project support for data processing applications like data modeling and designing physical database. A Metadata Administrator would provide approval authority over repository policies and design. A Metadata Administrator would work with the Data Administrator. Additional responsibilities for the Data Administrator Page 38 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

include developing and enforcing standards for metadata through the review of definitions. Also, assist in developing procedures and data management policies that ensure the integrity, consistency and control of the enterprise's data resource. 35. Answer: C. Project Management experience. When hiring a Data Administrator the skill that may be desirable but least preferred in the above list is Project Management experience. Typically, skills for Data Administrators are: Relational Database experience, logical and physical data modeling, strong written and oral communication skills, strong analysis skills and prior work experience. 36. Answer: B. CEO. The Data Administrator would not typically interact with the CEO in an organization. The Data Administrator would interact with Business Analysts, Repository Administrator, Management and Users/Customers of the data. 37. Answer: B. Manager, Data Administration. The Manager, Data Administration is responsible for planning, organizing, directing and controlling data definition, data use, and ensure data availability for the enterprise. The Data Administrator identifies and analyzes information needs for the enterprise or business area, and develops and maintains data architecture plus the strategic data plan. The Metadata Administrator is responsible for creating, administrating and enforcing of standards, guidelines and procedures for the use of metadata plus metadata query and analysis tools. 38. Answer: C. Metadata Administrator. The Metadata Administrator is responsible for creating, administrating and enforcing of standards, guidelines and procedures for the use of metadata plus metadata query and analysis tools. The Manager, Data Administration is responsible for planning, organizing, directing and controlling data definition, data use, and ensure data availability for the enterprise. A Database Administrator conducts data store modeling, design, implementation, and operation. 39. Answer: D. Provide final review and approval authority over data design for an application system. The Metadata Administrator role would: Establish and maintain the metadata architecture; Provide approval authority over metadata policies and design; and Maintain repository security profiles in addition to Metadata tool administration. 40. Answer: C. Metadata Analyst. In a company with a Metadata team, the Metadata Analyst would collect the requirements and design the metadata solution. The Metamodeler would convert the requirements into metamodels. The Metadata Administrator is responsible for creating, administrating and enforcing of standards, guidelines and procedures for the use of metadata plus metadata query and analysis tools. The Manager, Data Administration is responsible for planning, organizing, directing and controlling data definition, data use, and ensuring data availability for the enterprise. 41. Answer: D. Perform application tuning and performance monitoring. A Database Administrator would be responsible for: Establish and maintain sound backup and recovery policies and procedures; Implement and maintain database security (create and maintain users and roles, assign privileges); Perform database tuning and performance Page 39 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

monitoring; Capacity planning; Perform general technical trouble shooting and give consultation to development teams. 42. Answer: B. CEO. The Database Administrator would not typically interact with the CEO in an organization. The Database Administrator would interact with Business Analysts, Data Administrator, Management and Users/Customers of the data. 43. Answer: D. Network security administration. When hiring a Database Administrator the skill that may be desirable but least preferred in the above list is Network security administration experience. Typically, skills for Database Administrators are: Relational Database; related utilities and tools experience, physical data modeling; ability to perform both Relational Database and Operating System performance tuning and monitoring; and prior work experience. 44. Answer: D. Database Administrator. In an organization, the Database Administrator would develop the referential integrity constraint scripts. The Data Analyst would work with the Database Administrator to link the logical to physical data model. 45. Answer: D. Database Administrator. In an organization, the Database Administrator would have the responsibility to recover the physical database in the event of a power disruption. 46. Answer: D. Installing the Operating System on the Data Warehouse server. A Data Warehouse Administrator would be responsible for: Data Warehouse data modeling and design; Data Warehouse implementation and refresh; Data Access administration; Perform application performance monitoring; Perform general technical trouble shooting and give consultation to development and metadata teams. 47. Answer: B. Ability to calculate Data Warehouse return on investment, costs and benefits. When hiring a Data Warehouse Administrator the skill that may be desirable but least preferred in the above list is the ability to calculate Data Warehouse return on investment, costs and benefits. Typically, skills for Data Warehouse Administrators are: Relational Database; related utilities and tools experience, logical and physical data modeling; Expert in data structure including parallel data structure; Extract Transform and Load tool experience; and prior work experience. 48. Answer: A. Business Information Steward. Business Information Steward would review and approve data definitions and domain value specifications for business data. Other responsibilities would include: validating business rules and keeping the domain values current across the Enterprise. Managerial Information Steward is responsible for setting information policy and creating information measures for either the organization or a specific department or business area. Physical Data Trustee is accountable for the integrity of the physical data assets. An Information Producer is accountable for the content of the information.

Page 40 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

49. Answer: B. Managerial Information Steward. The Managerial Information Steward has the authority to select and mandate Business Information Stewards. Physical Data Trustee is accountable for the integrity of the physical data assets. An Information Producer is accountable for the content of the information. 50. Answer: D. Everyone is responsible for data quality. Everyone in an organization is responsible for data quality. While some roles, like a CEO may not be involved in data entry or usage of transactional systems, they would offer support both financial and policy. 51. Answer: A. Creation of data standards. The Physical data trustees are responsible for enforcement of physical security, performance tuning of physical databases, and backup and recovery of physical databases. 52. Answer: B. End User. The End User would be responsible for creating the data dictionary entries. The Repository Administrator would set up the structure and implement the data dictionary. The Data Librarian would assist in cataloguing and categorizing the data. 53. Answer: B. Configuration Management. A new data model is created and rolled into Production. The process that is responsible for registering the modification in the Configuration Management Database (CMDB) is Configuration Management. Configuration and Data Management organizations are responsible for defining, controlling, integrating and implementing essential policies and procedures that provide Configuration Management (CM) and Data Management (DM) discipline on Programs and contracts. Configuration Management Database is a database, which contains all relevant details of each Configuration Item (CI) and details of the important relationships between CIs. 54. Answer: B. Data Model. An example of a Configuration Item is data model. A Configuration item is a component of an infrastructure, or an item associated with infrastructure that needs to be managed and controlled by Configuration Management. 55. Answer: A. Status Accounting. The disciplines of Data Management Configuration are: Planning, Collection, Approval, Distribution, Storage, and Retrieval of data that are implemented through standard procedures and address specific customer or contractual data management requirements. 56. Answer: A. Management of all documentation and specifications. The only documents that should be managed through Configuration Management are those that relate to: hardware, software, firmware, data model, documentation, test, test fixtures and test documentation of an automated information system, throughout the life of a system.

Page 41 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

2.0 Data & Metadata Infrastructures Creation / Maintenance Overview An overall planning function needs to occur to govern the creation and maintenance of data and metadata infrastructures. The scope of the data and metadata infrastructure includes the definition of all activities and processes involved in the definition, creation, formatting, storage, access and maintenance of data and metadata. Data Architecture Methods defines the process to create infrastructure like using Information Engineering. Information Engineering has many purposes, including organization planning, business re-engineering, application development, information systems planning and systems reengineering. There are a number of data and metadata tools that perform various tasks associated with managing and deploying Data Management that provide flexibility needed to support the activities and processes defined in the infrastructure. Data Management tools include Database Management Systems (relational and object), Data modeling tools and Extract, transform, and load (ETL) tools. Metadata tools include Data dictionaries, Data directories, Data encyclopaedias, Metadata registries (e.g. ISO/IEC 11179) and Metadata repositories. The ISO/IEC 11179 Information Technology: Metadata Registries (MDR) specification developed by the ISO (International Standards Organization) and the IEC (the International Electrotechnical Commission) defines a number of fields and relationships for Metadata Registries including a detailed metamodel for defining and registering administered items, of which the primary component is a Data Element. Each data and metadata tool has a different purpose and usage.

Topics Data Architecture Methods Architectures Data Architecture Methods Tools and Technology Types Data . Metadata & Descriptive Information . Data Issues

Page 42 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Questions 2.1

Planning for Data & Metadata

2.1.1

Architectures

1. Which of the following is not a valid benefit of Enterprise Data Architecture? A. Organizes data around the enterprise's data subjects to create a shared data resource. B. Integrated view of enterprise data. C. Economies of scale in purchasing Case tools. D. Enables organizational change. 2. What is the best answer to why would Enterprise Data Architecture be created? A. Enterprise Data Architecture can be created in one iteration. B. Diagram application-specific databases. C. Information is an asset of the entire organization. D. Design stability and data object abstraction and generalization. 3. Which of the following is not a reason to architect Source Data? A. Determine sources of data needed. B. Determine the index for the data mart. C. Diagram application-specific source data for extraction. D. Determine methods for extraction and delivery. 4. Which one of the following is not a goal of Source Data Architecture? A. Ensure that the source data is extracted only once. B. Define the scope and implementation of the data warehouse. C. Oversee the construction of the enterprise data warehouse. D. Determine the monthly flat file transmission protocol. 5. Which one of the following is targeted towards the efficient delivery of the proper information to the proper recipients? A. Data Sourcing B. Data Distribution C. Data Integration D. Enterprise Data

Page 43 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

6. Which one of the following requires combining and matching information from different sources, and resolving a variety of conflicts? A. Data Sourcing B. Data Distribution C. Data Integration D. Enterprise Data 7. What is a fundamental principle in Change Authorization of Architectures? A. Single point of authorization. B. Single point of access. C. Private key encryption on data. D. Standard for communication. 8. Which of the following is not a feature of the client in client-server architecture? A. Passive B. Active C. Sending requests D. Waits until reply arrives 9. Which of the following is not a feature of the server in client-server architecture? A. Passive B. Active C. Waiting for requests D. On requests serves them and send a reply 10. What is the best name for a network called if the networks consists clients, application servers which process data for the clients, and database servers, which store data for the application servers? A. 2-tier Architecture B. 3-tier Architecture C. n-tier Architecture D. Multi-tier Architecture 11. Which one of the following is not a source of metadata? A. Case Tools B. Applications C. Physical Database D. Company Directory

Page 44 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

12. Which type of analysis is needed when beginning Metadata solution architecture? A. Metadata record B. Metadata flows C. Metadata Categorization D. Metadata user 13. What is best definition of the Zachman Framework? A. A 36-cell Matrix. B. A Normalized schema. C. A good analytical tool. D. Specific to methods/tools.

2.1.2. Data Architecture Methods 14. Which one is not a benefit of Enterprise Architecture Planning? A. Consistency and compatibility of systems. B. Interoperability between systems and databases. C. Economies of scale in purchasing and developing systems. D. Greater accounting staff effectiveness. 15. Which one of the following does Enterprise Architecture Planning does not address? A. Data management B. Application environment and development toolsets C. Maintain a secure infrastructure and IT support for networks and distributed systems D. Middle-ware and transaction management 16. An Enterprise Architecture Plan is usually part of what overall strategy? A. Information Technology Strategy B. Information Technology Infrastructure Strategy C. Application Infrastructure Plan D. Information Technology Architecture Plan 17. Which one of the following phases is not part of the data life cycle? A. Create/Store B. Modify/Update C. Delete D. Shred

Page 45 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

18. What is the definition of data life cycle management? A. Data Life cycle management is a product. B. Data Life cycle management is an approach to managing an organization's data. C. Data Life cycle describes the CRUD matrix of data elements. D. Data Life cycle is the storage used to store active and inactive data. 19. What is defined as "An integrated and evolutionary set of tasks and techniques that enhance business communication throughout an enterprise enabling it to develop people, procedures and systems to achieve its vision". A. Information Engineering B. Enterprise Architecture Planning C. Data Reengineering D. Prototyping 20. Which one is not a purpose of Information Engineering? A. Organization planning. B. Business re-engineering. C. Application development. D. Data Warehousing. 21. In which situation would a data reengineering apply? A. Developing a migration strategy from one application environment to another. B. Determining active and inactive data. C. Enhance business communication throughout the enterprise. D. Assist in developing the information management strategy. 22. What is the process of quickly putting together a working model in order to test various aspects of the design, illustrate ideas or features and gather early user feedback? A. Information Engineering B. Enterprise Architecture Planning C. Data Reengineering D. Prototyping

Page 46 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

2.2.

Tools and Technology Types

2.2.1. Data 23. What is the definition of a Database Management System (DBMS)? A. Controls the organization, storage and retrieval of data in a database. B. A modeling language to define the schema. C. Inverted list management. D. Supports the database query language to interactively access data. 24. Which one is not a common DBMS model? A. Hierarchical B. Network C. Relational D. File 25. Which one is not a function of a DBMS? A. A modeling language to define the schema B. A database query language C. Transaction method that ensures Atomicity, Consistency, Isolation, and Durability (ACID) D. RAID Disk arrays. 26. What is the definition of an Object Database Management System (ODBMS)? A. Controls the organization, storage and retrieval of data in a database. B. A modeling language to define the schema C. Inverted list management D. Supports the database query language to interactively access data. 27. Which one is not a function of an ODBMS? A. Object Definition Language (ODL) B. Object Query Language (OQL) C. C++ and Java Binding. D. Structured Query Language (SQL). 28. Which function does Extract, Transform, and Load (ETL) tool does not involve in the process in data warehousing? A. Extracting data from data sources. B. Transforming data to fit business requirements. C. Transforming metadata to fit business requirements. D. Loading data into the data warehouse. Page 47 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

29. The first part of an ETL process is to extract the data from what type of common data source formats? A. Relational database B. Flat File C. IMS D. C++ 30. Which one is not a typical function of the transformation process in ETL tools? A. Translating code values B. Deriving new calculated values C. Joining or merging data from multiple sources D. DDL SQL statements with SQL variations 31. Which of the following is not a type of load function of an ETL tool in the data warehouse? A. Overwrite old information B. Insert new records C. Update old record and Insert new record D. Insert audit trail records. 32. Which of the following is not true of a Data Modeling Tool? A. Specific to a DBMS B. Produce a diagram summarizing the results of your data modeling efforts C. Generate a database schema from a model. D. Diagram of referential integrity constraints.

2.2.2. Metadata & Descriptive Information 33. What is the definition of a data dictionary? A. Set of cleansed, organized, and transaction level data. B. Database that tracks data element definitions. C. Instances of characters. D. Internal database that store information tracked, developed, and maintained by a predefined set of single-vendor tools.

Page 48 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

34. What is the definition of a data dictionary? A. An entity in a file system that contains a group of files. B. LDAP directory services are examples of general-purpose distributed hierarchical object-oriented directory technologies. C. A repository or database of information. D. Directory technology is often used in white page applications and network information services. 35. What is the definition of a data encyclopaedia? A. Set of cleansed, organized, and transaction level data. B. Database that tracks data element definitions. C. Instances of characters. D. Internal database that store information tracked, developed, and maintained by a predefined set of single-vendor tools. 36. What data is not typically held in a data registry? A. Standardized information in a pre-defined model. B. Metadata, system metadata, system engineering C. Reference information D. XML 37. Which one of the following is not a federated Service-Oriented Architecture (SOA) standards-based registry? A. UDDI registry without a repository. B. UDDI registry with a proprietary repository C. ebXML registry-repository D. Combination of UDDI registry and ebXML registry-repository 38. What is defined as an automated resource “used to describe, document, protect, control and access informational representations of an enterprise”. A. Data dictionaries B. Data directories C. Metadata registry D. Metadata repositories

Page 49 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

39. What is “an integrated, virtual holding area with vendor-independent input, access, and structure; used to directly store metadata and/or metadata-based gateways to external metadata”? A. Data dictionaries B. Data directories C. Data encyclopaedias D. Metadata repositories

2.2.3. Data Issues 40. What is the goal of business intelligence tools? A. Detect patterns in data that explain the present and predict the future. B. Visually representing the data. C. Interactive “online” data exploration D. “Slice-and-dice” analysis 41. Which one of the following is not a new way of interacting with data in extended OLAP models? A. Pivot tables B. Small Multiples C. Geospatial Analysis D. Predictive Analytics 42. Which one of the following does not data mining tools have issues in analyzing? A. E-mails. B. Memos C. Marketing material D. Spreadsheet 43. What is a technique to increase searching of unstructured data? A. Data Semantics B. Data Ontologies C. Classification and Taxonomy D. Classes and Relations

Page 50 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Quick Answers 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 21. 22.

C D B D B C A A B B D D B D C A D B A D A D

23. 24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. 40. 41. 42. 43.

A D D A D C D D D A B C D D B C D A A D C

Page 51 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Detailed Answers 1. Answer: C. Economies of scale in purchasing Case tools. The benefits of an Enterprise Data Architecture are that it organizes data around the enterprise’s data subject to create shared data resources, an integrated view of enterprise data that enables organizational change. 2. Answer: D. Design stability and data object abstraction and generalization. An Enterprise Data Architecture should be created for design stability and data object abstraction and generalization. Enterprise Data Architecture treats data like an information asset; it is not application specific but enterprise specific. Enterprise Data Architecture is typically created in more than one iteration. 3. Answer: B. Determine the index for the data mart. The source data should be architected to determine the source of data needed, diagram the source data, and determine the method for extraction and delivery. 4. Answer: D. Determine the monthly flat file transmission protocol. The goal of Source Data Architecture is to ensure that the source data is extracted only once, define the scope and implementation of the data warehouse and oversee the construction of the enterprise data warehouse. 5. Answer: B. Data Distribution. Data Distribution is targeted towards the efficient delivery of the proper information to the proper recipients. In Data Distribution data can be streamed or supplied depending on the requirements of the communication. 6. Answer: C. Data Integration. Data Integration requires combining and matching information in different sources, and resolving a variety of conflicts. XML is becoming a de facto data integration standard. 7. Answer: A. Single point of authorization. A fundamental principle in Change Authorization of Architectures is a single point of change authorization. Every change must run the same process and authorization prior to changes are implemented. 8. Answer: A. Passive. In a client-server architecture, the features of the client are: Active (Master), sending request, and waiting until reply arrives. 9. Answer: B. Active. In a client-server architecture, the features of the client are: Passive (Slave), waiting for requests, and on requests serves them and send a reply. 10. Answer: B. 3-tier Architecture. If the networks consists of clients, application servers which process data for the clients, and database servers which store data for the application servers it is known as a 3-tier Architecture. 11. Answer: D: Company Directory. Metadata has many sources including but not limited to: Tools, Applications and Software Packages. Page 52 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

12. Answer: D. Metadata user. When beginning Metadata solution architecture, the following type of analysis should be carried out: Process Flow, Metadata Flow, Metadata Record Identification, and Metadata Categorization. Metadata occurs at the input and output of processes the tools like case tools and applications gather the metadata. The Metadata Record Identification identifies the data needed in a metadata solution and their origins. The last step is to categorize the metadata for display and usage. 13. Answer: B. A Normalized schema. The Zachman Framework is a normalized schema that is a good analytical tool. It is not a 36-cell matrix. The Zachman Framework logic is neutral to methods and tools. Each cell of the Framework is unique and primitive. 14. Answer: D. Greater accounting staff effectiveness. Enterprise Architecture Planning has the following benefits: consistency and compatibility of systems, interoperability between systems and databases, economies of scale in purchasing and developing systems, reduced overall system costs, and greater IT staff effectiveness. 15. Answer: C. Maintain a secure infrastructure and IT support for networks and distributed systems. Enterprise Architecture Planning addresses: data management, application environment and development toolsets, middle-ware and transaction management, Web delivery environment, operating systems and other system software, network environment, and hardware server and client environments. 16. Answer: A. Information Technology Strategy. An Enterprise Architecture Plan is usually part of the overall Information Technology Strategy. The Information Technology Strategy leads to Infrastructure Strategy, Information Technology Architecture Plan and Application Infrastructure Plan. 17. Answer: D. Shred. The data life cycle phases are: Create/Store, Retrieve, Modify/Update, Read/Use, Transport, Archive and Delete. The data lifecycle is the process of managing data throughout its lifecycle from conception until disposal, within the constraints of the data policy. 18. Answer: B. Data Life cycle management is an approach to managing an organization's data. The definition of data life cycle management is an approach to managing an organization's data that involves procedures and practices as well as applications. 19. Answer: A. Information Engineering. Information Engineering is defined as "an integrated and evolutionary set of tasks and techniques that enhance business communication throughout an enterprise enabling it to develop people, procedures and systems to achieve its vision". 20. Answer: D. Data Warehousing. Information Engineering has many purposes, including organization planning, business re-engineering, application development, information systems planning and systems re-engineering. Page 53 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

21. Answer: A. Developing a migration strategy from one application environment to another. Data reengineering is structured application redevelopment technique that applies to the structure, function and meaning of data. 22. Answer: D. Prototyping. Prototyping is the process of quickly putting together a working model in order to test various aspects of the design, illustrate ideas or features and gather early user feedback. 23. Answer: A. Controls the organization, storage and retrieval of data in a database. A Database Management System controls the organization, storage and retrieval of data in a database. Two of the functions that a DBMS has are a modeling language to define the schema and a database query language. 24. Answer: D. File. The common DBMS models are Hierarchical, Network and Relational. 25. Answer: D. RAID Disk arrays. The functions of a DBMS are: A modeling language to define the schema, a database query language and transaction method that ensures Atomicity, Consistency, Isolation, and Durability (ACID). Many DBMS also support the Open Database Connectivity API that supports a standard way for programmers to access the DBMS. DBMS do not include a storage device. 26. Answer: A. Controls the organization, storage and retrieval of data in a database. A Database Management System controls the organization, storage and retrieval of data in a database. Two of the functions that a DBMS has are a modeling language to define the schema and a database query language. 27. Answer: D. Structured Query Language. The functions of an ODBMS are: Object Definition Language (ODL), Object Query Language (OQL), C++ and Java Binding. Structured Query Language is a function of a Relational Database Management System. 28. Answer: C. Transforming metadata to fit business requirements. Metadata is created when using ETL tools not manipulated by ETL tools. ETL tools involve the functions of extracting data from data sources, transforming data to fit business requirements and loading data into the data warehouse. 29. Answer: D. C++. The first part of an ETL process is to extract the data from the common data source formats like Relational database, Flat Files, IMS or other data structures such as VSAM or ISAM. Extraction extracts the data into a format for transformation processing. C++ is a programming language not data. 30. Answer: D. DDL SQL statements with SQL variations. Typical functions of the transformation process in ETL tools include: Translating code values (e.g. M for male or F for Female), Deriving new calculated values (A+B = C), Joining or merging data from

Page 54 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

multiple sources, generating surrogate key values and summarizing values. The DDL SQL statement with SQL Variations is a DB mechanism to create tables for example. 31. Answer: D. Insert audit trail records. Typical load functions of an ETL tool in a data warehouse are to: Overwrite old information, insert new records and update old record and insert a new record. 32. Answer: A. Specific to a DBMS. Data modeling tools are RDBMS-independent. Data Modeling Tools: Produce a diagram summarizing the results of your data modeling efforts; Generate a database schema from a model; and Diagram of referential integrity constraints. 33. Answer: B. Database that tracks data element definitions. A data dictionary is a database that tracks data element definitions. An encyclopedia is an internal database that store information tracked, developed, and maintained by a predefined set of singlevendor tools. A data staging area is a set of cleansed, organized, and transaction level data. Instances of characters could represent any time of file or data store and is not specific to a data dictionary. 34. Answer: C. A repository or database of information. A data dictionary is a repository or database of information. Data dictionaries can be used as a white page application and network information service or as an LDAP directory service. 35. Answer: D. Internal database that store information tracked, developed, and maintained by a predefined set of single-vendor tools. A data encyclopedia is an Internal database that store information tracked, developed, and maintained by a predefined set of single-vendor tools. A data dictionary is a database that tracks data element definitions. A data staging area is a set of cleansed, organized, and transaction level data. 36. Answer: D. XML. XML is a format. A Data Registry is defined as an automated resource “used to describe, document, protect, control and access informational representations of an enterprise”. Typically the following is held in a data registry: Standardized information in a pre-defined model, Metadata, system metadata, system engineering; and Reference information. Standards for models and templates for data and metadata registries already exist – for example, the ISO 11179 standard for Metadata Registries, and ebXML for XML registries. 37. Answer: B. UDDI registry with a proprietary repository. A federated SOA deployment requires a standards-based registry-repository the choices involve two standards, UDDI and ebXML Registry. A UDDI registry offers a subset of capabilities offered by an ebXML Registry. Published in a UDDI registry are pointers to service artifacts such as WSDL. Published in an ebXML Registry is not just pointers to service artifacts, but also the actual artifact. Thus, an ebXML registry-repository can be used for governance of any type of service artifacts throughout their life cycles.

Page 55 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

38. Answer: C. Metadata registry. A metadata registry is defined as an automated resource “used to describe, document, protect, control and access informational representations of an enterprise”. 39. Answer: D. Metadata repositories. A Metadata repositories are “an integrated, virtual holding area with vendor-independent input, access, and structure; used to directly store metadata and/or metadata-based gateways to external metadata”. 40. Answer: A. Detect patterns in data that explain the present and predict the future. The goal of business intelligence tools are to help end-users detect patterns in data that explain the present and predict the future. End users use a number of techniques to visually represent the data and slice and dice analysis while interactively exploring the data online. 41. Answer: A. Pivot tables. New methods of interacting with data in extended OLAP models are small multiples or multidimensional matrix of related graphs, Geospatial analysis that combines cartographic elements and data information and Predictive Analytics. 42. Answer: D. Spreadsheet. Data mining tools have issues in analyzing unstructured data like e-mails, memos, marketing material, notes from internal groups, news, user groups, chats, and whitepapers. 43. Answer: C. Classification and Taxonomy. A technique to increase searching of unstructured data is classification and taxonomy. An Ontology Is “A specification of a representational vocabulary for a shared domain of discourse -- definitions of classes, relations, functions, and other objects -- is called ontology.” – T.R. Gruber

Page 56 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

3.0 Data Analysis and Modeling Overview Planning and requirements analysis are conducted as part of a project life cycle. The scope of the project is typically defined during planning. The scope can be refined using fact finding techniques like interviews or questionnaires. During requirements analysis, the analyst discovers the client’s requirements and detailed information needed to build the application and data model. The requirements are gathered using fact-finding techniques like Surveys, questionnaires, JAD sessions or Legacy systems analysis. The purpose of data modeling is to develop an accurate model, or graphical representation, of the client's data requirements at different levels of abstraction. The data model acts as a framework for the development of the new or enhanced application. There are many data models that may be used in an organization. In a typical organization, the order of creation of data models is Conceptual, Enterprise, Logical, and Physical. There are also Dimensional Data Models that support on-line analytical processing applications and Object Oriented Models that structures systems around the data. Regardless of the data models used, they are useful for confirming requirements and leading to development of the application.

Topics Data / Metadata Analysis & Design Fact Finding Techniques Requirements Definition and Management Data Model Components Logical Data Model Dimensional Warehouse Object Oriented / UML Data Representations in Process Models Data / Metadata Model Management Types of Data Models Scope of Model and Metadata Data Model Support

Page 57 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Questions 3.1 Data / Metadata Analysis & Design 3.1.1 Fact Finding Techniques 1. What type of fact-finding technique works best when dealing with numerous divisions? A. Interviewing B. Surveys, questionnaires C. JAD session D. Legacy systems analysis 2. What are the benefits of a JAD Workshop? 1. Communication and combined knowledge 2. Build consensus and ownership 3. Improve design quality 4. Design cross-functional solutions A. B. C. D.

1&2 2&3 3&4 All of the above

3. What type of fact-finding technique minimizes time and assists in narrowing scope? A. Interviewing B. Surveys, questionnaires C. JAD session D. Legacy systems analysis 4. What type of fact-finding technique is a systematic attempt to collect information from a person? A. Interviewing B. Surveys, questionnaires C. JAD session D. Legacy systems analysis 5. What type of fact-finding technique is always used in data warehousing projects? A. Interviewing B. Surveys, questionnaires C. JAD session D. Legacy systems analysis

Page 58 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

6. What is the most appropriate type of question to ask in an interview? A. Closed questions B. Open-ended questions C. Leading questions D. None 7. When is an unstructured interview more appropriate than a structured interview? A. When the interviewer wants to gain a broad based view on an issue that needs to be explored. B. Interviewer identifies gaps in the knowledge, which acts as a basis for questions . C. When the interview needs to be goal-oriented. D. It is never appropriate to be unstructured as you always need to be prepared.

3.1.2 Requirements Definition and Management 8. What is the next best step after gathering User Requirements for a new system? A. Evaluation of current environment and documentation B. Gap analysis / Future state creation C. Business rules discovery, validation and documentation. D. Data / process matrices creation. 9. What is the next best step after a current state environment evaluation? A. Gap analysis B. Future state creation C. Business rules discovery, validation and documentation. D. Data / process matrices creation. 10. What is the next best step after a current state environment evaluation and future state creation? A. Gap analysis B. Business rules discovery, validation and documentation. C. Data / process matrices creation. D. Requirements tracking and management to implementation 11. Which statement does not describe a business rule? A. It is a statement that defines some facet of the business. B. It asserts business structure, or controls or influences performance of the business. C. It is at the lowest level and cannot be decomposed further or it would lose business meaning. D. It specifies a Pre or Post condition of an entity. Page 59 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

12. Which of the following statements is a business rule? A. When a failure is reported, an expeditor is assigned by the maintenance department who sends the failure form to the service desk for scheduling. B. If Acct_num is between 0 and 5000 then the customer is a member of the branch that may deposit money. C. A customer places an order D. A customer with preferred status should have its orders filled as soon as possible. 13. Which one is not part of a typical business rule creation process? A. Discovery, B. Validation C. Documentation D. Rule Engine 14. Which of the following is not an example language to express business rules? A. UML B. Z notation C. BPEL D. ABAP 15. What is the benefit of Requirements tracking and management to implementation? A. To provide a matrix that has a listing of the requirements for the entire project. B. To ensure the system performs as it should. C. The information in the matrix includes the number assigned to the requirement, a brief description, the date submitted to project, and the tracking of the requirement as it relates to development. D. Determines multi-modal requirements. 16. What is the best benefit of a Data / process matrix? A. The matrix is used to represent relationships between data and process. B. Shorten design lead times while maintaining product quality. C. To clarify dependencies of information flow. D. To document the meaningful relationships between both process and data as it was related to the authors during the interviews.

Page 60 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

3.2 Data Model Components 3.2.1 Logical Data Modeling 17. What are the major components in an Entity-Relationship diagram? A. Attributes, relationships, and associations; B. Object types, relationships, associations, and supertype/subtypes; C. Object types and transitions, associations, and supertype/subtypes; D. States and transitions; 18. What does the following Entity-relationship diagram describe? Negotiates Price

Buyer

Agent

Seller 1. Real estate agent negotiates price between buyer and seller. 2. Buyer negotiates price with seller, through real estate agent A. B. C. D.

1. 2. 1&2 Neither 1 or 2

Page 61 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

19. In the following diagram, what are the subtypes? Employee

Manager

1. 2. 3. 4. A. B. C. D.

Contract Employee

Hourly Employee

Employee Manager Contract Employee Hourly Employee

1 2 3&4 2, 3 & 4

20. What is the definition of an Attribute? A. An atomic fact or characteristic, which describes an entity. B. A description of an entity occurrence. C. A decomposable fact or characteristic, which describes an entity. D. An association between entities. 21. Attributes roles do the following: a. Identify an occurrence of an entity type (primary key) b. Relate an occurrence of one entity type to an occurrence of another entity type (foreign key) c. Describe the entity or association d. Derives values from other data values in the model. A. B. C. D.

1&2 1&3 1, 2, & 3 1, 2, 3, & 4

Page 62 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

22. What is the definition of cardinality? A. Relate an occurrence of one entity type to an occurrence of another entity type; B. The relative number of occurrences which may exist between a pair of entities; C. A characteristic that describes something about an entity; D. Identify an occurrence of an entity type;

23. What is a Mandatory relationship in Optionality? A. At least one or many; B. None, one, or many C. One and only one D. None or one 24. Describe the following data model: Employee works for employs

A. B. C. D.

Organization Unit

Each Employee “works-for” one and only one Organization Unit Each Employee “works-for” at least one or many Organization Unit Each Employee “works-for” none, one or many Organization Unit Each Employee “works-for” none or one Organization Unit

25. What is the relationship between the Primary Key and Foreign Keys? A. One-to-one; B. One-to-many; C. Many-to-Many; D. Foreign Keys do not relate entities; 26. What is the Attribute that uniquely identifies an entity is called? A. Entity Type B. Entity Occurrence C. Primary Key D. Foreign Key

Page 63 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

27. What is the difference between an Entity Type and Entity Occurrence? A. An Entity Type is something that exists and is capable of being described and an entity occurrence is a relationship; B. An Entity Type is the definition and the entity occurrence is an instance of the Entity; C. An Entity Type is a physical object type and the entity occurrence is the project. D. There is no difference. 28. What data model is the Normalization Process applied? A. Conceptual Data Model B. Logical Data Model C. Physical Data Model D. Metadata Data Model 29. What is the objective of the Normalization process? A. To identify the one best place an attribute belongs. B. To organize the physical design of the data model into tables and columns. C. To organize columns based on the mathematical principles of set theory. D. To assign an attribute to multiple entities. 30. What are the attributes of a data model in First Normal Form? A. All repeating groups have been eliminated B. Every attribute describes completely that entity and not an entity identified by only part of the primary identifier. C. Data items that do not describe the entire primary key of the entity are eliminated. D. Identified restrictions that apply to the data and its relationships. 31. What are the attributes of a data model in Second Normal Form? A. All repeating groups have been eliminated B. Every attribute describes completely that entity and not an entity identified by only part of the primary identifier. C. Data items that do not describe the entire primary key of the entity are eliminated. D. Identified restrictions that apply to the data and its relationships. 32. What are the attributes of a data model in Third Normal Form? A. All repeating groups have been eliminated B. Every attribute describes completely that entity and not an entity identified by only part of the primary identifier. C. Data items that do not describe the entire primary key of the entity are eliminated. D. Identified restrictions that apply to the data and its relationships.

Page 64 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

33. What are the two principle types of static relationships in a class diagram? A. Primary Key and Foreign Key. B. Association and Subtype. C. Cardinality and Optionality. D. One-to-One and One-to-Many.

3.2.2 Dimensional Warehouse 34. When using Dimensional Modeling, a logical design technique, data is presented in which manner? A. Processes, data stores, flows and terminators. B. Processes, data stores, relationships and flows. C. Fact tables and dimension tables. D. Entities, data, and relationships. 35. Which one of the following is best described as a fact: A. promotion_name B. clerk_grade C. address D. dollars_sold 36. In Dimensional Modeling, snowflaking is a technique that does the following: A. Aggregates fact tables into summary tables for easier querying. B. Adding more attributes into a single dimension table to make it easier for the business user understandability. C. Removes low-cardinality textual attributes from dimension tables and places them in joined “secondary” dimension table. D. Identifies cross-dimensional attributes for easier browsing performance. 37. Which application is the best fit for Dimensional Modeling? A. OLTP B. OLAP C. HPC D. Web

Page 65 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

38. In a Dimensional Model, when tracking changes in a slowly changing dimensional table, the old value is discarded and has no significance is regarded as what type? A. type 1 B. type 2 C. type 3 D. cross-dimensional attribute 39. In a Dimensional Model, when tracking changes in a slowly changing dimensional table, the old value is recorded and has significance is regarded as what type? A. type 1 B. type 2 C. type 3 D. cross-dimensional attribute 40. In a Dimensional Model, when tracking changes in a slowly changing dimensional table, the old value and new value are equally important is regarded as what type? A. type 1 B. type 2 C. type 3 D. cross-dimensional attribute 41. In the Customer Dimension, a Slowly Changing Type 2 Dimension, the key would be described as the following? A. Customer_Key which is a new generated, meaningless, key; B. Customer_Key – made of Customer_Number appended with Time_Key; C. Customer_Key – made of Customer_Number appended with version number; D. Customer_Key from the old record because it will be overwritten. E. Customer_Key – made of the Operational Key

3.2.3 Object Oriented / UML 42. What term in the basis of modularity and structure in object-oriented modeling? A. Entity B. Class C. Object D. Relationship.

Page 66 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

43. What is the definition of an object? A. An instance of a class. B. Behavior that varies depending on the class. C. A type of privacy applied to the data and some of the methods. D. A person, place or thing. 44. What is the term for a type of privacy applied to the data and some methods of a class? A. Encapsulation B. Inheritance C. Abstractions D. Polymorphism 45. When modeling a class diagram, what is an association? A. An association represents relationship between instances of classes. B. A state of being associated. C. Primary and Foreign Key relationship between objects. D. There is no concept of an association in a class diagram. 46. Which one is not one of the relationship types in UML Modeling? A. Generalization B. Association C. Aggregation, Composition D. Polymorphism 47. What is the term for how many objects may participate in a given relationship in a class diagram? A. Cardinality B. Optionality C. Multiplicity D. Pairing 48. In the following diagram, what does the association show in terms of multiplicity? 1 A. B. C. D.

Exactly one. Many. Optional. Number specified.

Page 67 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

49. In the following diagram, what does the association show in terms of multiplicity? * A. B. C. D.

Exactly one. Many. Optional. Number specified

50. In the following diagram, what does the association show in terms of multiplicity? 0..1 A. B. C. D.

Exactly one. Many. Optional. Number specified

51. In the following diagram, what does the association show in terms of multiplicity? 1..10 A. B. C. D.

Exactly one. Many (zero or more) Optional (zero or one) Number specified

52. What Association does the following diagram represent?

A. B. C. D.

Aggregation. Composition. Ordered Role. Not an association type.

53. What does the marking + on an attribute mean? A. Public attribute B. Protected attribute C. Private attribute D. Package attribute

Page 68 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

54. Which of the following is false in describing Attributes in UML? A. Attributes are always single-valued. B. Attributes may be optional or mandatory C. Optional or mandatory may be drawn on the class diagram. D. Attributes have name, type, and default values. 55. What is the term that defines the processes that a class can carry out? A. Operations. B. Specifications. C. Multiplicity. D. Notation.

3.2.4 Data Representations in Process Models 56. Which one of the following is not a true statement about triggers? A. Triggers can accept parameters B. Triggers are code that automatically execute on a table or data in response to certain events. C. Triggers can be used to enforce Referential Integrity between tables. D. Triggers can be executed after SQL command type statements of: INSERT, UPDATE, or DELETE. 57. Which one of the following is not a typical type of referential integrity trigger? A. Identifying B. Non-identifying C. Subtype D. Supertype 58. What is the difference between a stored procedure and a trigger? A. Stored procedures can accept parameters while Triggers cannot. B. Stored procedures are stored in the database while Triggers are not. C. Stored procedures and procedural while Triggers can be used to enforce Referential Integrity between tables. D. Stored procedures and Triggers can simplify data management. 59. What is the benefit of the Unified Modeling Language (UML)? A. UML standardizes representation of object oriented analysis and design. B. In UML, requirements gathering comprise of Use Case and Activity diagrams. C. In UML, design comprise Class and Object diagrams. D. In UML, deployment comprises of Package and Subsystem diagrams.

Page 69 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

3.3 Data / Metadata Model Management 3.3.2 Types of Data Models 60. What data model is composed of subject areas, relationships, and subject area definitions? A. Conceptual Data Model B. Logical Data Model C. Physical Data Model D. Dimensional Data Model 61. What is the difference between a Conceptual Data Model and Enterprise Data Model? A. A Conceptual Data Model is a type of Business Model while an Enterprise Data Model is a physical data model. B. A Conceptual Data Model describes the whole enterprise business subject areas while an Enterprise Data Model is a decomposition of subject area entities. C. A Conceptual Data Model is a logical data model while an Enterprise Data model is a physical data model. D. A Conceptual Data Model is a concept applied to the enterprise while an Enterprise Data Model is applied to databases. 62. In a typical organization, the order of creation of data models is in which of the following orders? A. Conceptual, Enterprise, Logical, Physical B. Enterprise, Conceptual, Logical, Physical C. Logical, Conceptual, Enterprise, Physical D. Physical, Logical, Enterprise, Conceptual 63. What data model is composed of tables and columns? A. Conceptual Data Model B. Logical Data Model C. Physical Data Model D. Dimensional Data Model 64. What data model is geared to a decision support environment? A. Conceptual Data Model B. Logical Data Model C. Physical Data Model D. Dimensional Data Model

Page 70 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

65. Which data model has an Enterprise Wide scope? A. Conceptual Data Model B. Logical Data Model C. Physical Data Model D. Dimensional Data Model 66. Which data model has all the entities for each subject area? A. Logical Data Model B. Physical Data Model C. Dimensional Data Model D. Enterprise Data Model 67. What are the elements in an Enterprise Data Model? A. Use Cases and Activity diagrams B. Entities and Relationships C. Class and Object Diagrams D. Tables and Relationships 68. What is the definition of a metamodel? A. Data models that specify one or more other data models. B. Models of unclassified metadata. C. Physical metadata storage. D. Models that describe the flow of metadata. 69. What is the difference between a metamodel and a meta-metamodel? A. Metamodels are data models that specify other data models, while metametamodels defines ontology. B. Metamodels are the definition of concepts while meta-metamodels defines the language. C. Metamodels describe the types of records, while meta-metamodels describe the type of records used in the database. D. There is no difference. 70. Which one of the following is not an industry standard? A. Case Data Interexchange Format (CDIF) B. Meta Data Coalition (MDC) C. Common Warehouse Model (CWM) D. Metadata Interchange Specification (MDIS)

Page 71 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

71. What is the benefit of universal / industry models? A. Standard data models for an industry that may be used off the shelf. B. Adapt to industry users' characteristics. C. The data requirements have already been gathered and provided in use cases. D. There is no benefit as customizations can take hundreds of hours. 72. Which one of the following phases is not part of the data life cycle? A. B. C. D.

Create/Store Modify/Update Delete Shred

3.2.2 Scope of Model and Metadata 73. Which of the following is not true about the scope of an enterprise wide data model? A. Supports a wide audience. B. Provides a data picture of the business. C. Causes internal gridlock and inconsistencies. D. Capable of being easily extended to capture new requirements. 74. What is the scope of the enterprise wide data model? A. Business Unit (i.e. Marketing) B. Corporation C. Geographic Unit (i.e. North America) D. Functional Area (Information Technology) 75. Which statement is not true about the enterprise wide data model? A. The corporate data architect owns the Enterprise wide data model. B. The Enterprise wide data model is driven by the business. C. Subject areas are areas of concern for the corporation. D. The enterprise data model will frequently change. 76. A data model represents Marketing data for an organization. What is scope of data model? A. Enterprise B. Business Area C. Project Oriented D. Subject Area

Page 72 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

77. A data model represents financial data for an organization. What is scope of data model? A. Enterprise B. Business Area C. Project Oriented D. Subject Area

78. A project has begun to track costs for starting up operations in Plant A. scope of data model? A. Enterprise B. Business Area C. Project Oriented D. Subject Area

What is

3.2.3 Data Model Support 79. When using a data modeling tool, which one of the following is not applicable when creating data models? A. Forward Engineering B. Reverse Engineering C. Create Logical and Physical data models D. Split models of older versions into separate logical and physical 80. What is creating a data model from an existing database is known as? A. Forward Engineering B. Reverse Engineering C. Create Logical and Physical data models D. Importing prior version data model 81. Which one of the following is not a benefit of Reverse Engineering Functionality? A. Maintain database B. Change Database Type (e.g. from SQLServer to Oracle ) C. Analyze differences in databases D. Create the Enterprise Data Model

Page 73 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

82. Which one is not a typical Data Model tool function? A. Forward Engineering B. Reverse Engineering C. Comparisons D. Improving 83. What is the benefit of the Comparisons in data modeling tools? A. Keep data model and database synchronized. B. Compare changes between data model and database. C. Select objects that want to compare. D. Selectively import or export changes. 84. The database administrator would like to create a database from an existing data model, which data modeling tool function would they use? A. Forward Engineering B. Reverse Engineering C. Comparisons D. Version Control 85. When a data modeler would like to roll back a change to a data model, which function they would use? A. Change Control B. Model Merge C. Versioning D. Submodeling 86. When a data modeler would like to create an enterprise model which data model function would be used? A. Change Control B. Model Merge C. Versioning D. Submodeling

87. Which of the following is not a reason for using Model Merge feature? A. To create an enterprise data model. B. To bring individual data models together in a group. C. Two previously unrelated projects have merged. D. Comparing two data models to detect changes.

Page 74 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

88. Which one of the following does not apply when importing Customer data model with New Customer data model? A. Create a new data model B. Merge New Customer into Customer data model C. Merge Customer into New Customer data model D. The functionality is not allowed. 89. Which one is not typically represented in the breadth of data models in data modeling tools? A. Enterprise Data Model B. Logical Data Model C. Physical Data Model D. Business Process Model 90. What is the benefit of linkages and mappings between enterprise, logical, and physical data models? A. To define the different purposes of the data models in the application development process. B. To maintain links between the different data models. C. Synchronize changes between data models. D. Applying transformation functions to the data models.

Page 75 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Quick Answers 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 21. 22. 23.

C D A A D B A A B A D D D D B C B C D A D B A

24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. 40. 41. 42. 43. 44. 45. 46.

A B C B B A A B C B C D C B A A A A B A A A D

47. 48. 49. 50. 51. 52. 53. 54. 55. 56. 57. 58. 59. 60. 61. 62. 63. 64. 65. 66. 67. 68. 69.

C A B C D A A C A A D A A A B A C D A D B A A

70. 71. 72. 73. 74. 75. 76. 77. 78. 79. 80. 81. 82. 83. 84. 85. 86. 87. 88. 89. 90.

B A D C B D B D C A B D D A A C B D D D C

Page 76 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Detailed Answers 1. Answer: C. JAD Session: Joint Application Design session is a method for performing analysis that brings specific parties together within a workshop environment. Surveys and questionnaires must be made up of closed questions and typically involve clarification and discussion after the results are tabulated. Interviews are best done in small groups. Legacy system analysis involves examining and probing the legacy systems. 2. Answer: D. All of the above are the benefits of a JAD Workshop. 3. Answer: A. Surveys, questionnaires. Surveys and questionnaire is an effective technique to get opinions from a wide variety of stakeholders in an organization. Surveys and questionnaires must be made up of closed questions and typically involve clarification and discussion after the results are tabulated. Joint Application Design session is a method for performing analysis that brings specific parties together within a workshop environment to collect requirements. Interviews are best done in small groups and can be used when scope is unknown. Legacy system analysis involves examining and probing the legacy systems. 4. Answer: A. Interviewing. Surveys and questionnaire are typically done electronically. Joint Application Design is a workshop environment to collect requirements. Interviews are best done in small groups where heuristic questions can be asked. Legacy system analysis involves examining and probing the legacy systems and may or may not involve discussion with staff. 5. Answer: D. Legacy systems analysis. Legacy system analysis is used in data warehousing projects to perform source(legacy system) to target(data warehouse) transformations. 6. Answer: B. Open-ended questions. Open-ended questions cannot be answered with a simple yes or no response and thus encourage the interviewee to provide more information. Closed-ended questions give a yes or no answer. Leading questions put the interviewee’s opinion into the question and do not give an opportunity for the interviewer to answer without bias. 7. Answer: A. When the interviewer wants to gain a broad based view on an issue that needs to be explored. The unstructured interview is used when the interviewer wants to explore an issue and facilitates description of domain in a way that is easy for the interviewee. 8. Answer: A. Evaluation of current environment and documentation. After gathering user requirements for a new system, the next best step is an evaluation of the current environment and documentation to complete the current state assessment. Next, a future state can be derived with the linkages to the current state if needed. A gap analysis is

Page 77 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

then conducted to compare its current state with its future state to determine the variance between business requirements and current capabilities. 9. Answer: B. Future state. After creating a current state environment evaluation, the next best step is to complete the future state assessment. Next, a gap analysis is then conducted to compare its current state with its future state to determine the variance between business requirements and current capabilities. 10. Answer: A. Gap analysis. After creating a current state environment evaluation and future state creation, the next best step is to complete the gap analysis. A gap analysis compares current state with future state to determine the variance between business requirements and current capabilities. 11. Answer: D. It specifies a Pre or Post condition of an entity. Business rules are put in business terms not in terms of conditions on entities. A business rule: • Is a statement that defines some facet of the business. • Asserts business structure or controls or influences performance of the business and can be applied across the organization. • Is at the lowest level and cannot be decomposed further or it would lose business meaning. 12. Answer: D. A customer with preferred status should have its orders filled as soon as possible. Business rules should contain the words: must; must not; should; should not; only; only if. The following did not qualify as business rules due to the following reasons: When a failure is reported, an expeditor is assigned by the maintenance department who sends the failure form to the service desk for scheduling. Ordering of events in the business rule is declarative. Business rules should be procedural. If Acct_num is between 0 and 5000 then the customer is a member of the branch that may deposit money. Business rules should not contain technology nomenclature but be solely about the business. A customer places an order. This rule can be decomposed into further rules. 13. Answer: D. Rule Engine. The typical business rule creation process is discovery, validation and documentation. Business rules are discovered as part of a formal requirement gathering process during the initial stages of design. Once they are documented, they are validated to ensure consistency and non-conflicting business rules. Finally, they are documented. In some organizations, software packages are used to store business rules. 14. Answer: D. ABAP. ABAP is a language used in SAP. If the situation merits, business rules can be expressed in a very formal language, for example: UML, Z notation, Business Rules Markup Language, Business Process Execution Language (BPEL) and Business Process Modeling Notation (BPMN).

Page 78 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

15. Answer: B. To ensure the system performs as it should. The benefit of Requirements tracking and management to implementation is to ensure the system performs as it should. The requirements should be classified in a matrix that has a listing of the requirements for the entire project. The type of information about each requirement in the matrix should consist of the following information: the number assigned to the requirement, a brief description, the date submitted to project, and the tracking of the requirement as it relates to development. The matrix can determine multi-modal requirements. 16. Answer: C. To clarify dependencies of information flow. The matrix has the following structure: the columns in the matrix represent the major processes, and the rows the Entities and Attributes. Elements of the matrix thus represent interactions between a Process and Data. Matrix elements marked with an X represent data/process interactions. 17. Answer: B. Entity-relationship diagrams consist of two major components: Object types and relationships. Object types represent a collection, or set, or objects that can be identified uniquely, and can be described by one or more facts (represented by a rectangular box). Relationships represent a set of connections, or associates, between the object types (represented by diamonds). Associations represent a relationship that we need to maintain information. The subtype/supertype indicator consists of an object type and one or more subcategories; connected by a relationship. 18. Answer: C. The diagram shows that both descriptions are valid. 19. Answer: D. Manager, Contract Employee and Hourly Employee are examples of subtypes. Employee is the general category and the subcategories are: Manager, Contract Employee and hourly employee. 20. Answer: A. Attributes are an atomic fact or characteristic, which describes an entity. It possible to differentiate between an entity and attribute by examining whether it can stand-alone and hold meaning. For example Street Name only makes sense when it resides in the context of Employee. Street Name is an attribute of Employee entity an as such modifies Employee. Employee is an entity. One occurrence of Employee entity might be the employee Johnson. Attributes such as Street Name, City, and Phone Number are attributes, which describe or modify Employee Johnson. 21. Answer: D: 1, 2, 3, & 4 Attributes roles identify, describe and relate attributes. 22. Answer: B. Cardinality can be defined as the relative number of occurrences, which may exist between a pair of entities. There are three kinds of relationships: one-to-one; one-to-many; and many-to-many. In a one-to-one relationship between two entities, at any one time there exists only one occurrence of the entity (“Customer” and “Employee”). In a one-to-many relationship between two entities, at any one time, there may exist multiple occurrences of the entity “Customer” for each entity of “Customer

Page 79 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Type”. In a many-to-many relationship between two entities, multiple occurrences of the entity “Employee” can exist for multiple occurrences of the entity “Skill”. 23. Answer: A. Mandatory means At least one or many and is denoted by a bar. 24. Answer: A. The Data Model describes the relationship that Each Employee “worksfor” one and only one Organization Unit. The model also states that Each Organization Unit can have many employees working for it. Bar indicates a Mandatory relationship between Employee and Organization Unit. The Circle indicates an Optional relationship between Organization Unit and Employee. 25. Answer: B. One-to-many. The primary key of the “one” entity becomes the foreign key of the “many” entity. Foreign key data columns are part of the “many” entity in a one-to-many relationship that is identified through the Normalization Process. 26. Answer: C. Primary Key. The Attribute that uniquely identifies an entity is called a Primary Key. Foreign key data columns are part of the “many” entity in a one-to-many relationship that is identified through the Normalization Process. An Entity Type is the definition and the entity occurrence is an instance of the Entity. An entity type is an identifiable person, organization, place, event, concept or thing that exists and is capable of being described like an Employee. An Entity Occurrence is a unique instance of the entity type like one Employee may have the last name of Johnson, Smith, and Lee. 27. Answer: B. An Entity Type is the definition and the entity occurrence is an instance of the Entity. An entity type is an identifiable person, organization, place, event, concept or thing that exists and is capable of being described like an Employee. An Entity Occurrence is a unique instance of the entity type like one Employee may have the last name of Johnson, Smith, and Lee. 28. Answer: B. Logical Data Model. Normalizing the Logical Data Model to the generally accepted 3rd normal form is required to make the model useful for translation into a physical model that can be implemented with little redundancy and remove potential data anomalies in the create, update and deletion of data. The Normalization process is not applied to the conceptual, physical or metadata data models. 29. Answer: A. To identify the one best place an attribute belongs. The process of normalizing data elements is a technique based on mathematical principles of set theory, first introduced by Dr. E. F. Codd. Normalization is a systematic process of grouping attributes into data entities. Normalization does not dictate any physical design like tables or columns. Normalization finds the one best place that an attributes belongs not multiple places. 30. Answer: A. All repeating groups have been eliminated. A data model in First Normal Form has all repeating groups eliminated. A data model in Second Normal Form has every attribute describes completely that entity and not an entity identified by only part of the primary identifier. A data model in Third Normal Form has data items that do not Page 80 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

describe the entire primary key of the entity are eliminated. A simple rhyme to remember the ordering is "the key, the whole key and nothing but the key", so help me Codd. 31. Answer: B. Every attribute describes completely that entity and not an entity identified by only part of the primary identifier. A data model in First Normal Form has all repeating groups eliminated. A data model in Second Normal Form has every attribute describes completely that entity and not an entity identified by only part of the primary identifier. A data model in Third Normal Form has data items that do not describe the entire primary key of the entity are eliminated. A simple rhyme to remember the ordering is "the key, the whole key and nothing but the key", so help me Codd. 32. Answer: C. Data items that do not describe the entire primary key of the entity are eliminated. A data model in First Normal Form has all repeating groups eliminated. A data model in Second Normal Form has every attribute describes completely that entity and not an entity identified by only part of the primary identifier. A data model in Third Normal Form has data items that do not describe the entire primary key of the entity are eliminated. A simple rhyme to remember the ordering is "the key, the whole key and nothing but the key", so help me Codd. 33. Answer: B. Association and Subtype. The two principle types of static relationships in a class diagram are association and subtype. Primary key and foreign key, cardinality and optionality, and one-to-one and one-to-many describe data models. 34. Answer: C. Fact tables and dimension tables. Every dimensional model is composed of one table with a multipart key called a fact table and a set of tables called dimension tables that describe the dimensions of the fact table. Examples of dimension tables are: Time, Store, Product, Customer, and Employee while the fact table could be: Sales. The Data Warehouse Bus Architecture may be defined as: A master suite of conformed dimensions and to standardize the definitions of facts. Process, data stores, flows and terminators are part of data flow diagrams. Entities, data and relationships are part of data modeling. Processes, data stores, relationships and flows are a combination of data flow diagrams and data modeling. 35. Answer: D. dollars_sold is a fact attribute. The most useful facts in a fact table are numeric, additive and continuously valued. Continuously valued means that every time the attribute is sampled, it can take on different values. Dimension tables, most often contain descriptive textual information. Clerk_grade, address and promotion_name would all be dimensions. 36. Answer: C. Snowflaking removes low-cardinality textual attributes from dimension tables and places them in joined “secondary” dimension table. Snowflaking may be used both logically and physically. For example Customer may be separated from Customer Type in two dimension tables. 37. Answer: B. OLAP. Dimensional Modeling is best used in OLAP type applications for browsing, performance, and user understandability. OLTP systems or online Page 81 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

transaction processing systems use a normalized data model. HPC or High Performance Computing uses a normalized model that typically resides in memory for faster transaction time. Web is similar to OLTP in the data model used. 38. Answer: A. type 1. In a Dimensional Model, when tracking changes in a slowly changing dimensional table, the old value is discarded and has no significance is regarded as type 1. Type 1 overwrites the old record and does not track changes. Type 2 tracks full changes and partitions history of the dimension table. Type 3 tracks old and new definitions on the same record. Cross-dimensional attribute is an attribute that describes an attribute but may also be counted and could reside in either a fact or dimension table. 39. Answer: A. type 2. In a Dimensional Model, when tracking changes in a slowly changing dimensional table, the old value is recorded and has significance is regarded as type 2. Type 2 tracks full changes and partitions history of the dimension table. Type 1 overwrites the old record and does not track changes. Type 3 tracks old and new definitions on the same record. Cross-dimensional attribute is an attribute that describes an attribute but may also be counted and could reside in either a fact or dimension table. 40. Answer: A. type 3. In a Dimensional Model, when tracking changes in a slowly changing dimensional table, the old value and new value are equally important is regarded as type 3. Type 1 overwrites the old record and does not track changes. Type 2 tracks full changes and partitions history of the dimension table. Type 3 tracks old and new definitions on the same record. Cross-dimensional attribute is an attribute that describes an attribute but may also be counted and could reside in either a fact or dimension table. 41. Answer: A. In the Customer Dimension, a Slowly Changing Type 2 Dimension, the key would be Customer_Key, which is a new generated, meaningless, key or surrogate key. The type 2 response requires the use of a surrogate key to fully track the changes of the record. Type One Dimension is overwritten. Intelligence in the key like Time or Version numbers should not be used when creating keys. 42. Answer: B. Class. The basis of modularity and structure in object-oriented modeling is called class. A class is a grouping of data and behaviour for a concept. Entity and Relationship are concepts from data modeling not object-oriented modeling. An object is an instance of a class. Each object has its own data, though the code within a class. 43. Answer: A. An instance of a class. An object is an instance of a class. Each object has its own data, though the code within a class. 44. Answer: A. Encapsulation. The term for a type of privacy applied to the data and some methods of a call is known as encapsulation. Inheritance is a method of generalization that creates subtypes. Abstraction is the ability of a function to have different specifications. Polymorphism is the ability of objects belonging to different types to respond to method calls of methods of the same name, each one according to the right type-specific behavior. Page 82 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

45. Answer: A. An association represents relationship between instances of classes. When modeling a class diagram, an association represents the relationship between instances of classes. Each relationship has two roles, both to and from each class. 46. Answer: D. Polymorphism. The relationship types in UML Modeling are Generalization, Association and Aggregation, Composition. Generalization typifies an inheritance relationship of subtype/supertype. Association represents the relationships between objects. Aggregation and Composition are a special type of association that specifies a HAS-A relationship. Polymorphism is the ability of objects belonging to different types to respond to method calls of methods of the same name, each one according to the right type-specific behavior. 47. Answer: C. Multiplicity. The term for how many objects that may participate in a given relationship in a class diagram is multiplicity. Cardinality and Optionality are terms in that describe participation in a relationship(optional, mandatory, etc.) and the number of times the entities can participate (one-to-one). 48. Answer: A. Exactly one. The number one specifies exactly one. 49. Answer: B. Many (zero or more). The * defines many (zero or more). 50. Answer: C. Optional. The diagram refers to a many (zero or one) association. 51. Answer: D. Number specified. 1..10 specify the number of items in multiplicity. 52. Answer: A. Aggregation. The diagram represents an aggregation association. A composition association would have a dark diamond. The ordered role is notated by a star. 53. Answer: A. Public attribute. The marking + on an attribute means it is a public attribute. The following is a listing for all the markings: "+" for public "#" for protected "−" for private "~" for package. 54. Answer: C. Optional or mandatory may be drawn on the class diagram. A diagram cannot specify optional or mandatory attributes on a class diagram. 55. Answer: A. Operations. The term that defines what processes a class can carry out is Operations. Operations that change the state of an object are called modifiers. Operations and modifiers are typically used interchangeably. 56. Answer: A. Triggers can accept parameters. Triggers are code that automatically execute on a table or data in response to certain events. Triggers can be used to enforce Page 83 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

referential integrity between tables that have a relationship. There are typically three triggering EVENTS that cause trigger to 'fire': • INSERT event (as a new record is being inserted into the database). • UPDATE event (as a record is being changed). • DELETE event (as a record is being deleted). 57. Answer: D. Supertype. The following are typical types of referential integrity triggers: Identifying, non-identifying (allowing nulls and non-null), and subtype. The typical actions that the triggers can conduct are: CASCADE, RESTRICT, SET NULL (non-identifying-nulls allowed), SET DEFAULT (non-identifying), and NONE. 58. Answer: A. Stored procedures can accept parameters while Triggers cannot. Stored Procedure is a program which, like a Trigger is physically stored in a database. A stored procedure and trigger can both be used to enforce Referential Integrity and simply data management. 59. Answer: A. UML standardizes representation of object oriented analysis and design. While all of the other statements are true, the overall benefit of UML is that it standardizes representation of the object oriented analysis and design. 60. Answer: A. A conceptual model is composed of subject areas, relationships, and subject area definitions. A Conceptual Model is a type of Business Model. 61. Answer: B. A Conceptual Data Model describes the whole enterprise business subject areas while an Enterprise Data Model is a decomposition of subject area entities. A Conceptual Model is a high-level starting-point for design and construction activities leading to implemented information systems that fulfill important business needs. The Enterprise Data Model is at a lower level of detail than the conceptual model. A typical conceptual model of a whole enterprise might consist of 7-9 subject areas, representing major business subject areas. An Enterprise Data Model (EDM) while still not containing all entities or all relationship will have many entities for each subject area. 62. Answer: A. Conceptual, Enterprise, Logical, Physical. In a typical organization, the order of creation of data models is Conceptual, Enterprise, Logical, Physical. 63. Answer: C. A Physical data model is composed of tables and columns. Physical data models are representations of models that specify database or file structure implementations. 64. Answer: D. A Dimension Data Model is geared to decision support environments. Data is modeled for retrieval of large amounts of data. Design for high volume retrieval is coupled with specialized administration skills and techniques, and often specialized dimensional database management systems. 65. Answer: A. Conceptual Data Model. A Conceptual Data Model describes the whole enterprise business subject areas.

Page 84 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

66. Answer: D. Enterprise Data Model. An Enterprise Data Model (EDM) while still not containing all entities or all relationship will have many entities for each subject area. 67. Answer: B. Entities and Relationships. An Enterprise Data Model is comprised of Entities and Relationships that are organized into Subject Areas. The model may include attributes of the Entities. 68. Answer: A. Data models that specify one or more other data models. Metamodels are the details behind the metadata that depict metadata relationships. 69. Answer: A. Metamodels are data models that specify other data models, while metametamodels defines ontology. 70. Answer: B. Meta Data Coalition (MDC). Meta Data Coalition is a group that defined the Metadata Interchange Specification. CDIF, CWM, MDIS are all industry standards. 71. Answer: A. Standard data models for an industry that may be used off the shelf. Standard data models are widely used in an industry and shared among different companies. 72. Answer: D. Shred. The data life cycle phases are: Create/Store, Retrieve, Modify/Update, Read/Use, Transport, Archive and Delete. The data lifecycle is the process of managing data throughout its lifecycle from conception until disposal, within the constraints of the data policy. 73. Answer: C. Causes internal gridlock and inconsistencies. The Enterprise wide data model known as a Conceptual Data Model describes the whole enterprise business subject areas, supports the entire enterprise, a wide audience, provides a data picture of the business that is capable of being easily extended to capture new requirements. The Enterprise wide data model leads to and integrated data picture that breaks down inconsistencies and promotes data knowledge, use and sharing. 74. Answer: B. Corporation. The Enterprise wide data model, known as the Conceptual Data Model describes the whole corporation not just a business unit, geographic unit or functional area. 75. Answer: D. The enterprise data model will frequently change as new requirements are determined. The enterprise data model should remain stable unless for example, a new company is acquired. The enterprise wide data model is driven by the business, which encompasses areas of concern or importance to the corporation. The corporate data architect owns the enterprise wide data model. 76. Answer: B. Business Area. The scope of a data model that represents Marketing data for an organization is business area. The data model may be specific to Marketing area needs or may have access restricted to the Marketing group. The data model would be a part of the enterprise data model. Page 85 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

77. Answer: D. Subject Area. The scope of a data model that represents financial data for an organization is subject area. Every area in the organization needs financial information of some sort, not just the Finance group. The Finance group will see the financial data for an entire organization. 78. Answer: C. Project Oriented. The scope of a data model that tracks costs for starting up operations in Plant A is Project Oriented. The data will only need to be collected while the project is in progress, and analyzed after the closure of the project. 79. Answer: A. Forward Engineering. Forward Engineering the data model is already created and the data modeling tool is creating the database scripts and will apply the scripts to create a database and/or tables in the database. Reverse Engineering captures information from a database or script file to create a physical data model. Creating Logical and Physical data model involves using the GUI to create a data model that conforms to business requirements. Splitting models of older version into separate logical and physical data models can create more manageable data models. 80. Answer: B. Reverse Engineering. Reverse Engineering creates a data model from an existing database. Forward Engineering the data model is already created and the data modeling tool is creating the database scripts and will apply the scripts to create a database and/or tables in the database. Creating Logical and Physical data model involves using the GUI to create a data model that conforms to business requirements. Importing a prior version data model is looking at an older version of the data model. 81. Answer: D. Create the Enterprise Data Model. Reverse Engineering benefits are: maintaining database, changing database types, analyzing the differences in databases. Reverse Engineering functionality cannot create the Enterprise Data Model unless the database has enterprise wide data. 82. Answer: D. Improving. Forward Engineering, Reverse Engineering and Comparisons are all typical Data Model tool functions. Improving is a function that does not exist in data modeling tools. Forward Engineering the data model is already created and the data modeling tool is creating the database scripts and will apply the scripts to create a database and/or tables in the database. Forward Engineering is also referred to as exporting. Reverse Engineering creates a data model from an existing database. The comparison function allows for comparisons between the data model and database; select the objects to compare and selectively import or export changes. 83. Answer: A. Keep data model and database synchronized. The benefit of Comparison function in data modeling tools is that it allows the data modeler or database administrator to keep the data model and database synchronized. The comparison function allows for comparisons between the data model and database; select the objects to compare and selectively import or export changes. These are features of the comparison function. The main benefit is the data model and database is synchronized.

Page 86 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

84. Answer: A. Forward Engineering. Forward Engineering the data model is already created and the data modeling tool is creating the database scripts and will apply the scripts to create a database and/or tables in the database. Forward Engineering is also referred to as exporting. 85. Answer: C. Versioning. Versioning records who made the change and when the change was made to provide full project audit trail thus enabling rollback options when comparing versions. Change control is a function that allows knowing the impact of change before saving. Model Merge is simply merging two data models together. Submodeling breaks a data model into smaller models (e.g. by subject area) for ease of use. 86. Answer: B. Model Merge. Model Merge supports merging of data models in an enterprise. Versioning records who made the change and when the change was made to provide full project audit trail thus enabling rollback options when comparing versions. Change control is a function that allows knowing the impact of change before saving. Submodeling breaks a data model into smaller models (e.g. by subject area) for ease of use. 87. Answer: D. Comparing two data models to detect changes. Model Merge features can be used to bring two data models together. Model Merge can: create an enterprise data model, bring individual data models together in a group, and merge two previously unrelated projects. If a data modeler wanted to compare two data models to detect changes, they would use a Comparison function if they were previously linked. 88. Answer: D. The functionality is not allowed. When importing and merging data models, data modelers can typically: create a new data model or update one of the existing data models. 89. Answer: D. Business Process Model. Although some tools offer linkages between data model and business process tools and have synchronization options, the tools are still typically separate. The benefit of this synchronization is it verifies the data models support the business processes and vice versa. Data Modeling tools have the breadth of data models that represent: Enterprise Data Model; Logical Data Model; and Physical Data Model. 90. Answer: C. Synchronize changes between data models. The benefit of linkages and mapping between enterprise, logical and physical data models is to synchronize changes between data models. The data models do not need to be stored in the same file, but need to maintain links between the different data models. Different data models are used as they serve different purposes in the application development process. A transformation function may include items like many-to-many where the relationship is dissolved by an identifying entity.

Page 87 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

4.0 Data / Metadata Infrastructure Management Overview Establishing data modeling standards provides a framework and guidance for new projects being implemented and provides a strong foundation for changes applications and the data. Standards, like entity naming, increase data sharing opportunities, reduce data redundancy, and improve application interoperability. Data security and privacy is an area of growing legal regulation and enforcement. Data security can be derived from the three principles of: Accountability, Authorization, and Availability. Accountability is the concept that every user must be responsible for their actions, so that in the event of any questionable activity or breach of policy, a specific user can be identified. Authorization is a concept that access to data and system resources should be limited to a need to know basis, and that specific users must be specifically allowed such access. Availability is the concept that system and data resources must be accessible whenever they are needed. Data privacy is being classified by the content, critically to business and availability requirements. OECD Guidelines define the Protection of Privacy and Transborder Flow of Data and provide privacy protection in relation to personal data.

Topics Standards, Policies, Procedures, Guidelines Standards Management Process Data Models Data Elements Data Security and Privacy Data Security Principles Data Security Policy Types

Page 88 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Questions 4.1 Standards, Policies, Procedures, Guidelines 1. What is the name of an entity? A. Customer B. Customers C. Customer_Name D. Name 2. Which is the best name for an entity describing employee information: A. EmployeeTable B. Employee_Table C. EmployeeTbl D. Employee 3. Which attributes follows best practises in naming? A. Social-insurance-number B. Social-insurance C. Social-insurance-code D. Social-insurance-numbers 4. When naming primary keys attributes follows best practises in naming? A. Customer-Ident B. Customer-Ids C. Customer-Id D. Customer-Identifier 5. Which of the following should be used when naming a relationship between an employee and a manager? A. Supervises B. Supervise C. Supervisor D. Is a 6. Which of the relationships capture a hierarchal relationship between employees: A. Has a B. Supervises C. Is a D. Manages

Page 89 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

7. Which of the following is the best description of a business rule: A. Repeat customers are categorized as type 01. B. Repeat customers are good for business. C. Repeat customers do not need a credit check. D. Repeat customers do not need a credit check if they are in good standing. 8. Data Integrity rules in a data warehouse environment include the following except: A. B. C. D.

Cleaning Redundancy resolution Business rule enforcement Random sampling.

9. What is the difference between ANSI and ISO/IEC? A. ISO/IEC form the specialized system for worldwide standardization, while ANSI that administers and coordinates the U.S. voluntary standardization and conformity assessment system. B. ISO/IEC cover the field of information technology while ANSI is concerned with certification. C. ISO/IEC cover the field of information technology while ANSI is concerned with Product Standards 10. After the creation of a Standards Management Process, which of the following is not considered: A. Approval B. Censure C. Enforcement D. Maintenance 11. Which of the following is not a notations typically used in logical and physical database design: A. Information Engineering B. IDEF1x C. Dimensional Modeling D. Conceptual Modeling

Page 90 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

12. Which one of the following is not a data element representation types: A. Amount B. Code C. Date D. Customer 13. A data element name that conforms to the ISO/IEC 11179 metadata registry naming convention does not have one of the following: A. Object B. Property C. Representation term D. Process Definition 14. In the ISO/IEC 11179 metadata is defined as: A. Data about data B. Data that defines and describes other data C. DNA of the data D. All information that is not the data itself. 15. What is the process for creating metadata when using a metadata registry? A. Attributing, Classifying, Defining, and Registering B. Attributing, Classifying, Defining, and Maintaining C. Creation, Approval, Enforcement, Registering D. Creation, Approval, Enforcement, Maintenance 16. Which of the following is not considered a sampling technique of data element audits: A. Random Sampling B. Systematic Sampling C. Cluster Sampling D. Standard Deviation. 17. In a data element audit, the data element must be: A. Valid and accurate B. Conformance of data values to its domain and business rules C. Complete D. Not null

Page 91 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

18. In a Data Warehouse, legacy element linkages are referred to as: A. Source System Metadata B. Data Staging Metadata C. DBMS Metadata D. Front Room Metadata 19. What should be standard for knowledge workers that analyze data in a data warehouse? A. Use data warehouse record and metadata to navigate back to the record in the source system. B. Complete the data in a data warehouse by adding postal information. C. Ability to consolidate the data in the data warehouse. D. Conduct a baseline assessment of the quality of the data. 20. Which of the following definition uses the best example of metadata principles? A. Employee ID Number - Number assigned to an employee. Employee - Person corresponding to the employee ID number. B. Employee – For the purpose of the data dictionary, employee is an internal stakeholder. C. Employee - an individual who has entered into or works under (or, where the employment has ceased, worked under) a contract of employment.

D. Employee - Someone who is hired by HR under the DMSP to provide services to a company.

4.2 Data Security and Privacy 21. The need for data security can be derived from three principles, which of the following does not apply? A. Agreement B. Accountability C. Authorization D. Availability 22. Logging into a system by supplying a user name and password is known as which data security principle? A. Accountability B. Authorization C. Availability D. Not a valid data security principle

Page 92 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

23. Authorization may be defined as: A. Authorization is a concept that access to data and system resources should be limited to a need to know basis, and that specific users must be specifically allowed such access. B. Authorization is the ability to grant power to make decisions from the data. C. Authorization is the policy that gives official instructions on the use of the data. D. Authorization is a digital document that describes a written permission from the issuer to use a service or a resource that the issuer controls or has access to use. The permission can be delegated. 24. Which data security principle defines that the system and data resources must be accessible whenever they are needed? A. Accountability B. Authorization C. Availability D. Not a valid data security principle 25. What is the main objective of the data steward responsibilities with respect to defining security policy? A. Define the security requirements, controls and mechanisms applicable to all data assets. B. Management of the data asset accessibility. C. Ensure compliance to the security policy. D. Define the data quality standards. 26. When defining the security policy, what is the trustee responsibility? A. Define the security requirements, controls and mechanisms applicable to all data assets. B. Management of the data asset accessibility. C. Ensure compliance to the security policy. D. Define the data quality standards. 27. What does the OECD Guidelines on the Protection of Privacy and Transborder Flows of Data protect? A. Provides privacy protection in relation to personal data. B. Determines how the data is transmitted between countries. C. Determines which countries can share data. D. Outlines the handling of data. 28. Which one of the following is not a typical Government security classified level? A. Top Secret / Secret Page 93 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

B. Confidential C. Restricted D. Classified 29. Which of the following applies the most when classifying data in a data warehouse? A. Only the summarized financial information needs to be confidential. B. Only detailed financial records need to be confidential. C. Data classification depends on the data content. D. All data in a data warehouse should be available. 30. Audit trail inspection is a classified under what type of security monitoring? A. Proactive B. Reactive C. Offensive D. Defensive 31. When data is considered mission critical, what is the class and data availability required, according to the Storage Networking Industry Association (SNIA) based upon Five 9s? A. Class 2 - 99% data availability B. Class 3 - 99.9% data availability C. Class 4 - 99.99% data availability D. Class 5 - 99.999% data availability 32. When data is classified as Class 1 90% data availability, what is the business classification of the data according to the Storage Networking Industry Association? A. Not important to Operations B. Important for Productivity C. Business Important Information D. Business Vital Information 33. Which of the following statements is the not true when data content is controlled in an organization? A. Business policies or regulatory rules require some/all data be retained “X” period of time. B. May need to prove records stored are “trustworthy” at later date C. High data value to an organization need to be accessible, available and protected D. Data availability is Class 1 – 90% data availability.

Page 94 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

34. Which of the following statements is the not true when data content is non-controlled in an organization? A. No business rules or regulations are requiring this data be kept for “X” period of time. B. Business just needs to keep the data archived and accessible C. High data value to an organization needs to be accessible, available and protected. D. Med-Low data value to an organization needs to be accessible and available.

Page 95 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Quick Answers 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 21. 22. 23.

A D A D A C C D A B D D D B A D A A A C A A A

24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34.

C B A A D C D D A D C

Page 96 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Detailed Answers 1. Answer: A. Customer. Naming an entity should follow a standard and uniform approach. Entity Names should be simple, clear, and expressed in business terms. Entities should be in noun or adjective noun format; singular; in business terms; and not process specific. Customers are plural. Customer_Name is not in business terms. Name is not clear to which name it is referring. 2. Answer: D. Employee. The best name for an entity describing employee information is simply Employee. Incorrect names include: EmployeeTable, Employee_Table, and EmployeeTbl. Employee_Table is not in business terms. EmployeeTbl is not clear. EmployeeTable is not in business terms. 3. Answer: A. Social-insurance-number. An attribute should be in the singular, consistent, and clearly defined in business terms. Properly defined attributes should define the data domain type like date, time, amount, code, name, quantity and description. Social-insure is unclear domain type. Social-insurance-numbers is plural. Social-insurance-code does not define the contents. 4. Answer: D. Customer-Identifier. An attribute should be in the singular, consistent, and clearly defined in business terms. Completely spelling out the attribute is preferred. If using an abbreviation, they should be as clear as possible and used consistently across the enterprise. If using abbreviations, whenever possible, use industry standard abbreviations. 5. Answer: A. Supervises. Relationships capture how two or more entities are related to one another. Relationships should be named after plural verbs. 6. Answer: C. Is a. Is a captures a hierarchal relationship between employees and the different classes of employees or subtypes. In this relationship, the new class or object has data or behavior aspects that are not part of the inherited class. 7. Answer: C. Repeat customers do not need a credit check. Business rules describe the operations, definitions and constraints for governing policies. The business rule needs to define in clear and concise business terms and easily applied. 8. Answer: D. Random sampling. Data integrity rules in a data warehouse environment include: cleaning, redundancy resolution and business rule enforcement. Random sampling is a technique and tool that may be used to conduct checks on the data. 9. Answer: A. ISO/IEC forms the specialized system for worldwide standardization, while ANSI that administers and coordinates the U.S. voluntary standardization and conformity assessment system. ISO (the International Organization for Standardization) and IEC (the International Electro technical Commission) form worldwide standards through technical committees. ANSI (The Page 97 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

American National Standards Institute), is the US representative to the ISO/IEC. 10. Answer: B. Censure. After the creation of a Standards Management Process, censure is not a considered activity. The four activities associated with the Standards Management Process are: Creation, Approval, Enforcement, and Maintenance. 11. Answer: D. Conceptual Modeling. The notations typically used in logical and physical database design are: information engineering, IDEF1x (ICAM DEFinition Language) and dimensional modeling. Conceptual Modeling is a high-level data model that leads to a logical model. 12. Answer: D. Customer. The following is not a data element type: Customer. Data element representation types are items like: Amount, Code, Date, Identifier, Name, Number, Text, Rate, and Year. Data elements are the fundamental units of data an Enterprise manages. 13. Answer: D. Process Definition. A data element name that conforms to the ISO/IEC 11179 metadata registry naming convention does not have of the following: Process Definition. It does have an object, property and representation term (Date, Time, etc) 14. Answer: B. Data that defines and describes other data. In the ISO/IEC 11179 metadata is defined as data that defines and describes other data. It specifies the kind and quality of metadata necessary to describe data, and it specifies the management and administration of that metadata in a metadata registry. 15. Answer: A. Attributing, Classifying, Defining, and Registering. The processes for creating metadata when using a metadata registry are: Attributing, Classifying, Defining, and Registering. 16. Answer D. Standard Deviation. The following is not considered a sampling technique of data element audits: Standard Deviation. Random, Systematic, and Cluster are all sampling techniques. Random chooses any record in the database. Systematic chooses every nth record. Cluster takes a subgroup of data based on a classification method like city. Standard deviation would be used to model the sample of data and calculate confidence that the data represents the whole. 17. Answer: A. Valid and accurate. In a data element audit, the data element must be valid and accurate. A data element could be complete (not null) and conform to its data values in its domain and business rules but still be inaccurate. 18. Answer: A. Source System Metadata. In a Data Warehouse, legacy element linkages are referred to as: Source System Metadata. DBMS Metadata is descriptions about the logical or physical data model. Data Staging Metadata is metadata around the staging area. Front Room Metadata is metadata for the front line users.

Page 98 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

19. Answer: A. Use data warehouse record and metadata to navigate back to the record in the source system. Prior to the source data being transformed to the data warehouse, it should be standardized, cleansed, completed, enhanced, consolidated and summarized where needed. The knowledge workers should be able to analyze the degree to which the data agrees with original source. Complete the data in a data warehouse by adding postal information and the ability to consolidate the data should be done prior to loading the data into the data warehouse. The knowledge worker should be told of the data quality in the data warehouse. In analyzing the data, they should not have to conduct a baseline assessment of the quality of the data. 20. Answer: C. Employee - an individual who has entered into or works under (or, where the employment has ceased, worked under) a contract of employment. Metadata principles are: state the essential meaning of the concept; be precise and unambiguous (Answer B); contain only commonly understood abbreviations(Answer C); be concise; be able to stand alone; be expressed without embedding rationale, functional usage, domain information, or procedural information; avoid circular reasoning (Answer A); and use the same terminology and consistent logical structure for related definitions. 21. Answer: A. Agreement. The need for data security can be derived from the three principles of: Accountability, Authorization, and Availability. 22. Answer: A. Accountability. Accountability is the concept that every user must be responsible for their actions, so that in the event of any questionable activity or breach of policy, a specific user can be identified. The specific security services that support accountability are identification, authentication, and auditing. Identification refers to a security service that recognizes a claim of identity by comparing a userid offered with stored security information. Authentication refers to a security service that verifies the claimed identity of the user, for example a password. Auditability refers to a security service that records information of potential security significance. Authorization is a concept that access to data and system resources should be limited to a need to know basis, and that specific users must be specifically allowed such access. Availability is the concept that system and data resources must be accessible whenever they are needed. 23. Answer: A. Authorization is a concept that access to data and system resources should be limited to a need to know basis, and that specific users must be specifically allowed such access. Access control refers to a security service that allows or denies a user request based on privilege, group information, or context. The specific security services that support authorization are access control and confidentiality. Confidentiality refers to a security service that prevents disclosure of information to unauthorized parties while the information is in use or in transit, or while the information is being stored or destroyed. Accountability is the concept that every user must be responsible for their actions, so that in the event of any questionable activity or breach of policy, a specific user can be identified. Availability is the concept that system and data resources must be accessible whenever they are needed.

Page 99 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

24. Answer: C. Availability. Availability is the concept that system and data resources must be accessible whenever they are needed. The necessity for availability is dependent upon your particular business proposition. The specific security service that supports availability is integrity. Integrity refers to a security service that guarantees data has not been altered, deleted, repeated, or rearranged during transmission, storage, processing, or recovery. Accountability is the concept that every user must be responsible for their actions, so that in the event of any questionable activity or breach of policy, a specific user can be identified. Authorization is a concept that access to data and system resources should be limited to a need to know basis, and that specific users must be specifically allowed such access. 25. Answer: B. Management of the data asset accessibility. The main objective of the data steward responsibilities with respect to defining the security policy is the management of the data asset accessibility. The data steward does not ensure compliance to the security policy, define the data quality standards, or define the security requirements, controls and mechanisms applicable to all data assets. 26. Answer: A. Define the security requirements, controls and mechanisms applicable to all data assets. The trustee is entrusted with the administration of the data assets. 27. Answer: A. Provides privacy protection in relation to personal data. OECD Guidelines on the Protection of Privacy and Transborder Flow of Data provides privacy protection in relation to personal data. The Guidelines apply to personal data, no matter if the company is public or private sectors, because of the potential detriment to civil liberties. The guidelines specify how to collect, store, process or disseminate personal information. 28. Answer: D. Classified. The typical security classified levels are top secret, secret, confidential, restricted and the lowest level of unclassified. The levels determine the impact on national security if the data was to be made public. Corporations typically have a similar type of classification structure. 29. Answer: C. Data classification depends on the data content. When classifying data in a data warehouse, the data content needs to be evaluated. In some cases the detailed data should be confidential and in others, the summarized data. The challenge in data privacy is to share data while protecting the identifiable information. 30. Answer: D. Defensive. Audit trail inspection is classified under a Defensive type of security monitoring. Other types of defensive monitoring are: Role Based Access Security and Process Rights Management. Offensive monitoring includes: Provisioning and Federated Identity Management. 31. Answer: D. Class 5 – 99.999% data availability. Mission critical data availability according to the Storage Networking Industry Association based upon Five 9s is Class 5 99.999%. The five classes are: Class 1 - 90% data availability: Not important to Operations Page 100 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Class 2 - 99% data availability: Important for Productivity Class 3 - 99.9% data availability: Business Important Information Class 4 - 99.99% data availability: Business Vital Information Class 5 - 99.999% data availability: Mission Critical Information 32. Answer: A. Not important to Operations. When data is classified as Class 1 – 90% data availability, the business classification of the data according to the Storage Networking Industry Association is Not Important to Operations. The five classes are: Class 1 - 90% data availability: Not important to Operations Class 2 - 99% data availability: Important for Productivity Class 3 - 99.9% data availability: Business Important Information Class 4 - 99.99% data availability: Business Vital Information Class 5 - 99.999% data availability: Mission Critical Information 33. Answer: D. Data availability is Class 1 – 90% data availability. When data content is controlled in an organization, the organization has a compliance or fiduciary responsibility for the data. Therefore, the controlled data has a high data value to an organization and needs to be accessible, available and protected in a “trustworthy” state and retained for a defined period of time. Data Value defines the business significance of different classes of data and the degree to which data has to be accessible, available, and protected. 34. Answer: C. High data value to an organization needs to be accessible, available and protected. When data content is non-controlled in an organization, the organization does not have a compliance or fiduciary responsibility for the data, only a business usage. Therefore, the controlled data has med-low data value to an organization and needs to be archived and accessible and the retention period is defined by the business and useful of the data, not an external entity.

Page 101 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

5.0 Information Quality Management Overview Information quality is as important as the data in an organization, but often takes a back seat. If the data cannot be depended upon, used and leveraged within the organization it holds no value, ergo “Garbage In, Garbage Out”. Information Quality Principles set the foundation for evaluating the quality of data through: Information Quality Characteristics, Data Definition (or Information Product Specification) and Quality Characteristics. Once the “parameters” of the data have been define, information quality assessment and ongoing information audits should occur. Audit and assessments measure the quality of data, either in physical form (file, database, spreadsheet) or output from a process. Accessing the data may take the form of Random sampling, Cluster sampling or Systematic sampling. All results would be written in an Information Quality Report for either further investigation or action. Once the Information Quality improvements have been identified, and their root cause determined, the process that produces the defective data must be fixed, plus the physical data. A cycle of continuous improvement of the data quality needs to be implemented in every organization. One approach, the Shewhart cycle, named for Walter Shewhart, discussed the concept in his 1939 book, "Statistical Method From the Viewpoint of Quality Control", is the continuous improvement cycle of Plan, Do, Check, Act. Overall, organizations who want to maximize data and data management by exposing, mitigating and managing quality within their business will reap the rewards to prevent first and restore second.

Topics Information Quality Principles Definition Information Quality Characteristics Data Definition (or Information Product Specification) Quality Characteristics Information Quality Assessment / Audit Quality Assessment Characteristics Quality /Cost Measurement Information Quality Improvement Data Corrective Maintenance Data Movement Control Information Quality Process Improvement Information Quality Culture Transformation

Page 102 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Questions 5.1 Information Quality Principles 1. What is the difference between data and information? A. Data is the representation of facts, information is data in context. B. Data is the context and Information is the content. C. Data is the value and Information is a valuable enterprise resource. D. Data is meaningful and information has context. 2. What is Information Quality? A. Quality data that enables knowledge workers to answer their questions. B. Correctness or accuracy of the data and the degree of usefulness and value data has to the organization. C. Quality data values in a data attribute field. D. Valid and meaningful information that the enterprise can make decisions upon. 3. Who benefits the most from Information Quality? A. Knowledge Workers B. Management C. Data Modeler D. Data Base Administrator 4. What type of data is required when the consequences of nonquality cause major process failure? A. Complete Quality data B. Accurate data C. Scientific data D. Zero-defect data 5. According to Larry English, what are the components that make up information? A. Information = Data + Definition B. Information = Data + Definition + Quality C. Information = Data + Definition + Presentation D. Information = Data + Definition + Content 6. What is Data Definition Quality? A. The definition, quality, and accuracy that govern data. B. The definition, domain value set, and business rules that govern data. C. The information about the data that meets manager expectations. D. The information that departmental knowledge workers depend on.

Page 103 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

7. What is needed to produce consistent high-quality information? A. Information Product Specification B. Data Definitions C. Customer Requirements D. Data Models 8. What are the features of Data Definition Quality? A. Data standards quality, data definition quality and information architecture quality. B. Data standards quality, data model quality and information architecture quality. C. Data models quality, domain type consistency, and definition quality. D. Business rule quality, data definition quality, and domain value quality. 9. What is Data Name quality? A. The name of the data plainly conveys the meaning of the objects named. B. The name of the data enables knowledge workers to easily define data completely and consistently across the organization. C. The name of the data that the enterprise needs to know about is consistent with the entity name. D. The name of the data indicates the type of data represented. 10. What is “Definition Conformance”? A. The measure of conformance of data its domain. B. The measure of conformance of data to the information product specification. C. To be consistent with the right level of granularity of the data values. D. To be consistent with the meaning of actual data values with its data definition. 11. Which of the following describes “the right level of granularity in the data values”? A. Definition Conformance B. Completeness C. Accuracy D. Precision 12. Which of the following describes “the characteristic of having all required values for the data fields”? A. Definition Conformance B. Completeness C. Accuracy D. Precision

Page 104 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

13. Which of the following depends on a reliable and precise recording of information? A. Definition Conformance B. Completeness C. Accuracy D. Precision 14. Which one of the following does nonduplication principles lead to the least? A. Identical values in multiple records. B. Duplicate records of a single event. C. Same data maintained in many independent, distributed databases. D. Its not a problem if there are standards and controls. 15. Which of the following describe when data is semantically equivalent? A. Consistency of Redundant data B. Timeliness C. Usefulness D. Objectivity 16. What does Information Float describe? A. The timetable to gather data. B. Length of time from when data is known, until it is available for a specific process or use. C. The average time required for data to be disseminated in the organization. D. The degree of variance in the data. 17. In a decision support system, which of the following is considered most useful? A. Tabular data B. Data list C. Graphic presentation (bar chart, etc) D. It depends on the type of data 18. What is presentation clarity? A. The degree the knowledge worker can understand the meaning of the data through the presentation. B. Statements of fact with a neutral point of view including reporting without bias, and emphasis on initutive presentation to the knowledge worker. C. The ability to see the legends, rows and columns in a bar chart. D. The redesign producing high quality data content or presentation.

Page 105 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

19. Which one is a statement of fact with a neutral point of view, reporting without bias, and emphasis on initutive presentation to the knowledge worker called? A. Consistency of Redundant data B. Timeliness C. Usefulness D. Objectivity 20. What are the primary inherent information quality characteristics? A. Information Product Specification, Data Definitions, Data Models. B. Definition, domain value set, and business rules that govern data. C. Definition conformance, completeness, validity, accuracy, precision, nonduplication, and consistency of redundant data. D. Accessibility, timeliness, contextual clarity, derivation integrity, usability, completeness. 21. What are the primary pragmatic information quality characteristics? A. Information Product Specification, Data Definitions, Data Models. B. Definition, domain value set, and business rules that govern data. C. Definition conformance, completeness, validity, accuracy, precision, nonduplication, and consistency of redundant data. D. Accessibility, timeliness, contextual clarity, derivation integrity, usability, completeness. 22. Which definition for Employee-start-date conforms best to the Attribute definition quality principles? A. The date an employee was hired. B. The date a new employee was hired. C. The date that an employee first started with the company regardless of location. D. Tells when the employee first came to work 23. Using data name and definition consistency principles, which attributes best describes: The date a service is started with the customer? A. Product-start-date B. Service-start-date C. Service-release-date D. Data-Service-release-date

Page 106 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

24. Which of the following is the most appropriate to demonstrate Entity Type Name Clarity? A. Retail Customer B. Customers C. Buying Customers D. RTL_CUST 25. Which of the following is the most appropriate to demonstrate Attribute Name Clarity? A. Retail Customer Name B. Customers Identities C. Buying Customers Names D. RTL_CUST_NAME 26. Which one of the following is not known as a domain type? A. Date B. Time C. Amount D. Customer 27. For a cross-reference attribute, what is the most appropriate abbreviation following acronym clarity principles? A. Xref B. Cross-ref C. X-reference D. Crss-rfrnc 28. Why should names be appropriate to the knowledge workers? A. Attributes names are consistent where facts context are equivalent. B. Attributes names are equivalent to ease dissemination of information. C. Attribute names are accurate to express the meaning of the fact being defined. D. Attribute names are clear and concise business terms. 29. What role in an organization is responsible for the enterprise wide glossary? A. Business Steward B. Knowledge Worker C. Data Modeler D. Database Administrator

Page 107 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

30. What role in an organization is responsible for the business term definition? A. Knowledge Worker B. Data Modeler C. Database Administrator D. Subject Matter Expert 31. Where should business terms be defined? A. Glossary B. Data Model C. Data Dictionary D. Database 32. Which of the following is not true for Business rules: A. Based on technical or existing system limitations. B. Expressed in business terms. C. Complete and specific. D. Defines an aspect of business to take action.

5.2 Information Quality Assessment / Audit 33. Why is an information quality assessment/audit conducted? A. To measure the quality of data and information of processes. B. To measure the quality of data, either in physical form(file, database, spreadsheet) or output from a process. C. To measure the quality of data and information for statistical quality control. D. To assess the stable processes. 34. Which of the following is not a purpose of information quality assessment? A. Check the inputs and outputs of processes to ensure accuracy. B. Declare the accuracy and reliability of data. C. Provide feedback to the technical people who create and maintain the data. D. Determine which processes should be retired. 35. Which data sampling technique is not valid when conducting an Information Quality Assessment? A. Linear Regression B. Random sampling C. Cluster sampling D. Systematic sampling

Page 108 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

36. Which of the following is not a data assessment test? A. Validity of business rule conformance. B. Timeliness and Nonduplication. C. Accuracy to surrogate source including derivation integrity. D. Usability 37. Which of the following is not a technique for Information Quality Report? A. Pareto Diagram. B. Bar Chart. C. Statistical Control Chart. D. Business Glossary List. 38. Which one of the following is not one of the redundant costs? A. Costs of inconsistent data. B. Cost to capture of interface. C. Data residing in multiple databases. D. Data residing in a single sharable database. 39. Which of the following do not determine the cost of data? A. Cost basis of developing and maintaining infrastructure. B. Cost to produce a product. C. Value basis uses the information to add value for the enterprise. D. Cost to define information requirements and design and build applications and databases. 40. Which of the following is not a direct result of process failure information costs: A. Duplicate catalogue mailings to a single customer. B. Knowledge workers re-verifying the data. C. Incorrect purchasing decision made. D. Unhappy supplier trying to make corrections several times through customer support. 41. What is the relationship between quality information and the business? A. Quality information costs the enterprise overall. B. Information enables the enterprise to accomplish its mission and goals. C. No relation between information and the bottom line. D. The relation between information and the bottom line cannot be quantified.

Page 109 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

5.3 Information Quality Improvement 42. Where is it best to correct defective data? A. In downstream databases. B. Fix the data at the source. C. Fix the process that produces the defective data by identifying root cause. D. Mark the defective data through metadata until it can be fixed. 43. What is the process that takes existing data, which is defective and brings the data to suitable levels of quality? A. Data Reengineering and cleansing process. B. Extract Transform and Load (ETL) process. C. Data Loading process. D. Metadata process. 44. Which one of the following will Data Architects not consider when embarking on an Information Product Improvement project? A. Source Data Cleansing B. Data Conversions C. Data Scrubbing D. Presenting data. 45. An analysis of the data revels that over 60% of the phone numbers in the database are the same number 000-0000. Further analysis reveals that the phone number cannot be null. A rule is added to the interface that the phone number cannot be zeros. This is known as: A. Source Data Cleansing B. Data Conversion C. Data Scrubbing D. Best Practice 46. In the source to target mapping of data, the data is standardized so that values depicting gender are the uniform. This is known as: A. Source Data Cleansing B. Data Conversion C. Data Scrubbing D. Best Practice

Page 110 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

47. What area is considered part of defining summary and derived data, adding data from external sources, consolidating data? A. Source Data Cleansing B. Data Conversion C. Data Scrubbing D. Best Practice 48. Which group should conduct the quality audit and control of data movement procedures? A. Internal Audit B. Knowledge Workers C. Data Conversion Specialists D. Information Steward 49. Which one of the following is not a cost of information quality? A. Nonquality information costs. B. Information quality assessments/audits. C. Information quality process improvements and prevention. D. Incorrect decisions made on poor quality information. 50. Which of the following is a proactive technique in Information Process Quality Improvement? A. Fix the symptoms. B. Ignore Problem signs until they become issues. C. Analyze root cause and eliminate the cause. D. Following Best Practise in IT System Management 51. Which one of the following techniques is not used in root cause analysis for determining quality issues? A. Cause-and-effect diagram B. Interrelationship diagram C. Current reality tree D. Value chain relationship diagram 52. Which step is not involved in the Shewhart cycle? A. Plan and Do B. Check C. Act D. Refine

Page 111 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

53. When implementing an Information Quality Culture Transformation which one of the following is most important to do well? A. Training B. Management Buy-in C. Define a Methodology D. Data Definition quality assessment process 54. Who is accountable for the integrity of the processes and quality of information? A. Knowledge Steward B. Managerial Information Steward C. Process Steward D. Business Information Steward 55. What is the first step an enterprise should take when embarking on an information quality program? A. Conduct an Information Quality Management Maturity Assessment and gap analysis. B. Create a vision, mission, and objectives for the information quality program. C. Appoint an Information Quality Leader. D. Conduct a customer satisfaction survey. 56. Which state is the least mature in the Information Quality Management Maturity Assessment? A. Uncertainty B. Awakening C. Enlightenment D. Wisdom and Certainty 57. Which state is characterized by knowing a data quality problem exists but not knowing what to do about it? A. Uncertainty B. Awakening C. Enlightenment D. Wisdom and Certainty 58. Which state has adopted a commitment to quality and implements the 14-point program? A. Uncertainty B. Awakening C. Enlightenment D. Wisdom and Certainty Page 112 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

59. Which state has prevention becoming a normal part of operations? A. Awakening B. Enlightenment C. Wisdom D. Certainty 60. Which state is the most mature in the Information Quality Management Maturity Assessment? A. Awakening B. Enlightenment C. Wisdom D. Certainty 61. When conducting an Information Quality Management Maturity Assessment, what are the five stages assessed across the Measurement Categories? A. Management understanding and attitude; Information Quality Organization Status; Information Quality Problem Handling; Cost of Information Quality as a Percent of Revenue or Operating Budget; Information Quality Improvement Actions; and Summary. B. Plan; Do; Check; Act C. Definition conformance, completeness, validity, accuracy, precision, nonduplication, and consistency of redundant data. D. Accessibility, timeliness, contextual clarity, derivation integrity, usability, completeness.

Page 113 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Quick Answers 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 21. 22. 23.

A B A D C B A A A D D B C C A B D A D C D C B

24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. 40. 41. 42. 43. 44. 45. 46.

A A D B A A D A A B D A D D D B B B C A D A B

47. 48. 49. 50. 51. 52. 53. 54. 55. 56. 57. 58. 59. 60. 61.

C D D C D D A B A A B C C D A

Page 114 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Detailed Answers 1. Answer: A. Data is the representation of facts, information is data in context. Data is the raw material and information is the finished product. 2. Answer: B. Correctness or accuracy of the data and the degree of usefulness and value data has to the organization. Information Quality needs to have inherent quality (data accuracy) and pragmatic quality (usefulness and value to support the enterprise process that enable accomplishing enterprise objectives). Data needs to support the knowledge workers decision-making process or else it holds no value to the organization. 3. Answer: A. Knowledge Workers. Knowledge Workers benefit the most from Information Quality because they require data to do their jobs to the benefit of the endcustomer. 4. Answer: D. Zero-defect data. Zero-defect data is required when the consequences of nonquality cause major process failures. An example of data that must be accurate is domain reference data like: Medical diagnosis codes. 5. Answer: C. Information = Data + Definition + Presentation. As defined by Larry English, the three components that make up information are meaning (definition) of a fact (data) in a context (presentation). 6. Answer: B. The definition, domain value set, and business rules that govern data. Data Definition quality is the degree to which the data definition describes the meaning of the data and meets the needs of all stakeholders to understand the data and the context. 7. Answer: A. Information Product Specification. To produce consistent high-quality information an information product specification is needed. The specification states clearly and definitely the requirements along with acceptable product variations. 8. Answer: A. Data standards quality, data definition quality and information architecture quality. Data Definition Quality or “Information Product Specification Quality” is the specification for building well designed information architecture like manufacturing a Product. 9. Answer: A. The name of the data plainly conveys the meaning of the objects named. 10. Answer: D. To be consistent with the meaning of actual data values with its data definition. Definition conformance is comprised of data definition quality, validity and accuracy. 11. Answer: D. Precision. Precision is the characteristic of having the right level of granularity in the data values. For example, measurement of uptime of a system(99.999 or 99.5 ) needs to allow for a finer breakdown of values instead of 99% available.

Page 115 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

12. Answer: B. Completeness. Completeness is the characteristic of having all required values for the data fields. The completeness is measured by the extent of sparsity of the data. For example, Resident_Age should be a number greater than zero and not null. 13. Answer: C. Accuracy. Accuracy depends on a reliable and precise recording of information or agreement to original source. 14. Answer: C. Same data maintained in many independent, distributed databases. Nonduplication represents a one-to-one correlation between a record and the event. An example of duplication may reside in customer databases in a company where each department may capture their own customer information. 15. Answer: A. Consistency of Redundant data. When data is semantically equivalent System A Parent has domain values of (Father, Mother); System B Parent_No (1,2), System C Parent_Abbr (F, M) all mean the same thing. 16. Answer: B. Information Float is the length of time from when data is known, until it is available for a specific process or use. The measure of information float is the average time required for data to be disseminated in the organization. Both concepts are related to the timeliness: the relative availability of data within the time required by the knowledge worker. For example, an order may be filled out on an order form and faxed to the company. Timeliness represents the time taken to enter the order form into the system. 17. Answer: D. It depends on the type of data. Usefulness refers the form of information presentation and the degree to which a knowledge worker can readily interpret the results. 18. Answer: A. Presentation clarity is the degree the knowledge worker can understand the meaning of the data through the presentation. Presentation clarity avoids misinterpretation of the results. May also be referred to as contextual clarity. 19. Answer: D. Objectivity. Objectivity is a statement of fact with a neutral point of view, reporting without bias, and emphasis on intuitive presentation to the knowledge worker. Objectivity may also be referred to rightness or fact completeness. 20. Answer: C. Definition conformance, completeness, validity, accuracy, precision, nonduplication, and consistency of redundant data. 21. Answer: D. Accessibility, timeliness, contextual clarity, derivation integrity, usability, and completeness. 22. Answer: C. The date that an employee first started with the company regardless of location. The attribute quality definition describes in business terms the definition of the attribute that clearly defines one and only one attribute. 23. Answer: B. Service-start-date. Data name and definition consistency measures how well the name and definition is understood in the organization. Page 116 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

24. Answer: A. Retail Customer. Retail Customer is the most appropriate to demonstrate Entity Type Name Clarity. Entity type name clarity is easily understood by the knowledge worker and represents the objects. The characteristics of Entity Name Type Clarity are: singular nouns, business terms and easily comprehended by the knowledge worker. 25. Answer: A. Retail Customer Name. Retail Customer Name is the most appropriate to demonstrate Attribute Name Clarity. Attribute Name Clarity is easily understood by the knowledge worker and represents the facts. The characteristics of Attribute Name Clarity are: business terms, easily associated with the entity type and easily comprehended by the knowledge worker. 26. Answer: D. Customer. Domain type consistency (also known as class word) in an attribute represents the type of data stored. For example, Start-Date attribute has a domain type of date and the valid values would be a subset of all possible dates. Typical domain types include: date, time, amount, identifier (id), amount, code, name, quantity, percent, rate, and description. 27. Answer: B. Cross-ref. The most appropriate abbreviation for acronym clarity for the term cross-reference is Cross-ref. When using abbreviations, they should be documented in a single, enterprise-wide standards abbreviation list that is used consistently throughout the enterprise. Rules of thumb for creating abbreviations are: use industry-standards or universally accepted abbreviations where applicable, use short abbreviations without loss of meaning, and always use the first letter of the term. 28. Answer: A. Attributes names are consistent where facts context are equivalent. Names should be appropriate or consistent across the enterprise even across different formats of presentation and storage formats. 29. Answer: A. Business Steward. A Business Steward in the enterprise is responsible for keeping the enterprise wide glossary current in an organization. The responsibilities include adding, changing and deleting the definitions as needed. 30. Answer: D. Subject Matter Expert. A Subject Matter Expert or Business Information Steward in the enterprise is responsible for keeping the business term definitions current in an organization. The responsibilities include adding, changing and deleting the definitions as needed. 31. Answer: A. The business terms should be defined in the glossary that is enterprise wide. The glossary can take on several forms. 32. Answer: A. Based on technical or existing system limitations. Business rules should not be due to a technical or existing system limitation. Business rules should be expressed in business terms, defines an aspect of the business to take action, complete and specific, plus defines who, what, when, why and how and identities any exceptions. Page 117 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

33. Answer: B. To measure the quality of data, either in physical form (file, database, spreadsheet) or output from a process. The information quality assessment is conducted for the benefit of the knowledge workers. 34. Answer: D. Determine which processes should be retired. The purpose of information quality assessments is evaluating the processes and data, certify the data, and providing feedback plus measuring against the baseline to calculate the costs of nonquality. 35. Answer: A. Linear Regression. Linear Regression is not a valid sampling technique when conducting an Information Quality assessment. 36. Answer: D. Usability. Usability is not a concern of the data assessment test. Data assessment tests measure: Validity of business rule conformance, Timeliness, Nondupliction, Accuracy to surrogate source including derivation integrity, and Consistency of data. 37. Answer: D. Business Glossary List. The Information Quality Report deduces and reports on the data assessment using Pareto Diagrams, Bar Chart, Statistical Control Charts and outputs from Information quality analysis software. 38. Answer: D. Data residing in a single sharable database. The cost of redundancy is part of the cost formula of information quality in the value basis component. Redundant costs occur when the data is contained in multiple databases. There is a cost to capture and control all the multiple databases and of the inconsistent or inaccurate data to the organization. Data residing in a database has only potential value. Information value occurs through usage only. 39. Answer: B. Cost to define the interface to acquire customer data. The cost of data is comprised of two areas: cost basis and value basis. Cost basis is the cost of developing and maintaining infrastructure. It is made up of the cost to define information requirements, develop information, application, and technology architectures; and to design and build applications and databases. Value basis uses the information to add value for the enterprise. 40. Answer: B. Knowledge workers re-verifying the data. Process failure information costs result in spent costs, liability and exposure costs, and recovery costs. 41. Answer: B. Information enables the enterprise to accomplish its mission and goals. The enterprise business performance objectives need to be aligned to measuring information quality. 42. Answer: C. Fix the process that produces the defective data by identifying root cause. Data cleansing fixes the data. Information process quality improvements fix the process that produces the defective data. The process is typically iterative that involves the Page 118 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

cycles of: planning, implementing, assessing and rollout (Shewhart cycle of Plan-DoCheck-Act or PDCA). 43. Answer: A. Data Reengineering and cleansing process. The process that takes existing data, which is defective and brings the data to suitable levels of quality, is known as data reengineering and cleansing. Data reengineering is similar to reverse engineering but only looks at the data not the application or system. It mainly focuses on how the data is used in an organization and can work backwards to the data models. Data reengineering often results in a deeper understanding of data assets of the enterprise and may lead to areas like: data consolidation, data architecture, and data acquisition strategies. Data Cleansing is the act of identifying and correcting data. Correcting data involves cleaning up data that is incorrect, out-of-date, redundant, incomplete, or formatted incorrectly. 44. Answer: D. Presenting data. Data Architects will consider source data cleansing, data conversions, and data scrubbing when embarking on an Information Product Improvement project. These are known as the three data cleansing areas: Source Data Cleansing, Data Conversions and Data Scrubbing. 45. Answer: A. Source Data Cleansing. Source data cleansing improves existing data quality where the data is initially stored. 46. Answer: B. Data Conversion. Data conversion is the act of mapping the source to target and improving the quality of data by correcting, standardizing, de-duplicating, completing and formatting. 47. Answer: C. Data Scrubbing. Data scrubbing is an act of defining summary and derived data, adding data from external sources, consolidating data. 48. Answer: D. Information Steward. Information Steward should conduct quality audit and control of data movement procedures. They should get input from internal audit, knowledge workers and data conversion specialists. 49. Answer: D. Incorrect decisions made on poor quality information. There are three categories of information quality costs: Nonquality information costs (process, rework, lost and missed opportunity costs); Information quality assessments/audits; and Information quality process improvements and prevention. 50. Answer: C. Analyze root cause and eliminate the cause. Information Process Quality Improvement has two elemental processes: reactive and proactive. Proactive process involves conducting root cause analysis. Root Cause Analysis identifies not only what and how an event occurred, but also why it happened. Only when an analysis of why an event or failure occurred will corrective measures are found. 51. Answer: D: Value chain relationship diagram. Value chain relationship diagram is not a type of root cause analysis technique. Cause-and-effect diagram (Ishikawa or Page 119 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

fishbone diagram) breaks down causes into detailed categories so they can be organized into related factors to identify root cause. Interrelationship diagram quantifies the relationships between factors and classifies causes leading to root cause. Current reality tree classifies interdependent relationships between effects leading to the determination of root cause. 52. Answer: D. Refine. The Shewhart cycle of Plan-Do-Check-Act or PDCA because known as the Deming cycle, is the foundation to improve information process quality. 53. Answer: A. Training. Training is most essential when implementing an Information Quality Culture Transformation for both management and staff. The training needs to cover why quality is fundamental to the enterprise and how to achieve quality. When defining training requirements, identify the role and their responsibilities toward information quality for example: general information, policies and processes, usage, and information management principles. Next define their training requirements and learning objectives for each role. 54. Answer: B. Managerial Information Steward. The Managerial Information Steward is accountable for the integrity of the processes and quality of information. Knowledge steward is accountable for the use of information. Process Steward is accountable for the definition of a business process. Business Information is accountable for validating the definition of data. 55. Answer: A. Conduct an Information Quality Management Maturity Assessment and gap analysis. When embarking on an information quality program, the first step an enterprise should take is to conduct an Information Quality Management Maturity Assessment and gap analysis to determine the current state of the organization and where they would like to be in the future. 56. Answer: A. Uncertainty. The Uncertainty state is Stage 1 and the least mature in the Information Quality Management Maturity Assessment. In the Uncertainty stage, Information quality is not considered a management tool. When issues occur they are dealt with in a reactive manner. 57. Answer: B. Awakening. The Awakening state is Stage 2 and is characterized by knowing a data quality problem exists but not knowing what to do about it in the Information Quality Management Maturity Assessment. In the Awakening stage, Information quality issues have been identified but management does not commit to their resolution. When issues occur they are cleaned up rather than fixed at the source and dealt with in a tight scope. 58. Answer: C. Enlightenment. The Enlightenment state is Stage 3 and is characterized by adopting a commitment to quality and implementing the 14-point program in the Information Quality Management Maturity Assessment. In the Enlightenment stage, Quality Improvement Program is implemented with communication and resolution.

Page 120 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

59. Answer: C. Wisdom. The Wisdom state is Stage 4 and is characterized by implemented a program of prevention in the Information Quality Management Maturity Assessment. In the Wisdom stage, Quality Improvement Program is mastered with quality being integrated into all areas and a routine part of enterprise. 60. Answer: D. Certainty. The Certainty state is Stage 5 and is the most mature in the Information Quality Management Maturity Assessment. In the Wisdom stage, Quality Improvement Program is an essential part of the enterprise. 61. Answer: A. Management understanding and attitude; Information Quality Organization Status; Information Quality Problem Handling; Cost of Information Quality as a Percent of Revenue or Operating Budget; Information Quality Improvement Actions; and Summary. To conduct an Information Quality Management Maturity Assessment, the five stages are assessed across the following Measurement Categories: Management understanding and attitude; Information Quality Organization Status(maturity of information quality in the enterprise); Information Quality Problem Handling (acts or reacts to issues); Cost of Information Quality as a Percent of Revenue or Operating Budget; Information Quality Improvement Actions; and Summary. Capability can be quantified using measurable criteria and is realized through an evolutionary not revolutionary process.

Page 121 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.

The Data Management Examination Guide

Selected Bibliography Brackett, Michael, DATA SHARING: USING A COMMON DATA ARCHITECTURE, John Wiley, 1994, ISBN 04711309931. DAMA International & DAMA Chicago Standards Committee, “DATA MANAGEMENT ASSOCIATION: GUIDELINES TO IMPLEMENTING DATA RESOURCE MANAGEMENT, DAMA International, 4th edition, 2002. English, Larry P., IMPROVING DATA WAREHOUSE AND BUSINESS INFORMATION QUALITY, John Wiley & Sons, 1999, ISBN: 0471253839. Fowler, Martin, UML DISTILLED: APPLYING THE STANDARD OBJECT MODELING LANGUAGE, Addison-Wesley, 1997, ISBN: 0-201-32563-2. Inmon, W.H. BUILDING THE DATA WAREHOUSE, John Wiley, 2002, ISBN 0-471081302. Kimball, Ralph, THE DATA WAREHOUSE TOOLKIT, John Wiley, 1996, ISBN: 0471-15337-0. Marco, David, BUILDING AND MANAGING THE META DATA REPOSITORY: A FULL LIFECYCLE GUIDE, John Wiley, 2002, ISBN: 0471355232. McFadden, Fred R., Hoffer, Jeffrey A., and Prescott, Mary B. MODERN DATABASE MANAGEMENT, Fifth Edition, Addison-Wesley, 1999, ISBN 0-8053-6054-9. Simsion, Graeme and Witt, Graham, DATA MODELING ESSENTIALS, Third Edition, Morgan Kaufman, 2004, ISBN: 0126445516. Tannenbaum, Adrienne, METADATA SOLUTIONS: USING METAMODELS, REPOSITORIES, XML, AND ENTERPRISE PORTALS TO ACHIEVE INFORMATION ON DEMAND, Addison-Wesley, 2001, ISBN: 0201719762. Watson, Richard T. DATA MANAGEMENT: DATABASES AND ORGANIZATIONS, John Wiley & Sons, 2002, ISBN 0-471-41845-5.

Page 122 of 122 Copyright © 2006 by DAMA International & DAMA International Foundation. All rights reserved.