In gathering requirements for DW/BI projects, begin with the data goals and strategies first.
Within projects, conceptual data modelling and logical data modelling are part of requirements planning and analysis activities, while physical data modelling is a design activity.
Resource Description Framework (RDF), a common framework used to describe information about any Web resource, is a standard model for data interchange in the Web.
A data warehouse deployment with multiple ETL, storage and querying tools often
suffers due to the lack of:
There are numerous methods of implementing databases on the cloud. The most common are:
Data asset valuation is the process of understanding and calculating the economic value of data to an organisation. Value comes when the economic benefit of using data outweighs the costs of acquiring and storing it, as
An implemented warehouse and its customer-facing BI tools is a technology product.
Assessment capabilities are evaluated against a pre-determined scale with established criteria. This is important because:
The data-vault is an object-orientated, time-based and uniquely linked set of normalized tables that support one or more functional areas of business.
The accuracy dimension of data quality refers to the degree that data correctly respresents ‘real-life’ entities.
Reference and Master data definition: Managing shared data to meet organizational goals, reduce risks associated with data redundancy, ensure higher quality, and reduce the costs of data integration.
Three classic implementation approaches that support Online Analytical Processing include:
Poorly managed Metadata leads to, among other, redundant data and data management processes.
Bold means doing something that might cause short term pain, not just something that looks good in a marketing email.
DAMA International’s Certified Data Management Professional (CDMP) certification required that data management professionals subscribe to a formal code of ethics, including an obligation to handle data ethically for the sake of society beyond the organization that employs them.
A change management program supporting formal data governance should focus communication on:
Obtaining buy-in from all stakeholders
Within each area of consideration mentioned in question 13, they should address morale adversity as per Ethical Risk Model for Sampling Projects.
With respect to health data, what is the difference between the privacy and the security of the data?
Measuring the effects of change management on in five key areas including: Awareness of the need to change; Desire to participate and support the change; Knowledge about how to change; Ability to implement new skills and behaviors; and Reinforcement to keep the change in place.
Confidentiality classification schemas might include two or more of the five confidentiality classification levels. Three correct classifications levels are:
A node is a group of computers hosting either processing or data as part of a distributed database.
Effective data management involves a set of complex, interrelated processes that enable an organisation to use its data to achieve strategic goals.
A roadmap for enterprise data architecture describes the architecture’s 3 to 5-year development path. The roadmap should be guided by a data management maturity assessment.
A staff member has been detected inappropriately accessing client records from
usage logs. The security mechanism being used is an:
The ethics of data handling are complex, but is centred on several core concepts. Please select the correct answers.
A sandbox environment can either be a sub-set of the production system, walled off from production processing or a completely separate environment.
When we consider the DMBoK2 definition of Data Governance, and the various practitioner definitions that exist in the literature, what are some of the key elements of Data Governance?
A goal of metadata management is to manage data related business terminology in
order toc
A controlled vocabulary is a defined list of explicitly allowed terms used to index, categorize, tag, sort and retrieve content through browsing and searching.
ANSI standard 859 has three levels of control of data, based on the criticality of the data and the perceived harm that would occur if data were corrupt or otherwise unavailable, including:
Real-time data integration is usually triggered by batch processing, such as historic data.
While the focus of data quality improvement efforts is often on the prevention of errors, data quality can also be improved through some forms of data processing.
For each subject area logical model: Decrease detail by adding attributes and less-significant entities and relationships.
XML provides a language for representing both structures and unstructured data and information.
The first two steps of the Reference data Change request process, as prescribed DMBOk2, include:
Data parsing is the process of analysing data using pre-determined rules to define its content or value.
An effective Data Governance communication program should include the following:
A goal of Data warehouse and business intelligence is to support and enable ineffective business analysis and decision making by knowledge workers.
Once the most critical business needs and the data that supports them have been identified, the most important part of the data quality assessment is actually looking data, querying it to understand data content and relationships, and comparing actual data to rules and expectations.
Please select the 2 frameworks that show high-level relationships that influence how an organization manages data.
You have completed analysis of a Data Governance issue in your organisation and have presented your findings to the executive management team. However, your findings are not greeted warmly and you find yourself being blamed for the continued existence of the issue. What is the most likely root cause for this?
Tools required to manage and communicate changes in data governance programs include
Use business rules to support Data Integration and Interoperability at various points, to:
A critical step in data management organization design is identifying the best-fit operating model for the organization.
Data security includes the planning, development and execution of security policies and procedures to provide authentication, authorisation, access and auditing of data and information assets.
Decentralized informality can be made more formal through a documented series of connections and accountabilities via a RACI matrix.
Archiving is the process of moving data off immediately accessible storage media and onto media with lower retrieval performance.
Data professionals involved in Business Intelligence, analytics and Data Science are often responsible for data that describes: who people are; what people do; where people live; and how people are treated. The data can be misused and counteract the principles underlying data ethics.
Select the areas to consider when constructing an organization’s operating model:
Different levels of policy are required to govern behavior to enterprise security. For example:
The Zachman Framweork’s communication interrogative columns provides guidance on defining enterprise architecture. Please select answer(s) that is(are) coupled correctly:
An image processing system captures, transforms and manages images of paper and electronic documents.
A dimensional physical data model is usually a star schema, meaning there is one structure for each dimension.
Achieving security risk reduction in an organisation begins with developing what?
When starting a Data Governance initiative it is important to understand what the Business cannot achieve due to data issues because:
Referential Integrity (RI) is often used to update tables without human intervention. Would this be a good idea for reference tables?
Access to data for Multidimensional databases use a variant of SQL called MDX or Multidimensional expression.
Data Governance is at the centre if the data management activities, since governance is required for consistency within and balance between functions.
Effective data management involves a set of complex, interrelated processes that disable an organization to use its data to achieve strategic goals.
To build models, data modellers heavily rely on previous analysis and modelling work.
Consistent input data reduces the chance of errors in associating records. Preparation processes include:
When constructing models and diagrams during formalisation of data architecture there are certain characteristics that minimise distractions and maximize useful information. Characteristics include:
Product Master data can only focus on an organization’s internal product and services.
ETL is the basic process which is central to all areas in Data Integration and Interoperability. It is an abbreviation for extract, transition and load.
Data modelling tools are software that automate many of the tasks the data modeller performs.
A deliverable in the data modelling and design context diagram is the logical data model.
CMDB provide the capability to manage and maintain Metdata specifically related to the IT assets, the relationships among them, and contractual details of the assets.
Data Storage and Operations: The design, implementation and support of stored data to maximize its value.
You are a reporting Data Analyst. A new Management Report has been requested. What is the most effective way to ensure you receive the appropriate data at the correct level of accuracy to meet the business need?
Your organization has many employees with official roles as data stewards and data custodians, but they don't seem to know exactly what they're supposed to be doing. Which of the following is most likely to be a root cause of this problem?
In Resource Description Framework (RDF) terminology, a triple store is composed of a subject that denotes a resource, the predicate that expresses a relationship between the subject and the object, and the object itself.
The best DW/BI architects will design a mechanism to connect back to transactional level and operational level reports in an atomic DW.
A deliverable in the data security context diagram is the data security architecture.
When doing reference data management, there many organizations that have standardized data sets that are incredibly valuable and should be subscribed to. Which of these organizations would be least useful?
Information gaps represent enterprise liabilities with potentially profound impacts on operational effectiveness and profitability.
Domains can be identified in different ways including: data type; data format; list; range; and rule-based.
Because Data Governance activities require coordination across functional areas, the DG program must establish an ___________ that defines accountabilities and intersections.
Enterprise data architecture influences the scope boundaries of project and system releases. An example of influence is data replication control.
An information maturity assessment provides a valuable input to Data Governance planning because it provides:
With reliable Metadata an organization does not know what data it has, what the data represents and how it moves through the systems, who has access to it, or what it means for the data to be of high quality.
Data handling ethics are concerned with how to procure, store, manage, use and dispose of data in ways that are aligned with ethical principles.
Misleading visualisations could be an example where a base level of truthfulness and transparency are not adhered to.
What are some of the business drivers for the ethical handling of data that Data Governance should satisfy?
Orchestration is the term used to describe how multiple processes are organized and executed in a system.
Which of the following is an activity for defining a Data Governance strategy?
Layers of data governance are often part of the solution. This means determining where accountability should reside for stewardship activities and who the owners of the data are.
In an information management context, the short-term wins and goals often arise from the resolution of an identified problem.
Location Master Data includes business party addresses and business party location, as well as facility addresses for locations owned by organizations.
A goal of reference and master data management is for data to ensure shared data is:
A Business Glossary forces a business to adopt a single definition of a business term.
Data management organizational constructs include the following type of model.
The best preventative action to prevent poor quality data from entering an organisation include:
Within the Data Handling Ethics Context Diagram a key deliverable is the Ethical Data Handling Strategy.
One of the key differences between operational systems and data warehouses is:
Data profiling also includes cross-column analysis, which can identify overlapping or duplicate columns and expose embedded value dependencies.
A Global ID is the MDM solution-assigned and maintained unique identifier attached to reconciled records.
One of the percentages to measure success of a records management system implantation is the percentage of the identified corporate records declared as such and put under records control.
A change management program supporting Data Governance should focus communication on what?
A metadata repository is essential to assure the integrity and consistent use of an enterprise data model across business processes.
Improving data quality requires a strategy that accounts for the work that needs to be done and the way people will execute it.
The IBM Data Governance Council model is organized around four key categories. Select the answer that is not a category.
When presenting a case for an organization wide Data Governance program to your Senior Executive Board, which of these potential benefits would be of LEAST importance?
What position is responsible for the quality and use of their organization's data
assets?
The failure to gain acceptance of a business glossary may be due to ineffective:
Data Management Professionals only work with the technical aspects related to data.
Data access control can be organized at an individual level or group level, depending on the need.
An organization will create an uncover valuable Metadata during the process of developing Data Integration and Interoperability solutions.
Examples of the ‘Who’ entity category include: employee; patient; player; and suspect.
Reference data management entails the preventative maintenance of undefined domain values, definitions and the relationship within and across domain values.
Primary deliverables of the Data Warehouse and Business Intelligence context diagram include:
Business people must be fully engaged in order to realize benefits from the advanced analytics.
Effectiveness metrics for a data governance programme includes: achievement of goals and objectives; extend stewards are using the relevant tools; effectiveness of communication; and effectiveness of education.
The most informal enterprise data model is the most detailed data architecture design document.
The database administrator (DBA) is the most established and the most widely adopted data professional role.
Communications are essential to the success of a DMM or Data Governance assessment. Communications are important because:
A System of Reference is an authoritative system where data consumers can obtain reliable data to support transactions and analysis, even if the information did not originate in the system reference.
Please select the user that best describes the following description: Uses the business glossary to make architecture, systems design, and development decisions, and to conduct the impact analysis.
The need to manage data movement efficiently is a primary driver for Data Integration and Interoperability.
Many people assume that most data quality issues are caused by data entry errors. A more sophisticated understanding recognizes that gaps in or execution of business and technical processes cause many more problems that mis-keying.
A e-discovery readiness assessment should examine and identify opportunities for the commercial response program.
Integration of ETL data flows will usually be developed within tools specialised to manage those flows in a proprietary way.
Which DMBok knowledge area is most likely responsible for a high percentage of
returned mail?
Gathering and interpreting results from a DMM or Data Governance assessment are important because:
A pensioner who usually receives a quarterly bill of around $300 was sent a
$100,000,000 electricity bill. They were a victim of poor data quality checks in
which dimension?
Small reference data value sets in the logical data model can be implemented in a physical model in three common ways:
Snowflaking is the term given to normalizing the flat, single-table, dimensional structure in a star schema into the respective component hierarchical or network structures.
Which Data Architecture artefact contains the names of key business entities, their
relationships, critical guiding business rules and critical attributes?
Data flows map and document relationships between data and locations where global differences occur.
The purpose of enterprise application architecture is to describe the structure and functionality of applications in an enterprise.
All assessments should include a roadmap for phased implementation of the recommendations. This is important because:
The repeated implementation of different CRM technologies with different data
structures is mostly a failure of:
Document and content management is defined as planning, implementation and control activities for storage management of data and information found in any form or medium.
Achieving near-real-time data replication, using a source accumulation technique,
triggers on: