Data occurs when the same data is duplicated in multiple files of a database

focusNode

Didn't know it?
click below

Knew it?
click below

Data occurs when the same data is duplicated in multiple files of a database

Embed Code - If you would like this activity on your web page, copy the script below and paste it into your web page.

  Normal Size     Small Size show me how

QuestionAnswer
A schematic of the entire database that describes the relationships in a database is called an: Entity-relationship diagram
Which of the following technologies would you use to analyze the social media data collected by a major online retailer: Hadoop
The DBMS programming language that end users and programmers use to manipulate data in the database is called CGI True
A physical view shows data as it is actually organized and structured on the data storage media. True
All of the following are technologies used to analyze and manage big data except: Cloud computing
Multiple data marts are combined and streamlined to create a data warehouse. False
Which of the following is not a method for performing a data quality audit: Surveying data definition and query files
In data mining, the technique of predicting uses a series of existing values to determine what other future values will be. False
The coupling of data with the software programs that manipulate the data, such that changes in programs requires changes to the data, is called program-data dependence. True
The specialized language programmers use to add and change data in the database is called: A data manipulation language
In terms of data relationships, associations refers to: Occurrences linked to a single event
Every record in a file should contain at least one key field. True
Big data does not refer to: Datasets with fewer than a billion records
Which of the following features enables a DBMS to reduce data redundancy and inconsistency: Minimizes isolated files with repeated data
The logical view of a database: Presents data as they would be perceived by end users
A ________ is a characteristic or quality that describes a database entity. Attribute
Data scrubbing is a more intensive corrective process than data cleansing. False
Data cleansing not only corrects errors but also: Enforces consistency among different sets of data
The project operation: Creates a subset consisting of columns in a table
NoSQL technologies are used to manage sets of data that don't require the flexibility of tables and relations. True
A database record is a grouping of characters into a word, a group of words, or a complete number. False
Legacy systems are used to populate and update data warehouses. True
A field identified in a table as holding the unique identifier of the table's records is called the: Primary key
________ tools are used to analyze large unstructured data sets, such as e-mail, memos, survey responses, etc. to discover patterns and relationships. Text mining
Databases record information about general categories of information referred to as entities. True
High-speed analytic platforms use both relational and non-relational tools to analyze large datasets. True
An application that transfers information from an organization's internal database to a web server for delivery as part of a web page to the user is called middleware. True
A one-to-many relationship between two entities is symbolized in a diagram by a line that ends with: A crow's foot topped by a short mark
In place of application server software to allow a web server to communicate with a back-end database, you can use: CGI
Complicated groupings of data in a relational database need to be adjusted to eliminate awkward many-t-many relationships. True
The DBMS programming language that end users and programmers use to manipulate data in the database is called CGI. True
Data _______ occurs when the same data is duplicated in multiple files of a database. Redundancy
To analyze unstructured data, such as memos and legal cases, you would use text mining tools. True
Which of the following is not a DBMS for midrange computers: Microsoft Access
A household appliances manufacturer has hired you to help analyze their social media datasets to determine which of their refrigerators are seen as the most reliable. Which of the following tools would you use to analyze this data? Sentiment analysis software
OLAP is unable to manage and handle queries with very large sets of data. False
A data warehouse is composed of: Historic and current internal data
DBMS have a data definition capability to specify the structure of the content of the database. True
_______ creates confusion that hampers the creation of information systems that integrate data from different sources.
The join operation: Combines relational tables to provide the user. . .
Relational DBMSs use key field rules to ensure that relationships between coupled tables remain consistent. False
The most prominent data manipulation language today is: SQL
A _______ organizes data in two-dimensional tables. Relational DBMS
In a relational database, the three basic operations used to develop useful sets of data are: Select, project, and join
Which common database challenge is illustrated by a person receiving multiple copies of a catalog, each addressed to a slightly different variation of his or her name? Data inconsistency
In a table for customers, the information about a single customer would reside in a single: Row
To perform multidemensional data analysis, you would use: OLAP
You work for a retail clothing chain whose primary outlets are in shopping malls, and you are conducting an analysis of your customers and their preferences. Identifying sequences
Data mining is a tool for allowing users to: Find hidden relationships in data
The tool that enables users to view the same data in different ways using multiple dimensions is: OLAP
The relationship between entities and attributes is best illustrated by: The entity customer with the attribute purchase
Detecting and correcting data in a database or file that are incorrect, incomplete, improperly formatted, or redundant is called: Data scrubbing
Implementing a web interface for an organization's internal database usually requires substantial changes to be made to the database. False
The traditional file environment does not typically have a problem with: Program-data independence
With in-memory computing, a computer's primary memory (RAM) is used for processing. True
The select operation: Creates a subset consisting of all records in the file that meet stated criteria.
HTML has become the preferred method of communicating with back-end databases because it is a cross-platform language. False
You work for a national car rental agency and want to determine what characteristics are shared among your most loyal customers. To do this, you will want to use data mining software that is capable of: Classification
DBMSs typically include report-generating tools in order to: Display data in an easier-to-read format
In a client/server environment, a DBMS is located on a dedicated computer called a Web server. False
A _______ is an example of non-digital data storage that is comparable to a database. Library card catalog
The organization's rules for sharing, disseminating, acquiring, standardizing, classifying and inventorying information is called: Information policy
The logical and physical views of data are separated in a DBMS. True
An automated or manual file that stores information about data elements and data characteristics such as usage, physical representation, ownership, authorization, and the security is the: Data dictionary
The process of streamlining data to minimize redundancy and awkward many-to-many relationships is called: Normalization
A DBMS makes the: Physical database available for different views
A field identified in a record as holding the unique identifier for that record is called the: Key field
When the same attribute in related data files has different values, this is called data: Inconsistency
In associations, a data mining tool discovers different groupings within data, such as finding affinity groups for bank cards. False
The policies and processes for managing the integrity and security of data in a firm are referred to as data governance. True
You can manipulate data on a web server by using a cgi script. True
OLAP is a tool for enabling: Users to obtain online answers to ad-hoc questions rapidly.
Microsoft SQL Server is a: DBMS for midrange computers
A database server is software that handles all application operations between browser-based computers and a company's back-end business applications or databases. False
The special organizational function whose responsibilities include physical database design and maintenance is called: Database administration
You are creating a database to store temperature and wind data from national airport locations. Airport code
The data dictionary serves as an important data management tool by: Creating an inventory of the data elements
The organizational function that is in charge of policies and procedures for managing internal data resources is called data administration. True

What is the presence of duplicate data in multiple data files?

The presence of duplicate data in multiple data files is called program-data dependence.

Which of these describes the scenario where data is duplicated in multiple files of a database?

Data redundancy refers to the practice of keeping data in two or more places within a database or data storage system. Data redundancy ensures an organization can provide continued operations or services in the event something happens to its data -- for example, in the case of data corruption or data loss.
Answer: Data redundancy occurs when different divisions, functional areas, and groups in an organization independently collect the same piece of information. Because it is collected andmaintained in so many different places, the same data item may have:1. different meanings in different parts of the organization,2.

What is database redundancy with example?

A common example of data redundancy is when a name and address are both present in different columns within a table. If the link between these data points is defined in every single new database entry it would lead to unnecessary duplication across the entire table.