Home »
MCQs
ETL Testing Multiple-Choice Questions (MCQs)
ETL tools extract the data from all the different data sources, transforms the data and and loads it into a data warehouse.
ETL Testing MCQs: This section contains multiple-choice questions and answers on the various topics of ETL Testing. Practice these MCQs to test and enhance your skills on ETL Testing.
List of ETL Testing MCQs
1. Using an ____ tool, data is extracted from multiple data sources, transformed, and loaded into a data warehouse after joining fields, calculating, and removing incorrect data fields.
- ETL
- TEL
- LET
- LTE
Answer: A) ETL
Explanation:
Using an ETL tool, data is extracted from multiple data sources, transformed, and loaded into a data warehouse after joining fields, calculating, and removing incorrect data fields.
Discuss this Question
2. After business ____, ETL testing ensures that the data has been loaded accurately from a source to a destination.
- Information
- Transformation
- Transfusion
- Transfiction
Answer: B) Transformation
Explanation:
After business transformation, ETL testing ensures that the data has been loaded accurately from a source to a destination.
Discuss this Question
3. During ETL, various stages of data are verified and used at ____.
- Source
- Destination
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
During ETL, various stages of data are verified and used both at source and destination.
Discuss this Question
4. What is the full form of ETL?
- Extract Transformation and Load
- Extract Transformation and Lead
- Extract Transfusion and Load
- Extract Transfusion and Lead
Answer: A) Extract Transformation and Load
Explanation:
The full form of ETL is Extract Transformation and Load.
Discuss this Question
5. To fetch data from one database and place it in another, ETL combines all ____ database functions into one tool.
- Two
- Three
- Four
- Five
Answer: B) Three
Explanation:
To fetch data from one database and place it in another, ETL combines all three database functions into one tool.
Discuss this Question
6. Getting information from a database is called ____ (reading).
- Extracting
- Transforming
- Loading
- None
Answer: A) Extracting
Explanation:
Getting information from a database is called extracting (reading).
Discuss this Question
7. The process of ____ data involves converting it from one form to another.
- Extracting
- Transforming
- Loading
- None
Answer: B) Transforming
Explanation:
The process of transforming data involves converting it from one form to another.
Discuss this Question
8. In addition to using ____, the data can be combined with other data to undergo transformation.
- Rules
- Lookup tables
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
In addition to using rules or lookup tables, the data can be combined with other data to undergo transformation.
Discuss this Question
9. Writing data into a database is called ____.
- Extracting
- Transforming
- Loading
- None
Answer: C) Loading
Explanation:
Writing data into a database is called loading.
Discuss this Question
10. Using ETL, you can ____ the data from multiple sources and blend them together according to your needs.
- Extract
- Transform
- Load
- All of the above
Answer: D) All of the above
Explanation:
Using ETL, you can extract, transform, and load the data from multiple sources and blend them together according to your needs.
Discuss this Question
11. ETL is often used to build a -
- Data Center
- Data Warehouse
- Data Set
- Data Care Center
Answer: B) Data Warehouse
Explanation:
ETL is often used to build a Data Warehouse.
Discuss this Question
12. As part of the ETL process, data from a source system is ____ into a format suitable for storing in a data warehouse or another system.
- Extracted
- Converted
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
As part of the ETL process, data from a source system is extracted and converted into a format suitable for storing in a data warehouse or another system.
Discuss this Question
13. In today's environment, ETL is becoming more and more necessary for many reasons, including:
- In order to make critical business decisions, companies use ETL to analyze their business data.
- Data warehouses are repositories where data is shared.
- In ETL, data is moved from a variety of sources into a data warehouse.
- All of the above
Answer: D) All of the above
Explanation:
In today's environment, ETL is becoming more and more necessary for many reasons, including:
- In order to make critical business decisions, companies use ETL to analyze their business data.
- Data warehouses are repositories where data is shared.
- In ETL, data is moved from a variety of sources into a data warehouse.
Discuss this Question
14. What is TRUE about ETL?
- The ETL process allows the source and target systems to compare sample data.
- As part of the ETL process, complex transformations can be performed and additional storage space is required.
- Earlier, we defined ETL as a process of converting source data into target data and manipulating it.
- All of the above
Answer: D) All of the above
Explanation:
The things that are TRUE about ETL are -
- The ETL process allows the source and target systems to compare sample data.
- As part of the ETL process, complex transformations can be performed and additional storage space is required.
- Earlier, we defined ETL as a process of converting source data into target data and manipulating it.
Discuss this Question
15. A ____ is loaded with data through the ETL process.
- Data Mart
- Data Warehouse
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
A data mart or data warehouse is loaded with data through the ETL process
Discuss this Question
16. To facilitate business analysis, our data warehouse needs to be ____ regularly
- Extracted
- Transformed
- Loaded
- None
Answer: C) Loaded
Explanation:
To facilitate business analysis, our data warehouse needs to be loaded regularly.
Discuss this Question
17. To avoid affecting the source system's performance, the ____ occurs on the ETL server or staging area.
- Loading
- Extraction
- Transformation
- None
Answer: C) Transformation
Explanation:
To avoid affecting the source system's performance, the transformation occurs on the ETL server or staging area.
Discuss this Question
18. Before data is moved to the warehouse, the extracted data can be validated in the ____ area.
- Staging
- Staggering
- Studying
- None
Answer: A) Staging
Explanation:
Before data is moved to the warehouse, the extracted data can be validated in the staging area.
Discuss this Question
19. How many methods are there to extract the data?
- Two
- Three
- Four
- Five
Answer: B) Three
Explanation:
There are three methods to extract the data.
Discuss this Question
20. Which of the following is/are the method(s) to extract the data?
- FULL Extraction
- Partial Extraction - Without Update Notification
- Partial Extraction - With Update Notification
- All of the above
Answer: D) All of the above
Explanation:
The following are the methods to extract the data -
- FULL Extraction
- Partial Extraction - Without Update Notification
- Partial Extraction - With Update Notification
Discuss this Question
21. Whatever extraction method we use, the source system should not be affected in terms of ____ time.
- Performance
- Response
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Whatever extraction method we use, the source system should not be affected in terms of performance or response time.
Discuss this Question
22. Which of the following is/are the validation(s) using the extraction(s)?
- Check the source data against the record
- Ensure that the data type is correct
- There will be a check to see if all the keys are there
- All of the above
Answer: D) All of the above
Explanation:
The following are the validations using the extractions -
- Check the source data against the record
- Ensure that the data type is correct
- There will be a check to see if all the keys are there
Discuss this Question
23. It is not possible to use the extracted data as originally formatted from the ____ server.
- Source
- Site
- Storage
- None
Answer: A) Source
Explanation:
It is not possible to use the extracted data as originally formatted from the source server.
Discuss this Question
24. ____ refers to data that does not need to be transformed.
- Direct Move
- Pass-through data
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Direct move or pass-through data refers to data that does not need to be transformed.
Discuss this Question
25. Which of the following is/are the validation point(s) during the transformation?
- Filtering
- Conversion of character sets and encodings
- Checking the threshold and validity of data
- All of the above
Answer: D) All of the above
Explanation:
The following are the validation points during the transformation -
- Filtering
- Conversion of character sets and encodings
- Checking the threshold and validity of data
Discuss this Question
26. Which of the following is the last step of the ETL process?
- Extraction
- Transformation
- Loading
- None
Answer: C) Loading
Explanation:
The following is last step of the ETL process is Loading.
Discuss this Question
27. Which of the following is/are the type(s) of loading?
- Initial Load
- Incremental Load
- Full Refresh
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of loading -
- Initial Load
- Incremental Load
- Full Refresh
Discuss this Question
28. Loads need to be ____ according to server performance by the admin of the data warehouse.
- Monitored
- Resumed
- Canceled
- All of the above
Answer: D) All of the above
Explanation:
Loads need to be monitored, resumed, and canceled according to server performance by the admin of the data warehouse.
Discuss this Question
29. With a ____, all tables are erased and reloaded with new information.
- Initial Load
- Incremental Load
- Full Refresh
- None of the above
Answer: C) Full Refresh
Explanation:
With a Full Refresh, all tables are erased and reloaded with new information.
Discuss this Question
30. The term is ETL now extended to ____ or Extract, Monitor, Profile, Analyze, Cleanse, Transform, and Load.
- E-MPAC-TL
- E-PAC-TL
- E-MAP-TL
- E-MPAA-TL
Answer: A) E-MPAC-TL
Explanation:
The term is now extended to E-MPAC-TL or Extract, Monitor, Profile, Analyze, Cleanse, Transform, and Load.
Discuss this Question
31. During ____, the main goal is to capture data as quickly as possible from a system while minimizing the inconvenience to the system.
- Extraction
- Transformation
- Loading
- None
Answer: A) Extraction
Explanation:
During extraction, the main goal is to capture data as quickly as possible from a system while minimizing the inconvenience to the system.
Discuss this Question
32. ____ the data requires integrating the data and finally presenting the combined data to the end-user community via the front-end tools.
- Transforming
- Loading
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Transforming and loading the data requires integrating the data and finally presenting the combined data to the end-user community via the front-end tools.
Discuss this Question
33. Compared to traditional ETL, ETL ____ the time it takes for sources and targets to develop.
- Extends
- Reduces
- Exceeds
- Manipulates
Answer: B) Reduces
Explanation:
Compared to traditional ETL, ETL reduces the time it takes for sources and targets to develop.
Discuss this Question
34. ____ consists of a process between stages that is defined according to the needs, and it can verify the quality of the product
- Quantity Assurance
- Quality Assurance
- Quantity Attribution
- Quality Attribution
Answer: B) Quality Assurance
Explanation:
Quality Assurance consists of a process between stages that is defined according to the needs, and it can verify the quality of the product.
Discuss this Question
35. In ____, analysis and validation of the data pattern and formats will be performed, as well as identification and validation of redundant data across data sources to determine the actual content, structure, and quality of the data.
- Data Profiling
- Data Analysis
- Source Analysis
- Cleansing
Answer: A) Data Profiling
Explanation:
In data profiling, analysis and validation of the data pattern and formats will be performed, as well as identification and validation of redundant data across data sources to determine the actual content, structure, and quality of the data.
Discuss this Question
36. It is important to focus on the surroundings of the sources in ____ analysis, so that the documentation of the sources can be obtained.
- Data
- Source
- Profile
- None
Answer: B) Source
Explanation:
It is important to focus on the surroundings of the sources in source analysis, so that the documentation of the sources can be obtained.
Discuss this Question
37. Based on the Metadata of a set of predefined rules, errors found can be fixed in ____.
- Data Analysis
- Source Analysis
- Cleansing
- Data Profiling
Answer: C) Cleansing
Explanation:
Based on the Metadata of a set of predefined rules, errors found can be fixed in cleansing.
Discuss this Question
38. An extended ETL concept, E-MPAC-TL is designed to meet the requirements while taking into account the realities of the systems, ____, constraints, and most importantly, the data itself.
- Tools
- Metadata
- Technical Issues
- All of the above
Answer: D) All of the above
Explanation:
An extended ETL concept, E-MPAC-TL is designed to meet the requirements while taking into account the realities of the systems, tools, metadata, technical issues, constraints, and most importantly, the data itself.
Discuss this Question
39. ETL testing is also known as -
- Table balancing
- Product Reconciliation
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
ETL testing is also known as Table balancing or Product Reconciliation
Discuss this Question
40. An ETL test ensures the data loaded after transformation is accurate after it has been ____ from a source to a destination.
- Added
- Loaded
- Deleted
- Altered
Answer: B) Loaded
Explanation:
An ETL test ensures the data loaded after transformation is accurate after it has been loaded from a source to a destination.
Discuss this Question
41. ETL testing is performed in ____ stages.
- Two
- Three
- Four
- Five
Answer: D) Five
Explanation:
ETL testing is performed in five stages.
Discuss this Question
42. Which of the following is/are the stage(s) the ETL testing?
- Data recovery
- Build and populate data
- Build reports
- All of the above
Answer: D) All of the above
Explanation:
The following are the stages of the ETL testing -
- Data recovery
- Build and populate data
- Build reports
Discuss this Question
43. Which of the following is/are the type(s) of ETL testing?
- New Data Warehouse Testing
- Production Validation Testing
- Application Upgrade
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of ETL testing -
- New Data Warehouse Testing
- Production Validation Testing
- Application Upgrade
Discuss this Question
44. Customer requirements and different sources of data are taken into account in ____.
- New Data Warehouse Testing
- Production Validation Testing
- Application Upgrade
- Metadata Testing
Answer: A) New Data Warehouse Testing
Explanation:
Customer requirements and different sources of data are taken into account in New Data Warehouse Testing.
Discuss this Question
45. Which of the following is/are the group that plays the responsibility in testing New Data Warehouses -
- Business Analyst
- Infrastructure People
- QA Testers
- All of the above
Answer: D) All of the above
Explanation:
The following are the group that plays responsibility in testing New Data Warehouses -
- Business Analyst
- Infrastructure People
- QA Testers
Discuss this Question
46. What is the responsibility of a Business Analyst?
- Requirements are gathered and documented by the business analyst.
- The test environment is set up by Business Analyst people.
- These plans and scripts are developed by Business Analysts and then executed by them.
- Each module is tested by a Business Analyst.
Answer: A) Requirements are gathered and documented by the business analyst.
Explanation:
Requirements are gathered and documented by the business analyst.
Discuss this Question
47. The ____ develops test plans and scripts and executes these plans and scripts.
- Infrastructure People
- QA Testers
- Developers
- Users
Answer: B) QA Testers
Explanation:
The QA tester develops test plans and scripts and executes these plans and scripts.
Discuss this Question
48. Performance and stress tests are conducted by ____.
- Infrastructure People
- Developers
- Users
- Database Administrators
Answer: D) Database Administrators
Explanation:
Performance and stress tests are conducted by Database Administrators.
Discuss this Question
49. What is the full form of UAT?
- User Analyst Testing
- User Acceptance Testing
- User Attribute Testing
- None
Answer: B) User Acceptance Testing
Explanation:
The full form of UAT is User Acceptance Testing.
Discuss this Question
50. Whenever data is moved into production systems, ____ tests are performed.
- Production Validation
- Source to Target
- Metadata
- Data Accuracy
Answer: A) Production Validation
Explanation:
Whenever data is moved into production systems, production validation tests are performed.
Discuss this Question
51. To ensure that the data don't compromise production systems, ____ automates ETL testing and management.
- Informatica Data Validation
- Irrelevant Data Validation
- Informatica Duration Validation
- Irrelevant Duration Validation
Answer: A) Informatica Data Validation
Explanation:
To ensure that the data don't compromise production systems, Informatica Data Validation automates ETL testing and management.
Discuss this Question
52. Validating the data values transformed to the expected data values is done through ____ testing.
- Source to target
- Metadata
- Data Accuracy
- Data Transformation
Answer: A) Source to target
Explanation:
Validating the data values transformed to the expected data values is done through source-to-target testing.
Discuss this Question
53. Tests are automatically generated for ____, which saves test developers' time.
- Data Accuracy
- Data Transformation
- Application Upgrades
- Data Quality
Answer: C) Application Upgrades
Explanation:
Tests are automatically generated for Application Upgrades, which saves test developers' time.
Discuss this Question
54. When an application is upgraded, the extracted data from the old application is checked against the new application's data to ensure that they are ____.
- Different
- Identical
- Similar
- Varied
Answer: B) Identical
Explanation:
When an application is upgraded, the extracted data from the old application is checked against the new application's data to ensure that they are identical.
Discuss this Question
55. As part of ____ testing, types of data, lengths of data, and indexes and constraints are measured.
- Metadata
- Data Accuracy
- Data Transformation
- Data Quality
Answer: A) Metadata
Explanation:
As part of metadata testing, types of data, lengths of data, and indexes and constraints are measured.
Discuss this Question
56. We test the data ____ to ensure that data loading and transformation is accurate.
- Accuracy
- Transformation
- Quality
- None
Answer: A) Accuracy
Explanation:
We test the data accuracy to ensure that data loading and transformation is accurate.
Discuss this Question
57. Which of the following testing(s) is/are included in Data Quality Testing?
- Syntax
- Reference
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
The following testings are included in Data Quality Testing -
- Syntax
- Reference
Discuss this Question
58. Invalid characters, invalid character patterns, or improper upper- or lower-case order will result in dirty data being reported by ____ tests.
- Syntax
- Reference
- Accuracy
- Transformation
Answer: A) Syntax
Explanation:
Invalid characters, invalid character patterns, or improper upper- or lower-case order will result in dirty data being reported by syntax tests.
Discuss this Question
59. A data integrity test is conducted for ____ testing when new data is added to old data.
- Incremental ETL
- GUI/Navigation
- Migration
- Report
Answer: A) Incremental ETL
Explanation:
A data integrity test is conducted for incremental ETL testing when new data is added to old data.
Discuss this Question
60. After data has been inserted and updated during an incremental ETL process, incremental testing verifies the system is ____ properly.
- Deadlocked
- Still Working
- Crashed
- Initiated
Answer: B) Still Working
Explanation:
After data has been inserted and updated during an incremental ETL process, incremental testing verifies the system is still working properly.
Discuss this Question
61. ____ reports are tested for navigation and GUI aspects by GUI/Navigation Testing.
- Front-end
- Back-end
- Both A and B
- None of the above
Answer: A) Front-end
Explanation:
Front-end reports are tested for navigation and GUI aspects by GUI/Navigation Testing.
Discuss this Question
62. An existing data warehouse is used in ____ Testing, and ETL is used to process the data.
- Migration
- Report
- Incremental ETL
- GUI
Answer: A) Migration
Explanation:
An existing data warehouse is used in Migration Testing, and ETL is used to process the data.
Discuss this Question
63. Which of the following steps is/are included in Migration Testing?
- Design and validation tests
- Setting up the test environment
- Executing the validation test
- All of the above
Answer: D) All of the above
Explanation:
The following steps are included in Migration Testing -
- Design and validation tests
- Setting up the test environment
- Executing the validation test
Discuss this Question
64. ____ validation should be done for reports.
- Data
- Layout
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data validation and layout validation should be done for reports.
Discuss this Question
65. Which of the following task(s) is/are performed in ETL testing?
- The ability to understand and report on data
- Source-to-target mapping
- Analyzes the source data for errors
- All of the above
Answer: D) All of the above
Explanation:
The following tasks are performed in ETL testing -
- The ability to understand and report on data
- Source-to-target mapping
- Analyzes the source data for errors
Discuss this Question
66. ____ testing is typically performed on transactional systems, whereas ____ testing is typically performed on data in a data warehouse.
- Database, ETL
- ETL, Database
- Database, ELT
- ELT, Database
Answer: A) Database, ETL
Explanation:
Database testing is typically performed on transactional systems, whereas ETL testing is typically performed on data in a data warehouse.
Discuss this Question
67. The following operations are involved in ETL testing:
- Data movement validation between the source and target systems.
- The source and target systems should be verified for data counts.
- During ETL testing, the transformations and extractions are verified according to requirements.
- All of the above
Answer: D) All of the above
Explanation:
The following operations are involved in ETL testing:
- Data movement validation between the source and target systems.
- The source and target systems should be verified for data counts.
- During ETL testing, the transformations and extractions are verified according to requirements.
Discuss this Question
68. During ETL testing, ____ are verified to ensure that they are preserved.
- Joins
- Keys
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
During ETL testing, joins and keys are verified to ensure that they are preserved.
Discuss this Question
69. In database testing, the focus is on ensuring that data is ____.
- Accurate
- Correct
- Valid
- All of the above
Answer: D) All of the above
Explanation:
In database testing, the focus is on ensuring that data is accurate, correct, and valid.
Discuss this Question
70. The following operations are performed during database testing:
- During database testing, data values in columns are verified to ensure they are valid.
- Tests are conducted on databases to determine whether primary or foreign keys are maintained.
- Testing the database verifies if the column has any missing data.
- All of the above
Answer: D) All of the above
Explanation:
The following operations are performed during database testing:
- During database testing, data values in columns are verified to ensure they are valid.
- Tests are conducted on databases to determine whether primary or foreign keys are maintained.
- Testing the database verifies if the column has any missing data.
Discuss this Question
71. A performance test is conducted to determine if ETL systems can handle ____ at the same time.
- Multiple Users
- Transactions
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
A performance test is conducted to determine if ETL systems can handle multiple users and transactions at the same time.
Discuss this Question
72. A ____ compares the data between a source system and a target system without transforming the data in either system.
- Value Comparison
- Value Compression
- Value Compromise
- Value Contraction
Answer: A) Value Comparison
Explanation:
A value comparison compares the data between a source system and a target system without transforming the data in either system.
Discuss this Question
73. The data accuracy of both the source and the target can be checked with a set of ____ operators.
- Relational
- Rational
- SQL
- Database
Answer: C) SQL
Explanation:
The data accuracy of both the source and the target can be checked with a set of SQL operators.
Discuss this Question
74. ____ the distinct values of critical data columns in the source and target systems is a good way to verify the integrity of critical data columns.
- Examining
- Comparing
- Differentiating
- None of the above
Answer: B) Comparing
Explanation:
Comparing the distinct values of critical data columns in the source and target systems is a good way to verify the integrity of critical data columns.
Discuss this Question
75. A single SQL query cannot convert data because ____.
- It can compare the output with the target.
- It can't compare the output with the target.
- It can't compare the input with the target.
- It can compare the input with the target.
Answer: B) It can't compare the output with the target.
Explanation:
A single SQL query cannot convert data because it can't compare the output with the target.
Discuss this Question
76. Data Transformation ETL testing involves writing ____ SQL queries for each row in order to confirm the rules of the transformation.
- Two
- Three
- Four
- Multiple
Answer: D) Multiple
Explanation:
Data Transformation ETL testing involves writing multiple SQL queries for each row in order to confirm the rules of the transformation.
Discuss this Question
77. It is imperative that we select sufficient and representative data from the source system to perform the successful ____ testing.
- Database
- ETL
- Both A and B
- None of the above
Answer: B) ETL
Explanation:
It is imperative that we select sufficient and representative data from the source system to perform successful ETL testing.
Discuss this Question
78. The document(s) that the ETL tester always uses during the testing process is/are:
- ETL Mapping Sheets
- DB Schema of Source (Target)
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
The document(s) that the ETL tester always uses during the testing process is/are:
- ETL Mapping Sheets
- DB Schema of Source (Target)
Discuss this Question
79. A ____ contains all the columns and their lookups in reference tables for both source and destination tables.
- Mapping Sheet
- DB Schema of Source
- DB Schema of Target
- None of the above
Answer: A) Mapping Sheet
Explanation:
A mapping sheet contains all the columns and their lookups in reference tables for both source and destination tables.
Discuss this Question
80. Which of the following is/are the type(s) of ETL Bugs?
- Calculation
- User Interface
- Load Condition
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of ETL Bugs -
- Calculation
- User Interface
- Load Condition
Discuss this Question
81. ____, spelling check, and other issues related to the Graphical User Interface of an application are examples of User Interface bugs.
- Color
- Font Style
- Navigation
- All of the above
Answer: D) All of the above
Explanation:
Color, font style, navigation, spelling check, and other issues related to the Graphical User Interface of an application are examples of User Interface bugs.
Discuss this Question
82. As a result of the ____ bug, invalid values are being taken by the application and valid values are being rejected.
- Input-output
- Boundary value analysis
- Calculation
- Load Condition
Answer: A) Input-output
Explanation:
As a result of the input-output bug, invalid values are being taken by the application and valid values are being rejected.
Discuss this Question
83. Bugs that check for minimums and maximums are called ____ bugs.
- Calculation
- Load Condition
- Boundary value analysis
- Race Condition
Answer: C) Boundary value analysis
Explanation:
Bugs that check for minimums and maximums are called boundary value analysis bugs.
Discuss this Question
84. Mathematical errors show up in ____ bugs, and the results are usually inaccurate.
- Load Condition
- Race Condition
- Hardware
- Calculation
Answer: D) Calculation
Explanation:
Mathematical errors show up in calculation bugs, and the results are usually inaccurate.
Discuss this Question
85. Invalid or invalid types are produced by ____ bugs.
- Load Condition
- Race Condition
- Equivalence Class Partitioning
- Version Control
Answer: C) Equivalence Class Partitioning
Explanation:
Invalid or invalid types are produced by Equivalence Class Partitioning bugs.
Discuss this Question
86. Regression Testing bugs do not indicate what version they came from, as they are usually caused by ____ bugs.
- Help Source
- Hardware
- Version Control
- Load Condition
Answer: C) Version Control
Explanation:
Regression Testing bugs do not indicate what version they came from, as they are usually caused by Version Control bugs.
Discuss this Question
87. It is the ETL tester's responsibility to validate the ____ and extract the data from the target table.
- Data Sources
- Apply Transformation
- Load the data into the target table
- All of the above
Answer: D) All of the above
Explanation:
It is the ETL tester's responsibility to validate the data sources, apply transformation logic, load the data into the target table, and extract the data from the target table.
Discuss this Question
88. ETL testers has/have the following responsibility/ies:
- In the source system, verify the table.
- Apply Transformation Logic
- Data Loading
- All of the above
Answer: D) All of the above
Explanation:
ETL testers have the following responsibilities:
- In the source system, verify the table.
- Apply Transformation Logic
- Data Loading
Discuss this Question
89. Following is/are the type(s) of operations involved in verifying the table in the source system:
- Count Check
- Data Type Check
- Remove Duplicate Data
- All of the above
Answer: D) All of the above
Explanation:
Following are the types of operations involved in verifying the table in the source system:
- Count Check
- Data Type Check
- Remove Duplicate Data
Discuss this Question
90. What is/are the advantage(s) of ETL Testing?
- During ETL testing, data can be extracted from or received from any data source simultaneously.
- In ETL, heterogeneous data sources can be loaded into a single generalized (frequent)/different target simultaneously.
- It is possible to load different types of goals simultaneously using ETL.
- All of the above
Answer: D) All of the above
Explanation:
The advantages of ETL Testing are -
- During ETL testing, data can be extracted from or received from any data source simultaneously.
- In ETL, heterogeneous data sources can be loaded into a single generalized (frequent)/different target simultaneously.
- It is possible to load different types of goals simultaneously using ETL.
Discuss this Question
91. An ____ procedure is capable of extracting business data from a variety of sources and loading it into a different target as desired.
- ETL
- Database
- Both A and B
- None of the above
Answer: A) ETL
Explanation:
An ETL procedure is capable of extracting business data from a variety of sources and loading it into a different target as desired.
Discuss this Question
92. What is/are the disadvantages of ETL Testing?
- ETL testing has the disadvantage of requiring us to be database analysts or developers with data-oriented experience.
- On-demand or real-time access is not ideal when we need a fast response.
- There will be a delay of months before any ETL testing can be done.
- All of the above
Answer: D) All of the above
Explanation:
The disadvantages of ETL Testing are -
- ETL testing has the disadvantage of requiring us to be database analysts or developers with data-oriented experience.
- On-demand or real-time access is not ideal when we need a fast response.
- There will be a delay of months before any ETL testing can be done.
Discuss this Question
93. What is/are the requisites provided by ETL tools?
- Multiple data structures and different platforms, such as mainframes, servers, and databases, can be collected, read, and migrated using ETL tools.
- Using an ETL tool is as easy as sorting, filtering, reformatting, merging, and joining data.
- A few ETL tools support BI tools and functionality such as transformation scheduling, monitoring, and version control.
- All of the above
Answer: D) All of the above
Explanation:
The requisites provided by ETL tools are -
- Multiple data structures and different platforms, such as mainframes, servers, and databases, can be collected, read, and migrated using ETL tools.
- Using an ETL tool is as easy as sorting, filtering, reformatting, merging, and joining data.
- A few ETL tools support BI tools and functionality such as transformation scheduling, monitoring, and version control.
Discuss this Question
94. What is/are the benefit(s) of ETL tools?
- Ease of Use
- Operational Resilience
- Visual Flow
- All of the above
Answer: D) All of the above
Explanation:
The benefits of ETL tools are -
- Ease of Use
- Operational Resilience
- Visual Flow
Discuss this Question
95. Data engineers can develop a successful and well-instrumented system with ETL tools that have ____ error handling.
- Artificial
- Built-in
- Natural
- None
Answer: B) Built-in
Explanation:
Data engineers can develop a successful and well-instrumented system with ETL tools that have built-in error handling.
Discuss this Question
96. An ETL tool simplifies the task of ____ and integrating multiple data sets when dealing with complex rules and transformations.
- Calculating
- String Manipulation
- Changing Data
- All of the above
Answer: D) All of the above
Explanation:
A ETL tool simplifies the task of calculating, string manipulating, changing data, and integrating multiple data sets when dealing with complex rules and transformations.
Discuss this Question
97. ____ can be simplified using ETL tools.
- Extraction
- Transformation
- Loading
- All of the above
Answer: D) All of the above
Explanation:
Extraction, transformation, and loading can be simplified using ETL tools.
Discuss this Question
98. Which of the following is/are the type(s) of ETL tools?
- Talend Data Integration
- Informatica
- Kettle
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of ETL tools -
- Talend Data Integration
- Informatica
- Kettle
Discuss this Question
99. Which of the following is a cloud-based tool?
- Clover ETL
- AWS Glue
- Jasper ETL
- Kettle
Answer: B) AWS Glue
Explanation:
The following is the cloud-based tool - AWS Glue.
Discuss this Question
100. ETL tool function is a ____-layered structure.
- One
- Two
- Three
- Four
Answer: C) Three
Explanation:
The ETL tool function is a three-layered structure.
Discuss this Question
101. Which of the following is/are the layer(s) in the ETL tool function?
- Staging Layer
- Data Integration Layer
- Access Layer
- All of the above
Answer: D) All of the above
Explanation:
The following are the layers in the ETL tool function -
- Staging Layer
- Data Integration Layer
- Access Layer
Discuss this Question
102. Data extracted from different sources is stored in a ____.
- Staging database
- Staging layer
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data extracted from different sources is stored in a staging database or staging layer.
Discuss this Question
103. A database is created from the data transformed by the ____ Layer.
- Data
- Staging
- Integration
- Access
Answer: C) Integration
Explanation:
A database is created from the data transformed by the Integration Layer.
Discuss this Question
104. Facts and aggregate facts are grouped into hierarchical groups in the database referred to as ____.
- Dimensions
- Data
- Dataset
- Deadlock
Answer: A) Dimensions
Explanation:
Facts and aggregate facts are grouped into hierarchical groups in the database referred to as dimensions.
Discuss this Question
105. Which of the following is TRUE about RightData?
- An online tool for testing ETL/Data integration, RightData is available as a self-service program.
- Data can be validated and coordinated between datasets despite differences in data models or types of sources with RightData's interface.
- Data platforms with high complexity and large volumes require RightData to work efficiently.
- All of the above
Answer: D) All of the above
Explanation:
The following are TRUE about RightData -
- An online tool for testing ETL/Data integration, RightData is available as a self-service program.
- Data can be validated and coordinated between datasets despite differences in data models or types of sources with RightData's interface.
- Data platforms with high complexity and large volumes require RightData to work efficiently.
Discuss this Question
106. ____ testing can be done with the QuerySurge tool.
- Data Warehouse
- Big Data
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data Warehouse and Big Data testing can be done with the QuerySurge tool.
Discuss this Question
107. Which of the following is/are the feature(s) of QuerySurge?
- A big data testing and ETL testing tool, QuerySurge automates the testing process.
- It automates the manual process and schedules tests for a specific date and time.
- Using this tool, you can create test scenarios and test suits along with configurable reports without knowing SQL.
- All of the above
Answer: D) All of the above
Explanation:
The following are the features of QuerySurge -.
- A big data testing and ETL testing tool, QuerySurge automates the testing process.
- It automates the manual process and schedules tests for a specific date and time.
- Using this tool, you can create test scenarios and test suits along with configurable reports without knowing SQL.
Discuss this Question
108. Data-centric projects, such as ____, etc., require automated ETL testing tools such as iCEDQ.
- Warehouses
- Data migrations
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data-centric projects, such as warehouses, data migrations, etc., require automated ETL testing tools such as iCEDQ.
Discuss this Question
109. Sources and systems are ____ by iCEDQ.
- Verified
- Validated
- Coordinated
- All of the above
Answer: D) All of the above
Explanation:
Sources and systems are verified, validated, and coordinated by iCEDQ.
Discuss this Question
110. What is/are the feature(s) of iCEDQ?
- We use iCEDQ to compare millions of files and rows of data when we do ETL testing.
- As a result, it is possible to identify exactly which columns and rows contain data errors.
- iCEDQ compares the data in memory based on the unique columns in the database.
- All of the above
Answer: D) All of the above
Explanation:
The features of iCEDQ are -
- We use iCEDQ to compare millions of files and rows of data when we do ETL testing.
- As a result, it is possible to identify exactly which columns and rows contain data errors.
- iCEDQ compares the data in memory based on the unique columns in the database.
Discuss this Question
111. ETL and end-to-end testing are offered by QualiDI's ____ testing platform.
- Non-automated
- Semi-automated
- Automated
- None of the above
Answer: C) Automated
Explanation:
ETL and end-to-end testing are offered by QualiDI's automated testing platform.
Discuss this Question
112. What is/are the feature(s) of QualiDI?
- Test cases can be created automatically with QualiDI, and the automated data can be compared with the manual data.
- Continuous integration is supported.
- Featuring a complex testing cycle, eliminating human error, and managing data quality, QualiDI manages complex BI testing cycles.
- All of the aboveA
Answer: D) All of the above
Explanation:
The features of QualiDI are -.
- Test cases can be created automatically with QualiDI, and the automated data can be compared with the manual data.
- Continuous integration is supported.
- Featuring a complex testing cycle, eliminating human error, and managing data quality, QualiDI manages complex BI testing cycles.
Discuss this Question
113. A data migration pipeline involves ____ data from an input source, transforming it, and loading it into an output destination for analysis, reporting, and synchronization (such as a datamart, database, and data warehouse).
- Adding
- Deleting
- Extracting
- Modifying
Answer: C) Extracting
Explanation:
A data migration pipeline involves extracting data from an input source, transforming it, and loading it into an output destination for analysis, reporting, and synchronization (such as a datamart, database, and data warehouse).
Discuss this Question
114. What is/are TRUE about ETL Pipelines?
- In addition to enterprise data warehouses, subject-specific data marts are also built using ETL pipelines.
- As a data migration solution, ETL pipelines are also used when replacing traditional applications with new ones.
- Industry-standard ETL tools are usually used to construct ETL pipelines that transform structured data.
- All of the above
Answer: D) All of the above
Explanation:
The things TRUE about ETL Pipelines are -
- In addition to enterprise data warehouses, subject-specific data marts are also built using ETL pipelines.
- As a data migration solution, ETL pipelines are also used when replacing traditional applications with new ones.
- Industry-standard ETL tools are usually used to construct ETL pipelines that transform structured data.
Discuss this Question
115. Using data pipelines, one can ____, create real-time data streaming applications, conduct data mining, and build data-driven digital products.
- Integrate data across applications
- Build data-driver web products
- Build predictive models
- All of the above
Answer: D) All of the above
Explanation:
Using data pipelines, one can integrate data across applications, build data-driven web products, build predictive models, create real-time data streaming applications, conduct data mining, and build data-driven digital products.
Discuss this Question
116. Logs of ETL contain information about disk access, ____, and the Microsoft Operating System's performance. They also contain the event of high-frequency events.
- Page Initials
- Page faults
- Pagination
- Page rows
Answer: B) Page faults
Explanation:
Logs of ETL contain information about disk access, page faults, and the Microsoft Operating System's performance. They also contain the event of high-frequency events.
Discuss this Question
117. ____ files are also used by the Eclipse Open Development Platform.
- .psd
- .etl
- .pdf
- .png
Answer: B) .etl
Explanation:
.etl files are also used by the Eclipse Open Development Platform.
Discuss this Question
118. What is/are TRUE about Trace Logs?
- By default, trace providers generate trace logs in their trace session buffers, which are stored by the operating system.
- A compressed binary format is then used to store trace logs in a log.
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
The things TRUE about Trace Logs are -
- By default, trace providers generate trace logs in their trace session buffers, which are stored by the operating system.
- A compressed binary format is then used to store trace logs in a log.
Discuss this Question
119. A product with an ETL ____ Mark has been independently tested to meet the applicable standard.
- Data
- Listed
- Stamp
- None
Answer: B) Listed
Explanation:
A product with an ETL Listed Mark has been independently tested to meet the applicable standard.
Discuss this Question
120. In order to maintain the certification for products with ETL listed marks, regular product and site inspections are conducted to ensure that the product is manufactured and matches the ____ product.
- Corrupted
- Original
- Copy
- Artificial
Answer: B) Original
Explanation:
In order to maintain the certification for products with ETL-listed marks, regular product and site inspections are conducted to ensure that the product is manufactured and matches the original product.
Discuss this Question
121. Operation(s) that is/are performed in Database Testing?
- Validating the values of columns in a table is the focus of database testing.
- Database testing is used to ensure the foreign key or primary key is maintained.
- During database testing, it is verified whether there is any missing data in a column.
- All of the above
Answer: D) All of the above
Explanation:
Operations that are performed in Database Testing -
- Validating the values of columns in a table is the focus of database testing.
- Database testing is used to ensure the foreign key or primary key is maintained.
- During database testing, it is verified whether there is any missing data in a column.
Discuss this Question
122. Which of the following tasks is NOT involved in ETL Transformation Process?
- Filtering
- Cleaning
- Joining
- Addressing
Answer: D) Addressing
Explanation:
Addressing is NOT any task involved in ETL Transformation Process.
Discuss this Question
123. The following point(s) is/are involved in building streaming ETL based on Kafka:
- Extracting data into Kafka
- Pulling data from Kafka
- Load data to other systems
- All of the above
Answer: D) All of the above
Explanation:
The following points are involved in building streaming ETL based on Kafka:
- Extracting data into Kafka
- Pulling data from Kafka
- Load data to other systems
Discuss this Question
124. The following operation(s) is/are performed in Applying Transformation Logic-
- Before and after checking the count record, transformation logic is applied.
- The intermediate table must be validated as the data flows from the staging area.
- Make sure the thresholds for the data are valid
- All of the above
Answer: D) All of the above
Explanation:
Following operations are performed in Applying Transformation Logic-
- Before and after checking the count record, transformation logic is applied.
- The intermediate table must be validated as the data flows from the staging area.
- Make sure the thresholds for the data are valid
Discuss this Question
125. ____ the data is loaded, transformation logic is applied.
- Before
- After
- While
- None
Answer: A) Before
Explanation:
Before the data is loaded, transformation logic is applied.
Discuss this Question