ETL tools extract the data from all the different data sources, transforms the data and and loads it into a data warehouse.
ETL Testing MCQs: This section contains multiple-choice questions and answers on the various topics of ETL Testing. Practice these MCQs to test and enhance your skills on ETL Testing.
List of ETL Testing MCQs
1. Using an ____ tool, data is extracted from multiple data sources, transformed, and loaded into a data warehouse after joining fields, calculating, and removing incorrect data fields.
- ETL
- TEL
- LET
- LTE
Answer: A) ETL
Explanation:
Using an ETL tool, data is extracted from multiple data sources, transformed, and loaded into a data warehouse after joining fields, calculating, and removing incorrect data fields.
2. After business ____, ETL testing ensures that the data has been loaded accurately from a source to a destination.
- Information
- Transformation
- Transfusion
- Transfiction
Answer: B) Transformation
Explanation:
After business transformation, ETL testing ensures that the data has been loaded accurately from a source to a destination.
3. During ETL, various stages of data are verified and used at ____.
- Source
- Destination
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
During ETL, various stages of data are verified and used both at source and destination.
4. What is the full form of ETL?
- Extract Transformation and Load
- Extract Transformation and Lead
- Extract Transfusion and Load
- Extract Transfusion and Lead
Answer: A) Extract Transformation and Load
Explanation:
The full form of ETL is Extract Transformation and Load.
5. To fetch data from one database and place it in another, ETL combines all ____ database functions into one tool.
- Two
- Three
- Four
- Five
Answer: B) Three
Explanation:
To fetch data from one database and place it in another, ETL combines all three database functions into one tool.
ADVERTISEMENT
6. Getting information from a database is called ____ (reading).
- Extracting
- Transforming
- Loading
- None
Answer: A) Extracting
Explanation:
Getting information from a database is called extracting (reading).
7. The process of ____ data involves converting it from one form to another.
- Extracting
- Transforming
- Loading
- None
Answer: B) Transforming
Explanation:
The process of transforming data involves converting it from one form to another.
8. In addition to using ____, the data can be combined with other data to undergo transformation.
- Rules
- Lookup tables
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
In addition to using rules or lookup tables, the data can be combined with other data to undergo transformation.
9. Writing data into a database is called ____.
- Extracting
- Transforming
- Loading
- None
Answer: C) Loading
Explanation:
Writing data into a database is called loading.
10. Using ETL, you can ____ the data from multiple sources and blend them together according to your needs.
- Extract
- Transform
- Load
- All of the above
Answer: D) All of the above
Explanation:
Using ETL, you can extract, transform, and load the data from multiple sources and blend them together according to your needs.
ADVERTISEMENT
11. ETL is often used to build a –
- Data Center
- Data Warehouse
- Data Set
- Data Care Center
Answer: B) Data Warehouse
Explanation:
ETL is often used to build a Data Warehouse.
12. As part of the ETL process, data from a source system is ____ into a format suitable for storing in a data warehouse or another system.
- Extracted
- Converted
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
As part of the ETL process, data from a source system is extracted and converted into a format suitable for storing in a data warehouse or another system.
13. In today’s environment, ETL is becoming more and more necessary for many reasons, including:
- In order to make critical business decisions, companies use ETL to analyze their business data.
- Data warehouses are repositories where data is shared.
- In ETL, data is moved from a variety of sources into a data warehouse.
- All of the above
Answer: D) All of the above
Explanation:
In today’s environment, ETL is becoming more and more necessary for many reasons, including:
- In order to make critical business decisions, companies use ETL to analyze their business data.
- Data warehouses are repositories where data is shared.
- In ETL, data is moved from a variety of sources into a data warehouse.
14. What is TRUE about ETL?
- The ETL process allows the source and target systems to compare sample data.
- As part of the ETL process, complex transformations can be performed and additional storage space is required.
- Earlier, we defined ETL as a process of converting source data into target data and manipulating it.
- All of the above
Answer: D) All of the above
Explanation:
The things that are TRUE about ETL are –
- The ETL process allows the source and target systems to compare sample data.
- As part of the ETL process, complex transformations can be performed and additional storage space is required.
- Earlier, we defined ETL as a process of converting source data into target data and manipulating it.
15. A ____ is loaded with data through the ETL process.
- Data Mart
- Data Warehouse
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
A data mart or data warehouse is loaded with data through the ETL process
ADVERTISEMENT
16. To facilitate business analysis, our data warehouse needs to be ____ regularly
- Extracted
- Transformed
- Loaded
- None
Answer: C) Loaded
Explanation:
To facilitate business analysis, our data warehouse needs to be loaded regularly.
17. To avoid affecting the source system’s performance, the ____ occurs on the ETL server or staging area.
- Loading
- Extraction
- Transformation
- None
Answer: C) Transformation
Explanation:
To avoid affecting the source system’s performance, the transformation occurs on the ETL server or staging area.
18. Before data is moved to the warehouse, the extracted data can be validated in the ____ area.
- Staging
- Staggering
- Studying
- None
Answer: A) Staging
Explanation:
Before data is moved to the warehouse, the extracted data can be validated in the staging area.
19. How many methods are there to extract the data?
- Two
- Three
- Four
- Five
Answer: B) Three
Explanation:
There are three methods to extract the data.
20. Which of the following is/are the method(s) to extract the data?
- FULL Extraction
- Partial Extraction – Without Update Notification
- Partial Extraction – With Update Notification
- All of the above
Answer: D) All of the above
Explanation:
The following are the methods to extract the data –
- FULL Extraction
- Partial Extraction – Without Update Notification
- Partial Extraction – With Update Notification
ADVERTISEMENT
21. Whatever extraction method we use, the source system should not be affected in terms of ____ time.
- Performance
- Response
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Whatever extraction method we use, the source system should not be affected in terms of performance or response time.
22. Which of the following is/are the validation(s) using the extraction(s)?
- Check the source data against the record
- Ensure that the data type is correct
- There will be a check to see if all the keys are there
- All of the above
Answer: D) All of the above
Explanation:
The following are the validations using the extractions –
- Check the source data against the record
- Ensure that the data type is correct
- There will be a check to see if all the keys are there
23. It is not possible to use the extracted data as originally formatted from the ____ server.
- Source
- Site
- Storage
- None
Answer: A) Source
Explanation:
It is not possible to use the extracted data as originally formatted from the source server.
24. ____ refers to data that does not need to be transformed.
- Direct Move
- Pass-through data
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Direct move or pass-through data refers to data that does not need to be transformed.
25. Which of the following is/are the validation point(s) during the transformation?
- Filtering
- Conversion of character sets and encodings
- Checking the threshold and validity of data
- All of the above
Answer: D) All of the above
Explanation:
The following are the validation points during the transformation –
- Filtering
- Conversion of character sets and encodings
- Checking the threshold and validity of data
26. Which of the following is the last step of the ETL process?
- Extraction
- Transformation
- Loading
- None
Answer: C) Loading
Explanation:
The following is last step of the ETL process is Loading.
27. Which of the following is/are the type(s) of loading?
- Initial Load
- Incremental Load
- Full Refresh
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of loading –
- Initial Load
- Incremental Load
- Full Refresh
28. Loads need to be ____ according to server performance by the admin of the data warehouse.
- Monitored
- Resumed
- Canceled
- All of the above
Answer: D) All of the above
Explanation:
Loads need to be monitored, resumed, and canceled according to server performance by the admin of the data warehouse.
29. With a ____, all tables are erased and reloaded with new information.
- Initial Load
- Incremental Load
- Full Refresh
- None of the above
Answer: C) Full Refresh
Explanation:
With a Full Refresh, all tables are erased and reloaded with new information.
30. The term is ETL now extended to ____ or Extract, Monitor, Profile, Analyze, Cleanse, Transform, and Load.
- E-MPAC-TL
- E-PAC-TL
- E-MAP-TL
- E-MPAA-TL
Answer: A) E-MPAC-TL
Explanation:
The term is now extended to E-MPAC-TL or Extract, Monitor, Profile, Analyze, Cleanse, Transform, and Load.
ADVERTISEMENT
31. During ____, the main goal is to capture data as quickly as possible from a system while minimizing the inconvenience to the system.
- Extraction
- Transformation
- Loading
- None
Answer: A) Extraction
Explanation:
During extraction, the main goal is to capture data as quickly as possible from a system while minimizing the inconvenience to the system.
32. ____ the data requires integrating the data and finally presenting the combined data to the end-user community via the front-end tools.
- Transforming
- Loading
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Transforming and loading the data requires integrating the data and finally presenting the combined data to the end-user community via the front-end tools.
33. Compared to traditional ETL, ETL ____ the time it takes for sources and targets to develop.
- Extends
- Reduces
- Exceeds
- Manipulates
Answer: B) Reduces
Explanation:
Compared to traditional ETL, ETL reduces the time it takes for sources and targets to develop.
34. ____ consists of a process between stages that is defined according to the needs, and it can verify the quality of the product
- Quantity Assurance
- Quality Assurance
- Quantity Attribution
- Quality Attribution
Answer: B) Quality Assurance
Explanation:
Quality Assurance consists of a process between stages that is defined according to the needs, and it can verify the quality of the product.
35. In ____, analysis and validation of the data pattern and formats will be performed, as well as identification and validation of redundant data across data sources to determine the actual content, structure, and quality of the data.
- Data Profiling
- Data Analysis
- Source Analysis
- Cleansing
Answer: A) Data Profiling
Explanation:
In data profiling, analysis and validation of the data pattern and formats will be performed, as well as identification and validation of redundant data across data sources to determine the actual content, structure, and quality of the data.
36. It is important to focus on the surroundings of the sources in ____ analysis, so that the documentation of the sources can be obtained.
- Data
- Source
- Profile
- None
Answer: B) Source
Explanation:
It is important to focus on the surroundings of the sources in source analysis, so that the documentation of the sources can be obtained.
37. Based on the Metadata of a set of predefined rules, errors found can be fixed in ____.
- Data Analysis
- Source Analysis
- Cleansing
- Data Profiling
Answer: C) Cleansing
Explanation:
Based on the Metadata of a set of predefined rules, errors found can be fixed in cleansing.
38. An extended ETL concept, E-MPAC-TL is designed to meet the requirements while taking into account the realities of the systems, ____, constraints, and most importantly, the data itself.
- Tools
- Metadata
- Technical Issues
- All of the above
Answer: D) All of the above
Explanation:
An extended ETL concept, E-MPAC-TL is designed to meet the requirements while taking into account the realities of the systems, tools, metadata, technical issues, constraints, and most importantly, the data itself.
39. ETL testing is also known as –
- Table balancing
- Product Reconciliation
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
ETL testing is also known as Table balancing or Product Reconciliation
40. An ETL test ensures the data loaded after transformation is accurate after it has been ____ from a source to a destination.
- Added
- Loaded
- Deleted
- Altered
Answer: B) Loaded
Explanation:
An ETL test ensures the data loaded after transformation is accurate after it has been loaded from a source to a destination.
41. ETL testing is performed in ____ stages.
- Two
- Three
- Four
- Five
Answer: D) Five
Explanation:
ETL testing is performed in five stages.
42. Which of the following is/are the stage(s) the ETL testing?
- Data recovery
- Build and populate data
- Build reports
- All of the above
Answer: D) All of the above
Explanation:
The following are the stages of the ETL testing –
- Data recovery
- Build and populate data
- Build reports
43. Which of the following is/are the type(s) of ETL testing?
- New Data Warehouse Testing
- Production Validation Testing
- Application Upgrade
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of ETL testing –
- New Data Warehouse Testing
- Production Validation Testing
- Application Upgrade
44. Customer requirements and different sources of data are taken into account in ____.
- New Data Warehouse Testing
- Production Validation Testing
- Application Upgrade
- Metadata Testing
Answer: A) New Data Warehouse Testing
Explanation:
Customer requirements and different sources of data are taken into account in New Data Warehouse Testing.
45. Which of the following is/are the group that plays the responsibility in testing New Data Warehouses –
- Business Analyst
- Infrastructure People
- QA Testers
- All of the above
Answer: D) All of the above
Explanation:
The following are the group that plays responsibility in testing New Data Warehouses –
- Business Analyst
- Infrastructure People
- QA Testers
46. What is the responsibility of a Business Analyst?
- Requirements are gathered and documented by the business analyst.
- The test environment is set up by Business Analyst people.
- These plans and scripts are developed by Business Analysts and then executed by them.
- Each module is tested by a Business Analyst.
Answer: A) Requirements are gathered and documented by the business analyst.
Explanation:
Requirements are gathered and documented by the business analyst.
47. The ____ develops test plans and scripts and executes these plans and scripts.
- Infrastructure People
- QA Testers
- Developers
- Users
Answer: B) QA Testers
Explanation:
The QA tester develops test plans and scripts and executes these plans and scripts.
48. Performance and stress tests are conducted by ____.
- Infrastructure People
- Developers
- Users
- Database Administrators
Answer: D) Database Administrators
Explanation:
Performance and stress tests are conducted by Database Administrators.
49. What is the full form of UAT?
- User Analyst Testing
- User Acceptance Testing
- User Attribute Testing
- None
Answer: B) User Acceptance Testing
Explanation:
The full form of UAT is User Acceptance Testing.
50. Whenever data is moved into production systems, ____ tests are performed.
- Production Validation
- Source to Target
- Metadata
- Data Accuracy
Answer: A) Production Validation
Explanation:
Whenever data is moved into production systems, production validation tests are performed.
51. To ensure that the data don’t compromise production systems, ____ automates ETL testing and management.
- Informatica Data Validation
- Irrelevant Data Validation
- Informatica Duration Validation
- Irrelevant Duration Validation
Answer: A) Informatica Data Validation
Explanation:
To ensure that the data don’t compromise production systems, Informatica Data Validation automates ETL testing and management.
52. Validating the data values transformed to the expected data values is done through ____ testing.
- Source to target
- Metadata
- Data Accuracy
- Data Transformation
Answer: A) Source to target
Explanation:
Validating the data values transformed to the expected data values is done through source-to-target testing.
53. Tests are automatically generated for ____, which saves test developers’ time.
- Data Accuracy
- Data Transformation
- Application Upgrades
- Data Quality
Answer: C) Application Upgrades
Explanation:
Tests are automatically generated for Application Upgrades, which saves test developers’ time.
54. When an application is upgraded, the extracted data from the old application is checked against the new application’s data to ensure that they are ____.
- Different
- Identical
- Similar
- Varied
Answer: B) Identical
Explanation:
When an application is upgraded, the extracted data from the old application is checked against the new application’s data to ensure that they are identical.
55. As part of ____ testing, types of data, lengths of data, and indexes and constraints are measured.
- Metadata
- Data Accuracy
- Data Transformation
- Data Quality
Answer: A) Metadata
Explanation:
As part of metadata testing, types of data, lengths of data, and indexes and constraints are measured.
56. We test the data ____ to ensure that data loading and transformation is accurate.
- Accuracy
- Transformation
- Quality
- None
Answer: A) Accuracy
Explanation:
We test the data accuracy to ensure that data loading and transformation is accurate.
57. Which of the following testing(s) is/are included in Data Quality Testing?
- Syntax
- Reference
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
The following testings are included in Data Quality Testing –
- Syntax
- Reference
58. Invalid characters, invalid character patterns, or improper upper- or lower-case order will result in dirty data being reported by ____ tests.
- Syntax
- Reference
- Accuracy
- Transformation
Answer: A) Syntax
Explanation:
Invalid characters, invalid character patterns, or improper upper- or lower-case order will result in dirty data being reported by syntax tests.
59. A data integrity test is conducted for ____ testing when new data is added to old data.
- Incremental ETL
- GUI/Navigation
- Migration
- Report
Answer: A) Incremental ETL
Explanation:
A data integrity test is conducted for incremental ETL testing when new data is added to old data.
60. After data has been inserted and updated during an incremental ETL process, incremental testing verifies the system is ____ properly.
- Deadlocked
- Still Working
- Crashed
- Initiated
Answer: B) Still Working
Explanation:
After data has been inserted and updated during an incremental ETL process, incremental testing verifies the system is still working properly.
61. ____ reports are tested for navigation and GUI aspects by GUI/Navigation Testing.
- Front-end
- Back-end
- Both A and B
- None of the above
Answer: A) Front-end
Explanation:
Front-end reports are tested for navigation and GUI aspects by GUI/Navigation Testing.
62. An existing data warehouse is used in ____ Testing, and ETL is used to process the data.
- Migration
- Report
- Incremental ETL
- GUI
Answer: A) Migration
Explanation:
An existing data warehouse is used in Migration Testing, and ETL is used to process the data.
63. Which of the following steps is/are included in Migration Testing?
- Design and validation tests
- Setting up the test environment
- Executing the validation test
- All of the above
Answer: D) All of the above
Explanation:
The following steps are included in Migration Testing –
- Design and validation tests
- Setting up the test environment
- Executing the validation test
64. ____ validation should be done for reports.
- Data
- Layout
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data validation and layout validation should be done for reports.
65. Which of the following task(s) is/are performed in ETL testing?
- The ability to understand and report on data
- Source-to-target mapping
- Analyzes the source data for errors
- All of the above
Answer: D) All of the above
Explanation:
The following tasks are performed in ETL testing –
- The ability to understand and report on data
- Source-to-target mapping
- Analyzes the source data for errors
66. ____ testing is typically performed on transactional systems, whereas ____ testing is typically performed on data in a data warehouse.
- Database, ETL
- ETL, Database
- Database, ELT
- ELT, Database
Answer: A) Database, ETL
Explanation:
Database testing is typically performed on transactional systems, whereas ETL testing is typically performed on data in a data warehouse.
67. The following operations are involved in ETL testing:
- Data movement validation between the source and target systems.
- The source and target systems should be verified for data counts.
- During ETL testing, the transformations and extractions are verified according to requirements.
- All of the above
Answer: D) All of the above
Explanation:
The following operations are involved in ETL testing:
- Data movement validation between the source and target systems.
- The source and target systems should be verified for data counts.
- During ETL testing, the transformations and extractions are verified according to requirements.
68. During ETL testing, ____ are verified to ensure that they are preserved.
- Joins
- Keys
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
During ETL testing, joins and keys are verified to ensure that they are preserved.
69. In database testing, the focus is on ensuring that data is ____.
- Accurate
- Correct
- Valid
- All of the above
Answer: D) All of the above
Explanation:
In database testing, the focus is on ensuring that data is accurate, correct, and valid.
70. The following operations are performed during database testing:
- During database testing, data values in columns are verified to ensure they are valid.
- Tests are conducted on databases to determine whether primary or foreign keys are maintained.
- Testing the database verifies if the column has any missing data.
- All of the above
Answer: D) All of the above
Explanation:
The following operations are performed during database testing:
- During database testing, data values in columns are verified to ensure they are valid.
- Tests are conducted on databases to determine whether primary or foreign keys are maintained.
- Testing the database verifies if the column has any missing data.
71. A performance test is conducted to determine if ETL systems can handle ____ at the same time.
- Multiple Users
- Transactions
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
A performance test is conducted to determine if ETL systems can handle multiple users and transactions at the same time.
72. A ____ compares the data between a source system and a target system without transforming the data in either system.
- Value Comparison
- Value Compression
- Value Compromise
- Value Contraction
Answer: A) Value Comparison
Explanation:
A value comparison compares the data between a source system and a target system without transforming the data in either system.
73. The data accuracy of both the source and the target can be checked with a set of ____ operators.
- Relational
- Rational
- SQL
- Database
Answer: C) SQL
Explanation:
The data accuracy of both the source and the target can be checked with a set of SQL operators.
74. ____ the distinct values of critical data columns in the source and target systems is a good way to verify the integrity of critical data columns.
- Examining
- Comparing
- Differentiating
- None of the above
Answer: B) Comparing
Explanation:
Comparing the distinct values of critical data columns in the source and target systems is a good way to verify the integrity of critical data columns.
75. A single SQL query cannot convert data because ____.
- It can compare the output with the target.
- It can’t compare the output with the target.
- It can’t compare the input with the target.
- It can compare the input with the target.
Answer: B) It can’t compare the output with the target.
Explanation:
A single SQL query cannot convert data because it can’t compare the output with the target.
76. Data Transformation ETL testing involves writing ____ SQL queries for each row in order to confirm the rules of the transformation.
- Two
- Three
- Four
- Multiple
Answer: D) Multiple
Explanation:
Data Transformation ETL testing involves writing multiple SQL queries for each row in order to confirm the rules of the transformation.
77. It is imperative that we select sufficient and representative data from the source system to perform the successful ____ testing.
- Database
- ETL
- Both A and B
- None of the above
Answer: B) ETL
Explanation:
It is imperative that we select sufficient and representative data from the source system to perform successful ETL testing.
78. The document(s) that the ETL tester always uses during the testing process is/are:
- ETL Mapping Sheets
- DB Schema of Source (Target)
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
The document(s) that the ETL tester always uses during the testing process is/are:
- ETL Mapping Sheets
- DB Schema of Source (Target)
79. A ____ contains all the columns and their lookups in reference tables for both source and destination tables.
- Mapping Sheet
- DB Schema of Source
- DB Schema of Target
- None of the above
Answer: A) Mapping Sheet
Explanation:
A mapping sheet contains all the columns and their lookups in reference tables for both source and destination tables.
80. Which of the following is/are the type(s) of ETL Bugs?
- Calculation
- User Interface
- Load Condition
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of ETL Bugs –
- Calculation
- User Interface
- Load Condition
81. ____, spelling check, and other issues related to the Graphical User Interface of an application are examples of User Interface bugs.
- Color
- Font Style
- Navigation
- All of the above
Answer: D) All of the above
Explanation:
Color, font style, navigation, spelling check, and other issues related to the Graphical User Interface of an application are examples of User Interface bugs.
82. As a result of the ____ bug, invalid values are being taken by the application and valid values are being rejected.
- Input-output
- Boundary value analysis
- Calculation
- Load Condition
Answer: A) Input-output
Explanation:
As a result of the input-output bug, invalid values are being taken by the application and valid values are being rejected.
83. Bugs that check for minimums and maximums are called ____ bugs.
- Calculation
- Load Condition
- Boundary value analysis
- Race Condition
Answer: C) Boundary value analysis
Explanation:
Bugs that check for minimums and maximums are called boundary value analysis bugs.
84. Mathematical errors show up in ____ bugs, and the results are usually inaccurate.
- Load Condition
- Race Condition
- Hardware
- Calculation
Answer: D) Calculation
Explanation:
Mathematical errors show up in calculation bugs, and the results are usually inaccurate.
85. Invalid or invalid types are produced by ____ bugs.
- Load Condition
- Race Condition
- Equivalence Class Partitioning
- Version Control
Answer: C) Equivalence Class Partitioning
Explanation:
Invalid or invalid types are produced by Equivalence Class Partitioning bugs.
86. Regression Testing bugs do not indicate what version they came from, as they are usually caused by ____ bugs.
- Help Source
- Hardware
- Version Control
- Load Condition
Answer: C) Version Control
Explanation:
Regression Testing bugs do not indicate what version they came from, as they are usually caused by Version Control bugs.
87. It is the ETL tester’s responsibility to validate the ____ and extract the data from the target table.
- Data Sources
- Apply Transformation
- Load the data into the target table
- All of the above
Answer: D) All of the above
Explanation:
It is the ETL tester’s responsibility to validate the data sources, apply transformation logic, load the data into the target table, and extract the data from the target table.
88. ETL testers has/have the following responsibility/ies:
- In the source system, verify the table.
- Apply Transformation Logic
- Data Loading
- All of the above
Answer: D) All of the above
Explanation:
ETL testers have the following responsibilities:
- In the source system, verify the table.
- Apply Transformation Logic
- Data Loading
89. Following is/are the type(s) of operations involved in verifying the table in the source system:
- Count Check
- Data Type Check
- Remove Duplicate Data
- All of the above
Answer: D) All of the above
Explanation:
Following are the types of operations involved in verifying the table in the source system:
- Count Check
- Data Type Check
- Remove Duplicate Data
90. What is/are the advantage(s) of ETL Testing?
- During ETL testing, data can be extracted from or received from any data source simultaneously.
- In ETL, heterogeneous data sources can be loaded into a single generalized (frequent)/different target simultaneously.
- It is possible to load different types of goals simultaneously using ETL.
- All of the above
Answer: D) All of the above
Explanation:
The advantages of ETL Testing are –
- During ETL testing, data can be extracted from or received from any data source simultaneously.
- In ETL, heterogeneous data sources can be loaded into a single generalized (frequent)/different target simultaneously.
- It is possible to load different types of goals simultaneously using ETL.
91. An ____ procedure is capable of extracting business data from a variety of sources and loading it into a different target as desired.
- ETL
- Database
- Both A and B
- None of the above
Answer: A) ETL
Explanation:
An ETL procedure is capable of extracting business data from a variety of sources and loading it into a different target as desired.
92. What is/are the disadvantages of ETL Testing?
- ETL testing has the disadvantage of requiring us to be database analysts or developers with data-oriented experience.
- On-demand or real-time access is not ideal when we need a fast response.
- There will be a delay of months before any ETL testing can be done.
- All of the above
Answer: D) All of the above
Explanation:
The disadvantages of ETL Testing are –
- ETL testing has the disadvantage of requiring us to be database analysts or developers with data-oriented experience.
- On-demand or real-time access is not ideal when we need a fast response.
- There will be a delay of months before any ETL testing can be done.
93. What is/are the requisites provided by ETL tools?
- Multiple data structures and different platforms, such as mainframes, servers, and databases, can be collected, read, and migrated using ETL tools.
- Using an ETL tool is as easy as sorting, filtering, reformatting, merging, and joining data.
- A few ETL tools support BI tools and functionality such as transformation scheduling, monitoring, and version control.
- All of the above
Answer: D) All of the above
Explanation:
The requisites provided by ETL tools are –
- Multiple data structures and different platforms, such as mainframes, servers, and databases, can be collected, read, and migrated using ETL tools.
- Using an ETL tool is as easy as sorting, filtering, reformatting, merging, and joining data.
- A few ETL tools support BI tools and functionality such as transformation scheduling, monitoring, and version control.
94. What is/are the benefit(s) of ETL tools?
- Ease of Use
- Operational Resilience
- Visual Flow
- All of the above
Answer: D) All of the above
Explanation:
The benefits of ETL tools are –
- Ease of Use
- Operational Resilience
- Visual Flow
95. Data engineers can develop a successful and well-instrumented system with ETL tools that have ____ error handling.
- Artificial
- Built-in
- Natural
- None
Answer: B) Built-in
Explanation:
Data engineers can develop a successful and well-instrumented system with ETL tools that have built-in error handling.
96. An ETL tool simplifies the task of ____ and integrating multiple data sets when dealing with complex rules and transformations.
- Calculating
- String Manipulation
- Changing Data
- All of the above
Answer: D) All of the above
Explanation:
A ETL tool simplifies the task of calculating, string manipulating, changing data, and integrating multiple data sets when dealing with complex rules and transformations.
97. ____ can be simplified using ETL tools.
- Extraction
- Transformation
- Loading
- All of the above
Answer: D) All of the above
Explanation:
Extraction, transformation, and loading can be simplified using ETL tools.
98. Which of the following is/are the type(s) of ETL tools?
- Talend Data Integration
- Informatica
- Kettle
- All of the above
Answer: D) All of the above
Explanation:
The following are the types of ETL tools –
- Talend Data Integration
- Informatica
- Kettle
99. Which of the following is a cloud-based tool?
- Clover ETL
- AWS Glue
- Jasper ETL
- Kettle
Answer: B) AWS Glue
Explanation:
The following is the cloud-based tool – AWS Glue.
100. ETL tool function is a ____-layered structure.
- One
- Two
- Three
- Four
Answer: C) Three
Explanation:
The ETL tool function is a three-layered structure.
101. Which of the following is/are the layer(s) in the ETL tool function?
- Staging Layer
- Data Integration Layer
- Access Layer
- All of the above
Answer: D) All of the above
Explanation:
The following are the layers in the ETL tool function –
- Staging Layer
- Data Integration Layer
- Access Layer
102. Data extracted from different sources is stored in a ____.
- Staging database
- Staging layer
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data extracted from different sources is stored in a staging database or staging layer.
103. A database is created from the data transformed by the ____ Layer.
- Data
- Staging
- Integration
- Access
Answer: C) Integration
Explanation:
A database is created from the data transformed by the Integration Layer.
104. Facts and aggregate facts are grouped into hierarchical groups in the database referred to as ____.
- Dimensions
- Data
- Dataset
- Deadlock
Answer: A) Dimensions
Explanation:
Facts and aggregate facts are grouped into hierarchical groups in the database referred to as dimensions.
105. Which of the following is TRUE about RightData?
- An online tool for testing ETL/Data integration, RightData is available as a self-service program.
- Data can be validated and coordinated between datasets despite differences in data models or types of sources with RightData’s interface.
- Data platforms with high complexity and large volumes require RightData to work efficiently.
- All of the above
Answer: D) All of the above
Explanation:
The following are TRUE about RightData –
- An online tool for testing ETL/Data integration, RightData is available as a self-service program.
- Data can be validated and coordinated between datasets despite differences in data models or types of sources with RightData’s interface.
- Data platforms with high complexity and large volumes require RightData to work efficiently.
106. ____ testing can be done with the QuerySurge tool.
- Data Warehouse
- Big Data
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data Warehouse and Big Data testing can be done with the QuerySurge tool.
107. Which of the following is/are the feature(s) of QuerySurge?
- A big data testing and ETL testing tool, QuerySurge automates the testing process.
- It automates the manual process and schedules tests for a specific date and time.
- Using this tool, you can create test scenarios and test suits along with configurable reports without knowing SQL.
- All of the above
Answer: D) All of the above
Explanation:
The following are the features of QuerySurge -.
- A big data testing and ETL testing tool, QuerySurge automates the testing process.
- It automates the manual process and schedules tests for a specific date and time.
- Using this tool, you can create test scenarios and test suits along with configurable reports without knowing SQL.
108. Data-centric projects, such as ____, etc., require automated ETL testing tools such as iCEDQ.
- Warehouses
- Data migrations
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
Data-centric projects, such as warehouses, data migrations, etc., require automated ETL testing tools such as iCEDQ.
109. Sources and systems are ____ by iCEDQ.
- Verified
- Validated
- Coordinated
- All of the above
Answer: D) All of the above
Explanation:
Sources and systems are verified, validated, and coordinated by iCEDQ.
110. What is/are the feature(s) of iCEDQ?
- We use iCEDQ to compare millions of files and rows of data when we do ETL testing.
- As a result, it is possible to identify exactly which columns and rows contain data errors.
- iCEDQ compares the data in memory based on the unique columns in the database.
- All of the above
Answer: D) All of the above
Explanation:
The features of iCEDQ are –
- We use iCEDQ to compare millions of files and rows of data when we do ETL testing.
- As a result, it is possible to identify exactly which columns and rows contain data errors.
- iCEDQ compares the data in memory based on the unique columns in the database.
111. ETL and end-to-end testing are offered by QualiDI’s ____ testing platform.
- Non-automated
- Semi-automated
- Automated
- None of the above
Answer: C) Automated
Explanation:
ETL and end-to-end testing are offered by QualiDI’s automated testing platform.
112. What is/are the feature(s) of QualiDI?
- Test cases can be created automatically with QualiDI, and the automated data can be compared with the manual data.
- Continuous integration is supported.
- Featuring a complex testing cycle, eliminating human error, and managing data quality, QualiDI manages complex BI testing cycles.
- All of the aboveA
Answer: D) All of the above
Explanation:
The features of QualiDI are -.
- Test cases can be created automatically with QualiDI, and the automated data can be compared with the manual data.
- Continuous integration is supported.
- Featuring a complex testing cycle, eliminating human error, and managing data quality, QualiDI manages complex BI testing cycles.
113. A data migration pipeline involves ____ data from an input source, transforming it, and loading it into an output destination for analysis, reporting, and synchronization (such as a datamart, database, and data warehouse).
- Adding
- Deleting
- Extracting
- Modifying
Answer: C) Extracting
Explanation:
A data migration pipeline involves extracting data from an input source, transforming it, and loading it into an output destination for analysis, reporting, and synchronization (such as a datamart, database, and data warehouse).
114. What is/are TRUE about ETL Pipelines?
- In addition to enterprise data warehouses, subject-specific data marts are also built using ETL pipelines.
- As a data migration solution, ETL pipelines are also used when replacing traditional applications with new ones.
- Industry-standard ETL tools are usually used to construct ETL pipelines that transform structured data.
- All of the above
Answer: D) All of the above
Explanation:
The things TRUE about ETL Pipelines are –
- In addition to enterprise data warehouses, subject-specific data marts are also built using ETL pipelines.
- As a data migration solution, ETL pipelines are also used when replacing traditional applications with new ones.
- Industry-standard ETL tools are usually used to construct ETL pipelines that transform structured data.
115. Using data pipelines, one can ____, create real-time data streaming applications, conduct data mining, and build data-driven digital products.
- Integrate data across applications
- Build data-driver web products
- Build predictive models
- All of the above
Answer: D) All of the above
Explanation:
Using data pipelines, one can integrate data across applications, build data-driven web products, build predictive models, create real-time data streaming applications, conduct data mining, and build data-driven digital products.
116. Logs of ETL contain information about disk access, ____, and the Microsoft Operating System’s performance. They also contain the event of high-frequency events.
- Page Initials
- Page faults
- Pagination
- Page rows
Answer: B) Page faults
Explanation:
Logs of ETL contain information about disk access, page faults, and the Microsoft Operating System’s performance. They also contain the event of high-frequency events.
117. ____ files are also used by the Eclipse Open Development Platform.
- .psd
- .etl
- .png
Answer: B) .etl
Explanation:
.etl files are also used by the Eclipse Open Development Platform.
118. What is/are TRUE about Trace Logs?
- By default, trace providers generate trace logs in their trace session buffers, which are stored by the operating system.
- A compressed binary format is then used to store trace logs in a log.
- Both A and B
- None of the above
Answer: C) Both A and B
Explanation:
The things TRUE about Trace Logs are –
- By default, trace providers generate trace logs in their trace session buffers, which are stored by the operating system.
- A compressed binary format is then used to store trace logs in a log.
119. A product with an ETL ____ Mark has been independently tested to meet the applicable standard.
- Data
- Listed
- Stamp
- None
Answer: B) Listed
Explanation:
A product with an ETL Listed Mark has been independently tested to meet the applicable standard.
120. In order to maintain the certification for products with ETL listed marks, regular product and site inspections are conducted to ensure that the product is manufactured and matches the ____ product.
- Corrupted
- Original
- Copy
- Artificial
Answer: B) Original
Explanation:
In order to maintain the certification for products with ETL-listed marks, regular product and site inspections are conducted to ensure that the product is manufactured and matches the original product.
121. Operation(s) that is/are performed in Database Testing?
- Validating the values of columns in a table is the focus of database testing.
- Database testing is used to ensure the foreign key or primary key is maintained.
- During database testing, it is verified whether there is any missing data in a column.
- All of the above
Answer: D) All of the above
Explanation:
Operations that are performed in Database Testing –
- Validating the values of columns in a table is the focus of database testing.
- Database testing is used to ensure the foreign key or primary key is maintained.
- During database testing, it is verified whether there is any missing data in a column.
122. Which of the following tasks is NOT involved in ETL Transformation Process?
- Filtering
- Cleaning
- Joining
- Addressing
Answer: D) Addressing
Explanation:
Addressing is NOT any task involved in ETL Transformation Process.
123. The following point(s) is/are involved in building streaming ETL based on Kafka:
- Extracting data into Kafka
- Pulling data from Kafka
- Load data to other systems
- All of the above
Answer: D) All of the above
Explanation:
The following points are involved in building streaming ETL based on Kafka:
- Extracting data into Kafka
- Pulling data from Kafka
- Load data to other systems
124. The following operation(s) is/are performed in Applying Transformation Logic-
- Before and after checking the count record, transformation logic is applied.
- The intermediate table must be validated as the data flows from the staging area.
- Make sure the thresholds for the data are valid
- All of the above
Answer: D) All of the above
Explanation:
Following operations are performed in Applying Transformation Logic-
- Before and after checking the count record, transformation logic is applied.
- The intermediate table must be validated as the data flows from the staging area.
- Make sure the thresholds for the data are valid
125. ____ the data is loaded, transformation logic is applied.
- Before
- After
- While
- None
Answer: A) Before
Explanation:
Before the data is loaded, transformation logic is applied.