Page three Teradata Architectural

A new application is being developed which will leverage data from customer-centric data. The new application's interface to the customer-centric data is a series of seven views which are referenced by several existing applications. The application will join three base tables to three of the seven views which provide access for the end users via SQL Assistant. Which performance concern might be a risk as the new application moves into production?
A. Nested views require excessive translation time in the database catalog tables.
B. There is no method for determining how often the underlying objects are accessed so lock contention may occur.
C. The underlying view DDL can be hidden from the end users and can change at any time without the users' knowledge.
D. Even with adequate locking access in the view layer, the other applications leveraging the views can impact the performance of the new application.
Answer: C
An active application requires uninterrupted access to tables. The Teradata support team wants an eight hour maintenance window every month. What two solutions will support these requirements? (Choose two.)
A. Use a dual systems approach using Query Director.
B. Set up a disaster-recovery system and allow user access.
C. Use a dual systems approach using Teradata Replication Server.
D. Use a hot standby node to replace the node undergoing maintenance.
Answer: B,C
Which benefit does report-based metadata provide to IT organizations when analyzing an application's capacity trends?
A. opportunity to improve system performance
B. opportunity to improve business decisions
C. opportunity to decrease IT-related problems
D. opportunity to deliver applications quickly
Answer: C
Which two can be done to protect confidential data? (Choose two.)
A. Use encryption on the data transmitted between the Teradata Database and the application.
B. Use access logging to determine the requesting user, type of access, and the referenced object.
C. Use stored procedures to limit the actions the user can take on the table and view columns and rows.
D. Use multi-factor authentication to provide strong access control and enhance the safety of data stored in the database.
Answer: B,C
A retailer would like to ensure that the latest sales data is available in their data warehouse so associates can identify additional up-sell opportunities and provide premium customer service. The data needs to be available within five minutes of a sales transaction. Both the base transaction table and the join index built upon it have NUPI duplicates being inserted and their primary index columns are different from one another. Which approach should be used for loading the transaction data into the data warehouse?
A. direct MultiLoad to the target
B. TPump with the SERIALIZE option on
C. periodic batch to staging, then apply with SQL
D. stored procedure fed via replication services
Answer: C
An analytical application uses session pooling to submit tactical requests. Session pooling hides the identity of the end user and the type of request, making it difficult to manage the requests and meet Service Level Agreements (SLAs). What can be done to identify the tactical requests and classify them into the correct workload?
A. Use SessionNo to classify the request.
B. Use query banding to classify the request.
C. Use filters to classify the request based on resource consumption.
D. Use local exceptions to move the query to the proper workload after it begins execution.
Answer: C
An event table was created without monthly partitions. Since implementation, the table has grown to an excessive amount of history. The application using this table has evolved into a 24 X 7 active application. The oldest month of data needs to be archived to improve table scan performance. How can a month of history be archived and deleted with minimal impact to the active application?
A. Use the ARC utility to archive the month and delete the month with a SQL DELETE request.
B. INSERT SELECT the month to an empty table, delete the month from the application table, then archive the month table.
C. Use the ARC utility to archive the month from the event table then delete the month from the application table using a MultiLoad DELETE task.
D. Issue a BT, INSERT SELECT the event table into a new table minus the month to be deleted, rename both tables, issue the ET, FastExport the month from the old event table, then drop the old event table.
Answer: A
A customer has a requirement to load a high volume of data from a single message queue into a Sales table that is part of an active data warehouse. The Sales table has both NUSIs and Join Indexes. The data latency requirements are less than 15 minutes and the arrival rate into the queue is continuous. What is the most effective approach to supporting this requirement?
A. Use TPump to continuously read from the queue and insert the data.
B. Pull the data from the queue using a CLI program and insert the data.
C. Pull the data from the queue into a file and use MultiLoad to insert the data from the file.
D. Pull the data from the queue into a file and use FastLoad to load the data into a staging table and then use SQL INSERT SELECT.
Answer: C
An insurance company loads data from multiple sources throughout the day with each source having a different time variance. The claims data comes in on an hourly basis from two independent data collection servers from a single source. One server is in the USA and the other one in Europe. It is possible for a claim to be opened in one location but updated in another. The corporate customer data comes in monthly from one source while the individual customer data comes in weekly from two sources. A set of referential data is loaded monthly from the CRM application which contains cleansed customer contact data. A new application is being developed which needs to leverage data from the most current customer data along with the combined claims data. Which two data inputs are critical for developing a load strategy for the new application? (Choose two.)
A. the frequency of update of the CRM application
B. the changed data capture strategy for the claims data
C. the changed data capture strategy for the customer data
D. a set of rules for determining the most authoritative claims source
E. a set of rules for determining the most authoritative customer source
Answer: A,E
The current perm usage by database DBC is growing. The administration team notices the growth in ResUsage and DBQL tables. Which practice should be employed to manage growth in these tables?
A. Delete the ResUsage and DBQL tables.
B. Terminate ResUsage and DBQL logging.
C. Extend the logging period on ResUsage and restrict DBQL logging to summary logging.
D. Move the ResUsage tables and DBQL tables to history tables then delete from the ResUsage and DBQL tables.
Answer: C
A strategic application seeks to execute 60 concurrent queries. Currently, the average response time of these queries is 90 seconds and the Service Level Agreement (SLA) is 60 seconds. Which workload management approach can be used to assist in meeting this SLA?
A. Decrease the workload concurrency throttle.
B. Decrease the concurrency of any ad hoc users.
C. Increase the Priority Scheduler weight for strategic queries.
D. Increase the number of sessions used by each strategic query.
Answer: C
An organization is evaluating data integration suites to replace the existing architecture. The old architecture is an ELT strategy, where data files are typically loaded into staging tables, then transformed into the target tables. What are three advantages of changing to an ETL strategy? (Choose three.)
A. uses less network bandwidth
B. uses less I/O in the database
C. uses less CPU on index maintenance
D. uses less perm space in the database
E. uses less spool space in the database
F. uses less memory on the data integration system
Answer: A,B,C
A number of different departments of a company are being integrated into a warehouse. Each department shares much of the data, but treat the shared data differently. When will this issue get resolved?
A. while loading the data
B. during the logical design
C. during the physical design
D. during application development
Answer: D
An account management application needs to have two additional columns added to the 200 million row TRANS table. The columns will be populated during the launch of a new application feature in two months. An INTEGER column ACC_LEVEL and a DECIMAL(38,2) column CUST_VALUE will need to be added to the table during an application maintenance window. The database administrator would like to add these columns as efficiently as possible and has enough space to create another copy of the table in the database. Which statement is true about the most efficient way to implement these columns in the TRANS table?
A. Alter the table and add the additional columns which will result in block and cylinder splits.


Popular posts from this blog

SQL for Quantile Function in Teradata

3 Uses of SAMPLE function in Teradata

All You Need NULL Value Functions