With Teradata MAPS Feature, you can expand the hardware configuration and choose to postpone the redistribution of tables from the old AMPs to the new AMPs. This delay in moving rows to the new AMPs can provide a significate reduction in down time during the expansion window. Postponing table redistributions is enabled by allowing multiple hash maps (the old hash map that covered the previous configuration and the new hash map for all of the AMPs in the new configuration) to co-exist at the same time. It will be up to the administrator to decide when to move tables into the new, larger map.
The move of a table from the old map to the new map is accomplished by a new type of ALTER TABLE command. When an ALTER TABLE statement that includes a “MAPS = map_name” clause is issued, that ALTER TABLE will be processed similar to an INSERT-SELECT.
The INSERT-SELECT moves tables from one map to a different map is essentially doing the same thing as the standard INSERT-SELECT. However, there are a few differences, which will be highlighted in this posting.
If you’ve got TASM for your workload management, and are demoting your very resource-intensive (potentially disruptive) queries into a penalty box workload in Timeshare Low, and that’s not working out as well as you’d like, here’s another approach to think about.
Move the penalty box workload to Tier 1 of the SLG Tier, give it a very low allocation percent, and run it there all by itself.
Sparse maps are one of the new functionalities available in the MAPS Architecture feature that is part of the Teradata Database 16.10 release. All Teradata deployment platforms will support the use of sparse maps as soon as the platform gets onto 16.10 software.
Sparse maps are simple to understand, easy to use, and can provide benefit when a very small table is frequently or repetitively accessed. Basically, sparse maps allows you to move very small tables onto one or a few AMPs on the system, rather than thinly spreading the table’s rows across all AMPs.
Most sites have a few (or many) very small tables with fewer rows than AMPs in the configuration. Having to perform an all-AMP operation every time those types of tables are read involves some level of activity across all AMPs. Some of those AMPs will be wasting their resources because they have no rows. That can add up, especially on systems that include a large number of AMPs.
This posting discusses key points to keep in mind before you begin moving your small tables into sparse maps.
I came across an interesting behavior related to Teradata workload management. This behavior is reflected in both TASM and TIWM. It has to do with using throttles to delay all queries that include access to a certain table (or set of tables) at times when those tables are being dropped and recreated.
Statistical information is vital for the optimizer when it builds query plans. But collecting statistics can involve time and resources. By understanding and combining several different statistics gathering techniques, users of Teradata can find the correct balance between good query plans and the time required to ensure adequate statistical information is always available.
Teradata Database workload management offers a new feature starting in Teradata Database 15.10 for managing the throttle delay queue. It’s called “Prioritized Delay Queue”. This posting looks at how the Prioritized Delay Queue feature works and some things to keep in mind if you begin to use it. The intent of this enhancement is to ensure that higher priority work is able to be released from the delay queue ahead of lower priority work.
The ResUsageSAWT logs detailed information about AMP worker tasks (AWTs). Because AWTs are a finite resource, most sites keep on eye when they are close to running out.
ResUsageSAWT reports in-use and max AWT counts for all of the 16 message work types on each AMP and reports them at the end of each logging interval. It also includes a column InuseMax that reflects the maximum number of AWTs in combination that were in use at any one time during the log period. This posting is intended to clear up any confusion over what InuseMax offers, when it’s useful, and when other AWT metrics are more suitable.
A new feature in Teradata Database 15.0 gives the optimizer the choice to bypass all-AMP row redistributions under certain circumstances. This can be a good thing for queries that are redistributing just a few rows and the system has many AMPs. Without this feature all row redistributions, no matter of what size, would require that an AMP worker task be spawned and started up on each AMP in preparation for receiving rows.
If you are someone who uses Database Query Log (DBQL) to keep on top of how much resource your users and applications are consuming, you are probably aware that there are different DBQL tables that collect information related to usage. Starting in Teradata Database 15.0, if you are tracking usage for utilities, that detail is carried in two places: The DBQLogTbl and the DBQLUtilityTbl.
Which should you use? Should you combine usage from both tables? Or only consider the usage in one of the two tables?