October 25, 2009

Teradata hardware strategy and tactics

In my opinion, the most important takeaways about Teradata’s hardware strategy from the Teradata Partners conference last week are:

In addition, some non-SSD componentry tidbits from Carson Schmidt include:

Let’s get back now to SSDs, because over the next few years they’re the potential game-changer. The big news on SSDs is that after last year’s Teradata Partners conference, a stealth supplier* introduced itself and convinced Teradata it offers really great SSD technology. For example, not a single SSD it has provided Teradata has ever failed. (In hardware, that is. There have of course been firmware bugs, suitably squashed.) I think SSD performance is also exceeding Teradata’s expectations. This supplier is where the 6-9 month time-to-market gain comes from.

*Based on how often the concept of “stealth” and “name is NDAed” came up, I do not believe this is the SSD company another vendor told me about that is going around claiming it has a Teradata relationship.

Teradata SSD highlights include:


13 Responses to “Teradata hardware strategy and tactics”

  1. Joe Harris on October 26th, 2009 4:06 am


    Great insights into the recent TD announcements. I haven’t seen any of this mentioned elsewhere, as usual.

    Two questions occur to me :

    First, if 10Gb ethernet improves query time versus Bynet, why on earth have they persisted with Bynet for all this time? Given the amount of data that TD shuffles between AMPs I would think this is their top priority. I wonder if the previous disk subsystems were the limiter, as opposed to the interconnect fabric.

    Second, if the SSD I/O is limited by SAS 1.0 speeds, why are they even using SAS at all. The flash cards in Exadata 2 are connected via PCI Express (AFAIK) which delivers 4Gb/s+. I’m out of my depth here a little bit but, simplistically, if Oracle can do it why can’t Teradata?


  2. Curt Monash on October 26th, 2009 10:22 am


    For Teradata, there’s a lag between “We think X is the right technology to use” and “We’re sure enough to use X.” Perhaps not coincidentally, few Teradata users are ever dissatisfied about much except price or, sometimes, complexity.

  3. FredZ on October 26th, 2009 11:30 am

    Open question;
    Doesn’t SSD technology greatly expand the size that SMP DW severs can scale to? Especially in the low to mid-range data warehouses? And with SSD prices dropping, won’t it soon make sense to spend dollars on SSD SAN versus MPP servers.

    For example; How would a 8 cpu/6 core server (48 cores) running SMP database software (take your pick Oracle, DB2 etc) with 256 Gig memory (total price ~$80K @ HP.com) with a 5TB SSD SAN (~ $200K RAMSAN) compete with a MPP players like Teradata or Netezza? Remember the lower energy costs of SSD and the OLTP functionality of SMP DB software. Which gives it more ODS flexibility. Again, for the low to mid-range DW it seems like a viable alternative?

  4. Curt Monash on October 26th, 2009 1:48 pm

    1. Kickfire claims to give great DW performance on any schema up to at least 5 TB of data on a single box for <$200K. And that 5 TB will go up a lot as the get compression ironed out.

    My point: There are lots of good alternatives.

    2. You can’t just add sufficiently well-performing long-running queries into a good OLTP DBMS and say you have a good mixed-used system. E.g., workload management gets a lot harder. A DW should ideally have more optimistic locking than an OLTP system. Etc.

    3. In theory, you can break tables into sufficiently small pieces and store them in sufficiently different ways that you can do everything with great performance and reliability in one system.

    That’s theory.

    4. The cost of hardware is often a small factor when compared with the cost of DBMS licenses. Just look at Oracle’s software cost per core (license and maintenance). Whether you’re paying $2,500/core for the hardware or (more likely) a lot less is, by comparison, pretty irrelevant.

  5. Joe Harris on October 27th, 2009 12:53 pm


    Just following up on my second point regarding the use of a SAS disk interface versus PCIe, a new post from James Hamilton (one of the genii behind AWS) appears to concur.

    “Expect flash to stay strong and relevant for the near term and expect it to be PCIe connected rather than SATA attached.”

    Admittedly this is in reference to a presentation given by a Sun employee. Nevertheless, disk interfaces are already too slow before TD even release their flash appliance and that seems like a problem to me.


  6. Greg Rahn on October 29th, 2009 1:31 pm

    @Joe Harris

    If you are reading between the lines here, you will see that Teradata’s I/O layer is very inefficient for a data warehouse. If you look around almost every DW DBMS vendor is focusing on maximizing sequential I/O so that they can get the most disk bandwidth out of each HDD. The quote of “15 MB/second on their fastest disks” is unfortunate as those drives are physically capable of doing well over 100MB/s of sequential I/O. This likely has to do with the small random read pattern that is common with Teradata. Michael McIntire commented on this back in April. This is also the reason that Teradata’s systems have a very large number of HDDs in them. For instance, the 5555 series generally consists of 3/5 cliques which is 3 dual socket quad core Intel Harpertown hosts (plus one spare host) and 5 fibre channel arrays each with 60 HDDs, so the ratio of HDDs to hosts is 100:1. The capacity of these drives has been 146GB for quite some time, and I hear they are starting to now offer the 300GB drives. Compare this to other vendors (say like Oracle Exadata) where the drive capacities are 600GB SAS (or 2TB Midline SAS) and the scan rates per disk are around 125MB/s – over 100MB/s more per HDD than Teradata. In comparison, a single Sun Oracle DB Machine has a physical scan capacity of 21GB/s with 168 SAS HDDs and a Teradata 12 node 1200 drive 5555 system has a physical scan capacity 18GB/s (1200*15MB/s), so what Exadata can do in 1 rack, it takes Teradata on the order of 8 or so racks of equipment. Knowing this, it now makes sense why Teradata is so “excited” about using SSD. By doing so, it removes the “latency penalty” for random I/O operations and has the opportunity to drastically shrink their data center footprint, though of course at the increased cost of SSD vs HDD.

  7. Teradata Transition On Course in Steady Quarter, With Exciting New Offerings Ahead « Market Strategies for IT Suppliers on November 6th, 2009 11:52 pm

    [...] summary of some of the issues and opportunities in an exeptionally useful post from Curt Monash here. But one hopes Teradata rationalizes the bewildering story, the wild mix of OSs, interconnects [...]

  8. Boston Big Data Summit keynote outline | DBMS2 -- DataBase Management System Services on November 23rd, 2009 2:26 am

    [...] Teradata, Oracle have both signaled moving to more modularity [...]

  9. Aster Data 4.0 and the evolution of “advanced analytic(s) servers” | DBMS2 -- DataBase Management System Services on December 4th, 2009 10:39 am

    [...] in its own software. (Other examples of this strategy would be Vertica, Oracle Exadata/ASM, and Teradata Fallback.) Prior to nCluster 4.0, this caused a problem, in that the block sizes for mirroring were so large [...]

  10. Research agenda for 2010 | DBMS2 -- DataBase Management System Services on December 31st, 2009 6:02 pm

    [...] Teradata has made large strides in making solid-state memory useful [...]

  11. Ashok Chand on March 3rd, 2010 11:11 pm

    Can I use Teradata for OLTP system?
    Please give me some pros and cons about the OLTP in Teradata.

  12. Curt Monash on March 4th, 2010 12:56 am

    Teradata is not optimized for cost-effective OLTP performance. And third-party OLTP apps generally don’t run on Teradata. On the other hand, formally it has everything you’d need to do OLTP.

    Bottom line: If you want to do a little OLTP on a Teradata box you buy for other reasons, fine. But if your main goal is OLTP, it’s the wrong thing to get.

    I should add that Teradata’s architecture is somewhat more OLTP-like that newer analytic DBMS. E.g., it is designed for random more than sequential reads, it assumes you’ll probably normalize your data, you can buy Teradata boxes with lots of small, fast disks, and so on.

  13. Flash, other solid-state memory, and disk | DBMS2 -- DataBase Management System Services on July 7th, 2010 12:44 am

    [...] and Teradata’s beliefs about the importance of solid-state [...]

Leave a Reply

Feed: DBMS (database management system), DW (data warehousing), BI (business intelligence), and analytics technology Subscribe to the Monash Research feed via RSS or email:


Search our blogs and white papers

Monash Research blogs

User consulting

Building a short list? Refining your strategic plan? We can help.

Vendor advisory

We tell vendors what's happening -- and, more important, what they should do about it.

Monash Research highlights

Learn about white papers, webcasts, and blog highlights, by RSS or email.