Data warehousing

Analysis of issues in data warehousing, with extensive coverage of database management systems and data warehouse appliances that are optimized to query large volumes of data. Related subjects include:

August 24, 2009

Teradata highlights some analytic use cases

A couple of slides in my recent briefing on Teradata’s Active Enterprise Data Warehouse Story contained long lists of analytic use cases, at a finer level of granualarity than I’m focusing on for a September speaking tour. I think they’re interesting to pass along. Read more

August 24, 2009

Teradata’s Active Enterprise Data Warehouse story

Teradata used to tell a one-size-fits-all Enterprise Data Warehouse (EDW) story. That’s no longer the case. Last year, Teradata introduced a range of products. I think Teradata is serious about selling its full product range, and by now has achieved buy-in from its sales force for that strategy. I base these beliefs on data points such as:

But that raises the question: How does Teradata pitch the advantages of its top-end product line these days? At least at the corporate level, the answer seems to focus less on the “EDW” concept than it used to, and more on “Active.” Teradata – which actually has been talking about “Active Data Warehousing” for about a decade — indeed calls its top-end 55xx series the “Teradata Active Enterprise Data Warehouse.”

Teradata proudly told me that it has >100 customers who have truly adopted an “Active” EDW. When we discussed what that meant, supported by a whole lot of named examples, it became clear that “Active” data warehousing: Read more

August 21, 2009

Bottleneck Whack-A-Mole

Developing a good software product is often a process of incremental improvement. Obviously, that can happen in the case of feature addition or bug-fixing. Less obviously, there’s also great scope for incremental improvement in how the product works at its core.

And it goes even further. For example, I was told by a guy who is now a senior researcher at Attivio: “How do you make a good speech recognition product? You start with a bad one and keep incrementally improving it.”

In particular, I’ve taken to calling the process of enhancing a product’s performance across multiple releases “Bottleneck Whack-A-Mole” (rhymes with guacamole). This is a reference to the Whack-A-Mole arcade game,* the core idea of which is:

Read more

August 21, 2009

Kickfire’s FPGA-based technical strategy

Kickfire’s basic value proposition is that, if you have a data warehouse in the 100s of gigabytes, they’ll sell you – for $32,000 – a tiny box that solves all your query performance problems, as per the Kickfire spec sheet. And Kickfire backs that up with a pretty cool product design. However, thanks in no small part to what was heretofore Kickfire’s penchant for self-defeating secrecy, the Kickfire story is not widely appreciated.

Fortunately, Kickfire is getting over its secrecy kick. And so, here are some Kickfire technical basics.

The new information there is that Kickfire relies on an FPGA; Read more

August 8, 2009

Sorting out Netezza and Oracle Exadata data warehouse appliance pricing

Netezza recently announced a new generation of data warehouse appliance called TwinFin. TwinFin’s clearest stated list price is “a little under $20,000 per terabyte of user data,” which in my opinion immediately became the new industry reference point for discussing prices in the data warehouse appliance category. Vigorous discussion ensued, especially in the comment thread to the first of the two posts linked above. Here’s some followup.

Netezza should not have claimed a “10-15X price/performance improvement,” based on a 3-5X performance improvement and a 3X decrease in price/terabyte, and I should have grilled Netezza harder when it first made the claim. In fact, there is no unit of performance that you can, in a reasonable blended average, get 10-15X more of per dollar in TwinFin than you can in the predecessor NPS series.

Read more

August 8, 2009

What does Netezza do in the FPGAs anyway, and other questions

The news of Netezza’s new TwinFin product family has generated a lot of comments and questions, some pretty reasonable, some quite silly. E.g., I’ve seen it suggested privately or publicly that

Netezza’s Phil Francisco addressed some points of this nature in a recent blog post.

More reasonable is the question:

Now that Netezza has changed its architecture, what are all those FPGAs (Field-Programmable Gate Arrays) being used for anyway?

The short answer is: Read more

August 5, 2009

Dataupia is officially for sale

Dataupia marketing VP Samantha Stone — who by the way has been one heck of a trooper through Dataupia’s troubles — is joining the exodus from the company.  General graciousness aside, the heart of Samantha’s farewell email reads:

Unfortunately, we have had to reduce our burn rate as we seek an acquirer for our technology.

We have a group of loyal employees remaining on staff focused on current production customers and the acquisition efforts.

As part of the most recent staff reductions I will be leaving Dataupia.

Two years ago I wrote:

[Dataupia would] make a great acquisition for a BI company or DBMS vendor who could then say “Oh, no, this isn’t a DBMS appliance – it’s merely a data warehouse accelerator.” When you look at it that way, their chances of prospering look distinctly higher.

But at this point I think there probably would be more appealing ways for those vendors to meet the same needs.

August 4, 2009

FlexStore and the rest of Vertica 3.5

Today, Vertica is announcing its 3.5 release, timed in line with a TDWI conference. Vertica 3.5 is scheduled to go into beta test in mid-August and be released to general availability in early October. Vertica 3.5 highlights include:

Read more

August 4, 2009

PAX Analytica? Row- and column-stores begin to come together

Column-store proponents are prone to argue, in effect, that the only reason to implement an analytic DBMS with row-based storage is laziness. Their case generally runs along the lines:

Pushbacks to this argument from row-based vendors include:

Read more

August 4, 2009

Vertica’s version of MapReduce integration

I talked with Omer Trajman of Vertica Monday night about Vertica’s MapReduce integration, part of its Vertica 3.5 release. Highlights included:

Apparently, the use cases for Vertica/Hadoop integration to date lie in algorithmic trading and two kinds of web analytics. Specifically: Read more

← Previous PageNext Page →

Feed: DBMS (database management system), DW (data warehousing), BI (business intelligence), and analytics technology Subscribe to the Monash Research feed via RSS or email:

Login

Search our blogs and white papers

Monash Research blogs

User consulting

Building a short list? Refining your strategic plan? We can help.

Vendor advisory

We tell vendors what's happening -- and, more important, what they should do about it.

Monash Research highlights

Learn about white papers, webcasts, and blog highlights, by RSS or email.