Partial overview of Ab Initio Software
Ab Initio is an absurdly secretive company, as per a couple of prior posts and the comment threads on same. But yesterday at TDWI I actually found civil people staffing an Ab Initio trade show booth. Based on that conversation and other tidbits, I think it’s fairly safe to say: Read more
Categories: Ab Initio Software, Analytic technologies, Benchmarks and POCs, Data integration and middleware, EAI, EII, ETL, ELT, ETLT, Expressor, Pricing, Talend | 14 Comments |
Introduction to Expressor Software
I’ve chatted a few times with marketing chief Michael Waclawiczek and others at data integration startup Expressor Software. Highlights of the Expressor story include:
- Expressor was founded in 2003 and funded in 2007. Two rounds of funding raised $16 million.
- Expressor’s first product release was in May, 2008; before that Expressor built custom integration tools for a couple of customers.
- Michael believes Expressor will have achieved 5 actual sales by the end of this quarter, as well being in 25 “highly active” sales cycles.
- Whatever Expressor’s long-term vision, right now it’s selling mainly on the basis of performance and affordability.
- In particular, Expressor believes it is superior to Ab Initio in both performance and ease of use.
- Expressor says that parallelism (a key aspect of data integration performance, it unsurprisingly seems) took a long time to develop. Obviously, they feel they got it right.
- Expressor is written in C, so as to do hard-core memory management for best performance.
- Expressor founder John Russell seems to have cut his teeth at Info USA, which he left in the 1990s. Other stops on his journey include Trilogy (briefly) and then Knightsbridge, before he branched out on his own.
Expressor’s real goals, I gather, have little to do with the performance + price positioning. Rather, John Russell had a vision of the ideal data integration tool, with a nice logical flow from step to step, suitable integrated metadata management, easy role-based UIs, and so on. But based on what I saw during an October visit, most of that is a ways away from fruition.
Categories: Analytic technologies, Data integration and middleware, EAI, EII, ETL, ELT, ETLT, Expressor, Market share and customer counts | 4 Comments |
Talend update
I chatted yesterday at TDWI with Yves de Montcheuil of Talend, as a follow-up to some chats at Teradata Partners in October. This time around I got more metrics, including:
- Talend revenue grew 6-fold in 2008.
- Talend revenue is expected to grow 3-fold in 2009.
- Talend had >400 paying customers at the end of 2008.
- Talend estimates it has >200,000 active users. This is based on who gets automated updates, looks at documentation, etc.
- ~1/3 of Talend’s revenue is from large customers. 2/3 is from the mid-market.
- Talend has had ~700,000 downloads of its core product, and >3.3 million downloads in all (including documentation, upgrades, etc.)
It seems that Talend’s revenue was somewhat shy of $10 million in 2008.
Specific large paying customers Yves mentioned include: Read more
Categories: Analytic technologies, Data integration and middleware, EAI, EII, ETL, ELT, ETLT, eBay, Market share and customer counts, Specific users, Talend | 5 Comments |
Microsoft SQL Server Fast Track
Stuart Frost of Microsoft (nee’ DATAllegro) checked in, with Microsoft’s TDWI-timed announcements. The news part was something called “SQL Server Fast Track“, which is the Microsoft SQL Server equivalent to Oracle’s “recommended configurations” or IBM’s “BCUs.” SQL Server Fast Track is further being portrayed as an incremental step toward Madison, Microsoft’s future high-end data warehousing offering.
Categories: Data warehousing, Microsoft and SQL*Server, Pricing | 5 Comments |
The questionable benefits of terabyte-scale data warehouse virtualization
Vertica is virtualizing via VMware, and has suggested a few operational benefits to doing so that might or might not offset VMware’s computational overhead. But on the whole,it seems virtualization’s major benefits don’t apply to the large-database MPP data warehousing. Read more
Categories: Columnar database management, Data warehousing, Database compression, Theory and architecture, Vertica Systems | 2 Comments |
Vertica Virtualizes Via VMware
(In other news, the sixth sick sheik’s sixth sheep is sick … but I digress.)
It seems that every analytic DBMS vendor feels compelled to issue at least one press release the week of winter TDWI. Vertica’s grand revelation this year is that you can use Vertica with VMware.* Of course, VMware working the way it does, you in fact have always been able to use Vertica with VMware. But now things are slightly improved, because Vertica has built install packages you can download, and has been working out recommended configuration settings as well.
Categories: Data warehousing, Vertica Systems | 2 Comments |
MapReduce user eHarmony chose Netezza over Aster or Greenplum
Depending on which IDG reporter you believe, eHarmony has either 4 TB of data or more than 12 TB, stored in Oracle but now analyzed on Netezza. Interestingly, eHarmony is a Hadoop/MapReduce shop, but chose Netezza over Aster Data or Greenplum even so. Price was apparently an important aspect of the purchase decision. Netezza also seems to have had a very smooth POC. Read more
Categories: Application areas, Aster Data, Benchmarks and POCs, Data warehousing, Greenplum, MapReduce, Netezza, Oracle, Predictive modeling and advanced analytics, Pricing | 5 Comments |
Microstrategy tidbits
I chatted with Microstrategy Wednesday in a call focused on the upcoming Microstrategy 9. There wasn’t a lot of technical content, but I did glean:
- In Microstrategy 9, virtual ROLAP cubes will be able to draw on multiple relational databases, not just one. (Frankly, I’ve never understood why BI vendors are so slow to put in features like that.)
- Actually, in Microstrategy 9 cubes won’t just be virtual. You’ll be able to instantiate parts of them in memory.
- The in-memory part requires manual intervention. However, that intervention can be as minor as pushing a button to accept the recommendations of a Cube Advisor.
- The Microstrategy Cube Advisor will examine workloads for a month or so to see which queries chew up the most resources.
- Another new feature is “complete” OLAP drilldown from any point in any chart or graph, without pre-programming or pre-specification.
- Microstrategy’s favorite DBMS partners are, in some order, Netezza and Teradata.
- Microstrategy 9 is currently scheduled for March 23 release.
The Netezza guys propose a POC checklist
The Netezza guys at “Data Liberators” are being a bit too cute in talking about FULL DISCLOSURE yet not actually saying they’re from Netezza — but only a bit, in that their identity is pretty clear even so. That said, they’ve proposed a not-terrible checklist of how to conduct POCs. Of course, vendor-provided as it is, it’s incomplete; e.g., there’s no real mention of a baseball-bat test.
Here’s the first part of the Netezza list, with my comments interspersed. Read more
Categories: Benchmarks and POCs, Buying processes, Data warehousing, Netezza | 1 Comment |
25 facts about Ingres, give or take a couple
Emma McGrattan of Ingres offers a “25 facts” post about Ingres. 24 really are about Ingres. Some are interesting (who knew Ingres still used a lot of Quel?). Some are if anything understated — e.g., there are lots of current CEOs who are Ingres alums (Dave Kellogg and Dennis Moore jump to mind). Only one is a real eyebrow-raiser.
Point 23 says “The average tenure of an Ingres Engineer is 15+ years.” On the other hand, Point 3 says “The longest serving member of Ingres staff is John Smedley who has been with us since June of 1987.” And most of Ingres’ technical staff left after Ingres was acquired by CA, which occurred a few months shy of 15 years ago. Reconciling all that is challenging.
Actually, I was dubious about a second claim too, namely that Ingres/Star was the first distributed DBMS; I thought that the distributed version of Tandem NonStop SQL actually predated it by a few years. But a somewhat contemporaneous article with a number of distributed DBMS dates shows my memory was wrong on that score.
Categories: Actian and Ingres | 3 Comments |