May 20, 2008
Top-end data warehouse sizes have grown hundreds-fold over the past 12 years
I just tripped across a link from February, 1996 in which NCR/Teradata:
- Bragged that it had half a dozen customers with >1 TB of raw user data
- Showed off a “record-breaking” 11 TB simulation
That represents roughly a 60-70% annual growth rate in top-end database sizes in the intervening 12 years.
Comments
4 Responses to “Top-end data warehouse sizes have grown hundreds-fold over the past 12 years”
Leave a Reply
Search our blogs and white papers
Monash Research blogs
- DBMS 2 covers database management, analytics, and related technologies.
- Text Technologies covers text mining, search, and social software.
- Strategic Messaging analyzes marketing and messaging strategy.
- The Monash Report examines technology and public policy issues.
- Software Memories recounts the history of the software industry.
User consulting
Building a short list? Refining your strategic plan? We can help.
Vendor advisory
We tell vendors what's happening -- and, more important, what they should do about it.
Monash Research highlights
Learn about white papers, webcasts, and blog highlights, by RSS or email. |
-
Recent posts
-
Categories
- About this blog
- Analytic glossary
- Analytic technologies
- Application areas
- Buying processes
- Companies and products
- 1010data
- Ab Initio Software
- Actian and Ingres
- Aerospike
- Akiban
- Aleri and Coral8
- Algebraix
- Alpha Five
- Amazon and its cloud
- ANTs Software
- Aster Data
- Ayasdi
- Basho and Riak
- Business Objects
- Calpont
- Cassandra
- Cast Iron Systems
- Cirro
- Citus Data
- ClearStory Data
- Cloudant
- Cloudera
- Clustrix
- Cogito and 7 Degrees
- Cognos
- Continuent
- Couchbase
- CouchDB
- Databricks, Spark and BDAS
- DATAllegro
- Datameer
- DataStax
- Dataupia
- dbShards and CodeFutures
- Elastra
- EMC
- Endeca
- EnterpriseDB and Postgres Plus
- Exasol
- Expressor
- FileMaker
- GenieDB
- Gooddata
- Greenplum
- Groovy Corporation
- Hadapt
- Hadoop
- HBase
- Hortonworks
- HP and Neoview
- IBM and DB2
- illuminate Solutions
- Infobright
- Informatica
- Information Builders
- Inforsense
- Intel
- Intersystems and Cache'
- Jaspersoft
- Kafka and Confluent
- Kalido
- Kaminario
- Kickfire
- Kognitio
- KXEN
- MapR
- MarkLogic
- McObject
- memcached
- MemSQL
- Metamarkets and Druid
- Microsoft and SQL*Server
- MicroStrategy
- MonetDB
- MongoDB
- MySQL
- Neo Technology and Neo4j
- Netezza
- NuoDB
- Nutonian
- Objectivity and Infinite Graph
- Oracle
- Oracle TimesTen
- ParAccel
- Pentaho
- Pervasive Software
- PivotLink
- Platfora
- PostgreSQL
- Progress, Apama, and DataDirect
- QlikTech and QlikView
- Rainstor
- Revolution Analytics
- Rocana
- salesforce.com
- SAND Technology
- SAP AG
- SAS Institute
- ScaleBase
- ScaleDB
- Schooner Information Technology
- SciDB
- SenSage
- SequoiaDB
- SnapLogic
- Software AG
- solidDB
- Splunk
- Starcounter
- StreamBase
- Sybase
- Syncsort
- Tableau Software
- Talend
- Teradata
- Tokutek and TokuDB
- Truviso
- VectorWise
- Vertica Systems
- VoltDB and H-Store
- WibiData
- Workday
- Xkoto
- XtremeData
- Yarcdata and Cray
- Zettaset
- Zoomdata
- Data integration and middleware
- Data types
- DBMS product categories
- Emulation, transparency, portability
- Fun stuff
- Market share and customer counts
- Memory-centric data management
- Michael Stonebraker
- Parallelization
- Presentations
- Pricing
- Public policy
- Software as a Service (SaaS)
- Specific users
- Storage
- Theory and architecture
- TransRelational
- Uncategorized
-
Date archives
-
Links
-
Admin
Note that the NCR person quoted in the article is Mark Hurd – now head of H-P (chairman, CEO, and president), if I am not mistaken.
You are quite right. HP’s data warehousing strategy is obviously influenced by Hurd’s background at Teradata.
CAM
For the mathematically challenged, what are you suggesting is top-end today? 1.7^12 = 582 so are you saying 500 TB to 5 PB is the top-end today? (Please don’t shred me on the math!)
Dave,
That would indeed be about correct.
Last year, Teradata had a handful of sites >1/2 terabytes in user data. Their marketing chief was referring to these as >1 petabyte, based on total disk, but that’s not my favored metric. eBay is even bigger, as I first noted in http://www.dbms2.com/2008/02/11/ebay-is-over-5-petabytes-now/ .
If you use the handy-dandy Google search box on the right, and search on “petabyte”, you’ll find much of the top-end discussion. (Yeah, I know it’s just Google custom search, but that’s a lot better than nothing. Besides, plans are afoot to upgrade that with technology from — well, from one of your quasi-competitors. 😉 )
Some of the uncertainty in my growth-rate range came from questioning what was meant in the original press release. Just as a petabyte isn’t always a petabyte now, all terabytes weren’t created equal back then.
Best,
CAM