Even casual technology watchers know that organizations are awash in huge volumes of data. But we may not recognize the sheer scale of the numbers involved. It is estimated that 2.5 exabytes of data are created every day. To put that into perspective, the amount of all digital information generated prior to 2003 was a mere 5 exabytes in total. In fact, 90% of the world’s data was created in the last two years alone, and that growth isn’t showing any signs of slowing. With all of this data, just 0.5% is currently analyzed, and that percentage is shrinking as more data is collected. This data is coming from many sources such as machine logs, mobile devices, physical sensors, and even cars and satellites.

Thanks to new big data frameworks like Hadoop and Spark, NoSQL data stores like Cassandra, and modern data warehouses like HP Vertica, this data can now be managed and analyzed for insight at huge scale. These applications start as small projects and quickly grow to large production clusters, and the data becomes more critical as companies begin to rely on it. With this growth and reliance of data within one application comes a need to protect and manage it, and companies are only now beginning to face up to that.

This is where Talena comes in. Talena is not just about back-up, which is generally regarded as an insurance policy by IT departments. Data is now regarded as a strategic asset but due to the sheer size, exploiting this asset has become difficult. That is where Talena’s story resonated with Canaan.

Talena is the first to focus on Big Data workloads, which have a fundamentally different set of requirements for data availability management including:

  • They are 100x the size of traditional environments, sometimes scaling to petabytes. A data management product must scale horizontally and elastically to support this volume.
  • They are often distributed across thousands of nodes. Deploying an agent on each node to copy data for backup, archiving, or even test/development management is clearly unmanageable. A far better approach is needed.
  • They can handle petabytes of data, so storage optimization is particularly important. A storage optimization process must accommodate the variety of data formats that now power a modern application portfolio. For example, compressing a Cassandra database is fundamentally different from compressing a Hive database.
  • They are data aware as well as file aware. For Big Data, a data management solution must also be data aware, not just file aware, so that it can effectively back up and recover data at a granular level.

The Talena solution delivers an entirely new architecture that overcomes the limitations inherent in traditional data management operations. Via the Talena software environment, companies can implement multiple workflows including test/dev management, backup and recovery, archiving, and disaster recovery.

The CEO/co-founder, Nitin Donde came from senior engineering roles at 3Par, EMC and AsterData. The rest of the team has a pedigree from great companies like Hortonworks, Couchbase, Sumo Logic, and NetApp. We are excited to welcome the Talena team to Canaan and look forward to a great journey together.

Deepak Kamra

General Partner, Iconoclast

Published

August 11, 2015

Share (v.) Enjoy jointly with others!