Enrico Signoretti of Juku writes:
Data grows (steadily… and exponentially) and nothing gets thrown away. Since data adds up, the concept of the “data lake” has taken shape. Even systems created for big data are starting to sense this problem and system architects are beginning to think differently about storage.
I’m going to take Hadoop as an example because this gives a good idea of a hyper converged infrastructure, doesn’t it?
Today, most Hadoop clusters are built on top of a HDFS (Hadoop Distributed File system). HDFS characteristics make this FS much cheaper, reliable and scalable than many other solutions but, at the same time it’s limited by the cluster design itself.
A great look at the types of convergence (or lack thereof) in the market. Hyperconvergence isn’t for everyone. Read on to find out what may work best for you.
- Migrating to Healthcare Cloud Apps With Acadia and Viptela - June 5, 2017
- FutureWAN – The SD-WAN Education You Need - May 26, 2017
- Linksys and the Resurgence of the SMB - May 8, 2017
- Review – Docker Networking Cookbook - February 13, 2017
- The Future of SD-WAN Is Now! - January 10, 2017
- How Kindred Healthcare Uses SD-WAN to Secure Patient Data - December 2, 2016
- The Power of ONUG And What It Means To You - November 30, 2016
- Enabling The Most Remote Offices With Viptela - November 4, 2016
- ONUG Day 2 Wrap Up – Thoughts on Monitoring 2.0 - October 26, 2016
- ONUG Day 1 Wrap Up: SD-WAN In The Spotlight - October 25, 2016