Storage – Connect Worldwide https://connect-community.org Your Independent HPE Technology User Community Wed, 09 Aug 2023 21:39:34 +0000 en-US hourly 1 https://wordpress.org/?v=6.7.1 https://connect-community.org/wp-content/uploads/2021/10/cropped-favicon-2-32x32.png Storage – Connect Worldwide https://connect-community.org 32 32 HPE GreenLake for File Storage: Storage for Splunk’s cold tier that’s fast and simple https://connect-community.org/hpe-greenlake-for-file-storage-storage-for-splunks-cold-tier-thats-fast-and-simple/ https://connect-community.org/hpe-greenlake-for-file-storage-storage-for-splunks-cold-tier-thats-fast-and-simple/#respond Wed, 09 Aug 2023 21:39:34 +0000 https://connect-community.org/?p=62572

Discover the recognizable benefits of deploying Splunk in a cloud-enabled environment based on HPE GreenLake for File Storage.

–By Keith Vanderford, Storage Solutions Engineer, HPE

Splunk deployments typically have a high performance storage tier for hot and warm data, and a cheaper but slower storage tier for cold data. With this implementation, you have to balance search times against capacity constraints. Splunk’s searches are fastest when searching over recent data in your hot and warm buckets. For the best response times, it’s desirable to have all your data on that fastest tier of storage. But the capacity of that tier is usually limited due to the cost of fast flash-based storage. In order to increase capacity while trying to hold total cost down, organizations usually implement a second tier of storage for Splunk’s cold data, using less costly but slower storage technologies. Managing this second tier of storage introduces more complexity to your infrastructure, causing your staff to spend more time administering storage and less time extracting value from the data contained in that storage.

Questions?

I know you’ve got questions, like:

Shouldn’t my business be more about discovering valuable insights than managing storage?

Why does the second tier of storage for my cold data have to be slower?

Why does storage management have to be complicated?

HPE GreenLake for File Storage provides answers

Here’s good news: With HPE GreenLake for File Storage, your cold storage tier can be fast as well as simple and intuitive to manage.

HPE GreenLake for File Storage provides the perfect infrastructure for a cold storage tier that is both fast and easy to manage. It lets you take advantage of Splunk’s ability to provide the insights you need with fast searches over not just your hot and warm data, but your older (cold) data as well. HPE GreenLake for File Storage is an ultra-efficient all-NVMe storage solution with a cloud-like operational experience for data lakes. It delivers sustained, predictable throughput for enterprise performance at scale. The intuitive cloud interface also helps you reduce operational overhead.

You can reduce the performance penalty normally associated with searching older data by using the fast file-based storage provided by HPE GreenLake for File Storage for Splunk’s cold buckets. Simply mount NFS or SMB file shares provided by the ultra-efficient all-NVMe HPE GreenLake for File Storage solution to your Splunk indexers for cold buckets. These shares can also be used for Splunk’s frozen buckets if you have compliance or archive requirements. With this high performance storage solution, searches over Splunk’s cold buckets are extremely fast, accelerating search response times over traditional implementations that use slower storage for cold data.

The unmatched data reduction and low overhead data protection of HPE GreenLake for File Storage decrease the overall capacity required to store your data. For example, in our internal lab testing the observed data reduction rate has been about 3:1. This is significantly better than the reduction typically achieved for Splunk’s indexed data with most other storage platforms. Low overhead erasure coding is implemented using up to 146 data drives with 4 parity drives. This enables HPE GreenLake for File Storage to provide complete data protection with as little as 3% overhead. The combined benefits of this unique data reduction and low overhead data protection help make the most efficient use of your cold storage tier without slowing down searches over your older data.

How using HPE GreenLake for File Storage with your Splunk deployment makes setup and configuration simple

Creating the SMB or NFS file shares you need is quick and easy, and the self-service console gives you an intuitive cloud experience you can access from anywhere. This empowers you to free up your staff to work on adding value to your business rather than managing the day-to-day operations of your infrastructure.

HPE GreenLake for File Storage is available using a pay-as-you-go pricing model that gives you even more value for your infrastructure investment. You only pay for what you use, without having to pay for excess capacity. More resources are always at the ready to allow you to expand when you need to, but you never have to pay for them until you use them. Thus you can maximize the agility and value of your Splunk storage without the costs associated with overprovisioning.

Free your cold data from a slow storage tier and complicated infrastructure

With Splunk and HPE GreenLake for File Storage, you can have extremely fast searches over older data in your cold data tier, while simplifying the management of your storage. Get faster time to insights, and enable your data analysts and data scientists to unlock more value from your data.

To learn more, read the technical brief: Maximize your Splunk investment with HPE GreenLake for File Storage

 
]]>
https://connect-community.org/hpe-greenlake-for-file-storage-storage-for-splunks-cold-tier-thats-fast-and-simple/feed/ 0
WEBINAR: Secure your organization by modernizing data protection with HPE storage solutions https://connect-community.org/webinar-secure-your-organization-by-modernizing-data-protection-with-hpe-storage-solutions/ https://connect-community.org/webinar-secure-your-organization-by-modernizing-data-protection-with-hpe-storage-solutions/#respond Wed, 07 Sep 2022 17:38:00 +0000 https://connect-community.org/?p=60510

Unlock the value of your data with a modern data protection strategy

With data at the heart of your organization, it’s essential to ensure that data is protected, recoverable, and accessible across the enterprise. To do that you need modern data protection solutions that provide strong defenses against ransomware alongside fast, reliable recovery, agile cloud backup, and cost-effective data archive. For all of that, there’s HPE.

 

September 13, 2022
8 am MDT

 


Learn how HPE data protection solutions help you:

  • Recover from ransomware attacks in minutes, at scale
  • Eliminate data protection complexity while meeting all your SLAs
  • Safeguard your data wherever it resides, from edge to cloud
  • Replace management overhead and capex expense with data protection as-a-service
  • Protect your on-prem VMware VMs or Amazon EBS and EC2 instances in minutes with HPE Backup and Recovery Service
]]>
https://connect-community.org/webinar-secure-your-organization-by-modernizing-data-protection-with-hpe-storage-solutions/feed/ 0
Listen to all the HPE GreenLake Around the IT Block podcasts here https://connect-community.org/listen-to-all-the-hpe-greenlake-around-the-it-block-podcasts-here/ https://connect-community.org/listen-to-all-the-hpe-greenlake-around-the-it-block-podcasts-here/#respond Wed, 31 Aug 2022 17:23:21 +0000 https://connect-community.org/?p=60493

In this blog article, you can find all of the Around the IT Block episodes focused on HPE GreenLake. There are a number of them and many of them are really insightful, if I don't say so myself! And as there are new episodes, I'll add them here.

I have mentioned on this blog and others that my Around the IT Block podcast is now an official HPE podcast. Since that happened, a majority of the episodes I’ve done have been HPE GreenLake focused so I thought it would make sense to point you to each of those episodes and order them in a way that makes sense to listen to them.

But before we jump into the episodes, please go to hpe.com/dmn/ATITB and subscribe to the podcast on your favorite podcast platform. It’s on iTunesGoogle PodcastAmazon Music PodcastSpotify Podcast and several others.

So with that, let’s get listening! Note the newest episodes are on top and some of the gems are the older ones so be sure to scroll down the list.

 

HPE making it easy to deploy, run, and manage with Project Edge Cluster
Edge computing is computing happening near or where data is being created and experts agree that edge computing will play a key role in the digital transformation of almost every business. Host Calvin Zito is joined by Denis Vilfort, Director of Edge Products and Strategy to discuss Project Edge Cluster that addresses edge computing.

HPE GreenLake for manufacturing with an AI focus
How can HPE GreenLake help manufacturing? In this episode, host Calvin Zito is joined by Jochen Mohr to talk about how manufacturing is focusing on data and how data can improve quality and the manufacturing lifecycle.

Customer choice with HPE GreenLake and Veeam
exia Clements, VP HPE GreenLake Alliances joined host Calvin Zito to discuss how HPE GreenLake and Veeam working together give customers choice around security, ransomware, data protection and backup.

Accelerate migration to HPE GreenLake through Asset Management
What is HPE Financial Services (HPEFS) and how can they help accelerate your migration to HPE GreenLake? In this episode, Calvin Zito is joined by Michael Swan of HPEFS to discuss how HPEFS can help you plan and fund your IT transformation and transition to aaS and HPE GreenLake. 

HPE GreenLake Management Services for NonStop
Host Calvin Zito is joined by Isaac Deeb to give an overview of HPE GreenLake Management Services for NonStop. Isaac is a Service Delivery Manager for HPE GreenLake with a deep NonStop background.

Introducing HPE GreenLake for Block Storage
Simon Watkins, product marketing manager for Cloud Data Infrastructure and Cloud Data Services joins me to give an overview of the just announced HPE GreenLake for Block Storage offering. We discuss what customer challenges it is addressing, how it’s different from public cloud block storage offerings and from the existing HPE GreenLake for Storage solution. 

Overview of HPE GreenLake managed services for Hybrid Cloud
In this episode, host Calvin Zito talks to Brian Ott, VP of Vice President GreenLake Hybrid Cloud Managed Services Business, to learn about the managed services for Hybrid Cloud and the benefits.

What is the HPE GreenLake customer experience team?
My guest in this episode is by Megs Suratkal, VP of Customer Experience and Customer Success for HPE GreenLake. We discuss what Megs’ team does to improve customers’ experience and ensure customer success with HPE GreenLake. 

Predictions and trends in AI and ML Ops
What does 2022 hold for AI and ML Ops? Host Calvin Zito is joined by Sorin Cheran, VP and HPE Fellow for AI Strategy and Solutions and Evan Sparks, founder of Determined AI and VP of AI at HPE. In this episode, they discuss deep learning and large language models that will help drive AI in 2022.

HPE GreenLake security, risk, and compliance managed services
his is a follow-up to the previous episode. One of the HPE GreenLake Management Services offerings is Security, Risk, and Compliance (SRC). In this episode, I talked to four experts on the HPE GreenLake team who are responsible for those security services.

How HPE GreenLake helps customers go faster with managed services
In this episode, I talked to Ron Irvine, Senior Director of the HPE GreenLake Management Services to learn about the managed services they offer, who it is for and the benefits. Another insightful episode because I learned a lot. You will too!

Ready to break-up with your datacenter but want to keep control?
What is a colo (co-location), what are the benefits, and why should customers consider them instead of public cloud or their own data centers? In this episode I talked to Malcolm Ferguson, a Distinguished Technologist with HPE about using HPE GreenLake cloud data services with a colo. I think this is one of my best podcasts ever. I learned a lot!

What is HPE GreenLake Central?
I was joined by Rajesh Mistry from the HPE GreenLake team to discuss what is and isn’t HPE GreenLake Central. 

What is the edge-to-cloud platform? HPE GreenLake!
In this episode, Flynn Maloy, VP Marketing of HPE GreenLake joined me to talk about what’s happened in IT and at HPE over the last 10 years that has positioned HPE GreenLake for success. 

How about a few Chalk Talks too? 

I get lots of positive reinforcement about my Chalk Talks so as long as I have your attention, how about if you point you to my HPE GreenLake related Chalk Talks. If all of this HPE GreenLake content doesn’t officially make me the HPE GreenLake Guy, I don’t know what else I can do! 

About Calvin Zito 

I’m Calvin Zito and November 2021 marked 38 years with HPE.  I was recognized as a 11 time VMware vExpert. As an early adopter of social media and active participant in communities, I’ve blogged for over 13 years. My Around the IT Block podcast is now an official HPE podcast. You can find ATITB at hpe.com/dmn/ATITB. Subscribe and listen from here. You can find me on Twitter as @CalvinZito. You can also contact me via email

]]>
https://connect-community.org/listen-to-all-the-hpe-greenlake-around-the-it-block-podcasts-here/feed/ 0
14,000 Days with HPE https://connect-community.org/14000-days-with-hpe/ https://connect-community.org/14000-days-with-hpe/#respond Wed, 16 Mar 2022 14:48:04 +0000 https://connect-community.org/?p=59654

I watched a movie recently called 500 Days of Summer. It chronicled a relationship that lasted 500 days between a guy and a woman named Summer. Watching that got me thinking how many days have I worked for HPE. So I asked Google how many days it was between my start date November 16th 1983 and I realized I was closing in on 14,000 days with HPE. That seems like quite a milestone, maybe more impressive than when I hit 35 years. So I thought I would do a chronicle of my experience of 14,000 days with one company.

In this post, I’ll switch between using HP and HPE (depending on the date of what I’m discussing) but my entire career was on the IT/data center side of the business, I look at it as a single company history.

Getting hired

I graduated from DeVry Institute of Technology in Phoenix, Arizona, with a Bachelor’s of Science in Electronic Engineering Technology. HP was one of only two companies that I interviewed with although there were tons of companies looking for potential new hires at DeVry.  I wasn’t too interested in working for anyone else besides HP.

I was hired as a Customer Engineer which was HP-speak for a field engineer. Not long after getting hired, HP was listed in fortune magazine as one of the top companies for hiring and retaining top employees. That stuck with me and certainly made me proud.

Early career in the field

I spent about a year of the first seven years at HP in the Bay Area either getting technical training or as a field on loan working at the Customer Response Center. My focus was commercial systems, the HP 3000 and all the peripherals that attached to it. From my earliest days, I worked on storage including large disk drives with removal platters and tape drives. What I disliked the most was working on chain drive printers and doing preventive maintenance on large datacenter laser printers (the HP 2680A). After working on one of those, I would have toner on my white dress shirts and black stuff when I blew my nose. YUCK! Most of my accounts were in Orange County and I spent a lot of time at Northrop, Hughes Aircraft, the Anaheim Convention Center and the Orange County Register. I really enjoyed my customer relationships.

While I liked what I did, I decided to get an MBA as that was talked about as a winning combination for someone with an engineering degree.

Moving into storage

I moved to Boise to work in HP Storage marketing right after getting my MBA at the end of 1990. HP Boise was a large site – with more 8 buildings. HP Storage started in the Bay Area but was moved to Boise in the mid 70s joining the printer business that was under development in Boise. I was given some products to manage as a product marketing manager. My first product was an enclosure that customers could customize with a mix of disks, CD drives, and tape drives. The forecasting Excel spreadsheet that I inherited from my colleague Rick Boss (who is still a great friend today) had a macro that took over an hour to finish but I had some of the best product forecasts and my planning and manufacturing people loved it!

In 1992, I was sent to Europe for 6 months to introduce RAID arrays into the European market. I had always wanted to live in Italy and while the job was in Germany, I jumped at the chance to live in Europe for an extended time. It was a great experience and it wasn’t long after the Berlin Wall had come down which made it a very interesting time to live in Germany. I got to see lots of different places including my favorite countries Spain and Italy.

One of the most interesting products I worked on was called AutoRAID which we launched in 1995. This was something based on technology developed in HP labs. It was the first virtualized array for mid range computer systems. AutoRAID had RAID 10 as a tier to write data and RAID 5 for more long term storage. It was a log-structured file system. To configure it, all a customer had to do was tell it what size volume they wanted from the front panel.

I remember when we gave a preview to Dave Vellante who at the time was a SVP at IDC and he was blown away by what we had developed. HP Storage was straddling the fence of selling storage to our server divisions running HP-UX and MPE and wanting to move into industry standard Microsoft based servers. While the technology was amazing, the way it was packaged was right in the middle of both of those customers so it wasn’t a great fit for either. In fact, the HP-UX division signed a resale agreement with EMC because the packaging of AutoRAID was too small for most of their customers.

Speaking of EMC, I was on the team that worked to OEM the XP Disk Array from Hitachi Japan. It was a super secret “dark site” project. The R&D team doing the work to release the first product was moved to an offsite location in Roseville to keep the project under wraps. For me to work on it, I had to sign an NDA that basically said I could get fired if I talked about it with anyone. The 5 months leading up to the announcement of the XP256 on May 5, 1999 was probably the most intense 5 months of my career. David Scott who became the CEO of 3PAR and later ran HPE Storage was the marketing leader for launching the XP.

Here’s a video interview I did in 2015 at the last HP Discover event with Jake Ludington (industry video blogger) that was looking back at my years in HP Storage as we were moving to become HPE. Another fun interview I did at the same time was with Dave Vellante of The Cube and this was more looking forward to HPE.

Getting Social

One more period of time to cover and that is when I first started using social media. In 2008. I was in a central marketing organization and was the marketing communications manager for storage. I had a budget between $250,000 to $400,000 a quarter. The storage business decided they wanted to manage that budget and so it was taken away from me. I had to figure out how I could contribute without any budget. I decided to focus on social media.

I took over our storage blog, Around the Storage Block. My goal was to get other people to write and I would act as an editor in chief managing the blog. However, in 2008 I came up against a lot of resistance from subject matter experts who either didn’t have time or didn’t believe writing a blog article was not worth their time. So I created a Twitter account, HPStorageGuy and started writing about 95% of the articles on Around the Storage Block.

A storage  marketing manager asked me what I was doing on Twitter and how was it helping. I had metrics for traffic I was driving to the storage blog and other hp.com websites and the light went on for her. At the time I was the only person in all of HP who was 100% focused on social media. There were no brand accounts, and no one else was focused on social media. Others across the company started to hear about the results I was getting and other businesses started to create a focus around social media. 

Also in 2009, I hosted the storage industry’s first storage focused “HP Tech Day”. We brought 12 bloggers from across the industry to what was then the center of HP Storage, Colorado Springs to hear from HP Storage execs and engineers about what we were doing. Stephen Foskett who now runs Tech Field Day wrote a blog post about his experience at our Tech Day.

I was the first person to use social media at HP and actively working with independent storage bloggers. I was given a companywide award that was called the HP Circle Award in 2009 for the work I had done in social media.

Of course, I have continued on that path of being focused on social media but over the years I kept trying to find new ways to expand what I did. I started podcasting in 2010 and have over 300 episodes (though many were lost when my hosting service moved to a cloud model). Late last year, Around the IT Block podcast became an official HPE podcast and I’ve only done a couple episodes focused on storage, expanding beyond all those years in storage.

One thing worth mentioning is Antonio Neri, current HPE CEO is the best of every CEO and there have been 7 since I’ve worked here. He is the perfect balance of caring about employees, creating a workplace that employees are motivated to give their best, and a rock-solid vision that has HPE well positioned in the market.

Approaching the end?

My LinkedIn network is filled with people from HP and HPE that have played a part in my 14,000 days. There are far too many to name and thank. Heck, I can’t remember half of them! But thank you to each and every one that has made my 14,000 days filled with fun and success. I guarantee I won’t work another 14,000 days but I could see myself working another 1,400 days. I’m having fun and still love what I do. 

 
]]>
https://connect-community.org/14000-days-with-hpe/feed/ 0
AN END TO ENDLESS COMPLEXITY https://connect-community.org/an-end-to-endless-complexity/ https://connect-community.org/an-end-to-endless-complexity/#respond Thu, 18 Nov 2021 16:42:55 +0000 https://connect-community.org/?p=56612
Think of all the ways traditional storage management lets you down — and now toss that aside because the  Cloud Experience of HPE GreenLake makes managing IT resources a breeze. See how HPE GreenLake means no more dealing with hardware and headroom — just effortless management of your new initiatives. http://hpe.to/6043J5SGf
]]>
https://connect-community.org/an-end-to-endless-complexity/feed/ 0
HPE Named a Leader in the 2021 Gartner® Magic Quadrant™ for Primary Storage podcast https://connect-community.org/hpe-named-a-leader-in-the-2021-gartner-magic-quadrant-for-primary-storage-podcast/ https://connect-community.org/hpe-named-a-leader-in-the-2021-gartner-magic-quadrant-for-primary-storage-podcast/#respond Thu, 11 Nov 2021 19:44:58 +0000 https://connect-community.org/?p=56299

It’s that time of year for the 2021 Gartner Magic Quadrant for Primary Storage and again, HPE was named a Leader. IGartner-Magic-Quadrant_2021_ HPE Storage.png recommend you read the Newsroom blog by Sandeep Singh, Marketing VP at HPE Storage. As the title of Sandeep’s blog posts states, this is the third year in a row that HPE is named a leader. You can download a copy of the Gartner Magic Quadrant for Primary Storage for yourself. Note it does require registration to download. 

 

A few highlights from Sandeep’s blog:

  • Primary Storage needs to do a lot more! Storing data is just the beginning. Customers need to ensure that data is: always on, always-fast, and powering their apps; secure and protected from ransomware and cyberattacks; accessible seamlessly to data innovators; and mobile across private, public, and multi-cloud environments.
  • Underpinning HPE’s achievement is a vision to simplify data management with a cloud operational experience. We deliver on this vision with the  HPE GreenLake edge-to-cloud platform.
  • The HPE GreenLake platform offers customers a broad portfolio of cloud services that provide greater choice and freedom for their business and IT strategies via an open and modern platform that delivers a cloud experience everywhere. And storage as a service is a very popular HPE GreenLake offering.
  • Recently announced innovations expand HPE GreenLake cloud services with HPE GreenLake for data protection,figure1.png providing disaster recovery and backup cloud services as seamless extensions to primary storage.
  • HPE InfoSight, our industry-leading AI Ops for infrastructure, has pioneered many innovations in AI-driven intelligence over the last decade as it has transformed the customer support experience.
  • We recently introduced HPE CloudPhysics, a SaaS service delivering data-driven insights for smarter IT decisions. HPE CloudPhysics enables IT to optimize application workload placement, procure right-sized infrastructure services, and lower costs.

 

Something to Talk About

Since the report hit without a lot of warning, I didn’t have time to get a podcast done when this blog article was published but, I have now. I have a podcast to share that I did with Sandeep Singh, VP of Marketing for HPE Storage.  The title of the podcast is: #12 Talking 2021 Gartner® Magic Quadrant™ for Primary Storage and HPE Storage vision.

Here are the places you can now subscribe to Around the IT Block presented by HPE! iTunesGoogle PodcastAmazon MusicStitcherPodcast AddictPlayer.fm, and Spotify podcast. Here’s a blog post with more details about the new home for Around the IT Block podcast

 

 

Here’s the agenda of what we discussed:

  • 0:00 – 01:00 Intros
  • 01:00 – 4:00 Importance of being in the leader’s quadrant and what has set the foundation for the overall leadership.
  • 04:00 – 08:15 HPE Storage strategy of Unified DataOps and the HPE vision of the HPE GreenLake edge-to-cloud platform coming together.
  • 08:15 – 11:30 HPE’s leadership delivering Storage-as-a-Service and data cloud services via HPE GreenLake.
  • 11:30 – 14:10 Leadership with HPE InfoSight vs the industry.
  • 14:10 – 17:00 HPE GreenLake for data protection extending cloud data services to simplify end-to-end data management.
  • 17:00 – 21:40 How does HPE CloudPhysics and Edge-to-Cloud Adoption Framework add value to what we’re doing?
  • 21:40 – 24:00  Wrap up

 

Chalk it up!

And you know what, I have Chalk Talks on each of these topics I want to share with you so you can see why we are very confident about our vision at HPE.

About Calvin Zito 

CJZ Headshot fixed 150 x 150.jpg

I’m Calvin Zito and November 2021 will mark 38 years with HP/HPE.  I was recognized as an 11 time VMware vExpert. As an early adopter of social media and active participant in communities, I’ve blogged for over 13 years. And my podcast is now an official HPE podcast. You can find me on Twitter as @CalvinZito. If you don’t follow me on Twitter, please do it now! You can also contact me via email.

Calvin Zito
Hewlett Packard Enterprise

twitter.com/HPE_Storage
linkedin.com/showcase/hpestorage/
hpe.com/storage

]]>
https://connect-community.org/hpe-named-a-leader-in-the-2021-gartner-magic-quadrant-for-primary-storage-podcast/feed/ 0
HPE Cloud Volumes: Secure, cloud-ready answer to holiday cyberattacks for SMBs https://connect-community.org/hpe-cloud-volumes-secure-cloud-ready-answer-to-holiday-cyberattacks-for-smbs/ https://connect-community.org/hpe-cloud-volumes-secure-cloud-ready-answer-to-holiday-cyberattacks-for-smbs/#respond Tue, 09 Nov 2021 20:21:28 +0000 https://connect-community.org/?p=56122 Martin_Oderinde
November 5, 2021

HPE Cloud Volumes is on-demand enterprise cloud storage delivering block and backup services solution that can protect operations and capital by leveraging cloud-based disaster recovery. 

 

Lights are going up, Thanksgiving turkey is being prepped, everyone is full of holiday cheer. As we approach the end of the year, businesses are ramping up production to meet the customers’ needs. In addition, people are scouring the internet for the best deals. As a result, there’s an influx of financial and email activity.

Now let’s take a look at the dark side of the season. As business and consumer activity spikes during the holidays, so do cyberattacks. And here I thought this was to be the season of goodwill.

 

Businesses rely on their customers, employees, and their data to keep themselves running. Sadly, hackers know this information as well and are likely to capitalize, especially during these times when guards are down; with an arsenal of tools to choose from malware, phishing, and DoS attacks.

 

The size of the company does not reduce risk. Your business is never too small to not be a target. In fact, one might argue SMBs are a prime target because they are easier to breach than larger corporations. These cyberattacks can ultimately destroy a business, whether it is reputationally, legally, or financially.

 

There’s an important saying that those that fail to plan should plan to fail. SMBs are adapting their IT infrastructure with the heightened risk of ransomware and exposed data vulnerabilities—and at the same time tackling challenges to manage costs. SMBs can meet these goals with a hybrid cloud approach. They want to protect their data without paying exorbitant fees for moving their data from cloud to on-prem.

 

HPE Cloud Volumes is the answer, offering an effortless, secure, and efficient solution. HPE Cloud Volumes is a suite of cloud data services that enable multi-cloud flexibility while delivering encrypted backups invisible to ransomware, eliminating costly egress fees, and providing pay-as-you-go pricing.

 

HPE Cloud Volumes is an excellent solution for security and disaster recovery. The data is encrypted in-flight and at rest for resilient protection. Your data sits separately from the operating systems of your applications, so hackers cannot access your data. In addition, disaster recovery capabilities to back up critical systems to secure cloud and restore with ease.

Last year we announced HPE Cloud Volumes Backup, a cloud backup service that allows you to start backing up to the cloud in minutes. This service is plug-and-play; it changes where – not how – you do backup, because it integrates with your favorite backup software.

Learn more about how Audio Acoustics leveraged a secure, cost-effective, cloud-based strategy with HPE Cloud Volumes.

Stay Safe, Stay Secure, and Happy Holidays,

Martin Oderinde
Hewlett Packard Enterprise

twitter.com/HPE_Servers
linkedin.com/showcase/hpe-servers-and-systems/
hpe.com/servers

]]>
https://connect-community.org/hpe-cloud-volumes-secure-cloud-ready-answer-to-holiday-cyberattacks-for-smbs/feed/ 0
Industrialize data science to attain mastery of repeatable intelligence delivery https://connect-community.org/2020-12-7-industrialize-data-science-to-attain-mastery-of-repeatable-intelligence-delivery/ https://connect-community.org/2020-12-7-industrialize-data-science-to-attain-mastery-of-repeatable-intelligence-delivery/#respond Mon, 07 Dec 2020 19:40:54 +0000 https://connect-community.org//2020-12-7-industrialize-data-science-to-attain-mastery-of-repeatable-intelligence-delivery/ Learn how data scientists make the enterprise analytics-oriented end-to-end.

]]>


top.jpg

Businesses these days are quick to declare their intention to become data-driven, yet the deployment of analytics and the use of data science remains spotty, isolated, and often uncoordinated.

To fully reach their digital business transformation potential, businesses large and small need to make data science more of a repeatable assembly line — an industrialization, if you will — of end-to-end data exploitation.

The next BriefingsDirect Voice of Analytics Innovation discussion explores the latest methods, tools, and thinking around making data science an integral core function that both responds to business needs and scales to improve every aspect of productivity.

Listen to the podcast. Find it on iTunes. Read a full transcript or download a copy.

To learn more about the ways that data and analytics behave more like a factory — and less like an Ivory Tower — please welcome Doug Cackett, EMEA Field Chief Technology Officer at Hewlett Packard Enterprise. The discussion is moderated by Dana Gardner, Principal Analyst at Interarbor Solutions.

Here are some excerpts:

Gardner: Doug, why is there a lingering gap — and really a gaping gap — between the amount of data available and the analytics that should be taking advantage of it?

Cackett: That’s such a big question to start with, Dana, to be honest. We probably need to accept that we’re not doing things the right way at the moment. Actually, Forrester suggests that something like 40 zettabytes of data are going to be under management by the end of this year, which is quite enormous.

And, significantly, more of that data is being generated at the edge through applications, Internet of Things (IoT), and all sorts of other things. This is where the customer meets your business. This is where you’re going to have to start making decisions as well.

So, the gap is two things. It’s the gap between the amount of data that’s being generated and the amount you can actually comprehend and create value from. In order to leverage that data from a business point of view, you need to make decisions at the edge.

You will need to operationalize those decisions and move that capability to the edge where your business meets your customer. That’s the challenge we’re all looking for machine learning (ML) — and the operationalization of all of those ML models into applications — to make the difference.

Gardner: Why does HPE think that moving more toward a factory model, industrializing data science, is part of the solution to compressing and removing this gap?

Data’s potential at the edge

Cackett: It’s a math problem, really, if you think about it. If there is exponential growth in data within your business, if you’re trying to optimize every step in every business process you have, then you’ll want to operationalize those insights by making your applications as smart as they can possibly be. You’ll want to embed ML into those applications.

Because, correspondingly, there’s exponential growth in the demand for analytics in your business, right? And yet, the number of data scientists you have in your organization — I mean, growing them exponentially isn’t really an option, is it? And, of course, budgets are also pretty much flat or declining.

There’s exponential growth in the demand for analytics in your business. And yet the number of data scientists in your organization, growing them, is not exponential. And budgets are pretty much flat or declining.

So, it’s a math problem because we need to somehow square away that equation. We somehow have to generate exponentially more models for more data, getting to the edge, but doing that with fewer data scientists and lower levels of budget.

Industrialization, we think, is the only way of doing that. Through industrialization, we can remove waste from the system and improve the quality and control of those models. All of those things are going to be key going forward.

Gardner: When we’re thinking about such industrialization, we shouldn’t necessarily be thinking about an assembly line of 50 years ago — where there are a lot of warm bodies lined up. I’m thinking about the Lucille Ball assembly line, where all that candy was coming down and she couldn’t keep up with it.

Perhaps we need more of an ultra-modern assembly line, where it’s a series of robots and with a few very capable people involved. Is that a fair analogy?

Industrialization of data science

Cackett: I think that’s right. Industrialization is about manufacturing where we replace manual labor with mechanical mass production. We are not talking about that. Because we’re not talking about replacing the data scientist. The data scientist is key to this. But we want to look more like a modern car plant, yes. We want to make sure that the data scientist is maximizing the value from the data science, if you like.

We don’t want to go hunting around for the right tools to use. We don’t want to wait for the production line to play catch up, or for the supply chain to catch up. In our case, of course, that’s mostly data or waiting for infrastructure or waiting for permission to do something. All of those things are a complete waste of their time.


colors.jpg

As you look at the amount of productive time data scientists spend creating value, that can be pretty small compared to their non-productive time — and that’s a concern. Part of the non-productive time, of course, has been with those data scientists having to discover a model and optimize it. Then they would do the steps to operationalize it.

But maybe doing the data and operations engineering things to operationalize the model can be much more efficiently done with another team of people who have the skills to do that. We’re talking about specialization here, really.

But there are some other learnings as well. I recently wrote a blog about it. In it, I looked at the modern Toyota production system and started to ask questions around what we could learn about what they have learned, if you like, over the last 70 years or so.

It was not just about automation, but also how they went about doing research and development, how they approached tooling, and how they did continuous improvement. We have a lot to learn in those areas.

For an awful lot of organizations that I deal with, they haven’t had a lot of experience around such operationalization problems. They haven’t built that part of their assembly line yet. Automating supply chains and mistake-proofing things, what Toyota called jidoka, also really important. It’s a really interesting area to be involved with.

Gardner: Right, this is what US manufacturing, in the bricks and mortar sense, went through back in the 1980s when they moved to business process reengineering, adopted kaizen principles, and did what Demingand more quality-emphasis had done for the Japanese auto companies.

And so, back then there was a revolution, if you will, in physical manufacturing. And now it sounds like we’re at a watershed moment in how data and analytics are processed.

Cackett: Yes, that’s exactly right. To extend that analogy a little further, I recently saw a documentary about Morgan cars in the UK. They’re a hand-built kind of car company. Quite expensive, very hand-built, and very specialized.


HPE building.jpg

And I ended up by almost throwing things at the TV because they were talking about the skills of this one individual. They only had one guy who could actually bend the metal to create the bonnet, the hood, of the car in the way that it needed to be done. And it took two or three years to train this guy, and I’m thinking, “Well, if you just automated the process, and the robot built it, you wouldn’t need to have that variability.” I mean, it’s just so annoying, right?

In the same way, with data science we’re talking about laying bricks — not Michelangelo hammering out the figure of David. What I’m really trying to say is a lot of the data science in our customer’s organizations are fairly mundane. To get that through the door, get it done and dusted, and give them time to do the other bits of finesse using more skills — that’s what we’re trying to achieve. Both [the basics and the finesse] are necessary and they can all be done on the same production line.

Gardner: Doug, if we are going to reinvent and increase the productivity generally of data science, it sounds like technology is going to be a big part of the solution. But technology can also be part of the problem.

What is it about the way that organizations are deploying technology now that needs to shift? How is HPE helping them adjust to the technology that supports a better data science approach?

Define and refine

Cackett: We can probably all agree that most of the tooling around MLOps is relatively young. The two types of company we see are either companies that haven’t yet gotten to the stage where they’re trying to operationalize more models. In other words, they don’t really understand what the problem is yet.

Forrester research suggests that only 14 percent of organizations that they surveyed said they had a robust and repeatable operationalization process. It’s clear that the other 86 percent of organizations just haven’t refined what they’re doing yet. And that’s often because it’s quite difficult.

Many of these organizations have only just linked their data science to their big data instances or their data lakes. And they’re using it both for the workloads and to develop the models. And therein lies the problem. Often they get stuck with simple things like trying to have everyone use a uniform environment. All of your data scientists are both sharing the data and sharing the computer environment as well.

Data scientists can be very destructive in what they’re doing. Maybe overwriting data, for example. To avoid that, you end up replicating terabytes of data, which can take a long time. That also demands new resources, including new hardware.

And data scientists can often be very destructive in what they’re doing. Maybe overwriting data, for example. To avoid that, you end up replicating the data. And if you’re going to replicate terabytes of data, that can take a long period of time. That also means you need new resources, maybe new more compute power and that means approvals, and it might mean new hardware, too.

Often the biggest challenge is in provisioning the environment for data scientists to work on, the data that they want, and the tools they want. That can all often lead to huge delays in the process. And, as we talked about, this is often a time-sensitive problem. You want to get through more tasks and so every delayed minute, hour, or day that you have becomes a real challenge.

The other thing that is key is that data science is very peaky. You’ll find that data scientists may need no resources or tools on Monday and Tuesday, but then they may burn every GPU you have in the building on Wednesday, Thursday, and Friday. So, managing that as a business is also really important. If you’re going to get the most out of the budget you have, and the infrastructure you have, you need to think differently about all of these things. Does that make sense, Dana?

Gardner: Yes. Doug how is HPE Ezmeral being designed to help give the data scientists more of what they need, how they need it, and that helps close the gap between the ad hoc approach and that right kind of assembly line approach?

Two assembly lines to start

Cackett: Look at it as two assembly lines, at the very minimum. That’s the way we want to look at it. And the first thing the data scientists are doing is the discovery.

The second is the MLOps processes. There will be a range of people operationalizing the models. Imagine that you’re a data scientist, Dana, and I’ve just given you a task. Let’s say there’s a high defection or churn rate from our business, and you need to investigate why.

First you want to find out more about the problem because you might have to break that problem down into a number of steps. And then, in order to do something with the data, you’re going to want an environment to work in. So, in the first step, you may simply want to define the project, determine how long you have, and develop a cost center.

You may next define the environment: Maybe you need CPUs or GPUs. Maybe you need them highly available and maybe not. So you’d select the appropriate-sized environment. You then might next go and open the tools catalog. We’re not forcing you to use a specific tool; we have a range of tools available. You select the tools you want. Maybe you’re going to use Python. I know you’re hardcore, so you’re going to code using Jupyter and Python.

And the next step, you then want to find the right data, maybe through the data catalog. So you locate the data that you want to use and you just want to push a button and get provisioned for that lot. You don’t want to have to wait months for that data. That should be provisioned straight away, right?

You can do your work, save all your work away into a virtual repository, and save the data so it’s reproducible. You can also then check the things like model drift and data drift and those sorts of things. You can save the code and model parameters and those sorts of things away. And then you can put that on the backlog for the MLOps team.

Then the MLOps team picks it up and goes through a similar data science process. They want to create their own production line now, right? And so, they’re going to seek a different set of tools. This time, they need continuous integration and continuous delivery (CICD), plus a whole bunch of data stuff they want to operationalize your model. They’re going to define the way that that model is going to be deployed. Let’s say, we’re going to use Kubeflow for that. They might decide on, say, an A/B testing process. So they’re going to configure that, do the rest of the work, and press the button again, right?

Clearly, this is an ongoing process. Fundamentally that requires workflow and automatic provisioning of the environment to eliminate wasted time, waiting for stuff to be available. It is fundamentally what we’re doing in our MLOps product.

But in the wider sense, we also have consulting teams helping customers get up to speed, define these processes, and build the skills around the tools. We can also do this as-a-service via our HPE GreenLake proposition as well. Those are the kinds of things that we’re helping customers with.

Gardner: Doug, what you’re describing as needed in data science operations is a lot like what was needed for application development with the advent of DevOps several years ago. Is there commonality between what we’re doing with the flow and nature of the process for data and analytics and what was done not too long ago with application development? Isn’t that also akin to more of a cattle approach than a pet approach?

Operationalize with agility

Cackett: Yes, I completely agree. That’s exactly what this is about and for an MLOps process. It’s exactly that. It’s analogous to the sort of CICD, DevOps, part of the IT business. But a lot of that tool chain is being taken care of by things like Kubeflow and MLflow Project, some of these newer, open source technologies.

I should say that this is all very new, the ancillary tooling that wraps around the CICD. The CICD set of tools are also pretty new. What we’re also attempting to do is allow you, as a business, to bring these new tools and on-board them so you can evaluate them and see how they might impact what you’re doing as your process settles down.

The way we’re doing MLOps and data science is progressing extremely quickly. So you don’t want to lock yourself into a corner where you’re trapped in a particular workflow. You want to have agility. It’s analogous to the DevOps movement.

The idea is to put them in a wrapper and make them available so we get a more dynamic feel to this. The way we’re doing MLOps and data science generally is progressing extremely quickly at the moment. So you don’t want to lock yourself into a corner where you’re trapped into a particular workflow. You want to be able to have agility. Yes, it’s very analogous to the DevOps movement as we seek to operationalize the ML model.

The other thing to pay attention to are the changes that need to happen to your operational applications. You’re going to have to change those so they can tool the ML model at the appropriate place, get the result back, and then render that result in whatever way is appropriate. So changes to the operational apps are also important.

Gardner: You really couldn’t operationalize ML as a process if you’re only a tools provider. You couldn’t really do it if you’re a cloud services provider alone. You couldn’t just do this if you were a professional services provider.

It seems to me that HPE is actually in a very advantageous place to allow the best-of-breed tools approach where it’s most impactful but to also start put some standard glue around this — the industrialization. How is HPE is an advantageous place to have a meaningful impact on this difficult problem?

Cackett: Hopefully, we’re in an advantageous place. As you say, it’s not just a tool, is it? Think about the breadth of decisions that you need to make in your organization, and how many of those could be optimized using some kind of ML model.

You’d understand that it’s very unlikely that it’s going to be a tool. It’s going to be a range of tools, and that range of tools is going to be changing almost constantly over the next 10 and 20 years.

This is much more to do with a platform approach because this area is relatively new. Like any other technology, when it’s new it almost inevitably to tends to be very technical in implementation. So using the early tools can be very difficult. Over time, the tools mature, with a mature UI and a well-defined process, and they become simple to use.

But at the moment, we’re way up at the other end. And so I think this is about platforms. And what we’re providing at HPE is the platform through which you can plug in these tools and integrate them together. You have the freedom to use whatever tools you want. But at the same time, you’re inheriting the back-end system. So, that’s Active Directory and Lightweight Directory Access Protocol (LDAP) integrations, and that’s linkage back to the data, your most precious asset in your business. Whether that be in a data lake or a data warehouse, in data marts or even streaming applications.

This is the melting point of the business at the moment. And HPE has had a lot of experience helping our customers deliver value through information technology investments over many years. And that’s certainly what we’re trying to do right now.

Gardner: It seems that HPE Ezmeral is moving toward industrialization of data science, as well as other essential functions. But is that where you should start, with operationalizing data science? Or is there a certain order by which this becomes more fruitful? Where do you start?

Machine learning leads change

Cackett: This is such a hard question to answer, Dana. It’s so dependent on where you are as a business and what you’re trying to achieve. Typically, to be honest, we find that the engagement is normally with some element of change in our custo
mers. That’s often, for example, where there’s a new digital transformation initiative going on. And you’ll find that the digital transformation is being held back by an inability to do the data science that’s required.

There is another Forrester report that I’m sure you’ll find interesting. It suggests that 98 percent of business leaders feel that ML is key to their competitive advantage. It’s hardly surprising then that ML is so closely related to digital transformation, right? Because that’s about the stage at which organizations are competing after all.

So we often find that that’s the starting point, yes. Why can’t we develop these models and get them into production in time to meet our digital transformation initiative? And then it becomes, “Well, what bits do we have to change? How do we transform our MLOps capability to be able to do this and do this at scale?”


data highway.jpg

Often this shift is led by an individual in an organization. There develops a momentum in an organization to make these changes. But the changes can be really small at the start, of course. You might start off with just a single ML problem related to digital transformation.

We acquired MapR some time ago, which is now our HPE Ezmeral Data Fabric. And it underpins a lot of the work that we’re doing. And so, we will often start with the data, to be honest with you, because a lot of the challenges in many of our organizations has to do with the data. And as businesses become more real-time and want to connect more closely to the edge, really that’s where the strengths of the data fabric approach come into play.

So another starting point might be the data. A new application at the edge, for example, has new, very stringent requirements for data and so we start there with building these data systems using our data fabric. And that leads to a requirement to do the analytics and brings us obviously nicely to the HPE Ezmeral MLOps, the data science proposition that we have.

Gardner: Doug, is the COVID-19 pandemic prompting people to bite the bullet and operationalize data science because they need to be fleet and agile and to do things in new ways that they couldn’t have anticipated?

Cackett: Yes, I’m sure it is. We know it’s happening; we’ve seen all the research. McKinsey has pointed out that the pandemic has accelerated a digital transformation journey. And inevitably that means more data science going forward because, as we talked about already with that Forrester research, some 98 percent think that it’s about competitive advantage. And it is, frankly. The research goes back a long way to people like Tom Davenport, of course, in his famous Harvard Business Reviewarticle. We know that customers who do more with analytics, or better analytics, outperform their peers on any measure. And ML is the next incarnation of that journey.

Gardner: Do you have any use cases of organizations that have gone to the industrialization approach to data science? What is it done for them?

Financial services benefits

Cackett: I’m afraid names are going to have to be left out. But a good example is in financial services. They have a problem in the form of many regulatory requirements.

When HPE acquired BlueData it gained an underlying technology, which we’ve transformed into our MLOps and container platform. BlueData had a long history of containerizing very difficult, problematic workloads. In this case, this particular financial services organization had a real challenge. They wanted to bring on new data scientists. But the problem is, every time they wanted to bring a new data scientist on, they had to go and acquire a bunch of new hardware, because their process required them to replicate the data and completely isolate the new data scientist from the other ones. This was their process. That’s what they had to do.

So as a result, it took them almost six months to do anything. And there’s no way that was sustainable. It was a well-defined process, but it’s still involved a six-month wait each time.

So instead we containerized their Cloudera implementation and separated the compute and storage as well. That means we could now create environments on the fly within minutes effectively. But it also means that we can take read-only snapshots of data. So, the read-only snapshot is just a set of pointers. So, it’s instantaneous.

They scaled out their data science without scaling up their costs or the number of people required. They are now doing that in a hybrid cloud environment. And they only have to change two lines of code to push workloads into AWS, which is pretty magical, right?

They were able to scale-out their data science without scaling up their costs or the number of people required. Interestingly, recently, they’ve moved that on further as well. Now doing all of that in a hybrid cloud environment. And they only have to change two lines of code to allow them to push workloads into AWS, for example, which is pretty magical, right? And that’s where they’re doing the data science.

Another good example that I can name is GM Finance, a fantastic example of how having started in one area for business — all about risk and compliance — they’ve been able to extend the value to things like credit risk.

But doing credit risk and risk in terms of insurance also means that they can look at policy pricing based on dynamic risk. For example, for auto insurance based on the way you’re driving. How about you, Dana? I drive like a complete idiot. So I couldn’t possibly afford that, right? But you, I’m sure you drive very safely.

But in this use-case, because they have the data science in place it means they can know how a car is being driven. They are able to look at the value of the car, the end of that lease period, and create more value from it.

These are types of detailed business outcomes we’re talking about. This is about giving our customers the means to do more data science. And because the data science becomes better, you’re able to do even more data science and create momentum in the organization, which means you can do increasingly more data science. It’s really a very compelling proposition.

Gardner: Doug, if I were to come to you in three years and ask similarly, “Give me the example of a company that has done this right and has really reshaped itself.” Describe what you think a correctly analytically driven company will be able to do. What is the end state?

A data-science driven future

Cackett: I can answer that in two ways. One relates to talking to an ex-colleague who worked at Facebook. And I’m so taken with what they were doing there. Basically, he said, what originally happened at Facebook, in his very words, is that to create a new product in Facebook they had an engineer and a product owner. They sat together and they created a new product.

Sometime later, they would ask a data scientist to get involved, too. That person would look at the data and tell them the results.

Then they completely changed that around. What they now do is first find the data scientist and bring him or her on board as they’re creating a product. So they’re instrumenting up what they’re doing in a way that best serves the data scientist, which is really interesting.


ezmeral.jpg

The data science is built-in from the start. If you ask me what’s going to happen in three years’ time, as we move to this democratization of ML, that’s exactly what’s going to happen. I think we’ll end up genuinely being information-driven as an organization.

That will build the data science into the products and the applications from the start, not tack them on to the end.

Gardner: And when you do that, it seems to me the payoffs are expansive — and perhaps accelerating.

Cackett: Yes. That’s the competitive advantage and differentiation we started off talking about. But the technology has to underpin that. You can’t deliver the ML without the technology; you won’t get the competitive advantage in your business, and so your digital transformation will also fail.

This is about getting the right technology with the right people in place to deliver these kinds of results.

Listen to the podcast. Find it on iTunes. Read a full transcript or download a copy. Sponsor: Hewlett Packard Enterprise.

You may also be interested in:

]]>
https://connect-community.org/2020-12-7-industrialize-data-science-to-attain-mastery-of-repeatable-intelligence-delivery/feed/ 0
How the Journey to Modern Data Management is Paved with an Inclusive Edge-to-Cloud Data Fabric https://connect-community.org/2020-11-20-how-the-journey-to-modern-data-management-is-paved-with-an-inclusive-edge-to-cloud-data-fabric/ https://connect-community.org/2020-11-20-how-the-journey-to-modern-data-management-is-paved-with-an-inclusive-edge-to-cloud-data-fabric/#respond Sun, 22 Nov 2020 18:03:32 +0000 https://connect-community.org//2020-11-20-how-the-journey-to-modern-data-management-is-paved-with-an-inclusive-edge-to-cloud-data-fabric/ Learn how edge-to-core-to-cloud dispersed data can be harmonized with a common fabric to make it accessible for use by more apps and across more analytics

]]>
The next BriefingsDirect Voice of Analytics Innovation discussion focuses on the latest insights into end-to-end data management strategies.

As businesses seek to gain insights for more elements of their physical edge — from factory sensors, myriad machinery, and across field operations — data remains fragmented. But a Data Fabric approach allows information and analytics to reside locally at the edge yet contribute to the global improvement in optimizing large-scale operations.

Stay with us now as we explore how edge-to-core-to-cloud dispersed data can be harmonized with a common fabric to make it accessible for use by more apps and across more analytics.

Listen to the podcast. Find it on iTunes. Read a full transcript or download a copy. 

To learn more about the ways all data can be managed for today’s data-rich but too often insights-poor organizations, we’re joined by Chad Smykay, Field Chief Technology Officer for Data Fabric at Hewlett Packard Enterprise (HPE). The interview is moderated by Dana Gardner, Principal Analyst at Interarbor Solutions.

Here are some excerpts:

Gardner: Chad, why are companies still flooded with data? It seems like they have the data, but they’re still thirsty for actionable insights. If you have the data, why shouldn’t you also have the insights readily available?

Smykay: There are a couple reasons for that. We still see today challenges for our customers. One is just having a common data governance methodology. That’s not just to govern the security and audits, and the techniques around that — but determining just what your data is.

I’ve gone into so many projects where they don’t even know where their data lives; just a simple matrix of where the data is, where it lives, and how it’s important to the business. This is really the first step that most companies just don’t do.

Gardner: What’s happening with managing data access when they do decide they want to find it? What’s been happening with managing the explosive growth of unstructured data from all corners of the enterprise?

Tame your data

Smykay: Five years ago, it was still the Wild West of data access. But we’re finally seeing some great standards being deployed and application programming interfaces (APIs) for that data access. Companies are now realizing there’s power in having one API to rule them all. In this case, we see mostly Amazon S3

There are some other great APIs for data access out there, but just having more standardized API access into multiple datatypes has been great for our customers. It allows for APIs to gain access across many different use cases. For example, business intelligence (BI) tools can come in via an API. Or an application developer can access the same API. So that approach really cuts down on my access methodologies, my security domains, and just how I manage that data for API access.

Gardner: And when we look to get buy-in from the very top levels of businesses, why are leaders now rethinking data management and exploitation of analytics? What are the business drivers that are helping technologists get the resources they need to improve data access and management?

Smykay: The business drivers gain when data access methods are as reusable as possible across the different use cases. It used to be that you’d have different point solutions, or different open source tools, needed to solve a business use-case. That was great for the short-term, maybe with some quarterly project or something for the year you did it in.

Gaining a common, secure access layer that can access different types of data is the biggest driver of our HPE Data Fabric. And the business drivers gain when the data access methods are as reusable as possible.

But then, down the road, say three years out, they would say, “My gosh, we have 10 different tools across the many different use cases we’re using.” It makes it really hard to standardize for the next set of use cases.

So that’s been a big business driver, gaining a common, secure access layer that can access different types of data. That’s been the biggest driver for our HPE Data Fabric. That and having common API access definitely reduces the management layer cost, as well as the security cost.

Gardner: It seems to me that such data access commonality, when you attain it, becomes a gift that keeps giving. The many different types of data often need to go from the edge to dispersed data centers and sometimes dispersed in the cloud. Doesn’t data access commonality also help solve issues about managing access across disparate architectures and deployment models?

Smykay: You just hit the nail on the head. Having commonality for that API layer really gives you the ability to deploy anywhere. When I have the same API set, it makes it very easy to go from one cloud provider, or one solution, to another. But that can also create issues in terms of where my data lives. You still have data gravity issues, for example. And if you don’t have portability of the APIs and the data, you start to see some lock-in with the either the point solution you went with or the cloud provider that’s providing that data access for you.

Gardner: Following through on the gift that keeps giving idea, what is it about the Data Fabric approach that also makes analytics easier? Does it help attain a common method for applying analytics?

Data Fabric deployment options

Smykay: There are
a couple of things there. One, it allows you to keep the data where it may need to stay. That could be for regulatory reasons or just depend on where you build and deploy the analytics models. A Data Fabric helps you to start separating out your computing and storage capabilities, but also keeps them coupled for wherever the deployment location is.

For example, a lot of our customers today have the flexibility to deploy IT resources out in the edge. That could be a small cluster or system that pre-processes data. They may typically slowly trickle all the data back to one location, a core data center or a cloud location. Having these systems at the edge gives them the benefit of both pushing information out, as well as continuing to process at the edge. They can choose to deploy as they want, and to make the data analytics solutions deployed at the core even better for reporting or modeling.

Gardner: It gets to the idea of act locally and learn globally. How is that important, and why are organizations interested in doing that?

Smykay: It’s just-in-time, right? We want everything to be faster, and that’s what this Data Fabric approach gets for you.

In the past, we’ve seen edge solutions deployed, but you weren’t processing a whole lot at the edge. You were pushing along all the data back to a central, core location — and then doing something with that data. But we don’t have the time to do that anymore.

Unless you can change the laws of physics — last time I checked, they haven’t done that yet — we’re bound by the speed of light for these networks. And so we need to keep as much data and systems as we can out locally at the edge. Yet we need to still take some of that information back to one central location so we can understand what’s happening across all the different locations. We still want to make the rearview reporting better globally for our business, as well as allow for more global model management.

Gardner: Let’s look at some of the hurdles organizations have to overcome to make use of such a Data Fabric. What is it about the way that data and information exist today that makes it hard to get the most out of it? Why is it hard to put advanced data access and management in place quickly and easily?

Track the data journey

Smykay: It’s tough for most organizations because they can’t take the wings off the airplane while flying. We get that. You have to begin by creating some new standards within your organization, whether that’s standardizing on an API set for different datatypes, multiple datatypes, a single datatype.

Then you need to standardize the deployment mechanisms within your organization for that data. With the HPE Data Fabric, we give the ability to just say, “Hey, it doesn’t matter where you deploy. We just need some x86 servers and we can help you standardize either on one API or multiple APIs.”

We now support more than 10 APIs, as well as the many different data types that these organizations may have.

We see a lot of data silos out there today with customers — and they’re getting worse. They’re now all over the place between multiple cloud providers. And there’s all the networking in the middle. I call it silo sprawl.

Typically, we see a lot of data silos still out there today with customers – and they’re getting worse. By worse, I mean they’re now all over the place between multiple cloud providers. I may use some of these cloud storage bucket systems from cloud vendor A, but I may use somebody else’s SQL databases from cloud vendor B, and those may end up having their own access methodologies and their own software development kits (SDKs).

Next you have to consider all the networking in the middle. And let’s not even bring up security and authorization to all of them. So we find that the silos still exist, but they’ve just gotten worse and they’ve just sprawled out more. I call it the silo sprawl.

Gardner: Wow. So, if we have that silo sprawl now, and that complexity is becoming a hurdle, the estimates are that we’re going to just keep getting more and more data from more and more devices. So, if you don’t get a handle on this now, you’re never going to be able to scale, right?

Smykay: Yes, absolutely. If you’re going to have diversity of your data, the right way to manage it is to make it use-case-driven. Don’t boil the ocean. That’s where we’ve seen all of our successes. Focus on a couple of different use cases to start, especially if you’re getting into newer predictive model management and using machine learning (ML) techniques.

But, you also have to look a little further out to say, “Okay, what’s next?” Right? “What’s coming?” When you go down that data engineering and data science journey, you must understand that, “Oh, I’m going to complete use case A, that’s going to lead to use case B, which means I’m going to have to go grab from other data sources to either enrich the model or create a whole other project or application for the business.”

You should create a data journey and understand where you’re going so you don’t just end up with silo sprawl.

Gardner: Another challenge for organizations is their legacy installations. When we talk about zettabytes of data coming, what is it about the legacy solutions — and even the cloud storage legacy — that organizations need to rethink to be able to scale?

Zettabytes of data coming

Smykay: It’s a very important point. Can we just have a moment of silence? Because now we’re talking about zettabytes of data. Okay, I’m in.

Some 20 years ago, we were talking about petabytes of data. We thought that was a lot of data, but if you look out to the future, we’re talking about some studies showing connected Internet of Things (IoT) devices generating this zettabytes amount of data.

If you don’t get a handle on where your data points are going to be generated, how they’re going to be stored, and how they’re going to be accessed now, this problem is just going to get worse and worse for organizations.

Look, Data Fabric is a great solution. We have it, and it can solve a ton of these problems. But as a consultant, if you don’t get ahead of these issues right now, you’re going to be under the umbrella of probably 20 different cloud solutions for the next 10 years. So, really, we need to look at the datatypes that you’re going to have to support, the access methodologies, and where those need to be located and supported for your organization.

Gardner: Chad, it wasn’t that long ago that we were talking about how to manage big data, and Hadoop was a big part of that. NoSQL and other open source databases in particular became popular. What is it about the legacy of the big data approach that also needs to be rethought?

Smykay: One common issue we often see is the tendency to go either/or. By that I mean saying, “Okay, we can do real-time analytics, but that’s a separate data deployment. Or we can do batch, rearview reporting analytics, and that’s a separate data deployment.” But one thing that our HPE Data Fabric has always been able to support is both — at the same time — and that’s still true.

So if you’re going down a big data or data lake journey — I think now the term now is a data lakehouse, that’s a new one. For these, basically I need to be able to do my real-time analytics, as well as my traditional BI reporting or rearview mirror reporting — and that’s what we’ve been doing for over 10 years. That’s probably one of the biggest limitations we have seen.

But it’s a heavy lift to get that data from one location to another, just because of the metadata layer of Hadoop. And then you had dependencies with some of these NoSQL databases out there on Hadoop, it caused some performance issues. You can only get so much performance out of those databases, which is why we have NoSQL databases just out of the box of our Data Fabric — and we’ve never run into any of those issues.

Gardner: Of course, we can’t talk about end-to-end data without thinking about end-to-end security. So, how do we think about the HPE Data Fabric approach helping when it comes to security from the edge to the core?

Secure data from edge to core

Smykay: This is near-and-dear to my heart because everyone always talks about these great solutions out there to do edge computing. But I always ask, “Well, how do you secure it? How do you authorize it? How does my application authorization happen all the way back from the edge application to the data store in the core or in the cloud somewhere?”

That’s what I call off-sprawl, where those issues just add up. If we don’t have one way to secure and manage all of our different data types, then what happens is, “Okay, well, I have this object-based system out there, and it has its own authorization techniques.” It has its own authentication techniques. By the way, it has its own way of enforcing security in terms of who has access to what, unless … I haven’t talked about monitoring, right? How do we monitor this solution?

So, now imagine doing that for each type of data that you have in your organization — whether it’s a SQL database, because that application is just a driving requirement for that, or a file-based workload, or a block-based workload. You can see where this starts to steamroll and build up to be a huge problem within an organization, and we see that all the time.

We’re seeing a ton of issues today in the security space. We’re seeing people getting hacked. It happens all the way down to the application layer, as you often have security sprawl that makes it very hard to manage all of the different systems.

And, by the way, when it comes to your application developers, that becomes the biggest annoyance for them. Why? Because when they want to go and create an application, they have to go and say, “Okay, wait. How do I access this data? Oh, it’s different. Okay. I’ll use a different key.” And then, “Oh, that’s a different authorization system. It’s a completely different way to authenticate with my app.”

I honestly think that’s why we’re seeing a ton of issues today in the security space. It’s why we’re seeing people get hacked. It happens all the way down to the application layer, as you often have this security sprawl that makes it very hard to manage all of these different systems.

Gardner: We’ve come up in this word sprawl several times now. We’re sprawling with this, we’re sprawling with that; there’s complexity and then there’s going to be even more scale demanded.

The bad news is there is quite a bit to consider when you want end-to-end data management that takes the edge into consideration and has all these other anti-sprawl requirements. The good news is a platform and standards approach with a Data Fabric forms the best, single way to satisfy these many requirements.

So let’s talk about the solutions. How does HPE Ezmeral generally — and the Ezmeral Data Fabric specifically — provide a common means to solve many of these thorny problems?

Smykay: We were just talking about security. We provide the same security domain across all deployments. That means having one web-based user interface (UI), or one REST API call, to manage all of those different datatypes.

We can be deployed across any x86 system. And having that multi-API access — we have more than 10 – allows for multi-data access. It includes everything from storing data into files and storing data in blocks. We’re soon going to be able to support blocks in our solution. And then we’ll be storing data into bit streams such as Kafka, and then into a NoSQL database as well.

Gardner: It’s important for people to understand that HPE Ezmeral is a larger family and that the Data Fabric is a subset. But the whole seems to be greater than the sum of the parts. Why is that the case? How has what HPE is doing in architecting Ezmeral been a lot more than data management?

Smykay: Whenever you have this “whole is greater than the sum of the parts,” you start reducing so many things across the chain. When we talk about deploying a solution, that includes, “How do I manage it? How do I update it? How do I monitor it?” And then back to securing it.

Honestly, there is a great report from IDC that says it best. We show a 567-percent, five-year return on investment (ROI). That’s not from us, that’s IDC talking to our customers. I don’t know of a better business value from a solution than that. The report speaks for itself, but it comes down to these paper cuts of managing a solution. When you start to have multiple paper cuts, across multiple arms, it starts to add up in an organization.

Gardner: Chad, what is it about the HPE Ezmeral portfolio and the way the Data Fabric fits in that provides a catalyst to more improvement?

All data put
to future use

Smykay: One, the HPE Data Fabric can be deployed anywhere. It can be deployed independently. We have hundreds and hundreds of customers. We have to continue supporting them on their journey of compute and storage, but today we are already shipping a solution where we can containerize the Data Fabric as a part of our HPE Ezmeral Container Platform and also provide persistent storage for your containers.

The HPE Ezmeral Container Platform comes with the Data Fabric, it’s a part of the persistent storage. That gives you full end-to-end management of the containers, not only the application APIs. That means the management and the data portability.

So, now imagine being able to ship the data by containers from your location, as it makes sense for your use case. That’s the powerful message. We have already been on the compute and storage journey; been down that road. That road is not going away. We have many customers for that, and it makes sense for many use cases. We’ve already been on the journey of separating out compute and storage. And we’re in general availability today. There are some other solutions out there that are still on a road map as far as we know, but at HPE we’re there today. Customers have this deployed. They’re going down their compute and storage separation journey with us.

Gardner: One of the things that gets me excited about the potential for Ezmeral is when you do this right, it puts you in a position to be able to do advanced analytics in ways that hadn’t been done before. Where do you see the HPE Ezmeral Data Fabric helping when it comes to broader use of analytics across global operations?

Smykay: One of our CMOs used to say it best, and which Jack Morris has said: “If it’s going to be about the data, it better be all about the data.”

When you improve automating data management across multiple deployments — managing it, monitoring it, keeping it secure — you can then focus on those actual use cases. You can focus on the data itself, right? That’s living in the HPE Data Fabric. That is the higher-level takeaway. Our users are not spending all their time and money worrying about the data lifecycle. Instead, they can now go use that data for their organizations and for future use cases.

HPE Ezmeral sets your organization up to use your data instead of worrying about your data. We are set up to start using the Data Fabric for newer use cases and separating out compute and storage, and having it run in containers. We’ve been doing that for years. The high-level takeaway is you can go focus on using your data and not worrying about your data.

Gardner: How about some of the technical ways that you’re doing this? Things like global namespaces, analytics-ready fabrics, and native multi-temperature management. Why are they important specifically for getting to where we can capitalize on those new use cases?

Smykay: Global namespaces is probably the top feature we hear back from our customers on. It allows them to gain one view of the data with the same common security model. Imagine you’re a lawyer sitting at your computer and you double-click on a Data Fabric drive, you can literally then see all of your deployments globally. That helps with discovery. That helps with bringing onboard your data engineers and data scientists. Over the years that’s been one of the biggest challenges, they spend a lot of time building up their data science and data engineering groups and on just discovering the data.

Global namespace means I’m reducing my discovery time to figure out where the data is. A lot of this analytics-ready value we’ve been supporting in the open source community for more than 10 years. There’s a ton of Apache open source projects out there, like PrestoHive, and Drill. Of course there’s also Spark-ready, and we have been supporting Spark for many years. That’s pretty much the de facto standard we’re seeing when it comes to doing any kind of real-time processing or analytics on data.

As for multi-temperature, that feature allows you to decrease your cost of your deployment, but still allows managing all your data in one location. There are a lot of different ways we do that. We use erasure coding. We can tear off to Amazon S3-compliant devices to reduce the overall cost of deployment.

These features contribute to making it still easier. You gain a common Data Fabric, common security layer, and common API layer.

Gardner: Chad, we talked about much more data at the edge, how that’s created a number of requirements, and the benefits of a comprehensive approach to data management. We talked about the HPE Data Fabric solution, what it brings, and how it works. But we’ve been talking in the abstract.

What about on the ground? Do you have any examples of organizations that have bitten off and made Data Fabric core for them? As an adopter, what do they get? What are the business outcomes?

Central view benefits businesses 

Smykay: We’ve been talking a lot about edge-to-core-to-cloud, and the one example that’s just top-of-mind is a big, tier-1 telecoms provider. This provider makes the equipment for your AT&Ts and your Vodafones. That equipment sits out on the cell towers. And they have many Data Fabric use cases, more than 30 with us. 

But the one I love most is real-time antenna tuning. They’re able to improve customer satisfaction in real time and reduce the need to physically return to hotspots on an antenna. They do it via real-time data collection on the antennas and then aggregating that across all of the different layers that they have in their deployments.

One example is real-time antennae tuning. They’re able to improve customer satisfaction in real time and reduce the need to physically return to hotspots on an antennae. They do it instead via real-time data collection and aggregating that across all of their deployments.

They gain a central view of all of the data using a modern API for the DevOps needs. They still centrally process data, but they also process it at the edge today. We replicate all of that data for them. We manage that for them and take a lot of the traditional data management tasks off the table for them, so they can focus on the use case of the best way to tune antennas.

Gardner: They have the local benefit of tuning the antenna. But what’s the global payback? Do we have a business quantitative or qualitative returns for them in doing that?

Smyk
ay:
 Yes, but they’re pretty secretive. We’ve heard that they’ve gotten a payback in the millions of dollars, but an immediate, direct payback for them is in reducing the application development spend everywhere across the layer. That reduction is because they can use the same type of API to publish that data as a stream, and then use the same API semantics to secure and manage it all. They can then take that same application, which is deployed in a container today, and easily deploy it to any remote location around the world.

Gardner: There’s that key aspect of the application portability that we’ve danced around a bit. Any other examples that demonstrate the adoption of the HPE Data Fabric and the business pay-offs?

Smykay: Another one off the top of my head is a midstream oil and gas customer in the Houston area. This one’s not so much about edge-to-core-to-cloud. This is more about consolidation of use cases.

We discussed earlier that we can support both rearview reporting analytics as well as real-time reporting use cases. And in this case, they actually have multiple use cases, up to about five or six right now. Among them, they are able to do predictive failure reports for heat exchangers. These heat exchangers are deployed regionally and they are really temperamental. You have to monitor them all the time.

But now they have a proactive model where they can do a predictive failure monitor on those heat exchangers just by checking the temperatures on the floor cameras. They bring in all real-time camera data and they can predict, “Oh, we think we’re having an issue with this heat exchanger on this time and this day.” So that decreases management cost for them.

They also gain a dynamic parts management capability for all of their inventory in their warehouses. They can deliver faster, not only on parts, but reduce their capital expenditure (CapEx) costs, too. They have gained material measurement balances. When you push oil across a pipeline, they can detect where that balance is off across the pipeline and detect where they’re losing money, because if they are not pushing oil across the pipe at x amount of psi, they’re losing money.

So they’re able to dynamically detect that and fix it along the pipe. They also have a pipeline leak detection that they have been working on, which is modeled to detect corrosion and decay.

The point is there are multiple use cases. But because they’re able to start putting those data types together and continue to build off of it, every use case gets stronger and stronger.

Gardner: It becomes a virtuous adoption cycle; the more you can use the data generally, then the more value, then the more you invest in getting a standard fabric approach, and then the more use cases pop up. It can become very powerful.

This last example also shows the intersection of operational technology (OT) and IT. Together they can start to discover high-level, end-to-end business operational efficiencies. Is that what you’re seeing?

Data science teams work together

Smykay: Yes, absolutely. A Data Fabric is kind of the Kumbaya set among these different groups. If they’re able to standardize on the IT and developer side, it makes it easier for them to talk the same language. I’ve seen this with the oil and gas customer. Now those data science and data engineering teams work hand in hand, which is where you want to get in your organization. You want those IT teams working with the teams managing your solutions today. That’s what I’m seeing. As you get a better, more common data model or fabric, you get faster and you get better management savings by having your people working better together.

Gardner: And, of course, when you’re able to do data-driven operations, procurement, logistics, and transportation you get to what we’re referring generally as digital business transformation.

Chad, how does a Data Fabric approach then contribute to the larger goal of business transformation?

Smykay: It allows organizations to work together through a common data framework. That’s been one of the biggest issues I’ve seen, when I come in and say, “Okay, we’re going to start on this use case. Where is the data?”

Depending on size of the organization, you’re talking to three to five different groups, and sometimes 10 different people, just to put a use case together. But as you create a common data access method, you see an organization where it’s easier and easier for not only your use cases, but your businesses to work together on the goal of whatever you’re trying to do and use your data for.

Listen to the podcast. Find it on iTunes. Read a full transcript or downloada copy. Sponsor:Hewlett Packard Enterprise.

You may also be interested in:

]]>
https://connect-community.org/2020-11-20-how-the-journey-to-modern-data-management-is-paved-with-an-inclusive-edge-to-cloud-data-fabric/feed/ 0
The IT intelligence foundation for digital business transformation rests on HPE InfoSight AIOps https://connect-community.org/2020-10-8-the-it-intelligence-foundation-for-digital-business-transformation-rests-on-hpe-infosight-aiops/ https://connect-community.org/2020-10-8-the-it-intelligence-foundation-for-digital-business-transformation-rests-on-hpe-infosight-aiops/#respond Thu, 08 Oct 2020 18:22:59 +0000 https://connect-community.org//2020-10-8-the-it-intelligence-foundation-for-digital-business-transformation-rests-on-hpe-infosight-aiops/ Learn about the expanding role and impact of AIOps in making IT resiliency broader and more inclusive than ever.

]]>


Top.jpg

The next BriefingsDirect podcast explores how artificial intelligence (AI) increasingly supports IT operations.

One of the most successful uses of machine learning (ML) and AI for IT efficiency has been the InfoSight technology developed at Nimble Storage, now part of Hewlett Packard Enterprise (HPE).

Initially targeting storage optimization, HPE InfoSight has emerged as a broad and inclusive capability for AIOps across an expanding array of HPE products and services.

Please welcome a Nimble Storage founder, along with a cutting-edge machine learning architect, to examine the expanding role and impact of HPE InfoSight in making IT resiliency better than ever.

Listen to the podcast. Find it on iTunes. Read a full transcript or download a copy.

To learn more about the latest IT operations solutions that help companies deliver agility and edge-to-cloud business continuity, we’re joined by Varun Mehta, Vice President and General Manager for InfoSight at HPE and founder of Nimble Storage, and David Adamson, Machine Learning Architect at HPE InfoSight. The discussion is moderated by Dana Gardner, Principal Analyst at Interarbor Solutions.

Here are some excerpts:

Gardner: Varun, what was the primary motivation for creating HPE InfoSight? What did you have in mind when you built this technology?

Mehta: Various forms of call home were already in place when we started Nimble, and that’s what we had set up to do. But then we realized that the call home data was used to do very simple actions. It was basically to look at the data one time and try and find problems that the machine was having right then. These were very obvious issues, like a crash. If you had had any kind of software crash, that’s what call home data would identify.

We found that if instead of just scanning the data one time, if we could store it in a database and actually look for problems over time in areas wider than just a single use, we could come up with something very interesting. Part of the problem until then was that a database that could store this amount of data cheaply was just not available, which is why people would just do the one-time scan.

The enabler was that a new database became available. We found that rather than just scan once, we could put everyone’s data into one place, look at it, and discover issues across the entire population. That was very powerful. And then we could do other interesting things using data science such as workload planning from all of that data. So the realization was that if the databases became available, we could do a lot more with that data.

Gardner: And by taking advantage of that large data capability and the distribution of analytics through a cloud model, did the scope and relevancy of what HPE InfoSight did exceed your expectations? How far has this now come?

Mehta: It turned out that this model was really successful. They say that, “imitation is the sincerest form of flattery.” And that was proven true, too. Our customers loved it, our competitors found out that our customers loved it, and it basically spawned an entire set of features across all of our competitors.

The reason our customers loved it — followed by our competitors — was that it gave people a much broader idea of the issues they were facing. We then found that people wanted to expand this envelope of understanding that we had created beyond just storage.

Data delivers more than a quick fix

And that led to people wanting to understand how their hypervisor was doing, for example. And so, we expanded the capability to look into that. People loved the solution and wanted us to expand the scope into far more than just storage optimization.

Gardner: David, you hear Varun describing what this was originally intended for. As a machine learning architect, how has HPE InfoSight provided you with a foundation to do increasingly more when it comes to AIOps, dependability, and reliability of platforms and systems?

The database is full of data that not only tracks everything longitudinally across the installed base, but also over time. The richness of that data gives us features we otherwise could not have conceived of. Many issues can now be automated away.

Adamson: As Varun was describing, the database is full of data that not only tracks everything longitudinally across the installed base, but also over time. The richness of that data set gives us an opportunity to come up with features that we otherwise wouldn’t have conceived of if we hadn’t been looking through the data. Also very powerful from InfoSight’s early days was the proactive nature of the IT support because so many simple issues had now been automated away.

That allowed us to spend time investigating more interesting and advanced problems, which demanded ML solutions. Once you’ve cleaned up the Pareto curve of all the simple tasks that can be automated with simple rules or SQL statements, you uncover problems that take longer to solve and require a look at time series and telemetry that’s quantitative in nature and multidimensional. That data opens up the requirement to use more sophisticated techniques in order to make actionable recommendations.

Gardner: Speaking of actionable, something that really impressed me when I first learned about HPE InfoSight, Varun, was how quickly you can take the analytics and apply them. Why has that rapid capability to dynamically impact what’s going on from the data proved so successful?

Support to succeed

Mehta: It turned out to be one of the key points of our success. I really have to compliment the deep partnership that our support organization has had with the HPE InfoSight team.

The support team right from the beginning prided themselves on providing outstanding service. Part of the proof of that was incredible Net Promoter scores (NPS), which is this independent measurement of how satisfied customers are with our products. Nimble’s NPS score was 86, which is even higher than Apple. We prided ourselves on providing a really strong support experience to the customer. 

Whenever a problem would surface, we would work with the support team. Our goal was for a customer to see a problem only once. And then we would rapidly fix that problem for every other customer. In fact, we would fix it preemptively so customers would never have to see it. So, we evolved this culture of identifying problems, creating signatures for these problems, and then running everybody’s data through the signatures so that customers would be preemptively inoculated from these problems. That’s why it became very successful.

Gardner: It hasn’t been that long since we were dealing with red light-green light types of IT support scenarios, but we’ve come a long way. We’re not all the way to fully automated, lights-out, machines running machines operations.

David, where do you think we are on that automated support spectrum? How has HPE InfoSight helped change the nature of systems’ dependability, getting closer to that point where they are more automated and more intelligent?

Adamson: The challenge with fully automated infrastructure stems from the variety of different components in the environments — and all of the interoperability among those components. If you look at just a simple IT stack, they are typically applications on top of virtual machines (VMs), on top of hosts — they may or may not have independent storage attached – and then the networking of all these components. That’s discounting all the different applications and various software components required to run them. 

There are just so many opportunities for things to break down. In that context, you need a holistic perspective to begin to realize a world in which the management of that entire unit is managed in a comprehensive way. And so we strive for observability models and services that collect all the data from all of those sources. If we can get that data in one place to look at the interoperability issues, we can follow the dependency chains.

But then you need to add intelligence on top of that, and that intelligence needs to not only understand all of the components and their dependencies, but also what kinds of exceptions can arise and what is important to the end users. 

So far, with HPE InfoSight, we go so far as to pull in all of our subject matter expertise into the models and exception-handling automation. We may not necessarily have upfront information about what the most important parts of your environment are. Instead, we can stop and let the user provide some judgment. It’s truly about messaging to the user the different alternative approaches that they can take. As we see exceptions happening, we can provide those recommendations in a clean and interpretable way, so [the end user] can bring context to bear that we don’t necessarily have ourselves.

Gardner: And the timing for these advanced IT operations services is very auspicious. Just as we’re now able to extend intelligence, we’re also at the point where we have end-to-end requirements – from the edge, to the cloud, and back to the data center.

And under such a hybrid IT approach, we are also facing a great need for general digital transformation in businesses, especially as they seek to be agile and best react to the COVID-19 pandemic. Are we able yet to apply HPE InfoSight across such a horizontal architecture problem? How far can it go?

Seeing the future: End-to-end visibility 

Mehta: Just to continue from where David started, part of our limitation so far has been from where we began. We started out in storage, and then as Nimble became part of HPE, we expanded it to compute resources. We targeted hypervisors; we are expanding it now to applications. To really fix problems, you need to have end-to-end visibility. And so that is our goal, to analyze, identify, and fix problems end-to-end.

That is one of the axis of development we’re pursuing. The other axis of development is that things are just becoming more-and-more complex. As businesses require their IT infrastructure to become highly adaptable they also need scalability, self-healing, and enhanced performance. To achieve this, there is greater-and-greater complexity. And part of that complexity has been driven by really poor utilization of resources.

Go back 20 years and we had standalone compute and storage machines that were not individually very well-utilized. Then you had virtualization come along, and virtualization gave you much higher utilization — but it added a whole layer of complexity. You had one machine, but now you could have 10 VMs in that one place.

Now, we have containers coming out, and that’s going to further increase complexity by a factor of 10. And right on the horizon, we have serverless computing, which will increase the complexity another order of magnitude.

Complexity is increasing, interconnectedness is increasing, and yet the demands on the business to stay agile, competitive, and scalable are also increasing. It’s really hard for IT administrators to stay on top of this. That’s why you need end-to-end automation.

So, the complexity is increasing, the interconnectedness is increasing, and yet the demands on businesses to stay agile and competitive and scalable are also increasing. It’s really hard for IT administrators to stay on top of this. And that’s why you need end-to-end automation and to collect all of the data to actually figure out what is going on. We have a lot of work cut out for us. 

There is another area of research, and David spends a lot of time working on this, which is you really want to avoid false positives. That is a big problem with lots of tools. They provide so many false positives that people just turn them off. Instead, we need to work through all of your data to actually say, “Hey, this is a recommendation that you really should pay attention to.” That requires a lot of technology, a lot of ML, and a lot of data science experience to separate the wheat from the chaff.

One of the things that’s happened with the COVID-19 pandemic response is the need for very quick response stats. For example, people have had to quickly set up web sites for contact tracing, reporting on the diseases, and for vaccines use. That shows an accelerated manner in how people need digital solutions — and it’s just not possible without serious automation.

Gardner: Varun just laid out the complexity and the demands for both the business and the technology. It sounds like a problem that mere mortals cannot solve. So how are we helping those mere mortals to bring AI to bear in a way that allows them to benefit – but, as Varun also pointed out, allows them to trust that technology and use it to its full potential?

Complexity requires automated assistance

Adamson: The point Varun is making is key. If you are talking about complexity, we’re well beyond the point where people could realistically expect to log-in to each machine to find, analyze, or manage exceptions that happen across this ever-growing, complex regime.

Even if you’re at a place where you have the observability solved, and you’re monitoring all of these moving parts together in one place — even then, it easily becomes overwhelming, with pages and pages of dashboards. You couldn’t employ enough people to monitor and act to spot everything that you need to be spotting.

You need to be able to trust automated exception [finding] methods to handle the scope and complexity of what people are dealing with now. So that means doing a few things.

People will often start with naïve thresholds. They create manual thresholds to give alerts to handle really critical issues, such as all the servers went down.

But there are often more subtle issues that show up that you wouldn’t necessarily have anticipated setting a threshold for. Or maybe your threshold isn’t right. It depends on context. Maybe the metrics that you’re looking at are just the raw metrics you’re pulling out of the system and aren’t even the metrics that give a reliable signal.


worker.jpg

What we see from the data science side is that a lot of these problems are multi-dimensional. There isn’t just one metric that you could set a threshold on to get a good, reliable alert. So how do you do that right?

For the problems that IT support provides to us, we apply automation and we move down the Pareto chart to solve things in priority of importance. We also turn to ML models. In some of these cases, we can train a model from the installed base and use a peer-learning approach, where we understand the correlations between problem states and indicator variables well enough so that we can identify a root cause for different customers and different issues.

Sometimes though, if the issue is rare enough, scanning the installed base isn’t going to give us a high enough signal to the noise. Then we can take some of these curated examples from support and do a semi-supervised loop. We basically say, “We have three examples that are known. We’re going to train a model on them.” Maybe it’s a few tens of thousands of data points, but it’s still in the three examples, so there’s co-correlation that we are worried about.

In that case we say: “Let me go fishing in that installed base with these examples and pull back what else gets flagged.” Then we can turn those back over to our support subject matter experts and say, “Which of these really look right?” And in that way, you can move past the fact that your starting data set of examples is very small and you can use semi-supervised training to develop a more robust model to identify the issues.

Gardner: As you are refining and improving these models, one of the benefits in being a part of HPE is to access growing data sets across entire industries, regions, and in fact the globe. So, Varun, what is the advantage of being part of HPE and extending those datasets to allow for the budding models to become even more accurate and powerful over time?

Gain a global point of view

Mehta: Being part of HPE has enabled us to leapfrog our competition. As I said, our roots are in storage, but really storage is just the foundation of where things are located in an organization. There is compute, networking, hypervisors, operating systems, and applications. With HPE, we certainly now cover the base infrastructure, which is storage followed by compute. At some point we will bring in networking. We already have hypervisor monitoring, and we are actively working on application monitoring.

HPE has allowed us to radically increase the scope of what we can look at, which also means we can radically improve the quality of the solutions we offer to our customers. And so it’s been a win-win solution, both for HPE where we can offer a lot of different insights into our products, and for our customers where we can offer them faster solutions to more kinds of problems. 

Gardner: David, anything more to offer on the depth, breadth, and scope of data as it’s helping you improve the models?

Adamson: I certainly agree with everything that Varun said. The one thing I might add is in the feedback we’ve received over time. And that is, one of the key things in making the notifications possible is getting us as close as possible to the customer experience of the applications and services running on the infrastructure.

Gaining additional measurements from the applications themselves is going to give us the ability to differentiate ourselves, to find the important exceptions to the end user, what they really want us to take action on, the events that are truly business-critical.

We’ve done a lot of work to make sure we identify what look like meaningful problems. But we’re fundamentally limited if the scope of what we measure is only at the storage or hypervisor layer. So gaining additional measurements from the applications themselves is going to give us the ability to differentiate ourselves, to find the important exceptions to the end user, what they really want to take action on. That’s critical for us — not sending people alerts they are not interested in but making sure we find the events that are truly business-critical.

Gardner: And as we think about the extensibility of the solution — extending past storage into compute, ultimately networking, and applications — there is the need to deal with the heterogeneity of architecture. So multicloud, hybrid cloud, edge-to-cloud, and many edges to cloud. Has HPE InfoSight been designed in a way to extend it across different IT topologies?

Across all architecture

Mehta: At heart, we are building a big data warehouse. You know, part of the challenge is that we’ve had this explosion in the amount of data that we can bring home. For the last 10 years, since InfoSight was first developed, the tools have gotten a lot more powerful. What we now want to do is take advantage of those tools so we can bring in more data and provide even better analytics. 

The first step is to deal with all of these use cases. Beyond that, there will probably be custom solutions. For example, you talked about edge-to-cloud. There will be locations where you have good bandwidth, such as a colocation center, and you can send back large amounts of data. But if you’re sitting as the only compute in a large retail store like a Home Depot, for example, or a McDonald’s, then the bandwidth back is going to be limited. You have to live within that and still provide effective monitoring. So I’m sure we will have to make some adjustments as we widen our scope, but the key is having a really strong foundation and that’s what we’re working on right now.

Gardner: David, anything more to offer on the extensibility across different types of architecture, of analyzing the different sources of analytics? 

Adamson: Yes, originally, when we were storage-focused and grew to the hypervisor level, we discovered some things about the way we keep our data organized. If we made it more modular, we could make it easier to write simple rules and build complex models to keep turnaround time fast. We developed some experience and so we’ve taken that and applied it in the most recent release of recommendations into our customer portal.

We’ve modularized our data model even further to help us support more use cases from environments that may or may not have specific components. Historically, we’ve relied on having Nimble Storage, they’re a hub for everything to be collected. But we can’t rely on that anymore. We want to be able to monitor environments that don’t necessarily have that particular storage device, and we may have to support various combinations of HPE products and other non-HPE applications.


HPE StoreFabric.jpg

Modularizing our data model to truly accommodate that has been something that we started along the path for and I think we’re making good strides toward.

The other piece is in terms of the data science. We’re trying to leverage longitudinal data as much as possible, but we want to make sure we have a sufficient set of meaningful ML offerings. So we’re looking at unsupervised learning capabilities that we can apply to environments for which we don’t have a critical mass of data yet, especially as we onboard monitoring for new applications. That’s been quite exciting to work on. 

Gardner: We’ve been talking a lot about the HPE InfoSight technology, but there also has to be considerations for culture. A big part of digital transformation is getting silos between people broken down. 

Is there a cultural silo between the data scientists and the IT operations people? Are we able to get the IT operations people to better understand what data science can do for them and their jobs? And perhaps, also allow the data scientists to understand the requirements of a modern, complex IT operations organization? How is it going between these two groups, and how well are they melding?

IT support and data science team up

Adamson: One of the things that Nimble did well from the get-go was have tight coupling between the IT support engineers and the data science team. The support engineers were fielding the calls from the IT operations guys. They had their fingers on the pulse of what was most important. That meant not only building features that would help our support engineers solve their escalations more quickly, but also things that we can productize for our customers to get value from directly.

Gardner: One of the great ways for people to better understand a solution approach like HPE InfoSight is through examples.  Do we have any instances that help people understand what it can do, but also the paybacks? Do we have metrics of success when it comes to employing HPE InfoSight in a complex IT operations environment? 

Mehta: One of the examples I like to refer to was fairly early in our history but had a big impact. It was at the University Hospital of Basel in Switzerland. They had installed a new version of VMware, and a few weeks afterward things started going horribly wrong with their implementation that included a Nimble Storage device. They called VMware and VMware couldn’t figure it out. Eventually they called our support team and using InfoSight, our support team was able to figure it out really quickly. The problem turned out to be a result of a new version of VMware. If there was a hold up in the networking, some sort of bottleneck in their networking infrastructure, this VMware version would try really hard to get the data through.

So instead of submitting each write once to the storage array once, it would try 64 times. Suddenly, their traffic went up by 64 times. There was a lot of pounding on the network, pounding on the storage system, and we were able to tell with our analytics that, “Hey this traffic is going up by a huge amount.” As we tracked it back, it pointed to the new version of VMware that had been loaded. We then connected with the VMware support team and worked very closely with all of our partners to identify this bug, which VMware very promptly fixed. But, as you know, it takes time for these fixes to roll out to the field.

We were able to preemptively alert other people who had the same combination of VMware on Nimble Storage and say, “Guys, you should either upgrade to this new patch that VMware has made or just be aware that you are susceptible to this problem.”

So that’s a great example of how our analytics was able to find a problem, get it fixed very quickly — quicker than any other means possible — and then prevent others from seeing the same problem.

Gardner: David, what are some of your favorite examples of demonstrating the power and versatility of HPE InfoSight?

Adamson: One that comes to mind was the first time we turned to an exception-based model that we had to train. We had been building infrastructure designed to learn across our installed base to find common resource bottlenecks and identify and rank those very well. We had that in place, but we came across a problem that support was trying to write a signature for. It was basically a drive bandwidth issue.

But we were having trouble writing a signature that would identify the issue reliably. We had to turn to an ML approach because it was fundamentally a multidimensional problem. If we looked across, we have had probably 10 to 20 different metrics that we tracked per drive per minute on each system. We needed to, from those metrics, come up with a good understanding of the probability that this was the biggest bottleneck on the system. This was not a problem we could solve by just setting a threshold.


HPE logo.png

So we had to really go in and say, “We’re going to label known examples of these situations. We’re going to build the sort of tooling to allow us to do that, and we’re going to put ourselves in a regime where we can train on these examples and initiate that semi-supervised loop.”

We actually had two to three customers that hit that specific issue. By the time we wanted to put that in place, we were able to find a few more just through modeling. But that set us up to start identifying other exceptions in the same way.

We’ve been able to redeploy that pattern now several times to several different problems and solve those issues in an automated way, so we don’t have to keep diagnosing the same known flavors of problems repeatedly in the future.

Gardner: What comes next? How will AI impact IT operations over time? Varun, why are you optimistic about the future?

Software eats the world

Mehta: I think having a machine in the loop is going to be required. As I pointed out earlier, complexity is increasing by leaps and bounds. We are going from virtualization to containers to serverless. The number of applications keeps increasing and demand on every industry keeps increasing.

Andreessen Horowitz, a famous venture capital firm once said, “Software is eating the world,” and really, it is true. Everything is becoming tied to a piece of software. The complexity of that is just huge. The only way to manage this and make sure everything keeps working is to use machines.

That’s where the challenge and opportunity is. Because there is so much to keep track of, one of the fundamental challenges is to make sure you don’t have too many false positives. You want to make sure you alert only when there is a need to alert. It is an ongoing area of research. 

There’s a big future in terms of the need for our solutions. There’s plenty of work to keep us busy to make sure we provide the appropriate solutions. So I’m really looking forward to it. 


HPE building.jpg

There’s also another axis to this. So far, people have stayed in the monitoring and analytics loop and it’s like self-driving cars. We’re not yet ready for machines to take over control of our cars. We get plenty of analytics from the machines. We have backup cameras. We have radars in front that alert us if the car in front is braking too quickly, but the cars aren’t yet driving themselves.

It’s all about analytics yet we haven’t graduated from analytics to control. I think that too is something that you can expect to see in the future of AIOps once the analytics get really good, and once the false positives go away. You will see things moving from analytics to control. So lots of really cool stuff ahead of us in this space.

Gardner: David, where do you see HPE InfoSight becoming more of a game changer and even transforming the end-to-end customer experience where people will see a dramatic improvement in how they interact with businesses?

Adamson: Our guiding light in terms of exception handling is making sure that not only are we providing ML models that have good precision and recall, but we’re making recommendations and statements in a timely manner that come only when they’re needed — regardless of the complexity.

A lot of hard work is being put into making sure we make those recommendation statements as actionable and standalone as possible. We’re building a differentiator through the fact that we maintain a focus on delivering a clean narrative, a very clear-cut, “human readable text” set of recommendations.

And that has the potential to save a lot of people a lot of time in terms of hunting, pecking, and worrying about what’s unseen and going on in their environments.

Gardner: Varun, how should enterprise IT organizations prepare now for what’s coming with AIOps and automation? What might they do to be in a better position to leverage and exploit these technologies even as they evolve?

Pick up new tools

Mehta: My advice to organizations is to buy into this. Automation is coming. Too often we see people stuck in the old ways of doing things. They could potentially save themselves a lot of time and effort by moving to more modern tools. I recommend that IT organizations make use of the new tools that are available.

HPE InfoSight is generally available for free when you buy an HPE product, sometimes with only the support contract. So make use of the resources. Look at the literature with HPE InfoSight. It is one of those tools that can be fire-and-forget, which is you turn it on and then you don’t have to worry about it anymore.

It’s the best kind of tool because we will come back to you and tell you if there’s anything you need to be aware of. So that would be the primary advice I would have, which is to get familiar with these automation tools and analytics tools and start using them.

Gardner: I’m afraid we’ll have to leave it there. We have been exploring how HPE InfoSight has emerged as a broad and inclusive capability for AIOps across an expanding array of edge-to-cloud solutions. And we’ve learned how these expanding AIOps capabilities are helping companies deliver increased agility — and even accelerated digital transformation.

So please join me in thanking our guests, Varun Mehta, Vice President and General Manager for InfoSight at HPE and a founder of Nimble Storage. Thanks so much, Varun.

Mehta: Thank you, Dana.

Gardner: And we’ve also been here with David Adamson, Machine Learning Architect at HPE. Thanks so much, David.

Adamson: Thank you. It’s been a pleasure.

Gardner: And a big thank you as well to our audience for joining this sponsored BriefingsDirect AIOps innovation discussion. I’m Dana Gardner, Principal Analyst at Interarbor Solutions, your host for this ongoing series of Hewlett Packard Enterprise-supported discussions. 

Thanks again for listening. Please pass this along to your IT community, and do come back next time.

Listen to the podcast. Find it on iTunesDownload the transcript. Sponsor: Hewlett Packard Enterprise.

A discussion on how HPE InfoSight has emerged as a broad and inclusive capability for AIOps across an expanding array of HPE products and services. Copyright Interarbor Solutions, LLC, 2005-2020. All rights reserved.

You may also be interested in:

·       Nimble Storage leverages big data and cloud to produce data performance optimization on the fly

·       How Digital Transformation Navigates Disruption to Chart a Better Course to the New Normal

·       How REI used automation to cloudify infrastructure and rapidly adjust its digital pandemic response

·       How the right data and AI deliver insights and reassurance on the path to a new normal

·       How IT modern operational services enables self-managing, self-healing, and self-optimizing

·       How HPE Pointnext Services ushers businesses to the new normal via an inclusive nine-step plan

·       As containers go mainstream, IT culture should pivot to end-to-end DevSecOps

·       AI-first approach to infrastructure design extends analytics to more high-value use cases

·       How Intility uses HPE Primera intelligent storage to move to 100 percent data uptime

·       As hybrid IT complexity ramps up, operators look to data-driven automation tools

]]>
https://connect-community.org/2020-10-8-the-it-intelligence-foundation-for-digital-business-transformation-rests-on-hpe-infosight-aiops/feed/ 0