Mike Leone

Mike Leone

ESG Senior Analyst Mike Leone covers converged and hyperconverged infrastructure, data platforms, data management, data analytics, and the emerging blockchain segment. Mike joined ESG as part of the validation team, where he provided third-party validation of claims made by vendors. In his analyst role, Mike draws upon his enthusiasm for bleeding edge technology as well as his engineering and marketing backgrounds to help IT vendors improve everything from product development and marketing to go-to-market strategies and roadmap adjustments.

Recent Posts by Mike Leone:

Things to Think About on the Day of Snowflake's IPO

The day has finally arrived. Today, Snowflake will IPO under the ticker SNOW. Many knew this was coming, even before Frank Slootman was brought in as CEO. But I don’t think it was expected to have as much fanfare as we’re seeing. Buffet (Berkshire Hathaway) is on board. Benioff (Salesforce) is on board. It’s being called the hottest tech IPO of the year. So, what’s the draw? Is it the differentiable technology? Not really. The number of customers? Nope. Customer growth? Impressive, but no. What about the total market size? Getting there. What about the value propositions it offers customers? Well yes, but I would argue there are several competitors that offer similar value propositions. So, what is it? The opportunity. The opportunity is simply massive.

Topics: Data Platforms, Analytics, & AI

6 Reasons BigQuery Omni Will Change the Game for Google Cloud

A couple weeks back, Google Cloud’s multi-week virtual event Next 20: OnAir started. There were a number of announcements, but the biggest was BigQuery Omni. By combining BigQuery and Anthos, BigQuery Omni enables organizations to embrace multi-cloud analytics by cost-effectively bringing Google Cloud's data warehouse to where the data resides across public cloud environments.

Topics: Data Platforms, Analytics, & AI

From Data Warehouse to Data Lake to Data Lakehouse

First came the traditional enterprise data warehouse (EDW). Structured data is integrated into an EDW from external data sources using ETLs (check out my recent blog post on this). The data can then be queried by end-users for BI and reporting. EDWs were purpose built for BI and reporting. But with the growing desire to incorporate more data, of different types, from different sources, of different change rates, the traditional EDW has fallen short. It does not support unstructured data (i.e., video, audio, unstructured text, etc.), streaming is for the most part out of the question, there is no data science or machine learning that can be done directly on the data, and because of their closed/proprietary nature, costs quickly skyrocket as organizations scale their deployments. Modern, cloud-based EDWs have looked to address several of these challenges and done a good job of it, but some challenges still remain, with the obvious being lack of unstructured data support.

Topics: Data Platforms, Analytics, & AI

7 Features of an Ultimate Data Integration Tool

Data integration is hard. Over the years, of all the technologies and processes that are part of an organization’s analytics stack/lifecycle, data integration continuously has been cited as a challenge. In fact, according to recent ESG research, more than 1 in 3 (36%) organizations say data integration is one of their top challenges with data analytics processes and technologies. The data silo problem is very real, but it’s about so much more than having data in a bunch of locations and needing to consolidate. It’s becoming more about the need to merge data of different types and change rates; the need to leverage metadata to understand where the data came from, who owns it, and how it’s relevant to the business; the need to properly govern data as more folks ask for access; and the need to ensure trust in data because if there isn’t trust in the data, how can you trust the outcomes derived from it?

Topics: Data Platforms, Analytics, & AI

IBM Drops Facial Recognition, AWS Follows Suit

The recent announcement from IBM to withdraw from all research, development, and offerings of facial recognition will not stop facial recognition from being used by law enforcement or government entities. There. I said it. Facial recognition will continue on its gray area trajectory with or without IBM. But what IBM has done, specifically Arvind Krishna, is bringing attention to a growing concern that needs far more national and global attention. The use of facial recognition needs to be scrutinized for bias and privacy concerns. It needs oversight. It needs guardrails. Usage, especially from law enforcement and governing entities, needs to be transparent. And frankly, the technology needs to be better for it to work the way people envision.

Topics: Data Platforms, Analytics, & AI Artificial Intelligence

The (Robotic Process) Automation Renaissance

Over the last several months, automation has seen a jump in interest. Operational efficiency has been a top priority for years, but as of late, it’s an even greater priority. For businesses, tasks or processes that used to be viewed as manageable but inefficient are now being scrutinized. The inefficiency aspect is being amplified and organizations don’t have a choice but to act. And one of those actions is to look into a trendy buzzword that is proving to be so much more: robotic process automation (RPA).

Topics: Data Platforms, Analytics, & AI

Vaccines: Data's Fight Against COVID-19

I’ve talked through a couple ways that data is helping fight COVID-19, from detecting and tracking an outbreak, to detecting within people. In this blog, I’ll be focused on research and elimination. While the world sets its sights on a hopeful slowdown similar to what is experienced with a seasonal flu, a true vaccination or cure is the ultimate goal. With more and more researchers throwing their collective hats in the ring, data sharing and collaboration is becoming key. What new information is available on the virus makeup? How is it evolving/mutating? Where are we making progress on vaccination development? What approaches have a higher likelihood of success? How can progress be shared across the globe to spur new ideas or rapid insight? All of these questions tie back to data, and using data science and AI to help answer them is rapidly becoming a go-to approach.

Topics: Data Platforms, Analytics, & AI COVID-19 Tech Effect

Detection: Data’s Fight Against COVID-19

My first blog in this series focused on using data, AI, and visualization to help detect and track the continued outbreak of COVID-19. The next topic in this blog series is focused on a different form of detection: detecting the virus in people. And it starts with understanding symptoms and ends with techniques to identify infected individuals at scale, which eventually helps shape individual treatment plans and resource allocation based on established virus hotspots.

Topics: COVID-19 Tech Effect

Outbreak: Data's Fight Against COVID-19

While the world continues its fight against COVID-19, data is becoming one of the most prominent weapons for humans. Whether tracking an outbreak, detecting the virus on a case by case basis, preventing further spread, or eventually eliminating the virus altogether, data is fueling decision making from governing bodies to personal households. As we evolve to deal with the difficult norms of what today brings, the technology community is rising to answer a desperate call to arms. The value of trusted data has never been higher as actions are being taken based purely on statistical models and predictions. The need for collaboration on a global scale is paramount. Being able to rapidly respond to new information is now a matter of life or death. Over the next few weeks I’ll be revisiting and highlighting some of the ways data is being used in the ongoing fight against this historical and scary virus.

Topics: COVID-19 Tech Effect

The CCPA is here. What does it mean for AI?

As of January 1st, the California Consumer Privacy Act is now in effect. The CCPA lets anyone in California request all the information a company has on them as a consumer, including what data has been sold to /accessed by other companies. And when it comes to penalties, if a company is notified of being out of compliance (i.e., unable to provide all the data of their consumers), they have 30 days to comply or they will get fined per record. And that “per record” component is important because it highlights how quickly a fine could balloon into billions of dollars in fines. The interesting component of this is that if a company doesn’t comply, it opens companies to face class action lawsuits from consumers.

Topics: Data Platforms, Analytics, & AI