There is no question that the enterprises that successfully leverage AI’s potential will be the ones to get ahead. While businesses are more and more certain about the goals they want to use AI for, how to execute and deploy it successfully remains a difficult question.Read More
Anyone with any doubts about the interest in AI and its use across enterprise technologies only needs to look at the example of the Intelligent Document Processing (IDP) market and the kind of verticals that are investing in it to quash those doubts. According to the Everest Group’s recently published report, Intelligent Document Processing (IDP) State of the Market Report 2021 (purchase required) the market for this segment alone is estimated at $700-750 million in 2020 and expected to grow at a rate of 55-65% over the next year.Read More
The key to operationalizing models is the ability to address the critical challenges centered on governance and scale required to effectively unlock the transformational value of enterprise AI and machine learning investments. As a result, there is a lot of buzz around new “ModelOps” platforms designed to help manage the many models floating around enterprises these days.Read More
A Gentle Guide to the complexities of model deployment, and integrating with the enterprise application and data pipeline. What the Data Scientist, Data Engineer, ML Engineer, and ML Ops do, in Plain English.Read More
What if companies could build more accurate ML models by training data across disparate clouds and multiple data platforms?
Amid market uncertainty and volatility, the ability for an enterprise to both scale and improve the performance of its machine learning (ML) models has become a vital endeavor. The trouble is that for many enterprises, the very architecture upon which their ML models run functions not as an ally but as an enemy of performance and availability.
Growing in tandem with the cloud-native architectures, the notion of integrating across disparate cloud platforms has certainly grown in recent years. This has been doubly true for the data, analytics, and AI enterprise community, as businesses work to make the most out of their existing technology investments in order to optimize operations in light of the current market uncertainties. Such integration, unfortunately, often entails investing in software that facilitates the use of on-premises, hybrid, edge, and multi-cloud deployments to most effectively and economically move and process data.
Stringing together this kind of loosely coupled architecture may indeed cut costs and engender a sense of agility and flexibility. But imagine moving ML model data at speed and at scale across this kind of landscape. Certainly, it can be done, but often asynchronously. With use cases like anomaly detection, the real-time accuracy of a model depends upon real-time access to live data. Take for example, organizations operating in highly regulated markets. For financial institutions, even a highly unified on-premises data architecture can stand in the way of progress, since financial regulations can limit the very mobility of data itself.
But what if companies could more rapidly improve the accuracy of their ML models by training data across multiple clouds or across multiple data sovereignty regions/networks, all without compromising data privacy or security?
First discussed in a paper published by Google in 2016, federated learning allows data scientists to architect a series of local data models that collaboratively learn from a shared model, while keeping all of the training data locally. Google went on to put this idea to work in 2017, equipping all Android mobile devices with a keyboard (Gboard) tool that could constantly improve type-ahead predictions, using data from all participating devices without having to actually move data from each device to the cloud. This notion provided Google (and Android users) with a number of important benefits:
More accurate and robust predictive modelling for personalized (per device) type-ahead recommendations
Improved data privacy, capitalizing on the data minimization principle, which is mentioned in General Data Protection Regulation (EU GDPR)
Increased performance with lower latency and less power consumption, which leads to an improved cost model
Unfortunately, even with this early success, federated learning has not yet found its way into mainstream enterprise solutions. But that is slowly changing, as vendors are looking for ways to productize this unique architecture. In 2019, Google launched TensorFlow Federated (TFF), an open source framework, which Google hopes will encourage experimentation among the broader industry. And it is likely that the company will seek to build a more operationalized or perhaps even productized application of this framework in the future.
But Google is not the only vendor investigating federated learning. Given IBM’s storied history as an early innovator across a number of potentially impactful but somewhat unpredictable ideas such as blockchain and quantum computing, it’s no surprise that the company would itself investigate the notion of federated learning.
Tucked away among an array of new features within the company’s recently updated Cloud Pak for Data (now at version 3.5), IBM slipped in a Tech Preview of federated learning as a package of Python algorithms. The basic idea is to equip companies with the ability to train and refine AI models without having to centralize the data supporting those models. But IBM doesn’t intend for its customers to put its implementation of federated learning to work on isolated use cases like type-ahead predictions. Rather, the company hopes enterprises will see federated learning as an opportunity to drive better AI outcomes through better model training without being impeded by data silos and enjoin collective parties to participate without having to worry about the privacy, compliance, security, or performance concerns mentioned above.
IBM’s approach has three important and interrelated consequences. First, it allows data scientists and AI developers to make use of a lot of data without having to maintain stable, performant data pipelines between the source and the model. Second, it enables data scientists to make use of source data without having to worry about creating and maintaining data privacy and security. Third, enterprises can save costs and risks of maintaining model training performance, obviating the need to re-architect or move data.
Does this mean IBM customers can simply port existing projects to a federated learning environment? Certainly not. Given the many complexities of federated learning itself, companies should carefully evaluate the value of adapting existing projects to leverage this new capability. For example, customers must carefully map out how they wish to share data, tune hyper-parameters, architect privacy models, etc. in order to most effectively utilize federated learning. Furthermore, because federated learning is itself a very immature technology, there are many unanswered questions that must be addressed by all community players such as how to overcome challenges in data heterogeneity, client model bias, inference attack prevention, etc.
The best strategy will include some joint experimentation between customers and technology partners that are looking to build bespoke implementations for specific business outcomes. To that end, IBM has a great start, as there are already 50 systems integrator and technology partners developing industry-specific content that’s sold as a part of Cloud Pak for Data. Even horizontal database, data management, and data science partners like Anaconda, MongoDB, and NetApp, respectively, may find new routes to market in directly supporting verticalized implementations.
Regardless of the route to market for federated learning, these early efforts from market leaders Google and IBM bode well for enterprise AI practitioners. While it is unlikely that federated learning will ever show up as a one-click, deployable service option, the fact that it’s now available as both a framework (from Google) and a set of libraries (from IBM) prove that there is a light at the end of the tunnel for companies struggling to solve tough architectural AI challenges.
Note: this post was taken from a post that originally appeared within the Omdia VisionAIres community.
Since skills, knowledge and experience are vital to a successful business and the pace in which it innovates, retaining existing institutional knowledge is an increasing priority. How can you guarantee that your company’s know-how won’t just walk out the door and jeopardise your brand and positioning? The short answer is: You can’t. But there are ways that utilising a combination of analytics, IoT, and AI techniques, along with corporate training and knowledge replacement strategies, can help.Read More
As an enterprise leader, you are busier than ever, ramping up your sales and marketing machinery to reach the right customers at the right times — with the right messages and offers. Despite the headwinds from COVID-19, the central business tenets of customer acquisition, retention, and expansion (up-sell /cross-sell) remain unchanged. What has changed is their priorities for the time being.Read More