The Futurum Group's Statement on Israel

Making AI’s Arcane Neural Networks Accessible

Making AI’s Arcane Neural Networks Accessible

Data Scientists Remain in Hot Demand, but they Will Give up More of their Core Functions this Year and Beyond to Automated Tools

We’re only a few months into the new year, but already we’re seeing signs that automated machine learning modeling, sometimes known as autoML, is rising to a new plateau of sophistication.

Specifically, it appears that a promising autoML approach known as “neural architecture search” will soon become part of data scientists’ core toolkits. This refers to tools and methodologies for automating creation of optimized architectures for convolutional, recurrent, and other neural network architectures at the heart of AI’s machine learning models.

Neural architecture search tools optimize the structure, weights, and hyperparameters of a machine learning model’s algorithmic “neurons” in order to make them more accurate, speedy, and efficient in performing data-driven inferences. This technology has only recently begun to emerge from labs devoted to basic research in AI tools and techniques. The research literature shows that neural architecture search tools have already outperformed manually designed neural nets in many AI R&D projects.

Commercialization Coming to Neural Architecture Search

Within the burgeoning autoML space, neural architecture search is showing signs of early commercialization.

At CES 2020 in Las Vegas in early January, I met with Montreal-based AI startup Deeplite. Its Lightweight Intelligence tool can automatically optimize a neural network for high-performance inferencing on a range of edge-device hardware platforms. It does this without requiring manual inputs or guidance from scarce, expensive data scientists.

To see how Deeplite’s tool accomplishes this, check out this discussion of the firm’s partnership with Taiwanese company Andes Technology. The RL (reinforcement learning) engine in Deeplite’s hardware-aware neural-architecture search engine automatically found, trained and deployed large neural network models to Andes’ RISC-V hardware. It compressed MobileNet models that were trained on a Visual Wake Words dataset from 13MB down to less than 188KB, a drop of almost 99 percent, with only a 1 percent drop in neural-net inferencing accuracy.

Amazon Launches Open Source autoML Toolkit with Neural Architecture Search

Another key milestone in the maturation of neural architecture search was Amazon’s recent launch of an open source autoML toolkit with this capability built in. Released the same week as CES, Amazon’s new AutoGluon tool enables AI developers of all skill levels to automate the optimization of new or existing models for high-performance inferencing on diverse target hardware platforms.

AutoGluon automates data preparation, model development, hyperparameter tuning, and training within the devops flow of an ML model. It can optimize existing PyTorch and MXNet ML models. It can also interface with existing AI devops pipelines via APIs to automatically tweak an existing ML model and thereby improve its performance of inferencing tasks.

Amazon currently has AutoGluon running on Linux platforms but has announced plans for MacOS and Windows support. Available from this project website or GitHub, AutoGluon can automatically generate a high-performance ML model from as few as three lines of Python code. It taps into available compute resources and uses reinforcement learning algorithms to search for the best fitting neural network architecture for its target environment.

AutoGluon uses RL to speed automated neural architecture searches using computing resources efficiently. Indeed, RL—as implemented both in AutoGluon and in Deeplite’s solution—is proving to be the most fruitful approach for recent advances in this area, using agent-centric actions and rewards to search the space of optimal neural architectures based on estimates of the performance of trained architectures on unseen data. If you truly want to get into the weeds of how AutoGluon works, check out this link.

RL is an up-and-coming alternative to evolutionary algorithms, which have been central to neural architecture search since the 1990s in AI R&D environments. Evolutionary algorithms are still widely used in lab environments such as OpenAI, Uber Labs, Sentient Labs (now Evolv), DeepMind, and Google Brain.

Still Early for Neural Architecture Search in Mainstream AI Devops

As autoML data science platforms become prevalent in the enterprise world, neural architecture search tools such as these will be a standard component. However, this capability is still scarcely evident in most AI devops environments.

By the end of this year, I predict that more than half of commercial and open source AI devops workbenches will add neural architecture search as an integrated feature. As autoML gains adoption and improves, it will boost data scientists’ productivity by guiding their decisions regarding whether to build their models on established machine learning algorithms, such as linear regression and random forest algorithms, or on any of the newer, more advanced neural-network algorithms.

As the decade proceeds, neural architecture search will reduce the need for data scientists to understand the neural-net guts of their ML models. This emerging approach will democratize AI by freeing developers to evolve their skillset away from tweaking arcane algorithms and toward developing powerfully predictive intelligent apps.

Futurum Research provides industry research and analysis. These columns are for educational purposes only and should not be considered in any way investment advice.

The original version of this article was first published on InfoWorld.

Image Credit: KTSimage / Getty Images

Author Information

James has held analyst and consulting positions at SiliconANGLE/Wikibon, Forrester Research, Current Analysis and the Burton Group. He is an industry veteran, having held marketing and product management positions at IBM, Exostar, and LCC. He is a widely published business technology author, has published several books on enterprise technology, and contributes regularly to InformationWeek, InfoWorld, Datanami, Dataversity, and other publications.


Latest Insights:

In this episode of Infrastructure Matters – Insider Edition, Krista Macomber and Randy Kerns are joined by Jim McGann, Index Engines’ Vice President of Strategic Partnerships, for a conversation focusing on the topic of the convergence of cyber security and storage.
New Cisco App Attention Index Indicates Elevated Digital Experience Expectations
Sherril Hanson, Senior Analyst at The Futurum Group, breaks down research results from Cisco focusing on digital and application experience and increased consumer expectations in this area.
An Assessment of Key Developments Across the 5G-IoT Ecosystem Including Netcracker Telco GenAI Debut, Amdocs NVIDIA Alliance, and SKT DT Telco-specific LLM
The Futurum Group’s Ron Westfall and Todd R Weiss review key telco GenAI developments such as Netcracker’s introduction of its GenAI Telco solution, the Amdocs NVIDIA GenAI partnership, and SKT DT jointly developing multilingual LLM aimed at global telcos.
Fourth Quarter Earnings for HP Indicate Continuing Sequential Growth in the Rebounding PC Market
Olivier Blanchard, Research Director at The Futurum Group, shares his insights on the HP Q4 2023 earnings, which suggest a rebound for PCs in 2024 despite persistent market headwinds.