AIMultiple ResearchAIMultiple Research

Future of Deep Learning according to top AI Experts of 2024

Deep learning is currently the most effective AI technology for numerous applications. However, there is still differing opinions on how capable deep learning can become. While deep learning researchers like Geoffrey Hinton believe 1  that all problems could be solved with deep learning, there are numerous scientists who point to flaws in deep learning where remedies are not clear. 2 

With increasing interest in deep learning from the general public as well as developer and research communities, there could be breakthroughs in the field. Experts such as recent Turing prize winners expect such breakthroughs come from areas such as capsule networks, deep reinforcement learning and other approaches that complement deep learning’s current limitations. For detailed answers:

What is the level of interest in deep learning?

General public

Interest in deep learning is continuing to increase. Reasons of this interest include deep learning’s capacity to

  • Improve accuracy of predictions, enabling improved data driven decisions
  • Learn from unstructured and unlabelled datasets, enable analysis of unstructured data

As a result of these, deep learning solutions provide operational and financial benefits to companies. In 2012, later Turing award recipient George Hinton’s team 3  demonstrated that deep learning could provide significant accuracy benefits in common AI tasks like image recognition. 4  After this, companies started investing into deep learning and interest in the area has exploded. Since 2017, interest in deep learning appears stable.

Number of times a phrase is searched on a search engine is a proxy for its popularity. You can see the frequency with which “deep learning” was searched on Google below.

 Google search interest for Deep Learning since 2015 has been increasing.
Figure 1: Google search interest for Deep Learning since 2015 has been increasing. Source: Google Trends

Research community

Number of deep learning publications on arXiv has increased almost 6 times in the last five years according to AI Index which provides globally sourced data to develop AI applications, ArXiv is an open-access platform for scientific articles in physics, mathematics, computer science etc. It includes both peer-reviewed and non-peer-reviewed articles.

Publications on deep learning has drastically  increased, meaning that there will be more publications in the future of deep learning.
Figure 2: Publications on deep learning has drastically increased. Source: AI Index

Developer community

TensorFlow and Keras are the most popular open source libraries for deep learning. Other popular libraries are PyTorch, Sckit-learn, BVL/caffe, MXNet and Microsoft Cognitive Toolkit (CNTK). These open source platforms help developers easily build deep learning models. As can be seen below, PyTorch, released by Facebook in 2016, is also rapidly growing in popularity.

Github most favored open source libraries since 2014 shows that Tensor flow has expanded the gap against other platforms exponentially.
Figure 3: Github most favored open source libraries since 2014, Source: AI Index

Open source libraries for deep learning are generally written in JavaScript, Python,  C++  and Scala.

What are the technologies that can shape deep learning?

Deep learning is a rapidly growing domain in AI. Due to its challenges about size and diversity of data, AI experts like Geoffrey Hinton, Yoshua Bengio, Yann LeCun who received 5 the Turing prize for their work on deep learning and Gary Marcus suggest new methods to improve deep learning solutions. These methods include introducing reasoning or prior knowledge to deep learning, self-supervised learning, capsule networks, etc.

Introduction of non-learning based AI approaches to deep learning

Gary Marcus, one of the pioneers in deep learning, highlights that deep learning techniques are data hungry, shallow, brittle, and limited in their ability to generalize

  • Gary Marcus states four possibilities for future of deep learning:
    • Unsupervised learning: If systems can determine their own objectives, do reasoning and problem-solving at a more abstract level, great improvements could be achieved
    • Symbol-manipulation & the need for hybrid models: Integration of deep learning with symbolic systems, which excel at inference and abstraction could provide better results
    • More insight from cognitive and developmental psychology: Better understanding the innate machinery in humans minds, gaining common sense knowledge and human understanding of narrative could be valuable for developing learning models
    • Bolder challenges: Generalized artificial intelligence could be multi-dimensional like natural intelligence to deal with the complexity of the world
  • He proposes a four-step program:
    • Hybrid neuro-symbolic architectures: Gary claims that we should embrace other AI approaches such as prior knowledge, reasoning, and rich cognitive models along with deep learning for transformational change
    • Construction of rich, partly-innate cognitive frameworks and large-scale knowledge databases
    • Tools for abstract reasoning for effective generalization
    • Mechanisms for the representation and induction of cognitive models
The schema summarizes the vendors with given capabilities.
Figure 4: Vendor diagram for systems, Source:ArXiv

For more on Gary Marcus’ ideas, feel free to read his articles: Deep Learning: A Critical Appraisal from 2018 and The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence from 2020.

Capsule networks

Capsule networks (CapsNets) is a new deep neural network architecture introduced by Geoffrey Hinton and his team in 2017. Capsules work with vectors and make calculations on the inputs. They encapsulate their results into a vector. So, when the orientation of the image is changed, the vector is moved. Geoffrey Hinton thinks that the approach of CNNs for object recognition is very different from human perception. CNNs need to be improved for dealing with some problems like rotation and scaling, and capsule networks can help to generalize better in deep learning architecture.

Deep reinforcement learning algorithms 

Deep reinforcement learning is a combination of reinforcement learning and deep learning. Reinforcement learning normally works on structured data. On the other hand, deep reinforcement learning makes decisions about optimizing an objective based on unstructured data.

Deep reinforcement learning models can learn to maximize cumulative reward. It is good for target optimization actions, such as complicated control problems. Yann LeCun thinks that reinforcement learning is good for simulations but it needs lots of trials and provides weak feedback. However, reinforcement learning models do not require large data sets compared to other supervised models.

Few-shot learning (FLS)

The advantage of few-shot learning (FSL) which is a subfield of machine learning is being able to work with a small amount of training data. Few-shot learning algorithms are useful to handle with data shortage and computational costs. Especially, few-shot learning models can be beneficial in healthcare to detect rare diseases with inadequate images into the training data. Few-shot learning models have potential to strengthen deep learning models with new researches and developments.

GAN-based data augmentation

Generative adversarial networks (GANs) are popular in data augmentation applications and they can create meaningful new data by using unlabelled original data. They work in these steps:

A study about insect pest classification shows that GAN-based augmentation method can help CNNs

  • perform better compared to classic augmentation method
  • reduce data collection needs.

Self-Supervised learning

According to Yann LeCun, self-supervised learning models would be a key component of deep learning models. Understanding how people learn quickly could allow to utilize full potential of self-supervised learning and reduce deep learning’s reliance on large, annotated training data sets. Self-supervised learning models can work without labeled data and make predictions if they have quality data and inputs of possible scenarios.

Other approaches

  • Imitation learning: If there are few rewards in reinforcement learning models, imitation learning is used as an alternative method. The agent can learn performing a task by imitating supervisor’s demonstrations including observations and actions. It is also called as Learning from Demonstration or Apprenticeship Learning.
  • Physics guided/informed machine learning: Physics laws are integrated into training process to induce interpretability and improve accuracy of predictions in deep learning models.
  • Transfer learning which is used to help machines transfer knowledge from one domain to another
  • Others: Motor learning and brain areas like cortical and subcortical neural circuits may be new fields of inspiration for machine learning models.

If you want to read more about deep learning, check our article on deep learning use cases.

If you are ready to use deep learning in your firm, we prepared a data driven list of companies offering deep learning platforms.

If you need help in choosing among deep learning vendors who can help you get started, let us know:

Find the Right Vendors

For more, you can watch 3 AI experts share their views during AAAI 20:

This article was drafted by former AIMultiple industry analyst Ayşegül Takımoğlu.

Access Cem's 2 decades of B2B tech experience as a tech consultant, enterprise leader, startup entrepreneur & industry analyst. Leverage insights informing top Fortune 500 every month.
Cem Dilmegani
Principal Analyst
Follow on

Cem Dilmegani
Principal Analyst

Cem has been the principal analyst at AIMultiple since 2017. AIMultiple informs hundreds of thousands of businesses (as per similarWeb) including 60% of Fortune 500 every month.

Cem's work has been cited by leading global publications including Business Insider, Forbes, Washington Post, global firms like Deloitte, HPE, NGOs like World Economic Forum and supranational organizations like European Commission. You can see more reputable companies and media that referenced AIMultiple.

Throughout his career, Cem served as a tech consultant, tech buyer and tech entrepreneur. He advised businesses on their enterprise software, automation, cloud, AI / ML and other technology related decisions at McKinsey & Company and Altman Solon for more than a decade. He also published a McKinsey report on digitalization.

He led technology strategy and procurement of a telco while reporting to the CEO. He has also led commercial growth of deep tech company Hypatos that reached a 7 digit annual recurring revenue and a 9 digit valuation from 0 within 2 years. Cem's work in Hypatos was covered by leading technology publications like TechCrunch and Business Insider.

Cem regularly speaks at international technology conferences. He graduated from Bogazici University as a computer engineer and holds an MBA from Columbia Business School.

To stay up-to-date on B2B tech & accelerate your enterprise:

Follow on

Next to Read

Comments

Your email address will not be published. All fields are required.

0 Comments