Re thinking Efficiency in Deep Learning under Accelerators and commodity and pro...
Re thinking Efficiency in Deep Learning under Accelerators and commodity and processors
In just a few short years, breakthroughs from the field of deep learning have transformed how computers perform a wide-variety of tasks such as recognizing a face, driving a car or translating a language. Not only has deep learnin...
ver más
¿Tienes un proyecto y buscas un partner? Gracias a nuestro motor inteligente podemos recomendarte los mejores socios y ponerte en contacto con ellos. Te lo explicamos en este video
Proyectos interesantes
MANOLO
Trustworthy Efficient AI for Cloud-Edge Computing
9M€
Cerrado
REAL
Reliable and cost effective large scale machine learning
1M€
Cerrado
A-B-C-Deep
Algorithmic Bias Control in Deep learning
2M€
Cerrado
TEC2015-69266-P
TECNOLOGIAS DE APRENDIZAJE PROFUNDO APLICADAS AL PROCESADO D...
247K€
Cerrado
Fecha límite de participación
Sin fecha límite de participación.
Descripción del proyecto
In just a few short years, breakthroughs from the field of deep learning have transformed how computers perform a wide-variety of tasks such as recognizing a face, driving a car or translating a language. Not only has deep learning become an everyday tool, it is also the most promising direction for tackling a number of still open problems in machine learning and artificial intelligence. However, routine deep learning activities (such as training a model) exert severe resource demands (e.g., memory, compute, energy) that are currently slowing the advancement of the field, and preventing full global participation in this research to only the largest of companies.
The goal of REDIAL is to solve core technical challenges that span the areas of machine learning and system research which collectively can enable a radical jump in the efficiency of deep learning. It aims to address both the challenge of high training costs and time, as well as the barrier to deploying models on constrained devices (like wearables, sensors) that currently require new efficiency techniques to be invented each time a deep learning innovation occurs. To accomplish this REDIAL takes two complementary approaches. First, it seeks to build a theoretical understanding of current approaches to deep learning efficiency, a desperately needed step given current over reliance on empirical observations. Second, it aims to develop new architectures and methods for training and inference that tackle core efficiency bottlenecks, such as: dependencies preventing parallelization and excessive on-chip data movement; while also opening new opportunities including the greater adoption of analog processing within accelerators. REDIAL aims to change the way the world trains its models, and deploys them to constrained devices, by producing a series of new deep architectures and algorithms with properties that promote high efficiency that can serve as a foundation for new machine learning innovation.