Opendata, web and dolomites

BigFastData SIGNED

Charting a New Horizon of Big and Fast Data Analysis through Integrated Algorithm Design

Total Cost €


EC-Contrib. €






Project "BigFastData" data sheet

The following table provides information about the project.


Organization address
postcode: 91128

contact info
title: n.a.
name: n.a.
surname: n.a.
function: n.a.
email: n.a.
telephone: n.a.
fax: n.a.

 Coordinator Country France [FR]
 Total cost 2˙472˙752 €
 EC max contribution 2˙472˙752 € (100%)
 Programme 1. H2020-EU.1.1. (EXCELLENT SCIENCE - European Research Council (ERC))
 Code Call ERC-2016-COG
 Funding Scheme ERC-COG
 Starting year 2017
 Duration (year-month-day) from 2017-09-01   to  2022-08-31


Take a look of project's partnership.

# participants  country  role  EC contrib. [€] 
1    ECOLE POLYTECHNIQUE FR (PALAISEAU CEDEX) coordinator 2˙472˙752.00


 Project objective

This proposal addresses a pressing need from emerging big data applications such as genomics and data center monitoring: besides the scale of processing, big data systems must also enable perpetual, low-latency processing for a broad set of analytical tasks, referred to as big and fast data analysis. Today’s technology falls severely short for such needs due to the lack of support of complex analytics with scale, low latency, and strong guarantees of user performance requirements. To bridge the gap, this proposal tackles a grand challenge: “How do we design an algorithmic foundation that enables the development of all necessary pillars of big and fast data analysis?” This proposal considers three pillars: 1) Parallelism: There is a fundamental tension between data parallelism (for scale) and pipeline parallelism (for low latency). We propose new approaches based on intelligent use of memory and workload properties to integrate both forms of parallelism. 2) Analytics: The literature lacks a large body of algorithms for critical order-related analytics to be run under data and pipeline parallelism. We propose new algorithmic frameworks to enable such analytics. 3) Optimization: To run analytics, today's big data systems are best effort only. We transform such systems into a principled optimization framework that suits the new characteristics of big data infrastructure and adapts to meet user performance requirements.

The scale and complexity of the proposed algorithm design makes this project high-risk, at the same time, high-gain: it will lay a solid foundation for big and fast data analysis, enabling a new integrated parallel processing paradigm, algorithms for critical order-related analytics, and a principled optimizer with strong performance guarantees. It will also broadly enable accelerated information discovery in emerging domains such as genomics, as well as economic benefits of early, well-informed decisions and reduced user payments.

Are you the coordinator (or a participant) of this project? Plaese send me more information about the "BIGFASTDATA" project.

For instance: the website url (it has not provided by EU-opendata yet), the logo, a more detailed description of the project (in plain text as a rtf file or a word file), some pictures (as picture files, not embedded into any word file), twitter account, linkedin page, etc.

Send me an  email ( and I put them in your project's page as son as possible.

Thanks. And then put a link of this page into your project's website.

The information about "BIGFASTDATA" are provided by the European Opendata Portal: CORDIS opendata.

More projects from the same programme (H2020-EU.1.1.)

VictPart (2019)

Righting Victim Participation in Transitional Justice

Read More  


The Mass Politics of Disintegration

Read More  


The Enemy of the Good: Towards a Theory of Moral Progress

Read More