Skip to main content


The paper "Estimating Join Selectivities using Bandwidth-Optimized Kernel Density Models" by Martin Kiefer, Max Heimel, Sebastian Breß and Volker Markl will be published in PVLDB Vol 10 and will be presented at VLDB 2018 in Rio de Janeiro, Brazil


The paper Optimized On-Demand Data Streaming from Sensor Nodes was accepted for publication at the ACM Symposium on Cloud Computing and will be presented at the conference in Sata Clara.


We are looking Forward to this year's Flink Forward Conference in Berlin which will take place September 11-13, 2017 at Kulturbrauerei. The conference offers a three days full program packed with talks and trainings. Tickets are available at


Three members of BBDC Management have submitted a paper to the TPCTC/VLDB 2017 together with two researchers of the Database Systems and Information Management Group DIMA. This paper titled PEEL: A Framework for benchmarking distributed systems and algorithms



Big Data Summer School 2017

The Big Data Competence Centers ScaDS Dresden/Leipzig and BBDC are hosting the third international big data summer school in Germany.

August, 21st - 25th 2017

read more

Big Data Excellence in Germany and UK

March 1st, 2017 in Berlin

A joint event by the Berlin Big Data Center and UK Science and Innovation Network took place at Smart Data Forum.

The First Berlin Big Data Center Symposium Held in Berlin

The Berlin Big Data Center held its first Symposium on November 8th at the Smart Data Forum, located in Berlin. At this event, members of the BBDC presented the project’s interim results from two years of research.

Newsletter published

The newsletter Big Data Research is a joint newsletter which reflects works done by Berlin Big Data Center (BBDC), Dresden/Leipzig Competence Center for Scalable Data Services and Solutions (ScaDS), Smart Data Innovation Lab (SDIL), and Smart Data Forum (SDF).

Edition #001 (Nov 2016)

Edition #002 (Feb 2017)

Edition #003 (July 2017)

Subscribe to Newsletter


Five Dimensions of Big Data

Big data is often defined as any data set that cannot be handled using today’s widely available mainstream techniques and technologies. The challenges of handling big data are often described using 3-Vs (volume, variety and velocity): high volume of data from a variety of data sources arriving with high velocity analysed to achieve an economic benefit. However, the 3-Vs fail to reflect complexity of “Big Data” in its entirety.

Read the full article

According to the Harvard Business Review, Data Scientist is “The Sexiest Job of the 21st Century”. Data scientists are often considered to be wizards that deliver value from big data. These wizards need to have knowledge in three very distinct subject areas, namely, scalable data management, data analysis and domain area expertise. However, it is a challenge to find these jacks-of-all-trades that cover all three areas. Or, as the Wall Street Journal puts it “Big Data’s Problem is Little Talent”. Naturally, finding talented data scientists is also a requirement, if we are to put big data to good use. If data analysis were specified using a declarative language, data scientists would not have to worry about low-level programming any longer. Instead, they would be free to concentrate on their data analysis problem. The goal of the Berlin Big Data Center is to help bridge the Talent Gap of Big Data through researching and developing novel technology.

Read more about it in the article of the VLDB keynote "Breaking the Chains: On Declarative Data Analysis and Data Independence in the Big Data Era" by Volker Markl.