Title :
Open data challenges at Facebook
Author :
Bronson, Nathan ; Lento, Thomas ; Wiener, Janet L.
Author_Institution :
Facebook Inc., Menlo Park, CA, USA
Abstract :
At Facebook, our data systems process huge volumes of data, ranging from hundreds of terabytes in memory to hundreds of petabytes on disk. We categorize our systems as “small data” or “big data” based on the type of queries they run. Small data refers to OLTP-like queries that process and retrieve a small amount of data, for example, the 1000s of objects necessary to render Facebook´s personalized News Feed for each person. These objects are requested by their ids; indexes limit the amount of data accessed during a single query, regardless of the total volume of data. Big data refers to queries that process large amounts of data, usually for analysis: trouble-shooting, identifying trends, and making decisions. Big data stores are the workhorses for data analysis at Facebook. They grow by millions of events (inserts) per second and process tens of petabytes and hundreds of thousands of queries per day. In this tutorial, we will describe our data systems and the current challenges we face. We will lead a discussion on these challenges, approaches to solve them, and potential pitfalls. We hope to stimulate interest in solving these problems in the research community.
Keywords :
Big Data; data analysis; query processing; social networking (online); Facebook; OLTP-like queries; big data; data analysis; personalized news feed; small data; Big data; Databases; Facebook; Feeds; Mobile handsets; Reliability; Time series analysis;
Conference_Titel :
Data Engineering (ICDE), 2015 IEEE 31st International Conference on
Conference_Location :
Seoul
DOI :
10.1109/ICDE.2015.7113415