SlideShare a Scribd company logo
Facebook’s Petabyte Scale Data
       Warehouse using Hive and Hadoop




Wednesday, January 27, 2010
Why Another Data Warehousing System?




                               Data, data and more data
                                    200GB per day in March 2008
                              12+TB(compressed) raw data per day today




Wednesday, January 27, 2010
Trends Leading to More Data




Wednesday, January 27, 2010
Trends Leading to More Data




                              Free or low cost of user services




Wednesday, January 27, 2010
Trends Leading to More Data




                              Free or low cost of user services



                Realization that more insights are derived from
                simple algorithms on more data




Wednesday, January 27, 2010
Deficiencies of Existing Technologies




Wednesday, January 27, 2010
Deficiencies of Existing Technologies



         Cost of Analysis and Storage on proprietary systems
         does not support trends towards more data




Wednesday, January 27, 2010
Deficiencies of Existing Technologies



         Cost of Analysis and Storage on proprietary systems
         does not support trends towards more data



                   Limited Scalability does not support trends
                   towards more data




Wednesday, January 27, 2010
Deficiencies of Existing Technologies



         Cost of Analysis and Storage on proprietary systems
         does not support trends towards more data



                   Limited Scalability does not support trends
                   towards more data


                              Closed and Proprietary Systems



Wednesday, January 27, 2010
Lets try Hadoop…

   Pros
        – Superior in availability/scalability/manageability
        – Efficiency not that great, but throw more hardware
        – Partial Availability/resilience/scale more important than ACID


   Cons: Programmability and Metadata
        – Map-reduce hard to program (users know sql/bash/python)
        – Need to publish data in well known schemas

   Solution: HIVE




Wednesday, January 27, 2010
What is HIVE?

   A system for managing and querying structured data built
    on top of Hadoop
        – Map-Reduce for execution
        – HDFS for storage
        – Metadata in an RDBMS


   Key Building Principles:
        –   SQL as a familiar data warehousing tool
        –   Extensibility – Types, Functions, Formats, Scripts
        –   Scalability and Performance
        –   Interoperability




Wednesday, January 27, 2010
Why SQL on Hadoop?
  hive> select key, count(1) from kv1 where key > 100 group by
      key;

  vs.

  $ cat > /tmp/reducer.sh
  uniq -c | awk '{print $2"t"$1}‘
  $ cat > /tmp/map.sh
  awk -F '001' '{if($1 > 100) print $1}‘
  $ bin/hadoop jar contrib/hadoop-0.19.2-dev-streaming.jar -input /user/hive/warehouse/kv1 -
      mapper map.sh -file /tmp/reducer.sh -file /tmp/map.sh -reducer reducer.sh -output /tmp/
      largekey -numReduceTasks 1
  $ bin/hadoop dfs –cat /tmp/largekey/part*




Wednesday, January 27, 2010
Data Flow Architecture at Facebook




Wednesday, January 27, 2010
Data Flow Architecture at Facebook




        Web Servers




Wednesday, January 27, 2010
Data Flow Architecture at Facebook




        Web Servers           Scribe MidTier




Wednesday, January 27, 2010
Data Flow Architecture at Facebook


                                                     Filers

        Web Servers           Scribe MidTier




                                               Scribe-Hadoop
                                               Cluster




Wednesday, January 27, 2010
Data Flow Architecture at Facebook


                                                     Filers

        Web Servers           Scribe MidTier




                                               Scribe-Hadoop
                                               Cluster




                                                Federated MySQL

Wednesday, January 27, 2010
Data Flow Architecture at Facebook


                                                                     Filers

        Web Servers                     Scribe MidTier




                                                               Scribe-Hadoop
                                                               Cluster




                              Production Hive-Hadoop Cluster
                                                                Federated MySQL

Wednesday, January 27, 2010
Data Flow Architecture at Facebook


                                                                     Filers

        Web Servers                     Scribe MidTier




                                                               Scribe-Hadoop
                                                               Cluster




                              Production Hive-Hadoop Cluster
          Oracle RAC                                            Federated MySQL

Wednesday, January 27, 2010
Data Flow Architecture at Facebook


                                                                      Filers

        Web Servers                      Scribe MidTier




                        Hive
                        replication                             Scribe-Hadoop
                                                                Cluster




                               Production Hive-Hadoop Cluster
          Oracle RAC                                             Federated MySQL

Wednesday, January 27, 2010
Data Flow Architecture at Facebook


                                                                      Filers

        Web Servers                      Scribe MidTier




                        Hive
                        replication                             Scribe-Hadoop
                                                                Cluster
 Adhoc Hive-Hadoop Cluster




                               Production Hive-Hadoop Cluster
          Oracle RAC                                             Federated MySQL

Wednesday, January 27, 2010
Scribe & Hadoop Clusters @ Facebook

   Used to log data from web servers

   Clusters collocated with the web servers

   Network is the biggest bottleneck

   Typical cluster has about 50 nodes.

   Stats:
        – ~ 25TB/day of raw data logged
        – 99% of the time data is available within 20 seconds



Wednesday, January 27, 2010
Hadoop & Hive Cluster @ Facebook

   Hadoop/Hive cluster
        –   8400 cores
        –   Raw Storage capacity ~ 12.5PB
        –   8 cores + 12 TB per node
        –   32 GB RAM per node
        –   Two level network topology
               1 Gbit/sec from node to rack switch
               4 Gbit/sec to top level rack switch


   2 clusters
        – One for adhoc users
        – One for strict SLA jobs




Wednesday, January 27, 2010
Hive & Hadoop Usage @ Facebook

   Statistics per day:
     – 12 TB of compressed new data added per day
     – 135TB of compressed data scanned per day
     – 7500+ Hive jobs per day
     – 80K compute hours per day

   Hive simplifies Hadoop:
    – New engineers go though a Hive training session
    – ~200 people/month run jobs on Hadoop/Hive
    – Analysts (non-engineers) use Hadoop through Hive
    – Most of jobs are Hive Jobs



Wednesday, January 27, 2010
Hive & Hadoop Usage @ Facebook

   Types of Applications:
        – Reporting
               Eg: Daily/Weekly aggregations of impression/click counts
               Measures of user engagement
               Microstrategy reports

        – Ad hoc Analysis
               Eg: how many group admins broken down by state/country

        – Machine Learning (Assembling training data)
               Ad Optimization
               Eg: User Engagement as a function of user attributes

        – Many others


Wednesday, January 27, 2010
More about HIVE




Wednesday, January 27, 2010
Data Model


                              Name                        HDFS Directory



   Table                      pvs                         /wh/pvs



   Partition                  ds = 20090801, ctry = US    /wh/pvs/ds=20090801/ctry=US




                                                          /wh/pvs/ds=20090801/ctry=US/
   Bucket                     user into 32 buckets
                                                          part-00000
                              HDFS file for user hash 0




Wednesday, January 27, 2010
Hive Query Language
   SQL
        –   Sub-queries in from clause
        –   Equi-joins (including Outer joins)
        –   Multi-table Insert
        –   Multi-group-by
        –   Embedding Custom Map/Reduce in SQL
   Sampling
   Primitive Types
        – integer types, float, string, boolean
   Nestable Collections
        – array<any-type> and map<primitive-type, any-type>
   User-defined types
        – Structures with attributes which can be of any-type


Wednesday, January 27, 2010
Optimizations
 Joins try to reduce the number of map/reduce jobs needed.
 Memory efficient joins by streaming largest tables.
 Map Joins
     – User specified small tables stored in hash tables on the mapper
     – No reducer needed

 Map side partial aggregations
     – Hash-based aggregates
     – Serialized key/values in hash tables
     – 90% speed improvement on Query
             SELECT count(1) FROM t;
 Load balancing for data skew



Wednesday, January 27, 2010
Hive: Open & Extensible

   Different on-disk storage(file) formats
        – Text File, Sequence File, …
   Different serialization formats and data types
        – LazySimpleSerDe, ThriftSerDe …
   User-provided map/reduce scripts
        – In any language, use stdin/stdout to transfer data …
   User-defined Functions
        – Substr, Trim, From_unixtime …
   User-defined Aggregation Functions
        – Sum, Average …
   User-define Table Functions
        – Explode …



Wednesday, January 27, 2010
Existing File Formats
                              TEXTFILE     SEQUENCEFILE   RCFILE

     Data type                text only    text/binary    text/binary

     Internal
                              Row-based    Row-based      Column-based
     Storage order

     Compression              File-based   Block-based    Block-based

     Splitable*               YES          YES            YES

     Splitable* after
                              NO           YES            YES
     compression

 * Splitable: Capable of splitting the file so that a single huge
      file can be processed by multiple mappers in parallel.


Wednesday, January 27, 2010
Map/Reduce Scripts Examples

     add file page_url_to_id.py;
     add file my_python_session_cutter.py;
     FROM
        (MAP uhash, page_url, unix_time
           USING 'page_url_to_id.py'
           AS (uhash, page_id, unix_time)
         FROM mylog
         DISTRIBUTE BY uhash
         SORT BY uhash, unix_time) mylog2
      REDUCE uhash, page_id, unix_time
        USING 'my_python_session_cutter.py'
        AS (uhash, session_info);




Wednesday, January 27, 2010
UDF Example
      add jar build/ql/test/test-udfs.jar;
      CREATE TEMPORARY FUNCTION testlength AS
       'org.apache.hadoop.hive.ql.udf.UDFTestLength';
      SELECT testlength(page_url) FROM mylog;
      DROP TEMPORARY FUNCTION testlength;


   UDFTestLength.java:
  package org.apache.hadoop.hive.ql.udf;
  public class UDFTestLength extends UDF {
    public Integer evaluate(String s) {
      if (s == null) {
        return null;
      }
      return s.length();
    }
  }

Wednesday, January 27, 2010
Comparison of UDF/UDAF/UDTF v.s. M/R scripts

                              UDF/UDAF/UDTF        M/R scripts

     language                 Java                 any language

     data format              in-memory objects    serialized streams

     1/1 input/output         supported via UDF    supported

     n/1 input/output         supported via UDAF   supported

     1/n input/output         supported via UDTF   supported

     Speed                    faster               slower




Wednesday, January 27, 2010
Interoperability: Interfaces

   JDBC
        – Enables integration with JDBC based SQL clients
   ODBC
        – Enables integration with Microstrategy
   Thrift
        – Enables writing cross language clients
        – Main form of integration with php based Web UI




Wednesday, January 27, 2010
Interoperability: Microstrategy

   Beta integration with version 8
   Free form SQL support
   Periodically pre-compute the cube




Wednesday, January 27, 2010
Operational Aspects on Adhoc cluster

   Data Discovery
        – coHive
               Discover tables
               Talk to expert users of a table
               Browse table lineage


   Monitoring
        – Resource utilization by individual, project, group
        – SLA monitoring etc.
        – Bad user reports etc.




Wednesday, January 27, 2010
HiPal & CoHive (Not open source)




Wednesday, January 27, 2010
Open Source Community

   Released Hive-0.4 on 10/13/2009
   50 contributors and growing
   11 committers
        – 3 external to Facebook
   Available as a sub project in Hadoop
        -   http://wiki.apache.org/hadoop/Hive (wiki)
        -   http://hadoop.apache.org/hive (home page)
        -   http://svn.apache.org/repos/asf/hadoop/hive (SVN repo)
        -   ##hive (IRC)
        -   Works with hadoop-0.17, 0.18, 0.19, 0.20
   Mailing Lists:
        – hive-{user,dev,commits}@hadoop.apache.org



Wednesday, January 27, 2010
Powered by Hive




Wednesday, January 27, 2010

More Related Content

Facebooks Petabyte Scale Data Warehouse using Hive and Hadoop

  • 1. Facebook’s Petabyte Scale Data Warehouse using Hive and Hadoop Wednesday, January 27, 2010
  • 2. Why Another Data Warehousing System? Data, data and more data 200GB per day in March 2008 12+TB(compressed) raw data per day today Wednesday, January 27, 2010
  • 3. Trends Leading to More Data Wednesday, January 27, 2010
  • 4. Trends Leading to More Data Free or low cost of user services Wednesday, January 27, 2010
  • 5. Trends Leading to More Data Free or low cost of user services Realization that more insights are derived from simple algorithms on more data Wednesday, January 27, 2010
  • 6. Deficiencies of Existing Technologies Wednesday, January 27, 2010
  • 7. Deficiencies of Existing Technologies Cost of Analysis and Storage on proprietary systems does not support trends towards more data Wednesday, January 27, 2010
  • 8. Deficiencies of Existing Technologies Cost of Analysis and Storage on proprietary systems does not support trends towards more data Limited Scalability does not support trends towards more data Wednesday, January 27, 2010
  • 9. Deficiencies of Existing Technologies Cost of Analysis and Storage on proprietary systems does not support trends towards more data Limited Scalability does not support trends towards more data Closed and Proprietary Systems Wednesday, January 27, 2010
  • 10. Lets try Hadoop…  Pros – Superior in availability/scalability/manageability – Efficiency not that great, but throw more hardware – Partial Availability/resilience/scale more important than ACID  Cons: Programmability and Metadata – Map-reduce hard to program (users know sql/bash/python) – Need to publish data in well known schemas  Solution: HIVE Wednesday, January 27, 2010
  • 11. What is HIVE?  A system for managing and querying structured data built on top of Hadoop – Map-Reduce for execution – HDFS for storage – Metadata in an RDBMS  Key Building Principles: – SQL as a familiar data warehousing tool – Extensibility – Types, Functions, Formats, Scripts – Scalability and Performance – Interoperability Wednesday, January 27, 2010
  • 12. Why SQL on Hadoop? hive> select key, count(1) from kv1 where key > 100 group by key; vs. $ cat > /tmp/reducer.sh uniq -c | awk '{print $2"t"$1}‘ $ cat > /tmp/map.sh awk -F '001' '{if($1 > 100) print $1}‘ $ bin/hadoop jar contrib/hadoop-0.19.2-dev-streaming.jar -input /user/hive/warehouse/kv1 - mapper map.sh -file /tmp/reducer.sh -file /tmp/map.sh -reducer reducer.sh -output /tmp/ largekey -numReduceTasks 1 $ bin/hadoop dfs –cat /tmp/largekey/part* Wednesday, January 27, 2010
  • 13. Data Flow Architecture at Facebook Wednesday, January 27, 2010
  • 14. Data Flow Architecture at Facebook Web Servers Wednesday, January 27, 2010
  • 15. Data Flow Architecture at Facebook Web Servers Scribe MidTier Wednesday, January 27, 2010
  • 16. Data Flow Architecture at Facebook Filers Web Servers Scribe MidTier Scribe-Hadoop Cluster Wednesday, January 27, 2010
  • 17. Data Flow Architecture at Facebook Filers Web Servers Scribe MidTier Scribe-Hadoop Cluster Federated MySQL Wednesday, January 27, 2010
  • 18. Data Flow Architecture at Facebook Filers Web Servers Scribe MidTier Scribe-Hadoop Cluster Production Hive-Hadoop Cluster Federated MySQL Wednesday, January 27, 2010
  • 19. Data Flow Architecture at Facebook Filers Web Servers Scribe MidTier Scribe-Hadoop Cluster Production Hive-Hadoop Cluster Oracle RAC Federated MySQL Wednesday, January 27, 2010
  • 20. Data Flow Architecture at Facebook Filers Web Servers Scribe MidTier Hive replication Scribe-Hadoop Cluster Production Hive-Hadoop Cluster Oracle RAC Federated MySQL Wednesday, January 27, 2010
  • 21. Data Flow Architecture at Facebook Filers Web Servers Scribe MidTier Hive replication Scribe-Hadoop Cluster Adhoc Hive-Hadoop Cluster Production Hive-Hadoop Cluster Oracle RAC Federated MySQL Wednesday, January 27, 2010
  • 22. Scribe & Hadoop Clusters @ Facebook  Used to log data from web servers  Clusters collocated with the web servers  Network is the biggest bottleneck  Typical cluster has about 50 nodes.  Stats: – ~ 25TB/day of raw data logged – 99% of the time data is available within 20 seconds Wednesday, January 27, 2010
  • 23. Hadoop & Hive Cluster @ Facebook  Hadoop/Hive cluster – 8400 cores – Raw Storage capacity ~ 12.5PB – 8 cores + 12 TB per node – 32 GB RAM per node – Two level network topology  1 Gbit/sec from node to rack switch  4 Gbit/sec to top level rack switch  2 clusters – One for adhoc users – One for strict SLA jobs Wednesday, January 27, 2010
  • 24. Hive & Hadoop Usage @ Facebook  Statistics per day: – 12 TB of compressed new data added per day – 135TB of compressed data scanned per day – 7500+ Hive jobs per day – 80K compute hours per day  Hive simplifies Hadoop: – New engineers go though a Hive training session – ~200 people/month run jobs on Hadoop/Hive – Analysts (non-engineers) use Hadoop through Hive – Most of jobs are Hive Jobs Wednesday, January 27, 2010
  • 25. Hive & Hadoop Usage @ Facebook  Types of Applications: – Reporting  Eg: Daily/Weekly aggregations of impression/click counts  Measures of user engagement  Microstrategy reports – Ad hoc Analysis  Eg: how many group admins broken down by state/country – Machine Learning (Assembling training data)  Ad Optimization  Eg: User Engagement as a function of user attributes – Many others Wednesday, January 27, 2010
  • 26. More about HIVE Wednesday, January 27, 2010
  • 27. Data Model Name HDFS Directory Table pvs /wh/pvs Partition ds = 20090801, ctry = US /wh/pvs/ds=20090801/ctry=US /wh/pvs/ds=20090801/ctry=US/ Bucket user into 32 buckets part-00000 HDFS file for user hash 0 Wednesday, January 27, 2010
  • 28. Hive Query Language  SQL – Sub-queries in from clause – Equi-joins (including Outer joins) – Multi-table Insert – Multi-group-by – Embedding Custom Map/Reduce in SQL  Sampling  Primitive Types – integer types, float, string, boolean  Nestable Collections – array<any-type> and map<primitive-type, any-type>  User-defined types – Structures with attributes which can be of any-type Wednesday, January 27, 2010
  • 29. Optimizations  Joins try to reduce the number of map/reduce jobs needed.  Memory efficient joins by streaming largest tables.  Map Joins – User specified small tables stored in hash tables on the mapper – No reducer needed  Map side partial aggregations – Hash-based aggregates – Serialized key/values in hash tables – 90% speed improvement on Query  SELECT count(1) FROM t;  Load balancing for data skew Wednesday, January 27, 2010
  • 30. Hive: Open & Extensible  Different on-disk storage(file) formats – Text File, Sequence File, …  Different serialization formats and data types – LazySimpleSerDe, ThriftSerDe …  User-provided map/reduce scripts – In any language, use stdin/stdout to transfer data …  User-defined Functions – Substr, Trim, From_unixtime …  User-defined Aggregation Functions – Sum, Average …  User-define Table Functions – Explode … Wednesday, January 27, 2010
  • 31. Existing File Formats TEXTFILE SEQUENCEFILE RCFILE Data type text only text/binary text/binary Internal Row-based Row-based Column-based Storage order Compression File-based Block-based Block-based Splitable* YES YES YES Splitable* after NO YES YES compression * Splitable: Capable of splitting the file so that a single huge file can be processed by multiple mappers in parallel. Wednesday, January 27, 2010
  • 32. Map/Reduce Scripts Examples  add file page_url_to_id.py;  add file my_python_session_cutter.py;  FROM (MAP uhash, page_url, unix_time USING 'page_url_to_id.py' AS (uhash, page_id, unix_time) FROM mylog DISTRIBUTE BY uhash SORT BY uhash, unix_time) mylog2 REDUCE uhash, page_id, unix_time USING 'my_python_session_cutter.py' AS (uhash, session_info); Wednesday, January 27, 2010
  • 33. UDF Example  add jar build/ql/test/test-udfs.jar;  CREATE TEMPORARY FUNCTION testlength AS 'org.apache.hadoop.hive.ql.udf.UDFTestLength';  SELECT testlength(page_url) FROM mylog;  DROP TEMPORARY FUNCTION testlength;  UDFTestLength.java: package org.apache.hadoop.hive.ql.udf; public class UDFTestLength extends UDF { public Integer evaluate(String s) { if (s == null) { return null; } return s.length(); } } Wednesday, January 27, 2010
  • 34. Comparison of UDF/UDAF/UDTF v.s. M/R scripts UDF/UDAF/UDTF M/R scripts language Java any language data format in-memory objects serialized streams 1/1 input/output supported via UDF supported n/1 input/output supported via UDAF supported 1/n input/output supported via UDTF supported Speed faster slower Wednesday, January 27, 2010
  • 35. Interoperability: Interfaces  JDBC – Enables integration with JDBC based SQL clients  ODBC – Enables integration with Microstrategy  Thrift – Enables writing cross language clients – Main form of integration with php based Web UI Wednesday, January 27, 2010
  • 36. Interoperability: Microstrategy  Beta integration with version 8  Free form SQL support  Periodically pre-compute the cube Wednesday, January 27, 2010
  • 37. Operational Aspects on Adhoc cluster  Data Discovery – coHive  Discover tables  Talk to expert users of a table  Browse table lineage  Monitoring – Resource utilization by individual, project, group – SLA monitoring etc. – Bad user reports etc. Wednesday, January 27, 2010
  • 38. HiPal & CoHive (Not open source) Wednesday, January 27, 2010
  • 39. Open Source Community  Released Hive-0.4 on 10/13/2009  50 contributors and growing  11 committers – 3 external to Facebook  Available as a sub project in Hadoop - http://wiki.apache.org/hadoop/Hive (wiki) - http://hadoop.apache.org/hive (home page) - http://svn.apache.org/repos/asf/hadoop/hive (SVN repo) - ##hive (IRC) - Works with hadoop-0.17, 0.18, 0.19, 0.20  Mailing Lists: – hive-{user,dev,commits}@hadoop.apache.org Wednesday, January 27, 2010
  • 40. Powered by Hive Wednesday, January 27, 2010