SlideShare une entreprise Scribd logo
1  sur  46
Télécharger pour lire hors ligne
Performance Management in
          ‘Big Data’ Applications
          It’s still about the Application


Michael Kopp, Technology Strategist          Edward Capriolo
michael.kopp@compuware.com                   edward@m6d.com
@mikopp                                      @edwardcapriolo
blog.dynatrace.com                           m6d.com/blog
BigData                    High Volume/Low Latency DBs




                            Web         Java




Key Challenges                                 Key Benefits
1) Even Distribution                           1) Fast Read/Write
2) Correct Schema and Access patterns          2) Horizontal Scalability
3) Understanding Application Impact            3) Redundancy and High Availability


 3
BigData                     Large Parallel Batch Processing




         Hive high-level
         map/reduce           JOB
         query                JOB
                               1
                               2
                               1
  batch                        3
                               2
  trigger                      4
                               3
                               .
                               .
                               .
Key Challenges Hive Server    754
                                      Key Benefits
1) Optimal Distribution               1) Massive Horizontal Batch Job
2) Unwieldy Configuration             2) Split big Problems into smaller ones
3) Can easily waste your resources
What is m6d?




5
Impressions look like…




 6
Map Reduce
    Performance

7
Typical MapReduce Job at m6d




8
Hadoop at m6d

    • Critical piece of infrastructure
    • Long Term Data Storage
        – Raw logs
        – Aggregations
        – Reports
        – Generated data (feed back loops)
    • Numerous ETL (Extract Transform Load)
    • Scheduled and adhoc processes
    • Used directly by Tech-Team, Ad Ops, Data Science



9
Hadoop at m6d

• Two deployments 'production' and 'research'
   – ~ 500 TB - 40+ Nodes
   – ~ 350 TB – 20+ Nodes
• Thousands of jobs
   – <5 minute jobs and 12 hour Job Flows
   – Mostly Hive Jobs
   – Some custom code and streaming jobs
Hadoop Design Tenants

• Linear scalability by adding more hardware
• HDFS Distributed file system
   – User space file system
   – Blocks are replicated across nodes
   – Limited semantics
• MapReduce
   – Paradigm which models using map/reduce
   – Data Locality
   – Split Job into Tasks by Data
       – Retry in failure
Schema Design Challenges

 • Partition data for good distribution
     – By time interval (optionally a second level)
        • Partition pruning with WHERE
     – Clustering (aka bucketing)
        • Optimized sampling and joins
     – Columnar
        • Column oriented
 • Raw Data Growth
 • Data features change (more distinct X)



12
Key Performance Challenges

 • Intermediate I/O
     – Compression codec
     – Block size
     – Split-table formats
 • Contentions between jobs
 • Data and Map/Reduce Distribution
     • Data Skew
     • Non Uniform Computation (long running tasks)
 • ‘Cost' of new feature – is this justified?
 • Tuning variables (spills, buffers, Etc, etc)

13
How to handle Performance Issues?

 • Profile the Job / Query?
     – Who should do this?
       (DBA, Dev, Ops, DevOps , NoOps, Big Data Guru)
     – How should we do this?
        • Look at job run times day over day?
        • Look at code and micro-benchmark?
        • Collect Job Counters?

 • Upgrade often for latest performance features?
 • Investigate/purchase newer better hardware
     – More cores? RAM? 10G Ethernet? SSD
                                                   Test Data is not like
 • Read blogs?                                          Real Data


14
But how to optimize
        the job itself?

15
Understanding Map/Reduce Performance




     Attention Data
       Maximum
       Parallelism
        Volume!


Actual Mapping
                      Also your own
                        Millions of
  Parallelism
                           Code
                      Executions!!!


    Attention
    Potential
   Choke Point!




        Maximum
         Reduce
        Parallelism


   Actual Reduce        Also your own
    Parallelism
    16
                             Code
Understanding Map/Reduce Performance
Map/Reduce Performance




18
Map/Reduce behind the scenes
                Serialize
                                               De-Serialize
                                               and Serialize
                                                  again




                                                  Potentionally
                                                   Inefficient

                                          Too Many Files,
                                            Same Key
                                          spread all over


De-Serialize                 Expensive
and Serialize               Synchronous
   again                      Combine




19
Map/Reduce Combine and Spill Performance




                      1) Pre Combine in Mapping Step
                      2) Avoid many intermediate files and combines




20
Map/Reduce “Map” Performance




                       Avoid Brute Force
                    Then on Big Hotspots
                    FocusOptimize Hadoop
                     Save a lot of Hardware




21
Map/Reduce to the Max!

 • Ensure Data Locality
 • Optimize Map/Reduce Hotspots
 • Reduce Intermediate Data and “Overhead”
 • Ensure optimal Data and Compute Distribution
 • Tune Hadoop Environment




22
Cassandra and
      Application
     Performance
23
A High Level look at RTB




     1.   Browsers visit Publishers and create impressions.
     2.   Publishers sell impressions via Exchanges.
     3.   Exchanges serve as auction houses for the impressions
     4.   On behalf of the marketer, m6d bids the impressions via
          the auction house. If m6d wins, we display our ad to the
          browser.
24
Cassandra at m6d for Real Time Bidding

 • RTB limited data is provided from exchange
 • System to store information on users
     – Frequency Capping
     – Visit History
     – Segments (product service affinity)
 • Low latency Requirements
     – Less then 100ms
     – Requires fast read/write on discrete data




25
Cassandra design




26
Key Cassandra Design Tennents

•   Swap/paging not possible
•   Mostly schema-less
•   Writes do not read
     –    Read/Write is an anti-pattern
•   Optimize around put and get
     –    Not for scan and query
•   De-Normalize data
     –    Attempt to get all data in single read*
Cassandra Design Challenges

 • De-normailize
     – Store data to optimize reads
     – Composite (multi-column) keys
 • Multi-column family and Multi-tenant scenarios
 • Compress settings
     – Disk and cache savings
     – CPU and JVM costs
 • Data/Compaction settings
     – Size tiered vs LevelDB
 • Caching, Memtable and other tuning

28
How to handle performance issues?

 • Monitor standard vitals (cpu,disk) ?
 • Read blogs and documentation?
 • Use Cassandra JMX to track req/sec
 • Use Cassandra JMX to track size of Column Families, rows and
   columns
 • Upgrade often to get latest performance enhancements? *


                               What about the Application?



29
APM for Cassandra


30
NoSQL APM is not so different after all…




                     Web     Java      Database




                             Key APM Problems Identified
                             1) Response Time Contribution
                             2) data access patterns
                             3) transaction to query
                                relationship (transaction flow)


31
Response Time Contribution




                               Contribution to
                             Business Transaction   Connection Pool




            Access Pattern



32
Statement Analysis

                        Executions per
                       Average and Total
                        Contribution to
                       Transactions and
                      Business Transaction
                        Execution Time
                             Total




33
Where, Why, How and which Transaction…

                      Which Business
                       Transaction




                                           Which Web Service




Where and why in my
   Transaction
                                       Single Statement
                                         Performance
34
How does this apply to NoSQL Databases?




                                     Web          Java




Key APM Problems Identified
1) Response Time Contribution
                                      1)   Data Access Distribution
2) data access patterns
                                      2)   End-to-End Monitoring
3) transaction to query
                                      3)   Storage (I/O, GC) Bottlenecks
   relationship (transaction flow)
                                      4)   Consistency Level

35
Real End-to-End Application Performance



                                 Our Application
                   Third Party
                                                        External

        End User

                    Services




                                        End User Response Time
                                             Contribution



37
Understanding Cassandra’s Contribution




     Which statements did the Transaction Execute?
     Which node where they executed against?
                  Contribution of each many calls?
                                   Too Statment      Data Access patterns
     Which Consistency Level was used?


38
Understand Response Time Contribution




                             5 Calls 4 Calls
                     ~50-80 ms ~15 ms Contribution
                               Contribution?

                                              Access and Data Distribution



39
Why and how was a statement executed?




                            45ms latency?   60ms waiting on
                                              the server?




40
Any Hotspots on the Cassandra Nodes?




                          Much more load on Node3?
                           Which Transactions are
                                responsible




41
Specific Cassandra Health Metrics




42
General Health of Cassandra


              Memory Issues?




                               Too many requests?




                                      Too much GC Suspensions?




43
Conclusion




44
Extend Performance Focus on Application




                    Web     Java




     A Fast Database doesn’t make
            a fast Application
45
Intelligent MapReduce APM



                                data/task node
       Hive high-level
       map/reduce         JOB   data/task node
       query              JOB
                           1
                           2
 batch                     1
                           3
                                master node
                           2
 trigger                   4
                           3
                           .
                           .
            Hive Server    .
                          754
                                data/task node


Simple Optimizations with big impact
Big Data is about solving
       Application Problems
      APM is about Application
     Performance and Efficiency

47
THANK YOU
     Michael Kopp, Technology Strategist   Edward Capriolo
     michael.kopp@compuware.com            edward@m6d.com
     @mikopp                               @edwardcapriolo
48
     blog.dynatrace.com                    m6d.com/blog

Contenu connexe

Tendances

Sql Performance Tuning For Developers
Sql Performance Tuning For DevelopersSql Performance Tuning For Developers
Sql Performance Tuning For Developerssqlserver.co.il
 
IBM Systems solution for SAP NetWeaver Business Warehouse Accelerator
IBM Systems solution for SAP NetWeaver Business Warehouse AcceleratorIBM Systems solution for SAP NetWeaver Business Warehouse Accelerator
IBM Systems solution for SAP NetWeaver Business Warehouse AcceleratorIBM India Smarter Computing
 
Real-Time Loading to Sybase IQ
Real-Time Loading to Sybase IQReal-Time Loading to Sybase IQ
Real-Time Loading to Sybase IQSybase Türkiye
 
Bank Data Frank Peterson DB2 10-Early_Experiences_pdf
Bank Data   Frank Peterson DB2 10-Early_Experiences_pdfBank Data   Frank Peterson DB2 10-Early_Experiences_pdf
Bank Data Frank Peterson DB2 10-Early_Experiences_pdfSurekha Parekh
 
SQLFire at VMworld Europe 2011
SQLFire at VMworld Europe 2011SQLFire at VMworld Europe 2011
SQLFire at VMworld Europe 2011Carter Shanklin
 
Hadoop World 2011: Practical HBase - Ravi Veeramchaneni, Informatica
Hadoop World 2011: Practical HBase - Ravi Veeramchaneni, InformaticaHadoop World 2011: Practical HBase - Ravi Veeramchaneni, Informatica
Hadoop World 2011: Practical HBase - Ravi Veeramchaneni, InformaticaCloudera, Inc.
 
LogisticsWired Offering from MavenWire
LogisticsWired Offering from MavenWireLogisticsWired Offering from MavenWire
LogisticsWired Offering from MavenWireMavenWire
 
My sql 5.6_replwebinar_may12
My sql 5.6_replwebinar_may12My sql 5.6_replwebinar_may12
My sql 5.6_replwebinar_may12Mat Keep
 
Complex Event Processing: What?, Why?, How?
Complex Event Processing: What?, Why?, How?Complex Event Processing: What?, Why?, How?
Complex Event Processing: What?, Why?, How?Fabien Coppens
 
DB2 – Differentiating Business Value
DB2 – Differentiating Business ValueDB2 – Differentiating Business Value
DB2 – Differentiating Business ValueIBM Sverige
 
An Active and Hybrid Storage System for Data-intensive Applications
An Active and Hybrid Storage System for Data-intensive ApplicationsAn Active and Hybrid Storage System for Data-intensive Applications
An Active and Hybrid Storage System for Data-intensive ApplicationsXiao Qin
 
Introduction to Hadoop - ACCU2010
Introduction to Hadoop - ACCU2010Introduction to Hadoop - ACCU2010
Introduction to Hadoop - ACCU2010Gavin Heavyside
 
Software im SAP Umfeld_IBM DB2
Software im SAP Umfeld_IBM DB2Software im SAP Umfeld_IBM DB2
Software im SAP Umfeld_IBM DB2IBM Switzerland
 
DB2 V10 Migration Guidance
DB2 V10 Migration GuidanceDB2 V10 Migration Guidance
DB2 V10 Migration GuidanceCraig Mullins
 
Windows Azure Design Patterns
Windows Azure Design PatternsWindows Azure Design Patterns
Windows Azure Design PatternsDavid Pallmann
 

Tendances (20)

Sql Performance Tuning For Developers
Sql Performance Tuning For DevelopersSql Performance Tuning For Developers
Sql Performance Tuning For Developers
 
IBM Systems solution for SAP NetWeaver Business Warehouse Accelerator
IBM Systems solution for SAP NetWeaver Business Warehouse AcceleratorIBM Systems solution for SAP NetWeaver Business Warehouse Accelerator
IBM Systems solution for SAP NetWeaver Business Warehouse Accelerator
 
Real-Time Loading to Sybase IQ
Real-Time Loading to Sybase IQReal-Time Loading to Sybase IQ
Real-Time Loading to Sybase IQ
 
Bank Data Frank Peterson DB2 10-Early_Experiences_pdf
Bank Data   Frank Peterson DB2 10-Early_Experiences_pdfBank Data   Frank Peterson DB2 10-Early_Experiences_pdf
Bank Data Frank Peterson DB2 10-Early_Experiences_pdf
 
SQLFire at VMworld Europe 2011
SQLFire at VMworld Europe 2011SQLFire at VMworld Europe 2011
SQLFire at VMworld Europe 2011
 
Hadoop World 2011: Practical HBase - Ravi Veeramchaneni, Informatica
Hadoop World 2011: Practical HBase - Ravi Veeramchaneni, InformaticaHadoop World 2011: Practical HBase - Ravi Veeramchaneni, Informatica
Hadoop World 2011: Practical HBase - Ravi Veeramchaneni, Informatica
 
LogisticsWired Offering from MavenWire
LogisticsWired Offering from MavenWireLogisticsWired Offering from MavenWire
LogisticsWired Offering from MavenWire
 
My sql 5.6_replwebinar_may12
My sql 5.6_replwebinar_may12My sql 5.6_replwebinar_may12
My sql 5.6_replwebinar_may12
 
Complex Event Processing: What?, Why?, How?
Complex Event Processing: What?, Why?, How?Complex Event Processing: What?, Why?, How?
Complex Event Processing: What?, Why?, How?
 
DB2 – Differentiating Business Value
DB2 – Differentiating Business ValueDB2 – Differentiating Business Value
DB2 – Differentiating Business Value
 
SQLFire Webinar
SQLFire WebinarSQLFire Webinar
SQLFire Webinar
 
An Active and Hybrid Storage System for Data-intensive Applications
An Active and Hybrid Storage System for Data-intensive ApplicationsAn Active and Hybrid Storage System for Data-intensive Applications
An Active and Hybrid Storage System for Data-intensive Applications
 
Greenplum hadoop
Greenplum hadoopGreenplum hadoop
Greenplum hadoop
 
Introduction to Hadoop - ACCU2010
Introduction to Hadoop - ACCU2010Introduction to Hadoop - ACCU2010
Introduction to Hadoop - ACCU2010
 
SQLFire at Strata 2012
SQLFire at Strata 2012SQLFire at Strata 2012
SQLFire at Strata 2012
 
An Hour of DB2 Tips
An Hour of DB2 TipsAn Hour of DB2 Tips
An Hour of DB2 Tips
 
Software im SAP Umfeld_IBM DB2
Software im SAP Umfeld_IBM DB2Software im SAP Umfeld_IBM DB2
Software im SAP Umfeld_IBM DB2
 
Hana Offerings Engl
Hana Offerings EnglHana Offerings Engl
Hana Offerings Engl
 
DB2 V10 Migration Guidance
DB2 V10 Migration GuidanceDB2 V10 Migration Guidance
DB2 V10 Migration Guidance
 
Windows Azure Design Patterns
Windows Azure Design PatternsWindows Azure Design Patterns
Windows Azure Design Patterns
 

En vedette

TLC Manchester - Russell Beck, Impellam
TLC Manchester - Russell Beck, ImpellamTLC Manchester - Russell Beck, Impellam
TLC Manchester - Russell Beck, ImpellamTALiNT Partners
 
HR Analytics and KPIs with LBi HR HelpDesk
HR Analytics and KPIs with LBi HR HelpDeskHR Analytics and KPIs with LBi HR HelpDesk
HR Analytics and KPIs with LBi HR HelpDeskLBi Software
 
EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...
EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...
EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...European Data Forum
 
EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...
EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...
EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...European Data Forum
 
Big data and employee engagement
Big data and employee engagementBig data and employee engagement
Big data and employee engagementTalkFreely
 
HR / Talent Analytics
HR / Talent AnalyticsHR / Talent Analytics
HR / Talent AnalyticsAkshay Raje
 

En vedette (7)

TLC Manchester - Russell Beck, Impellam
TLC Manchester - Russell Beck, ImpellamTLC Manchester - Russell Beck, Impellam
TLC Manchester - Russell Beck, Impellam
 
HR Analytics and KPIs with LBi HR HelpDesk
HR Analytics and KPIs with LBi HR HelpDeskHR Analytics and KPIs with LBi HR HelpDesk
HR Analytics and KPIs with LBi HR HelpDesk
 
EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...
EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...
EDF2014: Talk of Inge Buffolo, Head of Institutional Relations ad Linguistic ...
 
EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...
EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...
EDF2014: Rüdiger Eichin, Research Manager at SAP AG, Germany: Deriving Value ...
 
Big data and employee engagement
Big data and employee engagementBig data and employee engagement
Big data and employee engagement
 
HR Analytics, Done Right
HR Analytics, Done RightHR Analytics, Done Right
HR Analytics, Done Right
 
HR / Talent Analytics
HR / Talent AnalyticsHR / Talent Analytics
HR / Talent Analytics
 

Similaire à Performance Management in ‘Big Data’ Applications

Intro to Big Data and NoSQL
Intro to Big Data and NoSQLIntro to Big Data and NoSQL
Intro to Big Data and NoSQLDon Demcsak
 
Starfish: A Self-tuning System for Big Data Analytics
Starfish: A Self-tuning System for Big Data AnalyticsStarfish: A Self-tuning System for Big Data Analytics
Starfish: A Self-tuning System for Big Data AnalyticsGrant Ingersoll
 
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...smallerror
 
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...xlight
 
Fixing twitter
Fixing twitterFixing twitter
Fixing twitterRoger Xia
 
In memory grids IMDG
In memory grids IMDGIn memory grids IMDG
In memory grids IMDGPrateek Jain
 
Large scale computing with mapreduce
Large scale computing with mapreduceLarge scale computing with mapreduce
Large scale computing with mapreducehansen3032
 
David Loureiro - Presentation at HP's HPC & OSL TES
David Loureiro - Presentation at HP's HPC & OSL TESDavid Loureiro - Presentation at HP's HPC & OSL TES
David Loureiro - Presentation at HP's HPC & OSL TESSysFera
 
Beyond The Data Grid: Coherence, Normalisation, Joins and Linear Scalability
Beyond The Data Grid: Coherence, Normalisation, Joins and Linear ScalabilityBeyond The Data Grid: Coherence, Normalisation, Joins and Linear Scalability
Beyond The Data Grid: Coherence, Normalisation, Joins and Linear ScalabilityBen Stopford
 
Building FoundationDB
Building FoundationDBBuilding FoundationDB
Building FoundationDBFoundationDB
 
Learning from google megastore (Part-1)
Learning from google megastore (Part-1)Learning from google megastore (Part-1)
Learning from google megastore (Part-1)Schubert Zhang
 
Introducing MongoDB into your Organization
Introducing MongoDB into your OrganizationIntroducing MongoDB into your Organization
Introducing MongoDB into your OrganizationMongoDB
 
Getting Started with MongoDB at Oracle Open World 2012
Getting Started with MongoDB at Oracle Open World 2012Getting Started with MongoDB at Oracle Open World 2012
Getting Started with MongoDB at Oracle Open World 2012MongoDB
 
Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013
Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013
Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013Christopher Curtin
 
John adams talk cloudy
John adams   talk cloudyJohn adams   talk cloudy
John adams talk cloudyJohn Adams
 
The elephantintheroom bigdataanalyticsinthecloud
The elephantintheroom bigdataanalyticsinthecloudThe elephantintheroom bigdataanalyticsinthecloud
The elephantintheroom bigdataanalyticsinthecloudKhazret Sapenov
 

Similaire à Performance Management in ‘Big Data’ Applications (20)

Intro to Big Data and NoSQL
Intro to Big Data and NoSQLIntro to Big Data and NoSQL
Intro to Big Data and NoSQL
 
Starfish: A Self-tuning System for Big Data Analytics
Starfish: A Self-tuning System for Big Data AnalyticsStarfish: A Self-tuning System for Big Data Analytics
Starfish: A Self-tuning System for Big Data Analytics
 
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...
 
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...Fixing Twitter  Improving The Performance And Scalability Of The Worlds Most ...
Fixing Twitter Improving The Performance And Scalability Of The Worlds Most ...
 
Fixing twitter
Fixing twitterFixing twitter
Fixing twitter
 
Fixing_Twitter
Fixing_TwitterFixing_Twitter
Fixing_Twitter
 
COBOL to Apache Spark
COBOL to Apache SparkCOBOL to Apache Spark
COBOL to Apache Spark
 
In memory grids IMDG
In memory grids IMDGIn memory grids IMDG
In memory grids IMDG
 
Large scale computing with mapreduce
Large scale computing with mapreduceLarge scale computing with mapreduce
Large scale computing with mapreduce
 
David Loureiro - Presentation at HP's HPC & OSL TES
David Loureiro - Presentation at HP's HPC & OSL TESDavid Loureiro - Presentation at HP's HPC & OSL TES
David Loureiro - Presentation at HP's HPC & OSL TES
 
Beyond The Data Grid: Coherence, Normalisation, Joins and Linear Scalability
Beyond The Data Grid: Coherence, Normalisation, Joins and Linear ScalabilityBeyond The Data Grid: Coherence, Normalisation, Joins and Linear Scalability
Beyond The Data Grid: Coherence, Normalisation, Joins and Linear Scalability
 
Building FoundationDB
Building FoundationDBBuilding FoundationDB
Building FoundationDB
 
Wolfgang Lehner Technische Universitat Dresden
Wolfgang Lehner Technische Universitat DresdenWolfgang Lehner Technische Universitat Dresden
Wolfgang Lehner Technische Universitat Dresden
 
Learning from google megastore (Part-1)
Learning from google megastore (Part-1)Learning from google megastore (Part-1)
Learning from google megastore (Part-1)
 
Introduction to map reduce
Introduction to map reduceIntroduction to map reduce
Introduction to map reduce
 
Introducing MongoDB into your Organization
Introducing MongoDB into your OrganizationIntroducing MongoDB into your Organization
Introducing MongoDB into your Organization
 
Getting Started with MongoDB at Oracle Open World 2012
Getting Started with MongoDB at Oracle Open World 2012Getting Started with MongoDB at Oracle Open World 2012
Getting Started with MongoDB at Oracle Open World 2012
 
Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013
Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013
Kafka 0.8.0 Presentation to Atlanta Java User's Group March 2013
 
John adams talk cloudy
John adams   talk cloudyJohn adams   talk cloudy
John adams talk cloudy
 
The elephantintheroom bigdataanalyticsinthecloud
The elephantintheroom bigdataanalyticsinthecloudThe elephantintheroom bigdataanalyticsinthecloud
The elephantintheroom bigdataanalyticsinthecloud
 

Plus de Michael Kopp

JAX 2011 - Garbage collection verstehen
JAX 2011 - Garbage collection verstehenJAX 2011 - Garbage collection verstehen
JAX 2011 - Garbage collection verstehenMichael Kopp
 
Leveraging your hadoop cluster better - running performant code at scale
Leveraging your hadoop cluster better - running performant code at scaleLeveraging your hadoop cluster better - running performant code at scale
Leveraging your hadoop cluster better - running performant code at scaleMichael Kopp
 
Application Performance Management in the Clouds - Lessons Learned
Application Performance Management in the Clouds - Lessons LearnedApplication Performance Management in the Clouds - Lessons Learned
Application Performance Management in the Clouds - Lessons LearnedMichael Kopp
 
SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...
SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...
SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...Michael Kopp
 
Releasing fast code - The DevOps approach
Releasing fast code - The DevOps approachReleasing fast code - The DevOps approach
Releasing fast code - The DevOps approachMichael Kopp
 
What does performance mean in the cloud
What does performance mean in the cloudWhat does performance mean in the cloud
What does performance mean in the cloudMichael Kopp
 

Plus de Michael Kopp (6)

JAX 2011 - Garbage collection verstehen
JAX 2011 - Garbage collection verstehenJAX 2011 - Garbage collection verstehen
JAX 2011 - Garbage collection verstehen
 
Leveraging your hadoop cluster better - running performant code at scale
Leveraging your hadoop cluster better - running performant code at scaleLeveraging your hadoop cluster better - running performant code at scale
Leveraging your hadoop cluster better - running performant code at scale
 
Application Performance Management in the Clouds - Lessons Learned
Application Performance Management in the Clouds - Lessons LearnedApplication Performance Management in the Clouds - Lessons Learned
Application Performance Management in the Clouds - Lessons Learned
 
SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...
SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...
SLAs and Performance in the Cloud: Because There is More Than "Just" Availabi...
 
Releasing fast code - The DevOps approach
Releasing fast code - The DevOps approachReleasing fast code - The DevOps approach
Releasing fast code - The DevOps approach
 
What does performance mean in the cloud
What does performance mean in the cloudWhat does performance mean in the cloud
What does performance mean in the cloud
 

Dernier

Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...BookNet Canada
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...
Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...
Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...Nikki Chapple
 
Landscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdfLandscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdfAarwolf Industries LLC
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...Wes McKinney
 
Accelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with PlatformlessAccelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with PlatformlessWSO2
 
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...itnewsafrica
 
Kuma Meshes Part I - The basics - A tutorial
Kuma Meshes Part I - The basics - A tutorialKuma Meshes Part I - The basics - A tutorial
Kuma Meshes Part I - The basics - A tutorialJoão Esperancinha
 
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)Mark Simos
 
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...Jeffrey Haguewood
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesThousandEyes
 
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sectoritnewsafrica
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integrationmarketing932765
 
A Glance At The Java Performance Toolbox
A Glance At The Java Performance ToolboxA Glance At The Java Performance Toolbox
A Glance At The Java Performance ToolboxAna-Maria Mihalceanu
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...itnewsafrica
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesThousandEyes
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...itnewsafrica
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesBernd Ruecker
 

Dernier (20)

Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...
Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...
Microsoft 365 Copilot: How to boost your productivity with AI – Part two: Dat...
 
Landscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdfLandscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdf
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 
Accelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with PlatformlessAccelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with Platformless
 
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
 
Kuma Meshes Part I - The basics - A tutorial
Kuma Meshes Part I - The basics - A tutorialKuma Meshes Part I - The basics - A tutorial
Kuma Meshes Part I - The basics - A tutorial
 
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
 
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
 
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
 
A Glance At The Java Performance Toolbox
A Glance At The Java Performance ToolboxA Glance At The Java Performance Toolbox
A Glance At The Java Performance Toolbox
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architectures
 

Performance Management in ‘Big Data’ Applications

  • 1. Performance Management in ‘Big Data’ Applications It’s still about the Application Michael Kopp, Technology Strategist Edward Capriolo michael.kopp@compuware.com edward@m6d.com @mikopp @edwardcapriolo blog.dynatrace.com m6d.com/blog
  • 2. BigData High Volume/Low Latency DBs Web Java Key Challenges Key Benefits 1) Even Distribution 1) Fast Read/Write 2) Correct Schema and Access patterns 2) Horizontal Scalability 3) Understanding Application Impact 3) Redundancy and High Availability 3
  • 3. BigData Large Parallel Batch Processing Hive high-level map/reduce JOB query JOB 1 2 1 batch 3 2 trigger 4 3 . . . Key Challenges Hive Server 754 Key Benefits 1) Optimal Distribution 1) Massive Horizontal Batch Job 2) Unwieldy Configuration 2) Split big Problems into smaller ones 3) Can easily waste your resources
  • 6. Map Reduce Performance 7
  • 8. Hadoop at m6d • Critical piece of infrastructure • Long Term Data Storage – Raw logs – Aggregations – Reports – Generated data (feed back loops) • Numerous ETL (Extract Transform Load) • Scheduled and adhoc processes • Used directly by Tech-Team, Ad Ops, Data Science 9
  • 9. Hadoop at m6d • Two deployments 'production' and 'research' – ~ 500 TB - 40+ Nodes – ~ 350 TB – 20+ Nodes • Thousands of jobs – <5 minute jobs and 12 hour Job Flows – Mostly Hive Jobs – Some custom code and streaming jobs
  • 10. Hadoop Design Tenants • Linear scalability by adding more hardware • HDFS Distributed file system – User space file system – Blocks are replicated across nodes – Limited semantics • MapReduce – Paradigm which models using map/reduce – Data Locality – Split Job into Tasks by Data – Retry in failure
  • 11. Schema Design Challenges • Partition data for good distribution – By time interval (optionally a second level) • Partition pruning with WHERE – Clustering (aka bucketing) • Optimized sampling and joins – Columnar • Column oriented • Raw Data Growth • Data features change (more distinct X) 12
  • 12. Key Performance Challenges • Intermediate I/O – Compression codec – Block size – Split-table formats • Contentions between jobs • Data and Map/Reduce Distribution • Data Skew • Non Uniform Computation (long running tasks) • ‘Cost' of new feature – is this justified? • Tuning variables (spills, buffers, Etc, etc) 13
  • 13. How to handle Performance Issues? • Profile the Job / Query? – Who should do this? (DBA, Dev, Ops, DevOps , NoOps, Big Data Guru) – How should we do this? • Look at job run times day over day? • Look at code and micro-benchmark? • Collect Job Counters? • Upgrade often for latest performance features? • Investigate/purchase newer better hardware – More cores? RAM? 10G Ethernet? SSD Test Data is not like • Read blogs? Real Data 14
  • 14. But how to optimize the job itself? 15
  • 15. Understanding Map/Reduce Performance Attention Data Maximum Parallelism Volume! Actual Mapping Also your own Millions of Parallelism Code Executions!!! Attention Potential Choke Point! Maximum Reduce Parallelism Actual Reduce Also your own Parallelism 16 Code
  • 18. Map/Reduce behind the scenes Serialize De-Serialize and Serialize again Potentionally Inefficient Too Many Files, Same Key spread all over De-Serialize Expensive and Serialize Synchronous again Combine 19
  • 19. Map/Reduce Combine and Spill Performance 1) Pre Combine in Mapping Step 2) Avoid many intermediate files and combines 20
  • 20. Map/Reduce “Map” Performance Avoid Brute Force Then on Big Hotspots FocusOptimize Hadoop Save a lot of Hardware 21
  • 21. Map/Reduce to the Max! • Ensure Data Locality • Optimize Map/Reduce Hotspots • Reduce Intermediate Data and “Overhead” • Ensure optimal Data and Compute Distribution • Tune Hadoop Environment 22
  • 22. Cassandra and Application Performance 23
  • 23. A High Level look at RTB 1. Browsers visit Publishers and create impressions. 2. Publishers sell impressions via Exchanges. 3. Exchanges serve as auction houses for the impressions 4. On behalf of the marketer, m6d bids the impressions via the auction house. If m6d wins, we display our ad to the browser. 24
  • 24. Cassandra at m6d for Real Time Bidding • RTB limited data is provided from exchange • System to store information on users – Frequency Capping – Visit History – Segments (product service affinity) • Low latency Requirements – Less then 100ms – Requires fast read/write on discrete data 25
  • 26. Key Cassandra Design Tennents • Swap/paging not possible • Mostly schema-less • Writes do not read – Read/Write is an anti-pattern • Optimize around put and get – Not for scan and query • De-Normalize data – Attempt to get all data in single read*
  • 27. Cassandra Design Challenges • De-normailize – Store data to optimize reads – Composite (multi-column) keys • Multi-column family and Multi-tenant scenarios • Compress settings – Disk and cache savings – CPU and JVM costs • Data/Compaction settings – Size tiered vs LevelDB • Caching, Memtable and other tuning 28
  • 28. How to handle performance issues? • Monitor standard vitals (cpu,disk) ? • Read blogs and documentation? • Use Cassandra JMX to track req/sec • Use Cassandra JMX to track size of Column Families, rows and columns • Upgrade often to get latest performance enhancements? * What about the Application? 29
  • 30. NoSQL APM is not so different after all… Web Java Database Key APM Problems Identified 1) Response Time Contribution 2) data access patterns 3) transaction to query relationship (transaction flow) 31
  • 31. Response Time Contribution Contribution to Business Transaction Connection Pool Access Pattern 32
  • 32. Statement Analysis Executions per Average and Total Contribution to Transactions and Business Transaction Execution Time Total 33
  • 33. Where, Why, How and which Transaction… Which Business Transaction Which Web Service Where and why in my Transaction Single Statement Performance 34
  • 34. How does this apply to NoSQL Databases? Web Java Key APM Problems Identified 1) Response Time Contribution 1) Data Access Distribution 2) data access patterns 2) End-to-End Monitoring 3) transaction to query 3) Storage (I/O, GC) Bottlenecks relationship (transaction flow) 4) Consistency Level 35
  • 35. Real End-to-End Application Performance Our Application Third Party External End User Services End User Response Time Contribution 37
  • 36. Understanding Cassandra’s Contribution Which statements did the Transaction Execute? Which node where they executed against? Contribution of each many calls? Too Statment Data Access patterns Which Consistency Level was used? 38
  • 37. Understand Response Time Contribution 5 Calls 4 Calls ~50-80 ms ~15 ms Contribution Contribution? Access and Data Distribution 39
  • 38. Why and how was a statement executed? 45ms latency? 60ms waiting on the server? 40
  • 39. Any Hotspots on the Cassandra Nodes? Much more load on Node3? Which Transactions are responsible 41
  • 41. General Health of Cassandra Memory Issues? Too many requests? Too much GC Suspensions? 43
  • 43. Extend Performance Focus on Application Web Java A Fast Database doesn’t make a fast Application 45
  • 44. Intelligent MapReduce APM data/task node Hive high-level map/reduce JOB data/task node query JOB 1 2 batch 1 3 master node 2 trigger 4 3 . . Hive Server . 754 data/task node Simple Optimizations with big impact
  • 45. Big Data is about solving Application Problems APM is about Application Performance and Efficiency 47
  • 46. THANK YOU Michael Kopp, Technology Strategist Edward Capriolo michael.kopp@compuware.com edward@m6d.com @mikopp @edwardcapriolo 48 blog.dynatrace.com m6d.com/blog

Notes de l'éditeur

  1. Map/Reduce Problem PatternsUneven DistributionOptimal splittingOptimizing Design Choke Point (between map and reduce)Complex Jobs aka. Hive QueriesData Locality (a customer of ours has the problem that while the job itself is distributed the data comes from only 3 HBase/Data Nodes)Too Many HBase CallsWasteful Job Code aka. Add more hardware instead of fixing hotspotsPremarture flushing  (see http://blog.dynatrace.com/2012/01/25/about-the-performance-of-map-reduce-jobs/) Cassandra/NoSQLThe theme here will be that from and App point of view the problem patterns haven’t really changes, but we actually have additional onesToo many callsToo much data readNon optimal data accessData driven Locking issuesSlow QueriesUneven DistributionUsing wrong Consistency LevelSlower NodesI/O IssuesGC?
  2. Purepath is the only solution spanning client and server (or edge and cloud)Keynote has no Real User MonitoringAppDynamics?New Relic
  3. Done by MikeExplain NoSQL on a high levelExplain Key Benefits and Key Challenges
  4. Done by MikeExplain MapReduce on a high levelExplain Key Benefits and Key Challenges
  5. Ed:Describe how hadoop works on a high levelDescribe a M6DUseCase as an example?Typical Performance Issues and why it is hard (different jobs different options)point towards developer and hive query, complicated, but most potentialMike:We are now starting doing things a little differentWhen you look at the typical Map/Reduce flow you’ll see the major parts, now we can monitor these areas for each jobTherefore we can decide on a job per job basis if we have one of the typical hadoop problems, or if it is worth our while to optimize things at the core of it, at the code level, and here we get pretty decent hotspots. After all when cutting down mapcode from 60 to 20% it helps a lot, after that it might be good enough or now that we spend most of our time in the framework it is time to look at hadoop itself againThe message however is, it’s the same as APM always has been, first identify on a Job basis if and what the problem is and then go for it, don’t just tune away, you’ll need an expert like Ed to go anywhere then.
  6. Ed:Describe how hadoop works on a high levelDescribe a M6DUseCase as an example?Typical Performance Issues and why it is hard (different jobs different options)point towards developer and hive query, complicated, but most potentialMike:We are now starting doing things a little differentWhen you look at the typical Map/Reduce flow you’ll see the major parts, now we can monitor these areas for each jobTherefore we can decide on a job per job basis if we have one of the typical hadoop problems, or if it is worth our while to optimize things at the core of it, at the code level, and here we get pretty decent hotspots. After all when cutting down mapcode from 60 to 20% it helps a lot, after that it might be good enough or now that we spend most of our time in the framework it is time to look at hadoop itself againThe message however is, it’s the same as APM always has been, first identify on a Job basis if and what the problem is and then go for it, don’t just tune away, you’ll need an expert like Ed to go anywhere then.
  7. Ed:Describe how hadoop works on a high levelDescribe a M6DUseCase as an example?Typical Performance Issues and why it is hard (different jobs different options)point towards developer and hive query, complicated, but most potentialMike:We are now starting doing things a little differentWhen you look at the typical Map/Reduce flow you’ll see the major parts, now we can monitor these areas for each jobTherefore we can decide on a job per job basis if we have one of the typical hadoop problems, or if it is worth our while to optimize things at the core of it, at the code level, and here we get pretty decent hotspots. After all when cutting down mapcode from 60 to 20% it helps a lot, after that it might be good enough or now that we spend most of our time in the framework it is time to look at hadoop itself againThe message however is, it’s the same as APM always has been, first identify on a Job basis if and what the problem is and then go for it, don’t just tune away, you’ll need an expert like Ed to go anywhere then.
  8. Done by MikeExplain MapReduce on a high levelExplain Key Benefits and Key Challenges
  9. Ed:Describe how hadoop works on a high levelDescribe a M6DUseCase as an example?Typical Performance Issues and why it is hard (different jobs different options)point towards developer and hive query, complicated, but most potentialMike:We are now starting doing things a little differentWhen you look at the typical Map/Reduce flow you’ll see the major parts, now we can monitor these areas for each jobTherefore we can decide on a job per job basis if we have one of the typical hadoop problems, or if it is worth our while to optimize things at the core of it, at the code level, and here we get pretty decent hotspots. After all when cutting down mapcode from 60 to 20% it helps a lot, after that it might be good enough or now that we spend most of our time in the framework it is time to look at hadoop itself againThe message however is, it’s the same as APM always has been, first identify on a Job basis if and what the problem is and then go for it, don’t just tune away, you’ll need an expert like Ed to go anywhere then.