skip to main | skip to sidebar

Yet Another Link Link

Wednesday, June 2, 2010

Hadoop, Sorting Terabyte Petabyte


  • [PDF] Large-scale social media analysis with Hadoop, 124 pages
  • Distributed data processing with Hadoop, Part 1: Getting startedInstall and configure a simple cluster
  • Distributed data processing with Hadoop, Part 2: Going further Install and configure a multinode cluster
  • Hadoop Sorts a Petabyte in 16.25 Hours and a Terabyte in 62 Seconds
  • Writing An Hadoop MapReduce Program In Python
  • A MapReduce Algorithm for Matrix Multiplication
  • Parallel Machine Learning for Hadoop/Mapreduce – A Python Example
Posted by chihungchan at 6:52 AM
Labels: Hadoop, MapReduce, Python

No comments:

Post a Comment

Newer Post Older Post Home
Subscribe to: Post Comments (Atom)

Followers

Blog Archive

  • ►  2011 (9)
    • ►  February (5)
    • ►  January (4)
  • ▼  2010 (37)
    • ►  November (3)
    • ►  October (6)
    • ►  September (2)
    • ►  August (4)
    • ►  July (7)
    • ▼  June (5)
      • Fixing script output, Facebook scalability
      • Solaris Auditing, Performance
      • UNIX Troubleshooting Tips
      • Automount via ssh
      • Hadoop, Sorting Terabyte Petabyte
    • ►  May (10)

About Me

My photo
chihungchan
The Scripting Guy in the Lion City
View my complete profile