Googlebot data processing
1. Googlebot 'crawler' programs visit billions of websites, recording the destinations of every link on every page
2. Blocks of Googlebot data arrive and are broken into equally-sized chunks
3. Framework distributes multiple copies of every block amongst nodes in the cluster
4. Map procedure compiles long lists of every link on every page in the data block
5. Shuffling sorts the list of links into alphabetical order of the destination page's address
6. Reducing calculates the total number of links pointing to a web page