From f6bc66f298c39076e66504d2884bf5278e8e1f96 Mon Sep 17 00:00:00 2001 From: Crista Lopes Date: Fri, 27 Dec 2013 14:52:04 -0800 Subject: [PATCH] Simplification of map-reduce, hadoop --- 30-double-inverse-multiplexer/tf-30.py | 68 +++++++------------------- 1 file changed, 17 insertions(+), 51 deletions(-) diff --git a/30-double-inverse-multiplexer/tf-30.py b/30-double-inverse-multiplexer/tf-30.py index 039fe2e..b6d112a 100755 --- a/30-double-inverse-multiplexer/tf-30.py +++ b/30-double-inverse-multiplexer/tf-30.py @@ -1,5 +1,4 @@ #!/usr/bin/env python - import sys, re, operator, string # @@ -7,7 +6,7 @@ import sys, re, operator, string # def partition(data_str, nlines): """ - Generator function that partitions the input data_str (a big string) + Partitions the input data_str (a big string) into chunks of nlines. """ lines = data_str.split('\n') @@ -16,55 +15,35 @@ def partition(data_str, nlines): def split_words(data_str): """ - Takes a string, filters non alphanumeric characters, normalizes to - lower case, scans for words, and filters the stop words. - It returns a list of pairs (word, 1), one for each word in the input, so + Takes a string, returns a list of pairs (word, 1), + one for each word in the input, so [(w1, 1), (w2, 1), ..., (wn, 1)] """ - def _filter_chars(str_data): - """ - Takes a string and returns a copy with all nonalphanumeric chars - replaced by white space - """ - pattern = re.compile('[\W_]+') - return pattern.sub(' ', str_data) - - def _normalize(str_data): - """ - Takes a string and returns a copy with all characters in lower case - """ - return str_data.lower() - def _scan(str_data): - """ - Takes a string and scans for words, returning - a list of words. - """ - return str_data.split() + pattern = re.compile('[\W_]+') + return pattern.sub(' ', str_data).lower().split() def _remove_stop_words(word_list): - f = open('../stop_words.txt') - stop_words = f.read().split(',') - f.close() - # add single-letter words + with open('../stop_words.txt') as f: + stop_words = f.read().split(',') stop_words.extend(list(string.ascii_lowercase)) return [w for w in word_list if not w in stop_words] # The actual work of the mapper result = [] - words = _remove_stop_words(_scan(_normalize(_filter_chars(data_str)))) + words = _remove_stop_words(_scan(data_str)) for w in words: result.append((w, 1)) return result def regroup(pairs_list): """ - Takes a list of a list of pairs of the form + Takes a list of lists of pairs of the form [[(w1, 1), (w2, 1), ..., (wn, 1)], [(w1, 1), (w2, 1), ..., (wn, 1)], ...] - and returns a dictionary mapping each unique word to the corresponding - list of pairs, so + and returns a dictionary mapping each unique word to the + corresponding list of pairs, so { w1 : [(w1, 1), (w1, 1)...], w2 : [(w2, 1), (w2, 1)...], ...} @@ -81,38 +60,25 @@ def regroup(pairs_list): def count_words(mapping): """ Takes a mapping of the form (word, [(word, 1), (word, 1)...)]) - and returns a pair (word, frequency), where frequency is the sum - of all the reported occurrences + and returns a pair (word, frequency), where frequency is the + sum of all the reported occurrences """ def add(x, y): return x+y return (mapping[0], reduce(add, (pair[1] for pair in mapping[1]))) - # # Auxiliary functions # - def read_file(path_to_file): - """ - Takes a path to a file and returns the entire - contents of the file as a string - """ - f = open(path_to_file) - data = f.read() - f.close() + with open(path_to_file) as f: + data = f.read() return data def sort(word_freq): - """ - Takes a collection of words and their frequencies - and returns a collection of pairs where the entries are - sorted by frequency - """ return sorted(word_freq, key=operator.itemgetter(1), reverse=True) - # # The main function # @@ -120,6 +86,6 @@ splits = map(split_words, partition(read_file(sys.argv[1]), 200)) splits_per_word = regroup(splits) word_freqs = sort(map(count_words, splits_per_word.items())) -for tf in word_freqs[0:25]: - print tf[0], ' - ', tf[1] +for (w, c) in word_freqs[0:25]: + print w, ' - ', c