Complexity Theory for MapReduce Algorithms
- Jeffrey Ullman | Stanford University
For many problems, algorithms implemented in MapReduce or similar two-ranks-of-parallel-tasks systems exhibit a tradeoff between memory size and communication. More precisely, the tradeoff is between “reducer size” (the number of inputs received by the second rank of parallel tasks) and the “replication rate” (the average number of key-value pairs generated by the first rank in response to a single input). We begin with the simple but common “all-pairs” problem, where there is some output associated with every pair of inputs. For this problem, the reducer size and replication rate are seen to vary inversely. We then look at the different relationships that exist between these two parameters for a number of other problems, including (dense) matrix multiplication, similarity joins, and computing marginals
-
-
Surajit Chaudhuri
Technical Fellow, Data Platforms and Analytics
-
-
Series: Microsoft Research Talks
-
Decoding the Human Brain – A Neurosurgeon’s Experience
- Dr. Pascal O. Zinn
-
-
-
-
-
-
Challenges in Evolving a Successful Database Product (SQL Server) to a Cloud Service (SQL Azure)
- Hanuma Kodavalla,
- Phil Bernstein
-
Improving text prediction accuracy using neurophysiology
- Sophia Mehdizadeh
-
Tongue-Gesture Recognition in Head-Mounted Displays
- Tan Gemicioglu
-
DIABLo: a Deep Individual-Agnostic Binaural Localizer
- Shoken Kaneko
-
-
-
-
Audio-based Toxic Language Detection
- Midia Yousefi
-
-
From SqueezeNet to SqueezeBERT: Developing Efficient Deep Neural Networks
- Forrest Iandola,
- Sujeeth Bharadwaj
-
Hope Speech and Help Speech: Surfacing Positivity Amidst Hate
- Ashique Khudabukhsh
-
-
-
Towards Mainstream Brain-Computer Interfaces (BCIs)
- Brendan Allison
-
-
-
-
Learning Structured Models for Safe Robot Control
- Subramanian Ramamoorthy
-