hey all, just getting started on hadoop and curious what the best way in mapreduce would be to count unique visitors if your logfiles looked like this...
DATE siteID action username
05-05-2010 siteA pageview jim
05-05-2010 siteB pageview tom
05-05-2010 siteA pageview jim
05-05-2010 siteB pageview bob
05-05-2010 siteA pageview mike
and for each site you wanted to find out the unique visitors for each site?
I was thinking the mapper would emit siteID \t username and the reducer would keep a set() of the unique usersnames per key and then emit the length of that set. However that would be potentially storing millions of usernames in memory which doesn't seem right. Anyone have a better way?
I'm using python streaming by the way
thanks
You could do it as a 2-stage operation:
First step, emit (username => siteID)
, and have the reducer just collapse multiple occurrences of siteID using a set
- since you'd typically have far less sites than users, this should be fine.
Then in the second step, you can emit (siteID => username)
and do a simple count, since the duplicates have been removed.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With