I don't think it has a depth sensor. That blur could have easily been applied in software based on the area where the person is detected...
Well, the blur *is* applied in software, however I believe it uses a depth sensor to aid in detecting the person and applying the blur, for three reasons:
1. Tech Crunch says "The camera has a built-in depth sensor to blur out everything in the photo."
2. Amazon says "The built-in LED lighting and depth-sensing camera let you blur the background..."
3. The amount and area of blurring in the photo is consistent with something being blurred based on distance, not merely the shape of the person detected.
It *could be* that Amazon means video analytics style "depth-sensing", and that tech crunch mistakenly assumed it was a "sensor", and that the picture is purposely blurred by distance, not by body outline, for aesthetics.
But given the recent cheapness of these sensors, and the possible benefits to judging a person in 3 dimensions, I'm thinking its got one.