scala, spark, hadoop
Identifying Data Opportunities
How do you move multiple TB of data per hour and optimally compress and serialize it? How do you ensure we are resilient to data loss in case of failure? We are looking for outstanding engineers who love solving interesting challenges related to data at scale to help us answer questions like these.