DATA ENGINEER | REDWOOD CITY, CA At Yewno, we help people research and understand the world. We build knowledge discovery software and services, leveraging machine learning, graph analytics, and some of the most important content providers in the world. If this excites you, we'd love for you to join us. We're a team of hands-on entrepreneurs, engineers and data scientists with track records of billion-dollar exits. We're alums from top-10 universities and industry leaders like Yahoo, Oracle, eBay, Accenture, BT, Springer Nature, and Merrill Lynch. We’re a well-funded startup with product-market fit. Our product is live at Stanford, Harvard, MIT, etc. We will continue to grow the education vertical, and apply our technology to finance and other verticals. We offer a fun working environment with sharp, motivated colleagues, and great benefits: •
Competitive salary and stock options
•
Company-covered healthcare
•
401(k)
•
Unlimited vacation policy
•
Build your dream workstation
•
Walking distance from the Redwood City Caltrain Station
THE ROLE Creating, designing and implementing the next generation of knowledge discovery tools and big data analytics. Successful candidates must have a strong background in distributed computing and experience in building and deploying big data solutions to production systems. A basic knowledge of machine learning algorithms and graph analysis would be highly beneficial.
KEY QUALIFICATIONS •
Master’s degree in Computer Science or related fields
-1-
•
4+ years experience working with big data
•
Experience with Hadoop 2.0 ecosystem (Spark, MapReduce), NoSQL databases, Amazon AWS and related technologies
•
Hands-on experience with real-time data ingestion/processing pipelines (Spark Streaming, Storm, and Apache Kafka)
•
Previous experience working with mission-critical systems
•
Proficiency in Scala, Python and Go
•
Basic knowledge of state-of-the-art algorithms in machine learning and graph analysis would be a plus
CORE RESPONSIBILITIES As part of our Data Science team, you will design, build, test and maintain scalable data processing pipelines. You will be deploying machine learning algorithms to big data platforms, both as batch jobs and real-time processes. You will closely interact and coordinate with the Engineering team. If you are interested in applying, send your CV to
[email protected].
-2-