👉 Hiring for a remote position?on the 🏆 #1 remote jobs board
Data Engineer Infrastructureverified closed
Data Engineer Infrastructure
This job post is closed and the position is probably filled. Please do not apply.
Why work at Doximity?\n\nDoximity is the leading social network for healthcare professionals with over 70% of U.S. doctors as members. We have strong revenues, real market traction, and we're putting a dent in the inefficiencies of our $2.5 trillion U.S. healthcare system. After the iPhone, Doximity is the fastest adopted product by doctors of all time. Our founder, Jeff Tangney, is the founder & former President and COO of Epocrates (IPO in 2010), and Nate Gross is the founder of digital health accelerator RockHealth. Our investors include top venture capital firms who've invested in Box, Salesforce, Skype, SpaceX, Tesla Motors, Twitter, Tumblr, Mulesoft, and Yammer. Our beautiful offices are located in SoMa San Francisco.\n\nYou will join a small team of data infrastructure engineers (4) to build and maintain all aspects of our data pipelines, ETL processes, data warehousing, ingestion and overall data infrastructure. We have one of the richest healthcare datasets in the world, and we're not afraid to invest in all things data to enhance our ability to extract insight.\n\nJob Summary\n\n-Help establish robust solutions for consolidating data from a variety of data sources.\n-Establish data architecture processes and practices that can be scheduled, automated, replicated and serve as standards for other teams to leverage. \n-Collaborate extensively with the DevOps team to establish best practices around server provisioning, deployment, maintenance, and instrumentation.\n-Build and maintain efficient data integration, matching, and ingestion pipelines.\n-Build instrumentation, alerting and error-recovery system for the entire data infrastructure.\n-Spearhead, plan and carry out the implementation of solutions while self-managing.\n-Collaborate with product managers and data scientists to architect pipelines to support delivery of recommendations and insights from machine learning models.\n\nRequired Experience & Skills\n\n-Fluency in Python, SQL mastery.\n-Ability to write efficient, resilient, and evolvable ETL pipelines. \n-Experience with data modeling, entity-relationship modeling, normalization, and dimensional modeling.\n-Experience building data pipelines with Spark and Kafka.\n-Comprehensive experience with Unix, Git, and AWS tooling.\n-Astute ability to self-manage, prioritize, and deliver functional solutions.\n\nPreferred Experience & Skills\n\n-Experience with MySQL replication, binary logs, and log shipping.\n-Experience with additional technologies such as Hive, EMR, Presto or similar technologies.\n-Experience with MPP databases such as Redshift and working with both normalized and denormalized data models.\n-Knowledge of data design principles and experience using ETL frameworks such as Sqoop or equivalent. \n-Experience designing, implementing and scheduling data pipelines on workflow tools like Airflow, or equivalent.\n-Experience working with Docker, PyCharm, Neo4j, Elasticsearch, or equivalent. \n\nOur Data Stack\n\n-Python, Kafka, Spark, MySQL, Redshift, Presto, Airflow, Neo4j, Elasticsearch\n\nFun Facts About the Team\n\n-We have one of the richest healthcare datasets in the world.\n-Business decisions at Doximity are driven by our data, analyses, and insights.\n-Hundreds of thousands of healthcare professionals will utilize the products you build.\n-Our R&D team makes up about half the company, and the product is led by the R&D team. \n-Our Data Science team is comprised of about 20 people.
See more jobs at Doximity
Visit Doximity's website
# How do you apply?\n\n This job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.