Remote Data Infrastructure Engineer at Stripe 📈 Open Startup
RSS
API
Post a Job

get a remote job
you can do anywhere

The largest collection of Remote Jobs for Digital Nomads online. Get a remote job you can do anywhere at Remote Companies like Buffer, Zapier and Automattic who embrace the future. There are 32,300+ jobs that allow you to work anywhere and live everywhere.

The largest collection of Remote Jobs for Digital Nomads online. Get a remote job you can do anywhere at Remote Companies like Buffer, Zapier and Automattic who embrace the future. There are 32,300+ jobs that allow you to work anywhere and live everywhere.

  Jobs

  People

👉 Hiring for a remote Fintech position?

Post a Job - $299
on the 🏆 #1 remote jobs board

Stripe

 

Data Infrastructure Engineer

verified
North America

Data Infrastructure Engineer  


Stripe

North America verified

fintech

finance

engineer

fintech

finance

engineer

North America7mo
As a platform company powering businesses all over the world, Stripe processes payments, runs marketplaces, detects fraud, helps entrepreneurs start an internet business from anywhere in the world. Stripe’s Data Infrastructure Engineers build the platform, tooling, and pipelines that manage that data.\n\nAt Stripe, decisions are driven by data. Because every record in our data warehouse can be vitally important for the businesses that use Stripe, we’re looking for people with a strong background in big data systems to help us build tools to scale while maintaining correct and complete data. You’ll be creating best in class libraries to help our users fully leverage open source frameworks like Spark and Scalding. You’ll be working with a variety of teams, some engineering and some business, to provide tooling and guidance to solve their data needs. Your work will allow teams to move faster, and ultimately help Stripe serve our customers more effectively.\n\n\n\n# Responsibilities\n **You will:**\n* Create libraries and tooling that make distributed batch computation easy to create and test for all users across Stripe\n* Become expert in and contribute to open source frameworks such as Scalding, Spark to address issues our users at Stripe encounter\n* Create APIs to help teams materialize data models from production services into readily consumable formats for all downstream data consumption\n* Create libraries that enable engineers at Stripe to easily interact with various serialization frameworks (e.g. thrift, bson, protobuf)\n* Create observability tooling to help our users easily debug, understand, and tune their Spark / Scalding jobs\n* Leveraging batch computation frameworks and our workflow management platform (Airflow) to assist other teams in building out their data pipelines\n* Own and evolve the most critical upstream datasets \n\n# Requirements\n**We’re looking for someone who has:**\n* A strong engineering background and are interested in data. You’ll be writing production Scala and Python code.\n* Experience developing and maintaining distributed systems built with open source tools.\n* Experience building libraries and tooling that provide beautiful abstractions to users\n* Experience optimizing the end-to-end performance of distributed systems.\n* Experience in writing and debugging ETL jobs using a distributed data framework (Spark/Hadoop MapReduce etc…)\n\n**Nice to haves:**\n* Experience with Scala\n* Experience with Spark or Scalding\n* Experience with Airflow or other similar scheduling tools\n* It’s not expected that you’ll have deep expertise in every dimension above, but you should be interested in learning any of the areas that are less familiar.\n\n**Some things you might work on:**\n* Create libraries that enable engineers at Stripe to easily interact with various serialization frameworks (e.g. thrift, bson, protobuf)\n* Write a unified user data model that gives a complete view of our users across a varied set of products like Stripe Connect and Stripe Atlas\n* Continuing to lower the latency and bridge the gap between our production systems and our data warehouse by rethinking and optimizing our core data pipeline jobs\n* Pair with user teams to optimize and rewrite business critical batch processing jobs in Spark\n* Create robust and easy to use unit testing infrastructure for batch processing pipelines\n* Build a framework and tools to re-architect data pipelines to run more incrementally\n \n\n#Location\n- North America

See more jobs at Stripe

# How do you apply? This job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.
Apply for this Job

👉 Please reference you found the job on Remote OK, this helps us get more companies to post here!

When applying for jobs, you should NEVER have to pay to apply. That is a scam! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. Scams in remote work are rampant, be careful! When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.