FeedbackIf you find a bug, or have feedback, put it here. Please no job applications in here, click Apply on the job instead.Thanks for the message! We will get back to you soon.

[Spam check] What is the name of Elon Musk's company going to Mars?

Send feedback
Open Startup
RSS
API
Health InsurancePost a job

find a remote job
work from anywhere

Get a  email of all new Remote Software Developer + Terraform Jobs

Subscribe
×

👉 Hiring for a Remote Software Developer + Terraform position?

Post a job
on the 🏆 #1 Remote Jobs board

Remote Health by SafetyWing


Global health insurance for freelancers & remote workers

🤓 Software Developer Remove this filter🌳 Terraform Remove this filter
Clear 11 results

Clevertech is hiring a Remote Senior DevOps Engineer

Your role as a DevOps Engineer at Clevertech will actively contribute to reliability and scalability. You will work alongside some of the best in a collaborative environment while focusing on your core skills. Be a master of your craft while being 100% remote and never have to worry about filling in timesheets.\n\n**Requirements**\n* 7+ years of professional experience (A technical assessment will be required)\n* Senior-level experience with AWS (EC2, RDS, S3, ECS, ELB)\n* Strong background in Linux and Mongo Atlas administration\n* Experience deploying Kubernetes in a production environment\n* Experience with CI/CD in Jenkins or CircleCi\n* Infrastructure as code (we use Terraform)\n* Experience with requirement gathering and presentation to executives\n* English fluency, verbal and written\n* Personality traits: Professional, problem solver, proactive, attention to detail.\n \n\n#Salary and compensation\n$50,000 — $100,000/year\n\n\n#Location\n🌏 Worldwide


See more jobs at Clevertech

Chainlink Labs


closed
🌏 Worldwide
 
💰 $100k - $200k

smart contracts

 

golang

 

typescript

This job post is closed and the position is probably filled. Please do not apply.
**All roles with Chainlink Labs are globally remote based. We encourage you to apply regardless of your location.**\n\nAs a Solutions Engineer, you will lead integrations with upcoming blockchains into the Chainlink platform, working directly, and building relationships with the founding engineers while learning about each new innovative blockchain platform. Today, most blockchains are their own ecosystems, but this is about to change: cross-blockchain connectivity is set to be the next growth opportunity, with network effects compounding when leveraging the strengths of each blockchain. You will be the driving force for owning this initiative and will make a large impact on the growth of a critical piece of blockchain infrastructure.\n\n**Your Impact**\n\n* Interface with leadership, customers, users, and the community to implement seamless integrations to allow for the exponential growth of the network\n* Lead cutting edge blockchain platform integrations\n* Manage the technical aspects of integrating Chainlink with blockchains\n* Build token bridges between blockchains utilizing the Chainlink network\n* Work closely with node operators to connect their Chainlink nodes to other blockchains\n* Develop best practices and documentation to scale integrations\n\n**Requirements**\n\n* 2+ years of professional experience working in a product-driven environment\n* Experience in Golang or Rust; otherwise experience in JS, Ruby, Python, C++, C, or Solidity and a willingness to learn\n* Resourceful developer and fast learner continually striving to discover more environments\n* Ability to write and communicate effectively\n* Eagerness to work openly and collaboratively with a diverse team\n* Experience in blockchain and other Web 3.0 technologies\n* Comfort and/or experience working as a remote employee as a part of a distributed team\n\n**Our Stack**\n\nGolang, TypeScript, Solidity, Postgres, Terraform, AWS\n\n**About Us**\n\nChainlink is the industry standard oracle network for connecting smart contracts to the real world. With Chainlink, developers can build hybrid smart contracts that combine on-chain code with an extensive collection of secure off-chain services powered by Decentralized Oracle Networks. Managed by a global, decentralized community of hundreds of thousands of people, Chainlink is introducing a fairer model for contracts. Its network currently secures billions of dollars in value for smart contracts across the decentralized finance (DeFi), insurance, and gaming ecosystems, among others. The full vision of the Chainlink Network can be found in the [Chainlink 2.0 whitepaper](https://research.chain.link/whitepaper-v2.pdf). Chainlink is trusted by hundreds of organizations—from global enterprises to projects at the forefront of the blockchain economy—to deliver definitive truth via secure, reliable data. \n\nThis role is location agnostic anywhere in the world, but we ask that you overlap some working hours with Eastern Standard Time (EST).\n\nWe are a fully distributed team and have the tools and benefits to support you in your remote work environment.\n\nChainlink Labs is an Equal Opportunity Employer. \n\n#Salary and compensation\n$100,000 — $200,000/year\n\n\n#Location\n🌏 Worldwide


See more jobs at Chainlink Labs

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.

Chainlink Labs

 This job is getting a pretty high amount of applications right now (12% of viewers clicked Apply)

closed
🌏 Worldwide
 
💰 $100k - $200k

smart contracts

 

golang

 

typescript

This job post is closed and the position is probably filled. Please do not apply.
**All roles with Chainlink Labs are globally remote based. We encourage you to apply regardless of your location.**\n\nIn this quality assurance role, you will oversee the design, development, and implementation of our test automation framework. You will have a large impact as we grow the Chainlink eco-system and ensure we continue to ship the highest quality product. As a key stakeholder, you will be able to influence how we will scale our product in a testable environment and guide decisions that will have an impact on the future and success of a cutting edge product. \n\n**Your Impact**\n\n* Lead the implementation and adoption of a test automation framework, improved release process, and overall testing plans\n* Oversee and develop test cases, plans, and testing procedures\n* Create automated tests for new and existing functionality\n* Develop test plans and perform multiple types of black-box, white-box, and grey-box testing including functional, system, end to end, load testing, performance testing, data integrity, smoke, regression, and ad hoc testing\n* Submit comprehensive defects using our bug tracking system\n* Regress and verify defects as needed, record test results, and provide metrics\n* Test software applications at all points in the life cycle on multiple environments\n\n**Requirements**\n\n* 2+ years proven work experience in Software Engineering, Test, Quality Assurance, or Deployment\n* BS degree or equivalent work experience in Computer Science or related field\n* Strong experience with modern languages such as JavaScript, TypeScript, Golang, etc.\n* Strong experience with open-source test frameworks, including CypressJS or Selenium/WebDriver\n* Strong experience with build automation technologies like Jenkins, CircleCI, GitHub Actions, etc. \n* Experience with AWS, Docker, or Terraform a definite plus\n* Experience with Agile-Scrum and other Agile methodologies\n* Experience with load testing tools\n* Excellent analytical skills to work with developers for troubleshooting and peer review\n* Effective written and verbal communications skills\n* The ideal candidate will be comfortable with shaping testing culture and socializing best practices\n\n**Our Stack**\n\nGolang, TypeScript, Solidity, Postgres, Terraform, AWS\n\n**About Us**\n\nChainlink is the industry standard oracle network for connecting smart contracts to the real world. With Chainlink, developers can build hybrid smart contracts that combine on-chain code with an extensive collection of secure off-chain services powered by Decentralized Oracle Networks. Managed by a global, decentralized community of hundreds of thousands of people, Chainlink is introducing a fairer model for contracts. Its network currently secures billions of dollars in value for smart contracts across the decentralized finance (DeFi), insurance, and gaming ecosystems, among others. The full vision of the Chainlink Network can be found in the [Chainlink 2.0 whitepaper](https://research.chain.link/whitepaper-v2.pdf). Chainlink is trusted by hundreds of organizations—from global enterprises to projects at the forefront of the blockchain economy—to deliver definitive truth via secure, reliable data. \n\nThis role is location agnostic anywhere in the world, but we ask that you overlap some working hours with Eastern Standard Time (EST).\n\nWe are a fully distributed team and have the tools and benefits to support you in your remote work environment.\n\nChainlink Labs is an Equal Opportunity Employer.\n \n\n#Salary and compensation\n$100,000 — $200,000/year\n\n\n#Location\n🌏 Worldwide


See more jobs at Chainlink Labs

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.

Shopify


verified closed
United States, Canada

staff software developer

 

data platform engineering

 

data engineering

 

spark

This job post is closed and the position is probably filled. Please do not apply.
**Company Description**\n\nShopify is the leading omni-channel commerce platform. Merchants use Shopify to design, set up, and manage their stores across multiple sales channels, including mobile, web, social media, marketplaces, brick-and-mortar locations, and pop-up shops. The platform also provides merchants with a powerful back-office and a single view of their business, from payments to shipping. The Shopify platform was engineered for reliability and scale, making enterprise-level technology available to businesses of all sizes. \n\n**Job Description**\n\nOur Data Platform Engineering group builds and maintains the platform that delivers accessible data to power decision-making at Shopify for over a million merchants. We’re hiring high-impact developers across teams:\n\n* The Engine group organizes all merchant and Shopify data into our data lake in highly-optimized formats for fast query processing, and maintaining the security and quality of our datasets.\n* The Analytics group leverages the Engine primitives to build and deliver simple and useful products that power scalable transformation of data at Shopify in batch, streaming, or for machine learning. This group is focused on making it really simple for our users to answer three questions: What happened in the past? What is happening now? And, what will happen in the future? \n* The Data Experiences group builds end-user experiences for experimentation, data discovery, and business intelligence reporting.\n* The Reliability group operates the data platform in a consistent and reliable manner. They build tools for other teams on Data Platform to leverage and encourage consistency as they champion reliability across the platform.\n\n**Qualifications**\n\n* An experienced technical leader with a proven track record of delivering impactful results.\n* Technical engineering background in one or more areas in the next section.\n* Experience with technical mentoring, coaching, and improving the technical output of the people around you.\n* Exceptional communication skills and ability to translate technical concepts into easy to understand language for our stakeholders. \n* Excitement for working with a remote team; you value collaborating on problems, asking questions, delivering feedback, and supporting others in their goals whether they are in your vicinity or entire cities apart.\n\n**A Staff Data Developer would typically have 6-10 years of experience in one or more of the following areas:**\n\n* Experience with the internals of a distributed compute engine (Spark, Presto, DBT, or Flink/Beam)\n* Experience in query optimization, resource allocation and management, and data lake performance (Presto, SQL)\n* Experience with cloud infrastructure (Google Cloud, Kubernetes, Terraform\n* Experience with security products and methods (Apache Ranger, Apache Knox, OAuth, IAM, Kerberos)\n* Experience deploying and scaling ML solutions using open-source frameworks (MLFlow, TFX, H2O, etc.)\n* Experience building full-stack applications (Ruby/Rails, React, TypeScript)\n* Background and practical experience in statistics and/or computational mathematics (Bayesian and Frequentist approaches, NumPy, PyMC3, etc.)\n* Modern Big-Data storage technologies (Iceberg, Hudi, Delta)\n\n**Additional information**\nAt Shopify, we are committed to building and fostering an environment where our employees feel included, valued, and heard. Our belief is that a strong commitment to diversity and inclusion enables us to truly make commerce better for everyone. We strongly encourage applications from Indigenous people, racialized people, people with disabilities, people from gender and sexually diverse communities and/or people with intersectional identities.\n\n#Location\nUnited States, Canada


See more jobs at Shopify

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.
This job post is closed and the position is probably filled. Please do not apply.
Doximity is transforming the healthcare industry. Our mission is to help doctors be more productive, informed, and connected. Achieving this vision requires a multitude of disciplines, expertises and perspective. One of our core pillars have always been data. As a software engineer focused on the infrastructure aspect of our data stack you will work on improving healthcare by advancing our data capabilities, best practices and systems. Our team brings a diverse set of technical and cultural backgrounds and we like to think pragmatically in choosing the tools most appropriate for the job at hand.\n\n**About Us**\n\nOur data teams schedule over 1000 Python pipelines and over 350 Spark pipelines every 24 hours, resulting in over 5000 data processing tasks each day. Additionally, our data endeavours leverage datasets ranging in size from a few hundred rows to a few hundred billion rows. The Doximity data teams rely heavily on Python3, Airflow, Spark, MySQL, and Snowflake. To support this large undertaking, the data infrastructure team uses AWS, Terraform, and Docker to manage a high-performing and horizontally scalable data stack. The data infrastructure team is responsible for enabling and empowering the data analysts, machine learning engineers and data engineers at Doximity. We provide and evole a foundation on which to build, and ensure that incidental complexites melt into our abstractions. Doximity has worked as a distributed team for a long time; pre-pandemic, Doximity was already about 65% distributed.\n\nFind out more information on the Doximity engineering blog\n* Our [company core values](https://work.doximity.com/)\n* Our [recruiting process](https://technology.doximity.com/articles/engineering-recruitment-process-doximity)\n* Our [product development cycle](https://technology.doximity.com/articles/mofo-driven-product-development)\n* Our [on-boarding & mentorship process](https://technology.doximity.com/articles/software-engineering-on-boarding-at-doximity)\n\n**Here's How You Will Make an Impact**\n\nAs a data infrastructure engineer you will work with the rest of the data infrastructure team to design, architect, implement, and support data infrastructure, systems, and processes impacting all other data teams at Doximity. You will solidify our CI/CD pipelines, reduce production impacting issues and improve monitoring and logging. You will support and train data analysts, machine learning engineers, and data engineers on new or improved data infrastructure systems and processes. A key responsibility is to encourage data best-practices through code by continuing the development of our internal data frameworks and libraries. Also, it is your responsibility to identify and address performance, scaling, or resource issues before they impact our product. You will spearhead, plan, and carry out the implementation of solutions while self-managing your time and focus.\n\n**About you**\n\n* You have professional data engineering or operations experience with a focus on data infrastructure\n* You are fluent in Python and SQL, and feel at home in a remote Linux server session\n* You have operational experience supporting data stacks through tools like Terraform, Docker, and continuous integration through tools like CircleCI\n* You are foremost an engineer, making you passionate about high code quality, automated testing, and engineering best practices\n* You have the ability to self-manage, prioritize, and deliver functional solutions\n* You possess advanced knowledge of Linux, Git, and AWS (EMR, IAM, VPC, ECS, S3, RDS Aurora, Route53) in a multi-account environment\n* You agree that concise and effective written and verbal communication is a must for a successful team\n\n**Benefits & Perks**\n\n* Generous time off policy\n* Comprehensive benefits including medical, vision, dental, generous paternity and maternity leave, Life/ADD, 401k, flex spending accounts, commuter benefits, equipment budget, and continuous education budget\n* Pre-IPO stock incentives\n* and much more! For a full list, see our career page\n\n**More info on Doximity**\n\nWe're thrilled to be named the Fastest Growing Company in the Bay Area, and one of Fast Company's Most Innovative Companies. Joining Doximity means being part of an incredibly talented and humble team. We work on amazing products that over 70% of US doctors (and over one million healthcare professionals) use to make their busy lives a little easier. We're driven by the goal of improving inefficiencies in our $3.5 trillion U.S. healthcare system and love creating technology that has a real, meaningful impact on people's lives. To learn more about our team, culture, and users, check out our careers page, company blog, and engineering blog. We're growing steadily, and there's plenty of opportunities for you to make an impact.\n\n*Doximity is proud to be an equal opportunity employer and committed to providing employment opportunities regardless of race, religious creed, color, national origin, ancestry, physical disability, mental disability, medical condition, genetic information, marital status, sex, gender, gender identity, gender expression, pregnancy, childbirth and breastfeeding, age, sexual orientation, military or veteran status, or any other protected classification. We also consider qualified applicants with criminal histories, consistent with applicable federal, state, and local law.*\n\n\n\n\n#Location\n🇺🇸 US


See more jobs at Doximity

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.
This job post is closed and the position is probably filled. Please do not apply.
# We're building the Data Platform of the Future\nJoin us if you want to rethink the way organizations interact with data. We are a **developer-first company**, committed to building around open protocols and delivering the best experience possible for data consumers and publishers.\n\nSplitgraph is a **seed-stage, venture-funded startup hiring its initial team**. The two co-founders are looking to grow the team to five or six people. This is an opportunity to make a big impact on an agile team while working closely with the\nfounders.\n\nSplitgraph is a **remote-first organization**. The founders are based in the UK, and the company is incorporated in both USA and UK. Candidates are welcome to apply from any geography. We want to work with the most talented, thoughtful and productive engineers in the world.\n# Open Positions\n**Data Engineers welcome!** The job titles have "Software Engineer" in them, but at Splitgraph there's a lot of overlap \nbetween data and software engineering. We welcome candidates from all engineering backgrounds.\n\n[Senior Software Engineer - Backend (mainly Python)](https://www.notion.so/splitgraph/Senior-Software-Engineer-Backend-2a2f9e278ba347069bf2566950857250)\n\n[Senior Software Engineer - Frontend (mainly TypeScript)](https://www.notion.so/splitgraph/Senior-Software-Engineer-Frontend-6342cd76b0df483a9fd2ab6818070456)\n\n→ [**Apply to Job**](https://4o99daw6ffu.typeform.com/to/ePkNQiDp) ← (same form for both positions)\n\n# What is Splitgraph?\n## **Open Source Toolkit**\n\n[Our open-source product, sgr,](https://www.github.com/splitgraph/splitgraph) is a tool for building, versioning and querying reproducible datasets. It's inspired by Docker and Git, so it feels familiar. And it's powered by PostgreSQL, so it works seamlessly with existing tools in the Postgres ecosystem. Use Splitgraph to package your data into self-contained\ndata images that you can share with other Splitgraph instances.\n\n## **Splitgraph Cloud**\n\nSplitgraph Cloud is a platform for data cataloging, integration and governance. The user can upload data, connect live databases, or "push" versioned snapshots to it. We give them a unified SQL interface to query that data, a catalog to discover and share it, and tools to build/push/pull it.\n\n# Learn More About Us\n\n- Listen to our interview on the [Software Engineering Daily podcast](https://softwareengineeringdaily.com/2020/11/06/splitgraph-data-catalog-and-proxy-with-miles-richardson/)\n\n- Watch our co-founder Artjoms present [Splitgraph at the Bay Area ClickHouse meetup](https://www.youtube.com/watch?v=44CDs7hJTho)\n\n- Read our HN/Reddit posts ([one](https://news.ycombinator.com/item?id=24233948) [two](https://news.ycombinator.com/item?id=23769420) [three](https://news.ycombinator.com/item?id=23627066) [four](https://old.reddit.com/r/datasets/comments/icty0r/we_made_40k_open_government_datasets_queryable/))\n\n- [Read our blog](https://www.splitgraph.com/blog)\n\n- Read the slides from our early (2018) presentations: ["Docker for Data"](https://www.slideshare.net/splitgraph/splitgraph-docker-for-data-119112722), [AHL Meetup](https://www.slideshare.net/splitgraph/splitgraph-ahl-talk)\n\n- [Follow us on Twitter](https://ww.twitter.com/splitgraph)\n\n- [Find us on GitHub](https://www.github.com/splitgraph)\n\n- [Chat with us in our community Discord](https://discord.gg/eFEFRKm)\n\n- Explore the [public data catalog](https://www.splitgraph.com/explore) where we index 40k+ datasets\n\n# How We Work: What's our stack look like?\n\nWe prioritize developer experience and productivity. We resent repetition and inefficiency, and we never hesitate to automate the things that cause us friction. Here's a sampling of the languages and tools we work with:\n\n- **[Python](https://www.python.org/) for the backend.** Our [core open source](https://www.github.com/splitgraph/splitgraph) tech is written in Python (with [a bit of C](https://github.com/splitgraph/Multicorn) to make it more interesting), as well as most of our backend code. The Python code powers everything from authentication routines to database migrations. We use the latest version and tools like [pytest](https://docs.pytest.org/en/stable/), [mypy](https://github.com/python/mypy) and [Poetry](https://python-poetry.org/) to help us write quality software.\n\n- **[TypeScript](https://www.typescriptlang.org/) for the web stack.** We use TypeScript throughout our web stack. On the frontend we use [React](https://reactjs.org/) with [next.js](https://nextjs.org/). For data fetching we use [apollo-client](https://www.apollographql.com/docs/react/) with fully-typed GraphQL queries auto-generated by [graphql-codegen](https://graphql-code-generator.com/) based on the schema that [Postgraphile](https://www.graphile.org/postgraphile) creates by introspecting the database.\n\n- [**PostgreSQL](https://www.postgresql.org/) for the database, because of course.** Splitgraph is a company built around Postgres, so of course we are going to use it for our own database. In fact, we actually have three databases. We have `auth-db` for storing sensitive data, `registry-db` which acts as a [Splitgraph peer](https://www.splitgraph.com/docs/publishing-data/push-data) so users can push Splitgraph images to it using [sgr](https://www.github.com/splitgraph/splitgraph), and `cloud-db` where we store the schemata that Postgraphile uses to autogenerate the GraphQL server.\n\n- [**PL/pgSQL](https://www.postgresql.org/docs/current/plpgsql.html) and [PL/Python](https://www.postgresql.org/docs/current/plpython.html) for stored procedures.** We define a lot of core business logic directly in the database as stored procedures, which are ultimately [exposed by Postgraphile as GraphQL endpoints](https://www.graphile.org/postgraphile/functions/). We find this to be a surprisingly productive way of developing, as it eliminates the need for manually maintaining an API layer between data and code. It presents challenges for testing and maintainability, but we've built tools to help with database migrations and rollbacks, and an end-to-end testing framework that exercises the database routines.\n\n- [**PostgREST](https://postgrest.org/en/v7.0.0/) for auto-generating a REST API for every repository.** We use this excellent library (written in [Haskell](https://www.haskell.org/)) to expose an [OpenAPI](https://github.com/OAI/OpenAPI-Specification)-compatible REST API for every repository on Splitgraph ([example](http://splitgraph.com/mildbyte/complex_dataset/latest/-/api-schema)).\n\n- **Lua ([luajit](https://luajit.org/luajit.html) 5.x), C, and [embedded Python](https://docs.python.org/3/extending/embedding.html) for scripting [PgBouncer](https://www.pgbouncer.org/).** Our main product, the "data delivery network", is a single SQL endpoint where users can query any data on Splitgraph. Really it's a layer of PgBouncer instances orchestrating temporary Postgres databases and proxying queries to them, where we load and cache the data necessary to respond to a query. We've added scripting capabilities to enable things like query rewriting, column masking, authentication, ACL, orchestration, firewalling, etc.\n\n- **[Docker](https://www.docker.com/) for packaging services.** Our CI pipeline builds every commit into about a dozen different Docker images, one for each of our services. A production instance of Splitgraph can be running over 60 different containers (including replicas).\n\n- **[Makefile](https://www.gnu.org/software/make/manual/make.html) and** [docker-compose](https://docs.docker.com/compose/) **for development.** We use [a highly optimized Makefile](https://www.splitgraph.com/blog/makefile) and `docker-compose` so that developers can easily spin-up a stack that mimics production in every way, while keeping it easy to hot reload, run tests, or add new services or configuration.\n\n- **[Nomad](https://www.nomadproject.io/) for deployment and [Terraform](https://www.terraform.io/) for provisioning.** We use Nomad to manage deployments and background tasks. Along with Terraform, we're able to spin up a Splitgraph cluster on AWS, GCP, Scaleway or Azure in just a few minutes.\n\n- **[Airflow](https://airflow.apache.org/) for job orchestration.** We use it to run and monitor jobs that maintain our catalog of [40,000 public datasets](https://www.splitgraph.com/blog/40k-sql-datasets), or ingest other public data into Splitgraph.\n\n- **[Grafana](https://grafana.com/), [Prometheus](https://prometheus.io/), [ElasticSearch](https://www.elastic.co/), and [Kibana](https://www.elastic.co/kibana) for monitoring and metrics.** We believe it's important to self-host fundamental infrastructure like our monitoring stack. We use this to keep tabs on important metrics and the health of all Splitgraph deployments.\n\n- **[Mattermost](https://mattermost.com/) for company chat.** We think it's absolutely bonkers to pay a company like Slack to hold your company communication hostage. That's why we self-host an instance of Mattermost for our internal chat. And of course, we can deploy it and update it with Terraform.\n\n- **[Matomo](https://matomo.org/) for web analytics.** We take privacy seriously, and we try to avoid including any third party scripts on our web pages (currently we include zero). We self-host our analytics because we don't want to share our user data with third parties.\n\n- **[Metabase](https://www.metabase.com/) and [Splitgraph](https://www.splitgraph.com) for BI and [dogfooding](https://en.wikipedia.org/wiki/Eating_your_own_dog_food)**. We use Metabase as a frontend to a Splitgraph instance that connects to Postgres (our internal databases), MySQL (Matomo's database), and ElasticSearch (where we store logs and DDN analytics). We use this as a chance to dogfood our software and produce fancy charts.\n\n- **The occasional best-of-breed SaaS services** **for organization.** As a privacy-conscious, independent-minded company, we try to avoid SaaS services as much as we can. But we still find ourselves unable to resist some of the better products out there. For organization we use tools like [Zoom](https://www.zoom.us) for video calls, [Miro](https://miro.com/) for brainstorming, [Notion](https://www.notion.so) for documentation (you're on it!), [Airtable for workflow management](https://airtable.com/), [PivotalTracker](https://www.pivotaltracker.com/) for ticketing, and [GitLab for dev-ops and CI](https://about.gitlab.com/).\n\n- **Other fun technologies** including [HAProxy](http://www.haproxy.org/), [OpenResty](https://openresty.org/en/), [Varnish](https://varnish-cache.org/), and bash. We don't touch them much because they do their job well and rarely break.\n\n# Life at Splitgraph\n**We are a young company building the initial team.** As an early contributor, you'll have a chance to shape our initial mission, growth and company values.\n\n**We think that remote work is the future**, and that's why we're building a remote-first organization. We chat on [Mattermost](https://mattermost.com/) and have video calls on Zoom. We brainstorm with [Miro](https://miro.com/) and organize with [Notion](https://www.notion.so).\n\n**We try not to take ourselves too seriously**, but we are goal-oriented with an ambitious mission.\n\n**We believe that as a small company, we can out-compete incumbents** by thinking from first principles about how organizations interact with data. We are very competitive.\n\n# Benefits\n- Fully remote\n\n- Flexible working hours\n\n- Generous compensation and equity package\n\n- Opportunity to make high-impact contributions to an agile team\n\n# How to Apply? Questions?\n[**Complete the job application**](https://4o99daw6ffu.typeform.com/to/ePkNQiDp)\n\nIf you have any questions or concerns, feel free to email us at [[email protected]](mailto:[email protected])\n\n#Location\n🌏 Worldwide


See more jobs at Splitgraph

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.

JetBridge


verified closed
North America and Eu-Time Zones Only
 
💰 $48k - $100k

backend

 

python django

 

java

 

postgres

This job post is closed and the position is probably filled. Please do not apply.
* You'll be working self-directed in a small agile team developing next-generation microservices to manage and deliver genetic health test reports.\n* Primary technologies: Python+Django, Kotlin/Java + Spring Boot, Kafka, Postgres, Jenkins, Kubernetes.\n* You will also make necessary changes to infrastructure such as Jenkins, Terraform, and Kubernetes to support your work.\n* Experience in all technologies is not a requirement, just that you are capable of learning and getting up to speed.\n* You will have ownership and be responsible for delivering secure and high-quality production-ready code and services.\n* Unit testing, integration testing, infrastructure as code, secure, modern, 12-factor development processes are the standard. Code reviews, documentation, and extensive CI/CD automation helps make your life easier and to help you apply industry best practices in everything you do.\n* The team is extremely professional and helpful and a pleasure to work with. \n\n#Salary and compensation\n$48,000 — $100,000/year\n\n\n#Location\nNorth America and Eu-Time Zones Only


See more jobs at JetBridge

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.

InReach Ventures

 This job is getting a pretty high amount of applications right now (11% of viewers clicked Apply)

verified closed
UK or Italy
 
💰 $55k - $70k

java

 

python

 

aws

 

docker

This job post is closed and the position is probably filled. Please do not apply.
InReach is changing how VC in Europe works, for good. Through data, software and Machine Learning, we are building an in-house platform to help us find, reach-out to and invest in early-stage European startups, regardless of the city or country they’re based in.\n\nWe are looking for a back-end developer to continue the development of InReach’s data services. This involves: \n* Cleaning / wrangling / merging / processing the data on companies and founders from across Europe\n* Building data pipelines with the Machine Learning engineers\n* Building APIs to support front-end investment product used by the Investment team (named DIG)\n\nThis role will involve working across the stack. From DevOps (Terraform) to web scraping and Machine Learning (Python) all the way to data pipelines and web-services (Java) and getting stuck into the front-end (Javascript). It’s a great opportunity to hone your skills and master some new ones.\n\nIt is important to us that candidates be passionate about helping entrepreneurs and startups. This is our bread-and-butter and we want you to be involved.\n\nInReach is a remote-first employer and we are looking to this hire to help us become an exceptional place to work for remote employees. Whether you are in the office or remote, we are looking for people with excellent written and verbal communication skills.\n\n### Background Reading:\n* [InReach Ventures, the 'AI-powered' European VC, closes new €53M fund](https://techcrunch.com/2019/02/11/inreach-ventures-the-ai-powered-european-vc-closes-new-e53m-fund/?guccounter=1)\n* [The Full-Stack Venture Capital](https://medium.com/entrepreneurship-at-work/the-full-stack-venture-capital-8a5cffe4d71)\n* [Roberto Bonanzinga starts InReach Ventures with DIG platform](https://www.businessinsider.com/roberto-bonanzinga-starts-inreach-ventures-with-dig-platform-2015-11?r=US&IR=T)\n* [Exceptional Communication Guidelines](https://www.craft.do/s/Isrjt4KaHMPQ)\n\n## Responsibilities\n\n* Creatively and quickly coming up with effective solutions to undefined problems\n* Choosing technology that is modern but not hype-driven\n* Developing features and tests quickly with good, clean code\n* Being part of the wider development team, reviewing code and participating in architecture from across the stack\n* Communicating exceptionally, both asynchronously (written) and synchronously (spoken)\n* Helping to shape InReach as a remote-first organization\n\n## Technologies\n\nGiven that this position touches so much of the stack, it will be difficult for a candidate that only has experience in Python or only in Java to be successful in being effective quickly. While we expect the candidate to be stronger in one or the other, some professional exposure is required.\n\nIn addition to the programming skills and the ability to write well designed and tested code, infrastructure within modern cloud platforms and sound architectural reasoning are expected.\n\nNone of these are a prerequisite, but help:\n* Functional Programming\n* Reactive Streams (RxJava2)\n* Terraform\n* Postgres\n* ElasticSearch\n* SQS\n* Dynamodb\n* AWS Lambda\n* Docker\n* Dropwizard\n* Maven\n* Pipenv\n* Javascript\n* React\n* NodeJS\n\n## Interview Process\n* 15m video chat with Ben, CTO to find out more about InReach and the role\n* 2h data pipeline technical test (Python)\n* 2h web service technical test (Java)\n* 30m architectural discussion with Ben, talking through the work you did\n* 2h interview with the different team members from across InReach. We’re a small company so it’s important we see how we’ll all work together - not just the tech team!\n \n\n#Salary and compensation\n$55,000 — $70,000/year\n\n\n#Location\nUK or Italy


See more jobs at InReach Ventures

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.
This job post is closed and the position is probably filled. Please do not apply.
Doximity is transforming the healthcare industry. Our mission is to help doctors be more productive, informed, and connected. As a software engineer, you'll work within cross-functional delivery teams alongside other engineers, designers, and product managers in building software to help improve healthcare.  \n\nOur team brings a diverse set of technical and cultural backgrounds and we like to think pragmatically in choosing the tools most appropriate for the job at hand.\n\n**About Us**\n* Here are some of the ways [we bring value to doctors](https://drive.google.com/file/d/1qimYh0mG3i1nTJe6jDCDepJt2i4o8MEB/view)\n* Our web applications are built primarily using Ruby, Rails, Javascript (Vue.js), and a bit of Golang\n* Our data engineering stack run on Python, MySQL, Spark, and Airflow\n* Our production application stack is hosted on AWS and we deploy to production on average 50 times per day\n* We have over 350 private repositories in Github containing our applications, forks of gems, our own internal gems, and [open-source projects](https://github.com/doximity)\n* We have worked as a distributed team for a long time; we're [currently about 65% distributed](https://blog.brunomiranda.com/building-a-distributed-engineering-team-85d281b9b1c)\n* Find out more information on the [Doximity engineering blog](https://engineering.doximity.com/)\n* Our [company core values](https://work.doximity.com/)\n* Our [recruiting process](https://engineering.doximity.com/articles/engineering-recruitment-process-doximity)\n* Our [product development cycle](https://engineering.doximity.com/articles/mofo-driven-product-development)\n* Our [on-boarding & mentorship process](https://engineering.doximity.com/articles/software-engineering-on-boarding-at-doximity)\n\n**Here's How You Will Make an Impact**\n* Improve the performance and scalability of services, optimize our REST and GraphQL APIs\n* Address security concerns and proficiently maintain our application stack\n* Troubleshoot issues across the whole stack, such as high-load, memory full, network issues and come up with temporary/long term solutions based on the root cause\n* Hands-on maintenance on our Ruby on Rails and Go (Golang) applications\n* Increase our automated test coverage and deployment infrastructure robustness \n* Manage infrastructure using Chef and Terraform\n* Active involvement in design, implementation, and maintenance of the development, staging, and production infrastructure and services your team is responsible for\n* Create concise postmortems in the event of an outage\n* Write and maintain run-books for other engineers to leverage\n* Ensure proper security, monitoring, alerting, and reporting for the applications your team is responsible for\n* Collaborate with other engineers to make sound infrastructure decisions, improve workflow, and deploy applications ready for production\n* Monitor capacity, cost and plan for upgrades\n* Participate in an on-call rotation\n\n**About you**\n* You are a Ruby engineer at heart, very familiar and passionate about the Rails ecosystem\n* You are knowledgeable of memory and CPU profiling tools to help adjust Ruby jobs and processes to use resources effectively\n* You have experience working with Terraform and Chef (or similar tooling) either in a DevOps or product support capacity\n* You have experience deploying, configuring, and maintaining NGINX\n* You are proficient with Unix, AWS, and Git\n* You are self-motivated and able to manage yourself and your own queue\n* You are a problem solver with a passion for simple, clean, and maintainable solutions\n* You agree that concise and effective written and verbal communication is a must for a successful team\n* You are able to maintain a minimum of 5 hours overlap with 9:30 to 5:30 PM Pacific time\n* You can dedicate about two weeks per year for travel to company events\n\n**Benefits**\n\nDoximity has industry leading benefits. For an updated list, see our career page\n\n**More info on Doximity**\nWe’re thrilled to be named the Fastest Growing Company in the Bay Area, and one of Fast Company’s Most Innovative Companies. Joining Doximity means being part of an incredibly talented and humble team. We work on amazing products that over 70% of US doctors (and over one million healthcare professionals) use to make their busy lives a little easier. We’re driven by the goal of improving inefficiencies in our $3.5 trillion U.S. healthcare system and love creating technology that has a real, meaningful impact on people’s lives. To learn more about our team, culture, and users, check out our careers page, company blog, and engineering blog. We’re growing steadily, and there’s plenty of opportunity for you to make an impact.\n\n*Doximity is proud to be an equal opportunity employer, and committed to providing employment opportunities regardless of race, religious creed, color, national origin, ancestry, physical disability, mental disability, medical condition, genetic information, marital status, sex, gender, gender identity, gender expression, pregnancy, childbirth and breastfeeding, age, sexual orientation, military or veteran status, or any other protected classification. We also consider qualified applicants with criminal histories, consistent with applicable federal, state and local law.*\n\n#Location\nNorth America


See more jobs at Doximity

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.
This job post is closed and the position is probably filled. Please do not apply.
Doximity is transforming the healthcare industry. Our mission is to help doctors be more productive, informed, and connected. As a software engineer, you'll work within cross-functional delivery teams alongside other engineers, designers, and product managers in building software to help improve healthcare.  \n\nOur team brings a diverse set of technical and cultural backgrounds and we like to think pragmatically in choosing the tools most appropriate for the job at hand.\n\n**About Us**\n\nHere are some of the ways we bring value to doctors\n* Our web applications are built primarily using Ruby, Rails, Javascript (Vue.js), and a bit of Golang\n* Our data engineering stack run on Python, MySQL, Spark, and Airflow\n* Our production application stack is hosted on AWS and we deploy to production on average 50 times per day\n* We have over 350 private repositories in Github containing our applications, forks of gems, our own internal gems, and open-source projects\n* We have worked as a distributed team for a long time; we're currently about 65% distributed\n* Find out more information on the Doximity engineering blog\n* Our[ company core values](https://work.doximity.com/)\n* Our [recruiting process](https://engineering.doximity.com/articles/engineering-recruitment-process-doximity)\n* Our [product development cycle](https://engineering.doximity.com/articles/mofo-driven-product-development)\n* Our [on-boarding & mentorship process](https://engineering.doximity.com/articles/software-engineering-on-boarding-at-doximity)\n\n**Here's How You Will Make an Impact**\n* Improve the performance and scalability of services, optimize our Rest and GraphQL APIs\n* Manage infrastructure using Chef and Terraform\n* Address security concerns and proficiently maintain our application stack\n* Active involvement in design, implementation, and maintenance of the development, staging, and production infrastructure and services your team is responsible for\n* Troubleshoot issues across the whole stack, such as high-load, memory full, network issues and come up with temporary/long term solutions based on the root cause\n* Create concise postmortems in the event of an outage\n* Write and maintain run-books for other engineers to leverage\n* Ensure proper security, monitoring, alerting, and reporting for the applications your team is responsible for\n* Collaborate with other engineers to make sound infrastructure decisions, improve workflow, and deploy applications ready for production\n* Hands-on maintenance on our Ruby on Rails and Go (Golang) applications\n* Monitor capacity, cost and plan for upgrades\n* Increase our automated test coverage and deployment infrastructure robustness \n* Participate in an on-call rotation\n\n**About you**\n* You are a problem solver with a passion for simple, clean, and maintainable solutions\n* You have extensive experience with Terraform and Chef (or equivalent)\n* You are knowledgeable of memory and CPU profiling tools to help adjust Ruby jobs and processes to use resources effectively\n* You have high familiarity with OOP and design principles to ensure well-architected services\n* You have significant experience deploying, configuring, and maintaining NGINX\n* You are proficient with Unix, AWS, and Git\n* You have experience writing automated tests and appreciate the benefit that tests offer\n* You are self-motivated and able to manage yourself and your own queue\n* You agree that concise and effective written and verbal communication is a must for a successful team\n* You have experience with web infrastructure, distributed systems, and performance optimizations\n* You are able to maintain a minimum of 5 hours overlap with 9:30 to 5:30 PM Pacific time\n* You can dedicate about two weeks per year for travel to company events\n\n**Benefits**\n\nDoximity has industry leading benefits. For an updated list, see our career page\n\n**More info on Doximity**\n\nWe’re thrilled to be named the Fastest Growing Company in the Bay Area, and one of Fast Company’s Most Innovative Companies. Joining Doximity means being part of an incredibly talented and humble team. We work on amazing products that over 70% of US doctors (and over one million healthcare professionals) use to make their busy lives a little easier. We’re driven by the goal of improving inefficiencies in our $3.5 trillion U.S. healthcare system and love creating technology that has a real, meaningful impact on people’s lives. To learn more about our team, culture, and users, check out our careers page, company blog, and engineering blog. We’re growing steadily, and there’s plenty of opportunity for you to make an impact.\n\n*Doximity is proud to be an equal opportunity employer, and committed to providing employment opportunities regardless of race, religious creed, color, national origin, ancestry, physical disability, mental disability, medical condition, genetic information, marital status, sex, gender, gender identity, gender expression, pregnancy, childbirth and breastfeeding, age, sexual orientation, military or veteran status, or any other protected classification. We also consider qualified applicants with criminal histories, consistent with applicable federal, state and local law.*\n\n#Location\nNorth America


See more jobs at Doximity

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.

Volders GmbH


closed

ruby

 

elixir

 

aws

 
This job post is closed and the position is probably filled. Please do not apply.
You do have a bunch of contracts, like phone, electricity, Internet, insurance, gym, etc. Have you ever tried to cancel or negotiate any of them? We are sure your experience was bad. Our mission here at Volders is to fix that.\n\nWe free our customers from dealing with contract management tasks so that they can enjoy the things that matter. More than 1 Million customers relied on Volders services, and for us, that’s just the beginning.\n\n**About you**\n* You have experience with Ruby and Rails (but believe Elixir is the next step in your career)\n* You have the desire to learn and use Elixir/Phoenix full-time\n* You know your way around Javascript + HTML/CSS\n* You prefer server rendered applications instead of SPAs\n* You are proud about creating and shipping simple and well-tested code\n* You feel comfortable working by yourself and take ownership\n* Your working time overlaps with Berlin time (at least 4 hours)\n* You have good spoken/written English\n\n**About us**\n* We have an informal, collaborative and ego-free environment\n* We have a number of Ruby and Elixir projects\n* Our products run on AWS\n* We use Terraform and Packer to build our infrastructure\n* We embrace Continuous Integration\n\n**Benefits**\n* Equipment to improve your home office (camera, microphone, etc)\n* Educational budget to buy the books you need\n* Work from Berlin once a year, with everything covered\n* 2-weeks paid vacation\n* Build a product that people love using


See more jobs at Volders GmbH

# How do you apply?\n\nThis job post is older than 30 days and the position is probably filled. Try applying to jobs posted recently instead.
129ms