This job post is closed and the position is probably filled. Please do not apply. Work for vidIQ and want to re-open this job? Use the edit link in the email when you posted the job!
**About vidIQ:**\n\nvidIQ helps YouTube creators and brands generate more views and subscribers, while saving time. With over 1 Million active weekly users, we are the #1 Chrome Extension for YouTube creators, with clients including Red Bull, Buzzfeed, PBS, TMZ, BBC as well as hundreds of thousands of the largest YouTube creators in the world. Weโre backed by top Silicon Valley investors including Scott Banister and Mark Cuban. vidIQ is profitable with a fully remote team over 25 employees and growing.\n\n**Role & Responsibilities**\n\nvidIQ is seeking a highly-motivated Senior Data Engineer with 5+ years of hands-on data engineering experience to join our growing team. The ideal candidate will be a go-getter with the ability to work independently. In this role, you will have oversight of partitioning data, building an ETL pipeline, data compaction, and AWS optimization. \n\n\nYou must be highly collaborative and a self-starter who is able to work in a fast-paced environment. Strong communication skills are essential in this role, as it will be integral in communicating to the back-end team where and how to implement data integration and persistence. You will also communicate to management the volumes of data we are gathering, as well as communicate the data access points and how to use this data, to the team and management.\n\n# Responsibilities\n
**You will be a good fit for this role if the following are true:**\n\n* You love building things. You like new challenges and strive to ship new features to customers on a regular basis.\n* You love to learn. You enjoy keeping up with the latest trends. If a project uses a tool thatโs new to you, you dive into the docs and tutorials to figure it out.\n* You act like an owner. When bugs appear, you document and fix them. When projects are too complex, you work with others to refine the scope until itโs something you believe can be built in a reasonable amount of time and maintained in the long run.\n* You care about code quality. You believe simple is better and strive to write code that is easy to read and maintain. \n* You consider edge cases and write tests to handle them. When you come across legacy code that is difficult to understand, you add comments or refactor it to make it easier for the next person.\n* You understand balance. Great products must balance performance, customer value, code quality, dependencies, and so on. You know how to consider all of these concerns while keeping your focus on shipping things.\n* You over-communicate by default. If a project is off-track, you bring it up proactively and suggest ways to simplify and get things going. You proactively share status updates without being asked and strive to keep things as honest and transparent as possible. \n\n# Requirements\n**Minimum experience:**\n\n* 5+ years experience using *Python* for internal data pipelines (moving data inside *AWS* account)\n* *numpy*, pandas.\n* Experience with *Scala* - for external data pipelines (moving data from outside of *AWS* account into *AWS* account) *FS2*, *http4s*.\n* Additional experience with *DynamoDB*, *Lambda*, *Athena*, *S3*, *AWS* *GlueFamiliar* with *Spark* (in the moment Scala only) preferred.\n* Hands-on experience with data workflow orchestration (*Airflow*). \n\nPlease mention the words **BELOW ENHANCE WASH** when applying to show you read the job post completely (#RMy4xMzkuMTA0LjIxNA==). This is a feature to avoid spam applicants. Companies can search these words to find applicants that read this and see they're human.\n\n \n\n#Salary and compensation\n
No salary data published by company so we estimated salary based on similar jobs related to Python, Scala, Senior and Engineer jobs that are similar:\n\n
$80,000 — $130,000/year\n
\n\n#Location\nTimezones: from UTC-2 to UTC+8
# How do you apply?\n\nThis job post has been closed by the poster, which means they probably have enough applicants now. Please do not apply.