Location: Remote, US
Senior Data Engineer
At Coinbase, our mission is to increase economic freedom around the world, and we couldn’t do this without hiring the best people. We’re a group of hard-working overachievers who are deeply focused on building the future of finance and Web3 for our users across the globe, whether they’re trading, storing, staking or using crypto. Know those people who always lead the group project? That’s us.
There are a few things we look for across all hires we make at Coinbase, regardless of role or team. First, we look for candidates who will thrive in a culture like ours, where we default to trust, embrace feedback, and disrupt ourselves. Second, we expect all employees to commit to our mission-focused approach to our work. Finally, we seek people who are excited to learn about and live crypto, because those are the folks who enjoy the intense moments in our sprint and recharge work culture. We’re a remote-first company looking to hire the absolute best talent all over the world.
Who you are:
- You’ve got positive energy. You’re optimistic about the future and determined to get there.
- You’re never tired of learning. You want to be a pro in bleeding edge tech like DeFi, NFTs, DAOs, and Web 3.0.
- You appreciate direct communication. You’re both an active communicator and an eager listener – because let’s face it, you can’t have one without the other. You’re cool with candid feedback and see every setback as an opportunity to grow.
- You can pivot on the fly. Crypto is constantly evolving, so our priorities do, too. What you worked on last month may not be what you work on today, and that excites you. You’re not looking for a boring job.
- You have a “can do” attitude. Our teams create high-quality work on quick timelines. Owning a problem doesn’t scare you, but rather empowers you to take 100% responsibility for achieving our mission.
- You want to be part of a winning team. We’re stronger together, and you’re a person who embraces being pushed out of your comfort zone.
The data engineering team develops and maintains robust data pipelines, builds trusted data sources, and creates analytics/data products that inject automation into data science processes, with a goal to empower users via self-serve analytics.
What we do:
- Trusted data sources: Build and maintain a foundational data layer (data marts), that serve as the single source of truth across Coinbase.
- Reliable data pipelines: Design and implement robust data pipelines, guaranteeing data quality and timely data delivery across our organization.
- Data science developer tools: Build developer tools that inject automation into data science processes, improving efficiency and productivity. For example, data transformation, data modeling, and data quality tooling.
- Self-serve analytics products: Deliver tailored data products, designed to empower users with self-serve capabilities and ensure accurate answers to their data inquiries.
What you’ll be doing:
Your primary responsibilities will include building tools and products that enhance data science productivity, enable self-serve analytics, and ensure data reliability and quality. Additional responsibilities may include contributing to the maintenance of critical data marts and data pipelines as needed..
More specifically:
- Enforce best practices: Introduce automation mechanisms to enforce data best practices and optimal design principles.
- Data transformation tooling: Enhance our data transformation layer by adding new features and migrating data pipelines from legacy tools.
- Data quality tooling: Build systems to automate monitoring, alerting and fixing data anomalies.
- Productivity tooling: Build internal tools aimed at streamlining workflows, minimizing manual interventions and driving operational efficiencies.
- Cross-Functional collaboration: Work alongside fellow data engineers and cross-functional partners from Data Science, Data Platform, Machine Learning, and various analytics teams to ensure alignment on priorities and deliverables.
- Reliable data pipelines: Develop new data pipelines and ETL processes, while enhancing the efficiency of existing ones to reduce costs, improve landing times, and ensure reliability. This may involve indexing, partitioning, and tuning the underlying data models.
- Semantic layer integration: Build the semantic layer to ensure it effectively translates complex data models from the data marts into business terms. This includes defining dimensions, measures, hierarchies, and business logic.
- Self-serve analytics: Contribute to the self-serve analytics chatbot and its infrastructure by adding features, fixing bugs and promoting its use across the company.
- Knowledge base: Maintain clear and comprehensive documentation of processes to facilitate smooth team collaboration and hand-offs.
What we look for in you:
- Python: Must be adept at scripting in Python, particularly for data manipulation and integration tasks, with strong experience in Object-Oriented Programming (OOP).
- SQL: Must have a solid grasp of advanced SQL techniques for querying, transformation, and performance optimization.
- ETL/ELT Processes: Experience in designing, building, and optimizing ETL/ELT data pipelines to process large datasets. Experience with both batch and streaming is a plus.
- Apache Airflow: Experience in building, deploying, and optimizing DAGs in Airflow or a similar pipeline orchestration tool.
- GitHub: Experience with version control, branching, and collaboration on GitHub is a must-have.
- Data Visualization: Experience with tools like Superset, Looker, or Python visualization libraries (Matplotlib, Seaborn, Plotly…etc)
- Data Modeling: Understanding of best practices for data modeling, including star schemas, snowflake schemas, and data normalization techniques
- Collaboration and Communication: Ability to work closely with data scientists, analysts, and other stakeholders to translate business requirements into technical solutions. Strong documentation skills for pipeline design and data flow diagrams.
- Fundamental DevOps Practices: Knowledge of unit testing, CI/CD, and repository management.
- Technologies: Familiarity with Docker, Kubernetes, and cloud platforms such as AWS or GCP.
- Prompt Engineering for LLMs: Expertise in crafting, testing, and refining prompts for effective communication with LLMs, like GPT, to generate accurate and contextually relevant responses is a plus.
Job #: P60857
Pay Transparency Notice: Depending on your work location, the target annual salary for this position can range from $[Zone 3 Pay] to $[Zone 1 Pay] + target bonus + target equity + benefits (including medical, dental, vision and 401(k)).
Pay Transparency Notice: Depending on your work location, the target annual salary for this position can range as detailed below. Full time offers from Coinbase also include target bonus + target equity + benefits (including medical, dental, vision and 401(k)).
Pay Range:
$175,100—$206,000 USD
Commitment to Equal Opportunity
Coinbase is committed to diversity in its workforce and is proud to be an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, creed, gender, national origin, age, disability, veteran status, sex, gender expression or identity, sexual orientation or any other basis protected by applicable law. Coinbase will also consider for employment qualified applicants with criminal histories in a manner consistent with applicable federal, state and local law. For US applicants, you may view the Know Your Rights notice here. Additionally, Coinbase participates in the E-Verify program in certain locations, as required by law.
Coinbase is also committed to providing reasonable accommodations to individuals with disabilities. If you need a reasonable accommodation because of a disability for any part of the employment process, please contact us at accommodations[at]coinbase.com to let us know the nature of your request and your contact information. For quick access to screen reading technology compatible with this site click here to download a free compatible screen reader (free step by step tutorial can be found here).
Global Data Privacy Notice for Job Candidates and Applicants
Depending on your location, the General Data Protection Regulation (GDPR) and California Consumer Privacy Act (CCPA) may regulate the way we manage the data of job applicants. Our full notice outlining how data will be processed as part of the application procedure for applicable locations is available here. By submitting your application, you are agreeing to our use and processing of your data as required. For US applicants only, by submitting your application you are agreeing to arbitration of disputes as outlined here.
APPLY FOR JOB
NEXT JOBS