Staff Software Engineer, Data Incrementalization
- $240k – $270k
- Full Time
Not Available
About the job
The Data Platform team collaborates with an experienced group of data scientists, engineers, and product managers to build highly available and scalable data infrastructure for TRM's products and services. As a Senior Engineer on the Core Data team, you will help to develop internal tools to empower the Data Science and Machine Learning Engineers to refine Blockchain Data into Blockchain Intelligence at petabyte scale.
The impact you’ll have here:
- Design and build our Cloud Data Warehouse with a focus on incremental updates to improve cost efficiency and scalability.
- Research innovative methods to incrementally optimize data processing, storage, and retrieval to support efficient data analytics and insights.
- Develop and maintain ETL pipelines that transform and incrementally process petabytes of structured and unstructured data to enable data-driven decision-making.
- Collaborate with cross-functional teams to design and implement new data models and tools focused on accelerating innovation through incremental updates.
- Continuously monitor and optimize the Data Platform's performance, focusing on enhancing cost efficiency, scalability, and reliability.
What we’re looking for:
- Bachelor's degree (or equivalent) in Computer Science or a related field.
- 5+ years of experience in building distributed system architecture, with a particular focus on incremental updates from inception to production.
- Strong programming skills in Python and SQL.
- Deep technical expertise in advanced data structures and algorithms for incremental updating of data stores (e.g., Graphs, Trees, Hash Maps).
- Comprehensive knowledge across all facets of data engineering, including:
- Implementing and managing incremental updates in data stores like BigQuery, Snowflake, RedShift, Athena, Hive, and Postgres.
- Orchestrating data pipelines and workflows focused on incremental processing using tools such as Airflow, DBT, Luigi, Azkaban, and Storm.
- Developing and optimizing data processing technologies and streaming workflows for incremental updates (e.g., Spark, Kafka, Flink).
- Deploying and monitoring scalable, incremental update systems in public cloud environments (e.g., Docker, Terraform, Kubernetes, Datadog).
- Expertise in loading, querying, and transforming large datasets with a focus on efficiency and incremental growth.
About the Team:
- The Data Platform team is the funnel between all of TRM's data world and product world. We care about all layers of stack including petabyte of data stores, pipelines, and processes.
- We have quite a big scope as a the team with new and exciting projects every quarter. As a result, we collaborate across the board with most teams at TRM.
- We believe in async communication and are also not afraid to jump on a quick huddle if that helps to move things faster. We are both scrappy when the situation demands and also process-oriented when we need to achieve our OKRs.
- We are always looking for people who can elevate the quality our tech and our execution. If you enjoy a remote-first and async friendly environment to achieve efficacy and efficiency at petabyte scale, our team could be a great pick for you!
- Team members are based in the US across almost all timezones! Our on-call tends to be in EST/PST shift, whatever suits you the best.
- We do try to reserve some overlap in the day for meetings. Our north star - no IC spends more than 3-4 hours/week in meetings.
Learn about TRM Speed in this position:
- Build scalable engines to optimize routine scaling and maintenance tasks like create self-serve automation for creating new pgbouncer, scaling disks, scaling/updating of clusters, etc.
- Enable tasks to be faster next time and reducing dependency on a single person.
- Identify ways to compress timelines using 80/20 principle. For instance, what does it take to be operational in a new environment? Identify the must have and nice to haves that are need to deploy our stack to be fully operation. Focus on must haves first to get us operational and then use future milestones to harden for customer readiness. We think in terms of weeks and not months.
- Identify first version, a.k.a., "skateboards" for projects. For instance, build an observability dashboard within a week. Gather feedback from stakeholders after to identify more needs or bells and whistles to add to the dashboard.
About TRM's Engineering Levels:
Engineer: Responsible for helping to define project milestones and executing small decision decisions independently with the appropriate tradeoffs between simplicity, readability, and performance. Provides mentorship to junior engineers, and enhances operational excellence through tech debt reduction and knowledge sharing.
Senior Engineer: Successfully designs and documents system improvements and features for an OKR/project from the ground up. Consistently delivers efficient and reusable systems, optimizes team throughput with appropriate tradeoffs, mentors team members, and enhances cross-team collaboration through documentation and knowledge sharing.
Staff Engineer: Drives scoping and execution of one or more OKRs/projects that impact multiple teams. Partners with stakeholders to set the team vision and technical roadmaps for one or more products. Is a role model and mentor to the entire engineering organization. Ensures system health and quality with operational reviews, testing strategies, and monitoring rigor.
The following represents the expected range of compensation for this role:
- The estimated base salary range for this role is $240,000 - $270,000.
- Additionally, this role may be eligible to participate in TRM’s equity plan.