Key Areas of Impact:
- Design, build, and maintain robust, scalable, and efficient ETL/ELT pipelines for structured and unstructured data.
- Implement best practices for data ingestion, transformation, and integration across multiple sources.
- Optimize and maintain data storage solutions (e.g. SQL/NoSQL databases, data lakes, warehouses) for performance and cost-effectiveness.
- Collaborate with data analysts and business stakeholders to understand data requirements and deliver solutions that meet their needs.
- Contribute to the development and implementation of the organization’s data architecture and strategy.
- Continuously evaluate and integrate new technologies and tools for automated data quality validation and to improve data engineering processes overall.
- Maintain and improve robust monitoring and alerting mechanisms for data pipelines and systems.
Ideal Candidate will Bring:
- 5+ years of experience designing and implementing scalable data pipelines and infrastructure.
- Strong technical expertise in modern tools and frameworks, including cloud platforms (we're built on AWS), distributed systems, and advanced SQL.
- Proficiency in programming languages (such as Ruby or Python) – you don't need to be a software engineer, but should understand our codebase.
- Hands-on experience with data warehousing solutions like Snowflake or Redshift and orchestration tools like Airflow or dbt. (We're migrating from Redshift to Snowflake.)
- A strong sense of ownership and accountability, with a track record of proactively improving systems for performance, reliability, and cost while addressing complex technical challenges.
- Excels at cross-functional collaboration, aligning technical solutions with business goals and delivering measurable improvements to data workflows and decision-making.
- Curiosity and a growth mindset, demonstrated through learning new tools, deepening technical expertise, and continuously improving processes.
- Startup/high-growth experience with the scrappiness required to be successful at AngelList!
How Success will be Measured:
- Delivery of reliable, scalable, and efficient data pipelines within agreed timelines, with minimal downtime and robust monitoring systems in place.
- Successful collaboration with analysts and stakeholders to enable seamless access to accurate and timely data for decision-making.
- Tangible improvements in infrastructure scalability, cost-effectiveness, and adaptability to evolving business needs.
