Senior Data Engineer - Data Corpus
SoundCloud empowers artists and fans to connect and share through music. Founded in 2007, SoundCloud is an artist-first platform empowering artists to build and grow their careers by providing them with the most progressive tools, services, and resources.
With over 400+ million tracks from 40+ million artists, the future of music is SoundCloud.
We are looking for an experienced Senior Data Engineer (Level 3) to join the Data Foundations department in our Data Corpus team.
You will help us continue building our unified and standardized data warehouse, Corpus, which is used across SoundCloud for reporting, data science, and analytics, building data products, or simply answering questions.
You will also help us build out our business-ready data layer to be used in dashboarding and for simplified data usage. You can read more about Data Corpus in our blog post.
About the role
The Data Corpus team is responsible for building and maintaining SoundCloud’s petabyte-scale GCP data warehouse. You will build and maintain the warehouse to abstract the complexity of SoundCloud’s vast data ecosystem to exploit and get value from data.
You will work closely with business reporting, data science, and product teams. You will gather and refine requirements, design data architecture and solutions, and build ETL pipelines using Airflow to land data in BigQuery and beyond.
You will model and build out the business-ready data layer, designed to feed dashboarding tools with KPIs and simplify data access for business and product stakeholders.
You will have knowledge and passion for analytics engineering. You are a strong individual contributor who thrives in a dynamic environment and is motivated by solving challenging data problems.
About you
As a Senior individual contributor, you will be able to manage stakeholders, plan and lead technical projects, and identify and focus the business value of projects.
Must have :
- Senior Level Data Professional
- 6 years experience ideal, +4 years minimum experience.
- Experience working with Cloud technologies
- GCP (required)
- AWS / Azure is also nice to have
- Experience working with BigQuery
- Advanced SQL knowledge
- Python experience
- Airflow experience
- Experience working with big data (terabyte / petabyte scale)
- Experience in Data Architecture / solution design
- Experience working in Agile methodology, ideally working with Jira.
- Experience in data warehousing and analytical data modelling.
- Experience working with CI / CD pipelines and code repositories (Git preferred)
- Experiencing in building complex and high-reliability ETL pipelines
- Experience building data for BI Tools (Looker preferred)
- Experience building datasets for product and business KPIs
- Knowledge of basic statistical concepts and methods
- Experience producing high-quality technical documentation
- Excellent English-speaking communicator and demonstrator
Nice to have :
- Experience in basic QA engineering and testing methodologies
- Experience with experimentation (A / B testing) tools and processes
- Experience working with subscription-based businesses
- Experience integrating third-party tools and systems
- Data visualisation skills
- Experience with terraform