As a part of the Attribution team you will design and implement systems for ingesting, transforming, connecting, storing, and delivering data from a wide range of sources with varying levels of complexity and scale that enable us to associate domains and IPs to companies on a continous basis. You will enable other engineers to deliver value rapidly with minimum duplication of effort. Automate the infrastructure supporting the data pipeline as code and deployments by improving CI/CD pipelines. Monitor, troubleshoot, and improve the data platform to maintain stability and optimal performance.
Requirements
Basic Qualifications
Bachelor's degree or higher in a quantitative/technical field such as Computer Science, Engineering.
5+ years of data pipeline software development experience.
Exceptional skills in at least one high-level programming language (Scala, Java, Go, Python or equivalent)
Actively using and a strong understanding of big data technologies such as Kafka, Spark, Databricks toolkit
Additional Qualifications
Experience with Dataflow orchestration in Google Cloud Flow, Airflow, or Conductor
Experience with AWS services including EMR, S3, Redshift, and RDS
Solid understanding of DNS, networking.
Excellent communication skills to collaborate with cross-functional partners and independently drive projects and decisions
Previous experience working in distributed teams. We are a remote-first company!
Benefits
Competitive salary, stock options, a comprehensive benefits package, including health and dental insurance, unlimited PTO, parental leave, tuition reimbursements, and much more!