Skip to main content

Data Engineer at Atlanta Hawks

Data Engineer
Atlanta Hawks
On-site
Atlanta, GA
Full-time
Salary not listed
Posted 4 May 2026
Apply on TeamWork
Share this job:

Job Description

Who are we: A professional basketball team and state-of-the-art arena/entertainment venue that specializes in creating memorable experiences for each guest we interact with. Some of our favorite things are live sports, concerts, comedy shows, family shows, and most any other world-class event you can think of, and we’re looking for someone who shares the same interests. We live for the fast-paced world of sports & live entertainment, and as such, we work hard, run fast, execute flawlessly, and party it up when it all comes together. Lastly, we strive to deliver wonderful experiences that create lasting memories, and we prefer to surround ourselves with those who are the best at what they do. Who are you: An enthusiastic lover of sports, live entertainment, and people. You have true passion for engaging in meaningful interactions and creating memorable experiences for all guests. You strive to be helpful, engaging, and knowledgeable of all things Atlanta Hawks and State Farm Arena. You enjoy being a part of an exciting and dynamic group, and you’re committed to continuously enhancing the productivity and effectiveness of your team. Lastly, you enjoy working hard and celebrating hard, and you’d be shocked if guests weren’t positively impacted by their interactions with you. Job Summary: The Data Engineer is responsible for the end-to-end data lifecycle—from orchestrating high-frequency ingestion pipelines to ensuring reliable delivery through the activation layer. This role focuses on ingesting, transforming, and delivering reliable, high-quality data from a variety of internal and external sources into our cloud data ecosystem. Working closely with Analytics Engineers, BI Engineers, and business stakeholders, the Data Engineer ensures that data is accessible, timely, and trustworthy—enabling analytics, reporting, marketing activation, and operational decision-making across the company. In this position, you will play a key role in developing and optimizing the data ingestion and processing layers of our platform. You will be responsible for building robust ETL/ELT pipelines, managing data workflows, and ensuring efficient movement and storage of data across systems. This includes integrating with APIs, streaming platforms, and third-party systems to support both batch and real-time data use cases. Key Responsibilities: Data Pipeline Development & Orchestration • Design, build, and maintain scalable and reliable data pipelines to ingest data from various sources, including APIs, databases, and SaaS platforms. • Develop and optimize ETL/ELT workflows to efficiently process large volumes of data. • Implement orchestration frameworks and Databricks-native workflows to manage dependencies and scheduling. • Ensure pipelines are fault-tolerant, observable, and recoverable with appropriate logging and alerting. • Continuously improve pipeline performance, scalability, and cost efficiency. Data Platform & Infrastructure • Design and maintain core components of the data platform, including data lakes, warehouses, and storage systems. • Optimize data storage formats and partitioning strategies for performance and cost. • Manage integrations with Databricks and other cloud data platform services. • Support both batch and streaming architectures where applicable. • Collaborate on infrastructure-as-code and environment management practices. Data Quality, Reliability & Monitoring • Implement data validation and monitoring at ingestion and pipeline stages. • Ensure data completeness, freshness, and accuracy through automated checks and alerts. • Troubleshoot pipeline failures and data discrepancies, driving root cause resolution. • Establish SLAs/SLIs for data availability and pipeline performance. Data Integration & Enablement • Partner with Analytics Engineers to ensure clean handoffs between raw and curated data layers. • Enable downstream consumers by delivering well-structured, reliable datasets. • Support integrations with downstream tools such as BI platforms, CDPs, and operational systems. • Contribute to improving data accessibility and usability across the organization. Documentation, Governance & Best Practices • Document data pipelines, data sources, and system architecture. • Maintain clear data lineage and metadata practices. • Contribute to standards around naming conventions, code quality, and deployment workflows. • Participate in code reviews and CI/CD processes. Requirements: • Strong proficiency in SQL and at least one programming language (Python preferred). • Deep understanding of Apache Spark architecture and hands-on experience with PySpark for large-scale data processing and performance optimization. • Deep understanding of dimensional data modeling concepts (Kimball) and practical experience building star schemas, including the definition of facts, dimensions, and slowly changing dimensions (SCDs). • ...

Similar Jobs

View all jobs →