Seven West Media is one of Australia’s leading integrated media companies, with major operations in broadcast television, publishing, and digital content. They own Channel 7, The West Australian newspaper, and 7plus, a streaming platform. Their portfolio includes partnerships with leading global media brands, reaching millions of Australians across various media channels.
As a Senior Data Engineer, you will proactively design and implement data solutions that support our business needs while adhering to data protection and privacy standards. In addition to this, you would also be required to manage the technical delivery of the project, lead the overall development effort, and ensure timely and quality delivery.
Responsibilities:
- Data Acquisition: Proactively design and implement processes for acquiring data from both internal systems and external data providers. Understand the various data types involved in the data lifecycle, including raw, curated, and lake data, to ensure effective data integration.
- SQL Development: Develop advanced SQL queries within database frameworks to produce semantic data layers that facilitate accurate reporting. This includes optimizing queries for performance and ensuring data quality.
- Linux Command Line: Utilize Linux command-line tools and functions, such as bash shell scripts, cron jobs, grep, and awk, to perform data processing tasks efficiently. This involves automating workflows and managing data pipelines.
- Data Protection: Ensure compliance with data protection and privacy requirements, including regulations like GDPR. This includes implementing best practices for data handling and maintaining the confidentiality of sensitive information.
- Documentation: Create and maintain clear documentation of designs and workflows using tools like Confluence and Visio. This ensures that stakeholders can easily communicate and understand technical specifications.
- API Integration and Data Formats: Collaborate with RESTful APIs and AWS services (such as S3, Glue, and Lambda) to facilitate seamless data integration and automation. Demonstrate proficiency in parsing and working with various data formats, including CSV and Parquet, to support diverse data processing needs.
Key Requirements:
- 5+ years of experience as a Data Engineer, focusing on ETL development.
- 3+ years of experience in SQL and writing complex queries for data retrieval and manipulation.
- 3+ years of experience in Linux command-line and bash scripting.
- Familiarity with data modelling in analytical databases.
- Strong understanding of backend data structures, with experience collaborating with data engineers (Teradata, Databricks, AWS S3 parquet/CSV).
- Experience with RESTful APIs and AWS services like S3, Glue, and Lambda
- Experience using Confluence for tracking documentation.
- Strong communication and collaboration skills, with the ability to interact effectively with stakeholders at all levels.
- Ability to work independently and manage multiple tasks and priorities in a dynamic environment.
- Bachelor’s degree in Computer Science, Engineering, Information Technology, or a related field.
Good to Have:
- Experience with Spark
- Understanding of data visualization tools, particularly Tableau.
- Knowledge of data clean room techniques and integration methodologies.