Capital
Data Engineer (Financial Analytics focus)
Apply Now
Date Posted
Department
Data
Location
Melbourne, Victoria, Australia
We are a leading trading platform that is ambitiously expanding to the four corners of the globe. Our top-rated products have won prestigious industry awards for their cutting-edge technology and seamless client experience. We deliver only the best, so we are always in search of the best people to join our ever-growing talented team.
We are seeking a Data Engineer to join our Financial Analytics team developing and maintaining data pipelines, transformation logic, and data quality checks in a complex multi-jurisdiction financial data platform (PostgreSQL DWH + Airflow orchestration).
Responsabilities
- Implement enhancements and changes to existing reporting processes to improve accuracy, performance, and usability
- Design and develop new reporting pipelines and datasets aligned with business requirements
- Automate of data delivery processes
- Identify and implementation of automated data quality checks
- Resolve of issues related to data quality
- Collaborate with business stakeholders to gather reporting requirements, clarify logic, and ensure outputs meet expectations
Requirements
- 4+ years in analytics engineering or similar data-focused roles
- Advanced PostgreSQL: stored procedures and functions, complex CTEs, window functions, SCD2 patterns, query plan analysis and optimisation
- Strong understanding of data warehouse architecture: staging, core, and data mart layers; incremental load patterns; slowly changing dimensions
- Hands-on experience with Apache Airflow: DAG authoring, scheduling, dependency management, and failure handling
- Proficiency with Git (GitLab or GitHub) and JIRA
- Experience designing and evolving data warehouse architecture and data models
- Track record of building robust, maintainable ELT/ETL pipelines in production
- Experience implementing automated data quality checks
- Domain fluency in financial and trading concepts, with the ability to understand requirements and clearly explain implemented logic to business stakeholders
- High degree of autonomy: able to reverse-engineer undocumented systems, identify root causes, and take end-to-end ownership of pipelines and calculation logic
- Comfortable using AI-assisted development tools (e.g., Claude, Copilot, Cursor) to improve productivity
Nice to have
- Hands-on experience with dbt, particularly in the context of migration or adoption initiatives
- Exposure to Snowflake or strong interest in working with it as part of a target data architecture
- Proficiency in Python for scripting, automation, and data pipeline tooling
- Background in fintech or financial services in any capacity
What you will get in return
- Competitive compensation
- A generous paid leave policy, supporting a healthy work–life balance
- Two additional paid days per year dedicated to volunteering and giving back
- Private medical insurance for your peace of mind
- An additional flexible benefits budget, allowing you to tailor benefits to your needs
- Flexible working mode
- The opportunity to work from almost anywhere in the world for up to 30 days per year
- Annual company-wide events held in locations around the globe
- In-office massages to support wellbeing
Be a key player at the forefront of the digital assets movement, propelling your career to new heights! Join a dynamic and rapidly expanding company that values and rewards talent, initiative, and creativity. Work alongside one of the most brilliant teams in the industry.
Interested in this job?
Apply for this position
Date Posted
Department
Data
Remote
Yes
Salary
N/A