Resume Keywords to Include
Make sure these keywords appear in your resume to improve ATS scoring
Sign up free to auto-tailor your resume with all these keywords and get a higher ATS score
Job Description
About this opportunity
We are looking to add an experienced Data Engineer to play a critical role in designing and scaling the core infrastructure that powers the Digital Health and Discovery Platform (DHDP)’s federated ecosystem. Reporting to the National Director, DHDP, this role will build and optimize data pipelines, support secure data ingestion across institutions, and help implement privacy-by-design architecture aligned with regulatory standards.
The Digital Health and Discovery Platform, a research initiative led by the Terry Fox Research Institute (TFRI), is a federated AI and data-sharing software product for advancing disease research. The DHDP operates by bringing its software to where the data is located, analyzing it there, and delivering the results back to the platform and user.
This is a full-time, temporary position until March 2027.
About the Terry Fox Research Institute
The Terry Fox Research Institute (TFRI), named for Canadian icon Terry Fox, is built on Terry's dream to end cancer through research. Established in 2007, TFRI is a registered charity and invests its funds in cutting-edge cancer research through highly collaborative, team-oriented programs and national research networks. Together with its many research and funding partners, TFRI empowers Canada’s brightest researchers to seek out new transformational and precision medicine discoveries to advance our understanding, diagnosis, and treatment of cancer with the goal of significantly improving outcomes for patients.
Key responsibilities
Data Pipeline Development
- Design, develop, and manage scalable ETL/ELT pipelines for ingesting, transforming, and validating structured and unstructured healthcare data.
- Build data ingestion frameworks that support federated learning environments across multiple institutions.
- Optimize data workflows for performance, reliability, and reproducibility.
Cloud & Infrastructure Automation
- Automate infrastructure provisioning and deployment using Infrastructure as Code (e.g., Terraform).
- Implement CI/CD pipelines to support continuous integration and deployment of data services.
- Ensure production-grade monitoring, logging, and observability for data pipelines.
Platform Architecture & Governance
- Collaborate with architecture and cyber security teams to implement secure Azure-based data standards.
- Apply privacy-by-design principles, ensuring compliance with healthcare data regulations.
- Contribute to data modeling, warehousing strategies, and standardized data schemas for multi-site interoperability.
Collaboration & Innovation
- Partner with research, clinical, and business stakeholders to align data solutions with strategic objectives.
- Evaluate and integrate emerging technologies in big data, distributed computing, and federated learning.
- Contribute to documentation, best practices, and scalable engineering standards across the platform.
Our successful candidate
- 3 - 4 years of experience as a Data Engineer in cloud or enterprise data environments.
- Strong working knowledge of Microsoft Azure services, including:
- Azure Data Lake Storage (ADLS Gen2)
- Azure Data Factory (ADF)
- Azure Databricks
- Azure Synapse Analytics
- SQL Server
- Hands-on experience with Python, PySpark, and T-SQL.
- Experience designing and maintaining ETL pipelines in Azure Data Factory and Databricks.
- Strong communications and teamworking skills.
- Familiarity with DevOps and CI/CD workflows (Git, Jenkins, Azure DevOps).
- Solid understanding of data warehousing concepts and relational data modeling.
- Experience automating infrastructure using Terraform or similar tools.
- Understanding of healthcare data privacy, security frameworks, and compliance best practices.
- Exposure to federated learning or distributed machine learning architectures is preferred.
- Experience working with clinical or biomedical datasets is preferred.
- Knowledge of interoperability standards (e.g., HL7/FHIR) is an asset.
- Experience supporting precision medicine or research data platforms is an asset.
Benefits of working with us
- Flexible schedule
- Paid vacation
- Paid sick days
- Extended medical and dental coverage
- A work culture that values excellence, creativity, teamwork, learning, fairness and integrity
- A passionate group of high-performing teammates across Canada
Application Deadline: job requisition will be open until filled.
We appreciate all applications, but only candidates selected for an interview will be contacted.
We recognize the importance of diversity and equity in our workforce and encourage all qualified applicants to apply, including Indigenous persons, women, persons with disabilities, minorities, and other underrepresented communities.
Pay: $110,000.00-$130,000.00 per year
Benefits
- Dental care
- Extended health care
- Paid time off
Application question(s):
- Please describe your experience designing ETL/ELT pipelines.
- Have you worked in a federated learning or healthcare
Similar Jobs
Azure Data Engineer | 4 to 10 years | Navi Mumbai | 18 to 30 LPA | Big4
Acme Services
Data Engineer – Specialist
Carrier
GCP Big Data Engineer (Full-time at a Fortune 500 tech MNC )
HARP
GCP Big Data Engineer (Full-time at a Fortune 500 tech MNC )
HARP
Gcp Big Data Engineer
Talentmatics
More Jobs at Terry Fox Research Institute
View all →Want AI-powered job matching?
Upload your resume and get every job scored, your resume tailored, and hiring manager emails found - automatically.
Get Started Free