
Expert Data Engineer – Contract-Driven Airflow Ingestion Pipeline
FreelanceJobsResume Keywords to Include
Make sure these keywords appear in your resume to improve ATS scoring
Sign up free to auto-tailor your resume with all these keywords and get a higher ATS score
Job Description
Job Title
Expert Data Engineer Needed – Contract-Driven Airflow Ingestion Pipeline (Python, CI/CD)
Project Scope
Small to Medium (initial dataset implementation, potential follow-up work)
Level of Expertise
Expert
Location
Global (must be comfortable working with a European-based team)
Contract Type
Fixed-price preferred (milestone-based per dataset)
Description
We are building a structured data platform and are looking for an experienced Data Engineer to implement a production-grade ingestion pipeline for a predefined dataset.
This is a standards-driven execution task. Architecture, repository structure, schema templates, and data contracts are already defined. Your role is implementation and integration — not architecture design.
For a given dataset, you will:
- Implement ingestion from the approved source (API or structured dataset)
- Land raw data into object storage (bronze layer)
- Transform data into a predefined canonical schema (contract-driven)
- Implement automated data quality checks
- Integrate into an existing CI/CD workflow
- Deliver a production-ready Apache Airflow DAG
- Ensure idempotent execution and safe reprocessing
- Emit structured logs and telemetry metrics
You will work independently but in coordination with our internal data and platform team. Code review is mandatory before merge.
Deliverables
- Working Airflow DAG (production-ready)
- Python ingestion module
- Contract-compliant transformation logic
- Data quality validation layer
- CI/CD integration
- Technical documentation (README + assumptions)
Required Experience
- Strong Python (production-grade code, not notebook scripts)
- Apache Airflow (real production DAGs required)
- Experience building ingestion pipelines from APIs and/or structured datasets
- Schema-driven or contract-driven data processing
- CI/CD integration
(GitHub Actions, GitLab CI, etc.)
- Cloud storage experience (S3 or equivalent)
- Understanding of idempotent pipeline design
Important
- Architecture and schema design are predefined and must not be modified.
- Clean, maintainable, and standards-compliant code is required.
- Please share examples of Airflow DAGs you have implemented.
This initial project covers one dataset. Strong performance may lead to additional domain implementations.
Contract duration of 1 to 3 months. with 30 hours per week.
Mandatory skills:
Python, Data Modeling, CI/CD, data contract, DQ, Parquet, schema validation, Apache Airflow, Data Engineering, Data Integration
Similar Jobs
Azure Data Engineer | 4 to 10 years | Navi Mumbai | 18 to 30 LPA | Big4
Acme Services
Data Engineer – Specialist
Carrier
GCP Big Data Engineer (Full-time at a Fortune 500 tech MNC )
HARP
GCP Big Data Engineer (Full-time at a Fortune 500 tech MNC )
HARP
Gcp Big Data Engineer
Talentmatics
More Jobs at FreelanceJobs
View all →Data analyst/BI expert to integrate multiple data sources into a single dashboard in power bi
FreelanceJobs
Business Intelligence Expert for Power BI, Tableau, SQL, and Cloud Data Solutions
FreelanceJobs
Data analyst/BI expert to integrate multiple data sources into a single dashboard in power bi
FreelanceJobs
Migration from G Suite 2 office 365
FreelanceJobs
Software Engineer Needed for Dynamic Projects
FreelanceJobs
Want AI-powered job matching?
Upload your resume and get every job scored, your resume tailored, and hiring manager emails found - automatically.
Get Started Free