Skip to main content
Wissen Technology logo

Databricks Engineer- GCP Cloud

Wissen Technology
Full Timesenior
INPosted April 8, 2026

Resume Keywords to Include

Make sure these keywords appear in your resume to improve ATS scoring

PythonSQLGCPApacheBigQueryGitGitHubRESTSparkAirflow

Sign up free to auto-tailor your resume with all these keywords and get a higher ATS score

Job Description

Role Overview:

You will be responsible for building and maintaining scalable, distributed, fault-tolerant data pipelines on GCP, including BigQuery-based lakehouse layers and Dataproc-driven Delta Lake workflows. Your role will also involve actively participating in meetings with various stakeholders across data engineering, compliance, and business teams globally.

Key Responsibilities:

  • Build and maintain scalable, distributed, fault-tolerant data pipelines on GCP, including BigQuery-based lakehouse layers and Dataproc-driven Delta Lake workflows
  • Actively participate in meetings with various stakeholders across data engineering, compliance, and business teams globally
  • Understand market data processing and transformation needs; build pipelines to acquire, normalize, transform, and release large volumes of financial data through the OMDP data factory
  • Design and implement bitemporal data models (valid-time + system-time) on BigQuery to support certified, regulatory-grade time-series datasets
  • Build, use, and maintain software testing frameworks (unit / non-regression / user acceptance) for data pipelines and transformation logic
  • Take complete ownership of solutions and assigned tasks, including ingestion pipelines, QA workflows, correction management, and audit trail implementation
  • Work in a collaborative manner with other team members and contribute to shared platform services rather than vertical-specific implementations
  • Have business acumen to understand financial concepts around reference data related to equities and other asset classes
  • Support teams across data and technology in implementing AI solutions and integrating their services with MSCI's data science products and platforms, including AI-assisted ingestion, anomaly detection, and semantic search over the lakehouse using Vertex AI

Qualification Required:

  • 6-8 years of experience in data engineering
  • Proficient in Python programming data pipeline development, transformation logic, and automation scripts
  • Proficient in data query and analysis using SQL, with strong hands-on experience in BigQuery partitioning, clustering, materialized views, and time-series query patterns at scale
  • Hands-on experience building and scheduling pipelines using Cloud Composer (Apache Airflow) DAG authoring, SLA alerting, retry logic, and dependency management
  • Working knowledge of Dataproc (Apache Spark) batch ingestion, Delta Lake merge operations, and incremental data processing
  • Proficient in AI-assisted development tools such as GitHub Copilot, Cursor, or others for accelerating code generation and enhancing developer productivity
  • Code versioning and collaboration using Git branching strategies, pull request workflows, and pipeline-as-code practices
  • Familiarity with REST APIs consuming external data vendor APIs and building service-layer integrations
  • Familiarity with GCP cloud technologies Cloud Storage, Pub/Sub, Datastream, Cloud Monitoring, IAM, and VPC Service Controls

Note: For more information about Wissen Technology, please visit their official website at www.wissen.com. Role Overview:

You will be responsible for building and maintaining scalable, distributed, fault-tolerant data pipelines on GCP, including BigQuery-based lakehouse layers and Dataproc-driven Delta Lake workflows. Your role will also involve actively participating in meetings with various stakeholders across data engineering, compliance, and business teams globally.

Key Responsibilities:

  • Build and maintain scalable, distributed, fault-tolerant data pipelines on GCP, including BigQuery-based lakehouse layers and Dataproc-driven Delta Lake workflows
  • Actively participate in meetings with various stakeholders across data engineering, compliance, and business teams globally
  • Understand market data processing and transformation needs; build pipelines to acquire, normalize, transform, and release large volumes of financial data through the OMDP data factory
  • Design and implement bitemporal data models (valid-time + system-time) on BigQuery to support certified, regulatory-grade time-series datasets
  • Build, use, and maintain software testing frameworks (unit / non-regression / user acceptance) for data pipelines and transformation logic
  • Take complete ownership of solutions and assigned tasks, including ingestion pipelines, QA workflows, correction management, and audit trail implementation
  • Work in a collaborative manner with other team members and contribute to shared platform services rather than vertical-specific implementations
  • Have business acumen to understand financial concepts around reference data related to equities and other asset classes
  • Support teams across data and technology in implementing AI solutions and integrating their services with MSCI's data science products and platforms, including AI-assisted ingestion, anomaly detection, and semantic search over the lakehouse using Vertex AI

Qualification Required:

  • 6-8 years of ex

Want AI-powered job matching?

Upload your resume and get every job scored, your resume tailored, and hiring manager emails found - automatically.

Get Started Free