Senior Data Engineer

apartmentAvacone placeBasel scheduleVollzeit calendar_month 

The Opportunity

We are supporting a major data platform transformation within a banking environment, moving from a legacy SQL Server and SSIS-based setup to a modern, scalable architecture built on dbt, Dagster, and OpenShift.

This role is not about maintaining existing systems. It is about rebuilding a critical data platform from the ground up, with direct impact on risk, trading PnL, and core financial data flows.

We are looking for a hands-on Senior Data Engineer who can take ownership of complex migration workstreams and deliver reliably in a regulated, high-stakes environment.

What You Will Do

You will play a central role in the end-to-end migration and modernisation of the data platform.

Platform Transformation
  • Translate legacy ETL logic from SSIS and stored procedures into modern ELT pipelines using dbt
  • Implement Data Vault 2.0 structures including Raw Vault and Business Vault
  • Build datamarts and curated datasets for downstream analytics and reporting
Orchestration & Infrastructure
  • Design and operate workflows using Dagster, including scheduling, dependencies, and recovery mechanisms
  • Deploy and run data workloads on OpenShift / Kubernetes environments
Event-Driven Data Processing
  • Enable near real-time data processing using Kafka-triggered pipelines
  • Integrate with upstream data lake environments and external data providers
Data Quality & Validation
  • Establish robust data validation and reconciliation processes
  • Implement automated testing and monitoring using dbt
Operational Ownership
  • Support production pipelines and resolve incidents when required
  • Create clear documentation and ensure operational readiness
  • Continuously improve performance, reliability, and maintainability

What You Will Do

You will play a central role in the end-to-end migration and modernisation of the data platform.

Platform Transformation
  • Translate legacy ETL logic from SSIS and stored procedures into modern ELT pipelines using dbt
  • Implement Data Vault 2.0 structures including Raw Vault and Business Vault
  • Build datamarts and curated datasets for downstream analytics and reporting
Orchestration & Infrastructure
  • Design and operate workflows using Dagster, including scheduling, dependencies, and recovery mechanisms
  • Deploy and run data workloads on OpenShift / Kubernetes environments
Event-Driven Data Processing
  • Enable near real-time data processing using Kafka-triggered pipelines
  • Integrate with upstream data lake environments and external data providers
Data Quality & Validation
  • Establish robust data validation and reconciliation processes
  • Implement automated testing and monitoring using dbt
Operational Ownership
  • Support production pipelines and resolve incidents when required
  • Create clear documentation and ensure operational readiness
  • Continuously improve performance, reliability, and maintainability

Requirements

What You Bring

Technical Expertise
  • Strong experience with SQL Server and T-SQL, including performance optimisation
  • Proven hands-on experience with dbt in production environments
  • Solid experience with workflow orchestration tools, ideally Dagster
  • Practical knowledge of Data Vault 2.0 modelling concepts
  • Experience working with container platforms such as OpenShift or Kubernetes
  • Familiarity with event-driven architectures and Kafka
Domain Experience
  • Experience working with financial data, ideally in banking or trading environments
  • Understanding of risk and PnL data structures is a strong advantage
Working Style
  • Strong ownership mindset with the ability to work independently
  • Structured, pragmatic, and delivery-focused
  • Comfortable operating in complex and regulated environments
  • Clear communicator across both technical and business stakeholders

What Success Looks Like

Within the first months, you will have:

  • Delivered initial Data Vault structures and migrated datasets into the new platform
  • Established stable, event-driven pipelines
  • Ensured data consistency and validation between legacy and new systems
  • Contributed to a production-ready, scalable data platform
apartmentBachem AGplaceBubendorf, 18 km von Basel
to turn process data into decisions with measurable business impact. We have delivered analytical insights with measurable top-line impact and applications that chemists and engineers describe as fundamentally changing how they work. Now, we are looking...
local_fire_departmentDringend gesucht

Senior Network Engineer

apartmentAvaconeplaceBasel
banking operations, trading systems, data centres, cloud connectivity, and branch infrastructure. This position is ideal for an engineer who combines strong architectural understanding with a practical delivery mindset and who remains technically engaged...
apartmentBotnar Institute of Immune Engineering - BIIEplaceBasel
in a research setting. Familiarity with Benchling is a strong plus.  •  Strong communication skills: able to explain technical workflows to scientists and document procedures clearly.  •  Eagerness to learn requirements engineering, data and process modelling...