Know ATS Score
CV/Résumé Score
  • Expertini Resume Scoring: Our Semantic Matching Algorithm evaluates your CV/Résumé before you apply for this job role: Senior/Architect Data Engineer / RR.
Poland Jobs Expertini

Urgent! Senior/Architect Data Engineer / RR Position in Warsaw - Inetum Polska

Senior/Architect Data Engineer / RR



Job description

Job Description

We are seeking a highly skilled and experienced Senior/Architect Data Engineer to lead the end-to-end architecture of a Databricks-centric multi-agent processing engine.

This role will leverage cutting-edge technologies such as Mosaic AI, Model Serving, MLflow, Unity Catalog, Delta Lake, and Feature Store to automate decoder processes at scale.

The ideal candidate will have a strong background in data engineering, cloud security, and MLOps, with a proven track record of architecting solutions on the Databricks Lakehouse platform.

Key responsibilities:

  • Architecture Leadership: Lead the design and implementation of a Databricks-centric multi-agent processing engine, utilizing Mosaic AI, Model Serving, MLflow, Unity Catalog, Delta Lake, and Feature Store for scalable decoder automation.
  • Data Ingestion and Processing: Design governed data ingestion, storage, and real-time processing workflows using Delta Lake, Structured Streaming, and Databricks Workflows, ensuring enterprise security and full data lineage.
  • Model Lifecycle Management: Own the model lifecycle with MLflow, including experiment tracking, registry/versioning, A/B testing, drift monitoring, and automated retraining pipelines.
  • Low Latency Model Serving: Architect low latency model serving endpoints with auto-scaling and confidence-based routing for sub-second agent decisioning.
  • Data Governance: Establish robust data governance practices with Unity Catalog, including access control, audit trails, data quality, and compliance across all environments.
  • Performance and Cost Optimization: Drive performance and cost optimization strategies, including auto-scaling, spot usage, and observability dashboards for reliability and efficiency.
  • Production Release Strategies: Define production release strategies (blue-green), monitoring and alerting mechanisms, operational runbooks, and Service Level Objectives (SLOs) for dependable operations.
  • Collaboration: Partner with engineering, MLOps, and product teams to deliver human-in-the-loop workflows and dashboards using Databricks SQL and a React frontend.
  • Change Management: Lead change management, training, and knowledge transfer while managing a parallel shadow processing path during ramp-up.
  • Project Coordination: Plan and coordinate phased delivery, success metrics, and risk mitigation across foundation, agent development, automation, and production rollout.

Qualifications

  • Proven experience architecting solutions on the Databricks Lakehouse using Unity Catalog, Delta Lake, MLflow, Model Serving, Feature Store, AutoML, and Databricks Workflows.
  • Expertise in real-time/low latency model serving architectures with auto-scaling, confidence-based routing, and A/B testing.
  • Strong knowledge of cloud security and governance on Azure or AWS, including Azure AD/AWS IAM, encryption, audit trails, and compliance frameworks.
  • Hands-on MLOps skills across experiment tracking, model registry/versioning, drift monitoring, automated retraining, and production rollout strategies.
  • Proficiency in Python and Databricks native tooling, with practical integration of REST APIs/SDKs and Databricks SQL in analytics products.
  • Familiarity with React dashboards and human-in-the-loop operational workflows for ML and data quality validation.
  • Demonstrated ability to optimize performance, reliability, and cost for large-scale analytics/ML platforms with strong observability.
  • Experience leading multi-phase implementations with clear success metrics, risk management, documentation, and training/change management.
  • Domain knowledge in telemetry, time series, or industrial data (aerospace a plus) and prior work with agentic patterns on Mosaic AI.
  • Databricks certifications and experience in enterprise deployments of the platform are preferred.


Additional Information

Work is conducted in a hybrid system: 1-2 days a week from our offices in Warsaw, Poznań, Lublin, Katowice or Rzeszów.


Required Skill Profession

Computer Occupations



Your Complete Job Search Toolkit

✨ Smart • Intelligent • Private • Secure

Start Using Our Tools

Join thousands of professionals who've advanced their careers with our platform

Rate or Report This Job
If you feel this job is inaccurate or spam kindly report to us using below form.
Please Note: This is NOT a job application form.


    Unlock Your Senior Architect Potential: Insight & Career Growth Guide