*Location:*
Baden, Madrid | Workload: 80–100%
Who We Are
Axpo is driven by a single purpose to enable a sustainable future through innovative energy solutions. As Switzerland's largest producer of renewable energy and a leading international energy trader, we leverage cutting-edge technologies to serve customers in over 30 countries. We thrive on collaboration, innovation, and a passion for driving impactful change.
About The Team
You'll report to the Head of Development and work closely with the Chief Data & Analytics Office (CDAO) as part of a cross-functional effort to build a secure, scalable, and business-aligned data platform. Our mission is to empower Axpo's decentralized business hubs with self-service analytics and AI capabilities, combining the strengths of engineering, governance, and business ownership
What You Will Do
As a Databricks Solution Architect, you will play a pivotal role in Axpo's enterprise data transformation by designing and governing scalable and secure solutions on the Databricks Lakehouse platform.
You will:
* Design performant, secure, and cost-effective Lakehouse architectures that adhere to enterprise data governance and domain modeling standards defined by the CDAO.
* Lead the design of performant, secure, and cost-effective Lakehouse architectures aligned with enterprise needs.
* Collaborate with business stakeholders, engineers, and data scientists to design end-to-end solutions that enable innovation and data-driven decision making.
* Guide engineering teams on implementing technical best practices, ensuring alignment with CDAO-defined data models and stewardship principles.
* Collaborate with the CDAO office to implement Unity Catalog policies for access control, lineage, and metadata management.
* Support platform observability, data quality monitoring, and operational excellence in partnership with data governance stakeholders.
* Evaluate new Databricks features (e.g., Delta Sharing, governance enhancements) and lead their integration into platform capabilities.
* Establish solution review processes and mentor engineers and analysts on architectural thinking and Databricks capabilities.
* Support security, compliance, and cost-optimization efforts in close collaboration with platform and cloud teams.
What You Bring & Who You Are
You are a strategic thinker with hands-on technical expertise and a strong focus on business value. You bring:
* A degree in Computer Science, Data Engineering, Information Systems, or related field.
* 5+ years in data engineering and 3+ years in architecture roles, with deep experience designing solutions on Databricks and Apache Spark.
* Strong grasp of Delta Lake, Lakehouse architecture, and Unity Catalog policy implementation in coordination with data governance functions.
* Expertise in Python, SQL, and optionally Scala; strong familiarity with dbt and modern ELT practices.
* Proven experience integrating Databricks with Azure services (e.g., Data Lake, Synapse, Event Hubs).
* Hands-on knowledge of CI/CD, GitOps, Terraform, and orchestration tools (e.g., Dragster, Airflow).
* Sound understanding of enterprise data architecture, data governance, and security principles (e.g., GDPR).
* Strong communication and stakeholder management skills, able to bridge technical and business domains.
* Fluency in English; other European languages a plus.
Technologies You'll Work With
* Core: Databricks, Spark, Delta Lake, Unity Catalog, dbt, SQL, Python
* Cloud: Microsoft Azure (Data Lake, Synapse, Storage, Event Hubs)
* DevOps: Bitbucket/GitHub, Azure DevOps, Terraform
* Orchestration & Monitoring: Dragster, Airflow, Datadog, Grafana
* Visualization: Power BI
* Other: Confluence, Docker, Linux
Nice to Have
* Knowledge of Microsoft Fabric or Snowflake
* Familiarity with Dataiku or similar low-code analytics platforms
* Experience with enterprise metadata and lineage solutions
* Background in energy trading or related industries