Sodexo-Senior Data Engineer
Sodexo is currently hiring a Senior Data Engineer to join their amazing team About the company: Sodexo is a global leader in quality of life services, founded in France in 1966. The company operates in over 50 countries, offering a wide range of services that include:On-site Services: Catering, cleaning, maintenance, and facilities management. Benefits & Rewards Services: Meal cards, gift cards, mobility solutions, and employee benefits. Personal & Home Services: Concierge, home care, and childcare (less prominent). Sodexo focuses on improving quality of life for employees, students, patients, and others across various sectors, including corporate offices, schools, healthcare, defense, and remote sites.
Sodexo in Portugal
Sodexo has been present in Portugal since 1996. Core Services, Employee Benefits & Rewards: Well known for offering meal cards like Sodexo Refeição Pass, as well as gift cards and fuel cards. On-site Services: Catering and facilities management in corporate, healthcare, and educational environments. Key Clients: Includes multinational corporations, public institutions, and local businesses. Mission in Portugal: Helping improve employee
- being and organizational performance through tailored workplace and benefits solutions. About the role/What you'll do: Tech Leadership: Privileged technical point of contact for the Product Owner and business sponsors; Ability to take data engineering responsibility of a data project; Listening skills and excellent communication skills; Ability to manage technical debt (variable according to use cases); Technical architect, lead design meetings and tech breakdowns; Ability to mentor and guide the work of a more junior Data Engineer Data Ingestion & Processing: Design and implement batch and
- time data ingestion pipelines. Use Databricks (with Py
Spark) for big data processing tasks Use DBT to transform data. Cleanse, transform, and enrich raw data to make it analytics ready. Optimize queries and data processing for speed and cost efficiency. Data Storage & Management: Design and implement database schemas, tables, and views. Optimize storage formats for querying, such as Parquet or Delta Lake. Enforce data quality checks and data lineage documentation. Implement partitioning, bucketing, and indexing strategies for efficient data retrieval. Collaboration with data experts (data analyst, data scientists): Work closely with data scientists to provide data in appropriate formats for machine learning and advanced analytics. Collaborate with Platform Engineer teams to comply with Platform good practices and escalate common needs. Assist data analysts with SQL queries, views, and report generation. Collaborate on the deployment of machine learning models to production. Security & Compliance (with Azure security experts): Implement
- based access controls and data encryption (at-rest and
- transit). Comply with industry and organizational data standards, privacy regulations, and best practices. Regularly audit data access and usage. Infrastructure & Configuration (with Azure infrastructure experts): Set up and maintain Azure cloud infrastructure. Configure and optimize Azure Data Lake Storage, Blob Storage, and other Azure storage solutions. Deploy and manage Databricks clusters for processing tasks. Implement and maintain data pipelines using Azure Data Factory. Monitor and troubleshoot
- related issues Documentation & Training: Onboard new team members, providing access and initial training on tools. Create documentation and knowledge bases for data pipelines, best practices, and tooling. Continuous Improvement: Stay updated with the latest advancements in data engineering technologies. Propose and implement optimizations for current workflows and systems. Proactively identify areas of improvement and automation. Regularly update team on new features or changes in Azure, Databricks, or related technologies. Requirements: Engineering Master’s degree or Ph
D; 5 years+ experience in a Data engineering role into large corporate organizations; Experience in a Data/AI environment into a Cloud ecosystem. Software Engineering & SQL: Python: Strong coding skills in Python, especially with libraries related to data manipulation (e. g. , Pandas) and interfacing with databases. Python ecosystem: Strong knowledge of tooling in Python ecosystem such as dependency management tooling (poetry, venv). Software engineering practices: Strong experience in putting in place good software engineering practices such as design patterns, testing (unit, integration, e2e), clean code, code formatting etc. SQL: Advanced knowledge of SQL for data querying, transformation, and aggregation. Data Architecture: Design: Ability to design scalable and robust data pipelines considering functional and
- functional requirements. Integration: Knowledge of data architectures to ensure reliable and efficient data flow. Cloud Platforms: Azure Services: Proficiency in Azure Data Lake, Azure Data Factory, Azure Blob Storage, Azure SQL Database, and other related Azure services. Azure Databricks: Proficiency in using Databricks for deploying & running Spark jobs including advanced usage (such as handling clusters, secret scopes, warehouses, unity catalog etc. ) Cloud Infrastructure: Familiarity with setting up, configuring, and managing virtual networks, VMs, security groups, and related components on Azure is preferable. Data Processing with Big Data Frameworks: Spark: Mastery of Py
Spark for data processing, particularly Data
Frame API. Delta Lake: Understanding of Delta Lake for reliable data lakes. Data Storage & Data Management. Database Management: Knowledge of both relational (SQL Server) and No
SQL databases (like Cosmos
DB). Data Formats: Familiarity with different data storage formats such as Parquet, JSON, CSV, and Delta. Data Integration Tools: Azure Data Factory: Skill in designing, deploying, and managing data integration solutions with ADF. Data Modeling: Schema Design: Ability to design efficient and scalable database schemas for both operational and analytical use cases. Should know Dimensional Modelling, 3NF (Third Normal Form) Data Modeling and Data Vault Modeling. ETL Design: Knowledge in designing Extract, Transform, Load (ETL) processes. DBT: Used to model, transform and test data on specific data products. Performance Optimization: Query Tuning: Skills in optimizing complex SQL queries. Pipeline Optimization: Knowledge in optimizing data processing pipelines, particularly in Databricks/Spark. Security & Compliance: Data Security: Knowledge of encryption techniques, both
- rest and
- transit. Access Control: Understanding of
- based access controls and integration with Azure Active Directory. Dev
Ops & Automation : CI/CD: Experience with continuous integration and continuous deployment tools like Azure Dev
Ops. Infrastructure as Code: Familiarity with tools like Azure Resource Manager (ARM) templates or Terraform is a plus Containerization: Basic understanding of Docker and Kubernetes, especially as they might integrate with Azure services is a plus.
Monitoring & Troubleshooting: Logging & Monitoring: Familiarity with tools like Azure Monitor, Log Analytics, or other monitoring solutions. Debugging: Ability to troubleshoot and debug issues within data pipelines and storage systems. What they offer: Guaranteed career paths after 12-24 months Bonus, can go up to 2 salaries, according to Employee Performance Health insurance(for employee and the family) and Life insurance (for the employee 100€ monthly, like Coverflex 24 annual leave days Training Platform- Over 7000 training courses Want to know more? Get in touch with us
- Informações detalhadas sobre a oferta de emprego
Empresa: Damia Localização: Porto
Porto, Porto, PortugalPublicado: 31. 5. 2025
Vaga de emprego atual
4 pessoas já viram esta oferta de emprego.