Bell Integration

Senior Data Engineer

Job Locations ES-Valencia
Posted Date 7 days ago(11/18/2025 10:14 AM)
ID
2025-4417
# of Openings
1

Overview

We are seeking a talented Senior Data Engineer specializing in Starburst (Trino) and Dell Data Lakehouse to join our AI & Data team. You will be responsible for deploying, maintaining and optimizing Starburst installations & Dell Data Lakehouse, enabling our clients to seamlessly access their data across multiple platforms. The ideal candidate will have excellent communication skills, an advanced understanding of Starburst & Dell Data Lakehouse, and proficiency with troubleshooting and root cause analysis. 

Responsibilities

  • Deploy and manage Starburst Enterprise/Galaxy and Dell Data Lakehouse installations, overseeing environment setup, configuration, maintenance, upgrades, and ensuring optimal performance. 
  • Configure various server and application settings and parameters. 
  • Integrate Starburst with various data sources to create a unified data platform. 
  • Design and tune the container solution for performance and scalability. 
  • Set up and configure data catalogs in various modes.  
  • Implement robust security controls for data access, ensure compliance with data regulations, and manage potential vulnerabilities. 
  • Coordinate with various support partners and vendor teams.  
  • Troubleshoot and investigate server related issues and provide root cause analysis for incidents. 
  • Perform daily server administration and monitoring, and leverage automation (such as Ansible) for efficient maintenance. 
  • Plan and execute disaster recovery testing. 
  • Create documentation and provide training on Starburst administration and best practices. 

Qualifications

  • Required Skills & Experience: 
  • Bachelor’s degree in Computer Science, Information Systems, Data Science, Engineering or related field (or equivalent work experience).  
  • Proven experience with Trino/Starburst Enterprise/Galaxy administration / CLI. 
  • Implementation experience with container orchestration solutions (Kubernetes/OpenShift). 
  • Knowledge of Big Data (Hadoop/Hive/Spark) and Cloud technologies (AWS, Azure, GCP). 
  • Understanding of distributed system architecture, high availability, scalability, and fault tolerance.  
  • Familiarity with security authentication systems such as LDAP, Active Directory, OAuth2, Kerberos. 
  • Excellent Unix/Linux skills. 
  • Familiarity with JDBC / ODBC 
  • Preferred Skills: 
  • Certification: Starburst Certified Practitioner. 
  • Experience Python and/or Java programming. 
  • Proficient with infrastructure automation tools such as Ansible. 
  • Knowledge of data requirements for AI and machine learning workloads. 
  • Familiarity with Data Federation and Cached Services  
  • Familiarity with Data pipeline (Series of steps that move and transform data from one source to another for analyses and storage) 
  • Experience with Dell Data Lakehouse administration. 
  • Experience in Demand Driven Adaptive Enterprise (DDAE) administration  
     
  • Working Conditions 
  • This position may require evening and weekend work for time-sensitive project implementations. 

Options

Sorry the Share function is not working properly at this moment. Please refresh the page and try again later.
Share on your newsfeed