Are you passionate about data platforms, high-performance systems, and large-scale data processing? Join our innovative team as a Big Data Administrator & Development Engineer, where you’ll be responsible for ensuring the stability, security, and efficiency of on-premise Big Data environments. You will manage and enhance clusters based on Hadoop, Kafka, and Druid, while also contributing to automation, system optimization, and strategic architectural decisions.
This role combines deep technical administration with development responsibilities, offering the opportunity to work on cutting-edge data infrastructure that powers critical business insights and decision-making.
• Manage and Optimize Big Data Clusters
Ensure the reliable, secure, and high-performance operation of on-premise Hadoop, Kafka, and Druid clusters, proactively monitoring performance and system health.
• Implement and Maintain Monitoring Solutions
Deploy and maintain monitoring and alerting tools such as Cloudera Manager and Splunk to track system availability, detect anomalies, and ensure proactive issue resolution.
• Administer Access and Security Controls
Manage role-based access control (RBAC), encryption, and compliance policies to safeguard data integrity and meet security standards.
• Develop and Automate Data Workflows
Design and implement ETL processes, data transfers, and analysis pipelines using SQL, NiFi, Python, or Spark to support evolving business requirements.
• Support Architectural and Technical Evolution
Collaborate with cross-functional teams on architecture design, system upgrades, and technology evaluations to improve scalability, reliability, and performance.
• Minimum 8 years of hands-on experience in administering on-premise Hadoop, Kafka, and Druid clusters.
• Proven expertise with HDFS, Hive, Spark, Hue, and NiFi in enterprise data environments.
• Strong SQL proficiency, including query tuning and performance optimization.
• Solid understanding of Linux system administration and cluster maintenance best practices.
• Experience with automation and scripting tools such as Ansible, Python, or Bash.
• Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
• Experience developing ETL pipelines and integrating diverse data sources.
• Familiarity with cloud-based Big Data platforms (Azure preferred).
• Knowledge of disaster recovery, high-availability architectures, and runbook creation.
• Excellent English communication skills and ability to collaborate effectively in agile, cross-functional teams.
|
#bethechange We look forward to receiving your application. ∗ SMA is committed to diversity and equal opportunity - unattached of gender, age, origin, religion, disability or sexual orientation. |
Take the next step in your career journey
Get matched with similar opportunities at top startups
This role is hosted on SMA Solar Technology AG's careers site.
Join our talent pool first to get notified about similar roles that match your profile.