Clera - Your AI talent agent
LoginStart
Start
Gradera logo
Gradera

Data Engineer

full-time•Hyderabad

Summary

Location

Hyderabad

Type

full-time

Experience

5-10 years

Company links

WebsiteLinkedInLinkedIn

About this role

About Gradera — Digital Twin & Physical AI Platform 

At Gradera, we are building a next-generation Digital Twin and Physical AI platform that enables enterprises to model, simulate, and optimize complex real-world systems. Our work brings together strategy, architecture, data, simulation, and experience design to power decision-making across large-scale operational environments such as manufacturing, logistics, and supply chain networks. 

 

This platform-led initiative applies AI-native execution, advanced simulation, and governed orchestration to help organizations test scenarios, predict outcomes, and continuously improve performance. We operate with an enterprise-first mindset prioritizing reliability, transparency, and measurable business impact as we build intelligent systems that scale beyond a single industry or use case. 


Data Engineer 
 

Overview 

We are seeking skilled Data Engineers to join our Data & Digital Twin Foundation team. You will design, build, and maintain data pipelines that power digital twin platforms, real-time operational systems, and AI/ML workloads. Working closely with data architects, simulation engineers, and ML teams, you will transform raw operational data into high-quality, governed datasets that drive intelligent decision-making. 


Our core data platform stack includes:
 

Data Platform & Lakehouse 

  • Databricks (PySpark, Databricks SQL) for unified analytics and data engineering 
  • Delta Lake for ACID-compliant lakehouse architecture 
  • Unity Catalog for data governance, lineage, and access control 

Stream & Event Processing 

  • Apache Kafka for real-time event ingestion 
  • Structured Streaming for continuous data processing 
  • Delta Live Tables for declarative, quality-enforced pipelines 

Specialized Data Stores 

  • Neo4j for graph data modeling and network topology 
  • Python and SQL for data transformation 

Data Quality 

  • Delta Live Tables expectations for data validation 
  • Data profiling and anomaly detection 

Key Responsibilities 

  • Design, develop, and maintain scalable data pipelines using Databricks, PySpark, and Delta Lake 
  • Build real-time and batch data ingestion pipelines from diverse operational systems 
  • Implement data transformations that serve digital twin platforms and operational analytics 
  • Develop and maintain graph data models in Neo4j for network topology and relationship modeling 
  • Integrate Kafka event streams with Databricks for real-time operational state updates 
  • Implement data quality checks using Delta Live Tables expectations 
  • Ensure data governance compliance through Unity Catalog (lineage, access control, metadata) 
  • Optimize pipeline performance, reliability, and cost efficiency 
  • Write clean, well-documented, and testable code following engineering best practices 
  • Collaborate with ML engineers to deliver feature-engineered datasets 
  • Participate in code reviews, knowledge sharing, and continuous improvement initiatives 
  • Support production data systems through monitoring, troubleshooting, and incident resolution 

Preferred Qualifications 

  • 7+ years of hands-on data engineering experience 
  • Track record of building and maintaining production-grade data pipelines 
  • Experience with Delta Live Tables for declarative pipeline development 
  • Experience working in agile, cross-functional teams 
  • Familiarity with time-series data patterns and operational data modeling 

Highly Desirable 

  • Experience building data pipelines for digital twin or simulation platforms 
  • Familiarity with operational state modeling for real-time systems 
  • Exposure to physics-informed or time-series ML feature engineering 
  • Experience working with distributed, multidisciplinary teams 
  • Exposure to industrial domains such as Manufacturing, Logistics, or Transportation is a plus 

 

Location: Hyderabad, Telangana 
Department: Engineering 
Employment Type: Full-Time



Location

Hyderabad, Telangana


Department

Engineering


Employment Type

Full-Time


Minimum Experience

Mid-level


What you'll do

  • Design, develop, and maintain scalable data pipelines utilizing Databricks, PySpark, and Delta Lake to build real-time and batch ingestion systems. Implement data transformations, develop graph data models in Neo4j, and integrate Kafka streams to serve digital twin platforms and operational analytics.

About Gradera

For decades, the technology services industry has been built on people and projects. But in an age of intelligent systems and adaptive learning, that model has reached its limit. At Gradera, we’re defining the next evolution of enterprise transformation — Software-Orchestrated Services™ (SoS™) — where software governs how work flows across humans, digital workers, and systems to deliver measurable, governed outcomes at scale. Our model unites advisory, platforms, and solution suites into one orchestrated system of intelligence — continuously learning, evolving, and compounding value across the enterprise. Through Software-Orchestrated Services™, human expertise is amplified, not replaced. Digital workers and intelligent systems operate through governance, feedback, and explainability to deliver outcomes with trust and precision. Transformation no longer ends; it evolves. From strategy to scale, Gradera turns enterprise operations into orchestrated, self-improving systems. Our frameworks — Neural IQ™, NexusFlow™, PhiSphere™, and Value360™ — bring together governance, orchestration, and measurable ROI to help organizations accelerate outcomes and sustain continuous innovation. Founded by the leadership behind PK Global, Gradera carries decades of enterprise modernization experience — now focused on replacing project-based transformation with a governed, software-orchestrated model of continuous enterprise evolution. The result is an enterprise that thinks for itself — governed, adaptive, and built to last. Gradera — defining the era of Software-Orchestrated Services™. #SoftwareOrchestratedServices #EnterpriseAI #AdaptiveIntelligence #HumanDigitalHarmony #Gradera

Ready to join Gradera?

Take the next step in your career journey

Frequently Asked Questions

What does a Data Engineer do at Gradera?

Toggle
As a Data Engineer at Gradera, you will: design, develop, and maintain scalable data pipelines utilizing Databricks, PySpark, and Delta Lake to build real-time and batch ingestion systems. Implement data transformations, develop graph data models in Neo4j, and integrate Kafka streams to serve digital twin platforms and operational analytics..

Is the Data Engineer position at Gradera remote?

Toggle
The Data Engineer position at Gradera is based in Hyderabad, India. Contact the company through Clera for specific work arrangement details.

How do I apply for the Data Engineer position at Gradera?

Toggle
You can apply for the Data Engineer position at Gradera directly through Clera. Click the "Apply Now" button above to start your application. Clera's AI-powered platform will help match your profile with this opportunity and guide you through the application process.
Clera - Your AI talent agent
© 2026 Clera Labs, Inc.TermsPrivacyHelp

Join Clera's Talent Pool

Get matched with similar opportunities at top startups

This role is hosted on Gradera's careers site.
Join our talent pool first to get notified about similar roles that match your profile.