Worker Type:
EmployeeApplication End Date:
12-02-2026We are using and creating technology to transform operations at sea to enable people and the planet to thrive.
We are open-minded and fearless in our approach to innovation and don't believe in boundaries. We challenge everything and have massive ambitions to drag aging industries into the tech era.
We take safety, equality and education very seriously, and our responsibilities don't stop at our front door. Our business is built on the belief that there's definitely a more environmentally responsible way to operate at sea.
We employ people who share our core values. We expect our people to be courageous, trustworthy, and conscientious, driven by a desire to do the right thing. We strive for excellence, work collaboratively, and are genuinely excited by our work.
We offer opportunities for our people to develop beyond their role and span a multitude of disciplines. These are open to all, regardless of background and experience level. Working with us means being part of a team that is harnessing technology and creativity to disrupt a traditional industry.
We are not your average workplace.
Ocean Infinity is seeking a Data Product Engineer to transform raw operational and sensor data into automated, high-value data products that power our AI initiatives and production systems. This role sits at the intersection of data engineering, analytics engineering, and product thinking—applying domain logic to complex datasets, automating manual workflows, and delivering curated, versioned data assets that enable faster insight and greater autonomy across the organisation.
You will work closely with engineering, AI/ML, and product teams to build trusted datasets and transformation pipelines that are reproducible, governed, and usable as real internal products—not one-off analysis outputs.
What will you do:
• Build data products by transforming raw operational and sensor data into curated, versioned, high-quality datasets that power analytics, ML training, and production services.
• Design and implement scalable transformation pipelines using a Bronze → Silver → Gold approach, embedding domain logic, repeatable semantics, and robust data quality checks.
• Automate manual workflows used by analysts or product teams, turning them into maintainable, orchestrated pipelines.
• Enforce data quality and governance practices, including schema evolution, data contracts, validation, and documentation of transformation logic.
• Publish and manage data assets in Unity Catalog, ensuring discoverability, lineage, and consistent reuse across teams.
• Partner with Data Engineers, AI/ML teams, product leads, and subject matter experts to align requirements, refine data definitions, and ensure data products are fit for purpose.
• Contribute to the evolution of the lakehouse platform by improving frameworks, tooling, and standard patterns for ingestion, transformation, and consumption.
• Support engineering best practices across the data lifecycle, including version control, testing, and deployment discipline.
Who you are:
You are an engineer who thinks in products. You care about data being usable, trusted, and repeatable—not just “available.” You are comfortable working with messy real-world data, translating domain knowledge into transformation logic, and partnering closely with cross-functional stakeholders to deliver assets that become foundational building blocks across the business.
Qualifications and skills:
Essential:
· Degree in Computer Science, Engineering, Mathematics, or a related field (or equivalent practical experience).
· 3+ years’ experience in Data Engineering, Analytics Engineering, or a closely related role with strong engineering fundamentals.
· Strong proficiency in Python and SQL.
· Hands-on experience building ETL/ELT pipelines and using orchestration tools (e.g. Airflow, Prefect).
· Solid understanding of data modelling, transformation patterns, schema management, and reproducible semantics.
· Experience working with data lakes / lakehouse architectures (e.g. Delta Lake).
· Strong grasp of versioning concepts and best practices for reproducible pipelines and datasets.
· Strong communication skills and ability to work effectively with cross-functional stakeholders.
· Comfortable translating domain or business logic into clear, testable transformation code.
Desirable:
· Experience with the Azure data ecosystem (e.g. ADLS, ADF, Azure Databricks / Delta).
· Familiarity with data quality/testing frameworks (e.g. Great Expectations, Soda, dbt tests).
· Experience with CI/CD pipelines (e.g. Azure Pipelines, GitHub Actions, ArgoCD).
· Experience with Kubernetes (e.g. AKS).
· Understanding of ML feature engineering, feature stores, or supporting ML workflows in production.
Salary: The salary varies for this position as we are recruiting in multiple regional locations and job grades. The salary process is based on skills, abilities, and experience required.
What you can expect:
At Ocean Infinity, we believe in creating equal opportunities for all, celebrating each and everyone’s differences. We are driven by transforming the industry, through our technology, thoughts, behaviours and actions. Being inclusive and respectful to all is fundamental to who we are. It is the right thing to do and enables innovation and creativity to thrive.
There is more work to be done, and we know that we aren’t perfect, but our commitment to these values is unwavering. They are central to our mission and the impact we have on the industry, meaning, we cannot live without them.
Take the next step in your career journey
Get matched with similar opportunities at top startups
This role is hosted on Ocean Infinity's careers site.
Join our talent pool first to get notified about similar roles that match your profile.