Summary
tripla is a fast-growing travel tech company, building products for hotels across the globe to help them streamline their processes, improve efficiency, deliver high-quality customer service, and drive sales growth. We have engineers, product managers, data scientists, and other team members with diverse backgrounds and skill sets trying to do unique and impactful work in this competitive industry.
Our data team at tripla has a data platform with a modern tech stack on GCS; BigQuery, Apache Beam (DataFlow), Spark, and dbt are used to maintain a source of truth for the company on how guests interact with our hotel clients and how hotel clients use our products. The data platform powers several products for tripla directly and is used by our machine learning engineers to train models that power our chatbot product triplaBot, recommendations for guests, and so on.
We are looking for Senior Data Engineers to join our team and continue building on our data platform. We have interesting new problems involving modeling and ingesting new external data sources that give the company a better picture of the hospitality industry, and we are working on data platform improvements that make it easier for AI agents to find data and construct accurate performant queries. An ideal addition to our team would be someone with demonstrated experience developing data pipelines and data models on a big data scale using software engineering best practices who is excited to keep up with the latest advancements in data and work with the hospitality industry’s most interesting datasets.
Job Description
- Design, develop, and maintain robust batch and streaming data pipelines using Apache Beam, Spark, and BigQuery to process large-scale hospitality industry datasets
- Improve data models for tripla’s critical data entities, ensuring our data platform’s consumers can query the right thing without running up compute costs
- Work with data producers to develop data platform metadata that humans and AI agents can use to discover data and execute accurate, high-performance queries.
- Collaborate with ML engineers to architect data features and pipelines that power triplaBot and recommendation engines powering other tripla products
Qualifications
- More than 3 years of combined experience as a data engineer, data architect, and/or software engineer working on data-intensive services
- Demonstrated leadership experience on technical projects that involve frequent stakeholder communication to understand business needs and nail down and deliver on requirements.
- Proficient writing production code (Python, Scala, Java) and following software engineering best practices to build data platform components and data pipelines using big data-capable frameworks like Spark, DuckDB, Apache Beam, etc.
- Comfortable and experienced with SQL and data orchestration tools like DBT, Airflow, Prefect
- Strong understanding of data modeling, schema design, and transformation best practices.
- Fluency in English. Fluency in additional languages is a plus.
Nice to Have
- Strong understanding of embeddings, knowledge graphs, and working with vector databases
- Experience supporting RAG, context‑aware AI, and agent orchestration through data platform design
Commuting allowance: paid in accordance with company regulations
Full social insurance (employment, health, and pension)
Saturdays, Sundays and national holidays
Year-end and New Year vacations: 6 days from December 29 to January 3
There would be a few opportunities to gather as a team each year, but they are generally able to work fully remotely.
tripla Co., Ltd. Human Resources
[email protected]
