Skip to content

DATA ANALYTICS ENGINEER

  • Remote
  • €4,000 - €6,000 per month
  • CLOUD ENGINEERING

Job description

ABOUT THE COMPANY

LITIT, a joint venture between NTT DATA and Reiz Tech, is a company with deep-rooted industry know-how, dedicated to innovation within the IT sector. Its primary focus is delivering high-quality solutions in the DACH region. With a commitment to excellence, LITIT combines the best of German precision, Japanese work ethics, and Lithuanian talent to provide unparalleled IT service and support to its clients.

ABOUT THE ROLE

We are looking for an experienced AWS Data Engineer to join the development of our IoT Insurance Data Platform (IDP). In this role, you will design, build, and optimize scalable, cloud-native data solutions that power analytics, data products, and machine learning use cases in an industrial IoT and insurance environment. You will work in a modern AWS ecosystem, contributing to a data platform built on lakehouse principles, enabling high-performance data processing and governed data access.

RESPONSIBILITIES

  • Design, implement, and operate scalable, cloud-native data pipelines on AWS

  • Build and maintain ETL/ELT workflows, data lakes, and data mesh components

  • Develop and optimize PySpark-based data processing jobs for large-scale and time-series datasets

  • Design and manage data schemas, tables, and metadata using AWS Glue Data Catalog and Lake Formation

  • Ensure efficient, reliable, and secure data processing across the platform

  • Integrate external and internal systems via APIs (e.g., AWS API Gateway)

  • Contribute to CI/CD pipelines and automated deployment processes

  • Collaborate closely with data platform, analytics, and product teams in an agile (Scrum) environment

REQUIREMENTS

  • Strong hands-on experience with AWS services, including:
    Glue, Lambda, S3, Athena, Lake Formation, Step Functions, DynamoDB

  • Proven experience in data engineering, including building and operating ETL/ELT pipelines and working with data lakes and/or data mesh architectures

  • Solid experience with Spark / PySpark for distributed data processing and performance optimization

  • Experience with modern data formats such as Apache Iceberg and Parquet

  • Experience with API integration (e.g., AWS API Gateway, data exchange APIs)

  • Experience with CI/CD pipelines and automated deployments

  • Experience working in cross-functional Agile (Scrum) teams

  • Proven ability to deliver production-grade, scalable data solutions

  • Agile mindset with experience working in cross-functional Scrum teams

  • Willingness and readiness to travel as required by project or client needs is expected. This may include occasional domestic or international travel, sometimes on short notice.

    Nice to have:

  • Experience with Infrastructure as Code (e.g., Terraform)

  • Basic understanding of machine learning concepts, ML lifecycle, and MLOps practices

WHAT WE OFFER

  • Salary range: €4000 - €6000 (GROSS) / month.

  • Learning opportunities with compensated certificates, learning lunches, and language lessons.

  • Chance to switch projects after one year.

  • Team building twice a year.

  • Office in Vilnius, Lithuania that offers themed lunches and a pet-friendly environment.

  • Remote work opportunities.

  • Flexible time off depending on a project.

  • Seasonal activities with colleagues.

  • Additional health insurance and loyalty days for Lithuanian residents.

  • Referral bonuses.

  • Recognition of important occasions of your life.

or

Apply with Linkedin unavailable
Apply with Indeed unavailable