Data Engineer

ABOUT CLIENT

Our client is using new technology to develop products for the banking industry

JOB DESCRIPTION

Develop and enhance data ingestion pipelines with Python and PySpark to gather and modify data from various sources such as transactions, KYC, AML, authentication, devices, and logs.
Proficiency in SQL, with a preference for PostGres.
Design and maintain data models tailored to support Financial Crime/Fraud detection, profiling, and entity resolution.
Implement data quality checks and ensure data reliability across all environments.
Work closely with Data Scientists, Analysts, Compliance, Operations, and Product/Feature teams to put models and rules into practice.
Utilize jobs, workflows, APIs, and streaming to manage extensive data processing workloads.
Integrate with external systems, for example, sanctions, ID&V, biometrics, and authentication systems, to enhance risk and identity data.
Support automation and monitoring of ETL processes for improved operational efficiency.

JOB REQUIREMENT

Bachelor's degree or equivalent qualification
Over 5 years of experience with strong proficiency in Python, PySpark, Scala and Advanced SQL (preferably PostGres)
Hands-on experience with Databricks, Snowflake, Fabric or similar platforms
Proven hands-on experience working with structured and unstructured data in a production environment.
Familiarity with Agentic AI, MLFlow, ML models, and Eval Secure Coding practices - testing/QA
Comfortable working with cloud-based data platforms (preferably AWS).
Effective communication skills in English for collaborating with cross-functional teams in an international environment.
Proficient in working with Text, Delta, Parquet, JSON, CSV, and XML data formats.
Working knowledge of Spark structured streaming.
Experience with AWS infrastructure and working specifically with S3.
Solid understanding of git-based version control, DevOps, and CI/CD.
Experience with Atlassian stack would be a plus. Knowledge of common web API frameworks and web services.
Strong teamwork, relationship, and client management skills, and the ability to influence peers and senior management to accomplish team goals.
Willingness to embrace modern technology, best practices, and methods of work.
Experience in Financial Crime/AML, KYC, or fraud detection systems.
Familiarity with Entity Resolution frameworks (e.g., Quantexa, Sensing, open source Entity Resolution tools).
Experience with data streaming frameworks (Kafka, Spark Streaming, MQ).

WHAT'S ON OFFER

Company offers meal and parking benefits.
Full benefits and probationary salary provided.
Insurance coverage as per Vietnamese labor law and premium health care for employees and their families.
Work environment is values-driven, international, and agile in nature.
Opportunities for overseas travel related to training and work.
Participation in internal Hackathons and company events such as team building, coffee runs, and blue card activities.
Additional benefits include a 13th-month salary and performance bonuses.
Employees receive 15 days of annual leave and 3 days of sick leave per year.
Work-life balance with a 40-hour workweek from Monday to Friday.

CONTACT

PEGASI – IT Recruitment Consultancy | Email: recruit@pegasi.com.vn | Tel: +84 28 3622 8666
We are PEGASI – IT Recruitment Consultancy in Vietnam. If you are looking for new opportunity for your career path, kindly visit our website www.pegasi.com.vn for your reference. Thank you!

Job Summary

Company Type:

Offshore

Technical Skills:

Data Engineering, Big Data

Location:

Ho Chi Minh - Viet Nam

Working Policy:

Hybrid

Salary:

Negotiation

Job ID:

J01710

Status:

Active

Related Job:

Backend Engineer (Python/Kotlin)

Ho Chi Minh, Ha Noi - Viet Nam


Outsource

Design and develop financial products built on top of our core banking platform - Thought Machine Vault Design and develop event driven micro services for enhancing the functionality of our core banking platform Maintain and improve the reliability of our services using effective simulation, e2e and performance tests Improve SRE processes and provide production support for our services Write high quality, maintainable code using TDD Use Kubernetes and Docker to schedule and run microservices Our technology stack is predominantly Python and Kotlin / Java, but our architecture allows for using the most appropriate language to solve a given problem. - PostgreSQL, Aurora and S3 for persistence Leverage our elastic AWS infrastructure Practice continuous integration and delivery You build it, you run it.

Negotiation

View details

Senior Full-stack Java Software Engineer

Ho Chi Minh - Viet Nam


Outsource

Develop and maintain full-stack web applications using Java (main core), Spring Boot on the backend and Angular (TypeScript) on the frontend. Design and implement $1ESTful APIs, ensuring scalability, security, and performance. Participate in system design, code review, and technical discussions. Work with AWS cloud services to deploy and operate applications in production. Write unit tests and integration tests to ensure high code quality. Collaborate in Agile/Scrum teams with Product Owner, QA, and DevOps. Support and mentor junior developers.

Negotiation

View details

Senior Data Engineer

Ho Chi Minh - Viet Nam


Product

We're seeking a Staff Data Engineer to own and evolve our data infrastructure as we scale globally. You'll design and build the data systems that power our platform - from real-time pipelines and analytics infrastructure to the AI/ML foundations enabling intelligent insurance products.#Data Architecture & Engineering Design and implement scalable, future-proof data architectures aligned with business objectives across multiple regions and regulatory environments Build and maintain data pipelines for ingestion, transformation, and delivery using modern orchestration tools (Airflow, Spark, Kafka) Architect data solutions spanning data warehousing, data lakes, and real-time analytics Create and maintain data models (conceptual, logical, physical) using recognized modeling approaches Develop and document the enterprise data landscape, mapping data stores and flows across our microservices architecture#AI/ML Infrastructure Build and maintain data infrastructure supporting ML model training, deployment, and monitoring (MLOps) Design and implement vector database solutions for AI-powered features (e.g., MongoDB Atlas Vector Search, Pinecone, Weaviate) Develop data pipelines feeding recommendation engines, claims processing automation, fraud detection, and other AI-driven capabilities Ensure AI infrastructure scales globally while meeting data residency and compliance requirements#Data Operations & Quality Implement DataOps practices ensuring data quality, lineage, and governance across the platform Define and enforce data strategy and architectural principles across engineering teams Build monitoring and alerting for pipeline health, data quality, and SLA compliance Optimize query performance and cost efficiency across data systems#Technical Leadership Collaborate with product and engineering teams to translate business requirements into data solutions Act as a change agent, driving adoption of modern data practices across the organization Contribute to architectural reviews and technical decision-making Own data problems through to resolution

Negotiation

View details