Data Engineer

ABOUT CLIENT

Our client is using new technology to develop products for the banking industry

JOB DESCRIPTION

Develop and enhance data ingestion pipelines with Python and PySpark to gather and modify data from various sources such as transactions, KYC, AML, authentication, devices, and logs.
Proficiency in SQL, with a preference for PostGres.
Design and maintain data models tailored to support Financial Crime/Fraud detection, profiling, and entity resolution.
Implement data quality checks and ensure data reliability across all environments.
Work closely with Data Scientists, Analysts, Compliance, Operations, and Product/Feature teams to put models and rules into practice.
Utilize jobs, workflows, APIs, and streaming to manage extensive data processing workloads.
Integrate with external systems, for example, sanctions, ID&V, biometrics, and authentication systems, to enhance risk and identity data.
Support automation and monitoring of ETL processes for improved operational efficiency.

JOB REQUIREMENT

Bachelor's degree or equivalent qualification
Over 5 years of experience with strong proficiency in Python, PySpark, Scala and Advanced SQL (preferably PostGres)
Hands-on experience with Databricks, Snowflake, Fabric or similar platforms
Proven hands-on experience working with structured and unstructured data in a production environment.
Familiarity with Agentic AI, MLFlow, ML models, and Eval Secure Coding practices - testing/QA
Comfortable working with cloud-based data platforms (preferably AWS).
Effective communication skills in English for collaborating with cross-functional teams in an international environment.
Proficient in working with Text, Delta, Parquet, JSON, CSV, and XML data formats.
Working knowledge of Spark structured streaming.
Experience with AWS infrastructure and working specifically with S3.
Solid understanding of git-based version control, DevOps, and CI/CD.
Experience with Atlassian stack would be a plus. Knowledge of common web API frameworks and web services.
Strong teamwork, relationship, and client management skills, and the ability to influence peers and senior management to accomplish team goals.
Willingness to embrace modern technology, best practices, and methods of work.
Experience in Financial Crime/AML, KYC, or fraud detection systems.
Familiarity with Entity Resolution frameworks (e.g., Quantexa, Sensing, open source Entity Resolution tools).
Experience with data streaming frameworks (Kafka, Spark Streaming, MQ).

WHAT'S ON OFFER

Company offers meal and parking benefits.
Full benefits and probationary salary provided.
Insurance coverage as per Vietnamese labor law and premium health care for employees and their families.
Work environment is values-driven, international, and agile in nature.
Opportunities for overseas travel related to training and work.
Participation in internal Hackathons and company events such as team building, coffee runs, and blue card activities.
Additional benefits include a 13th-month salary and performance bonuses.
Employees receive 15 days of annual leave and 3 days of sick leave per year.
Work-life balance with a 40-hour workweek from Monday to Friday.

CONTACT

PEGASI – IT Recruitment Consultancy | Email: recruit@pegasi.com.vn | Tel: +84 28 3622 8666
We are PEGASI – IT Recruitment Consultancy in Vietnam. If you are looking for new opportunity for your career path, kindly visit our website www.pegasi.com.vn for your reference. Thank you!

Job Summary

Company Type:

Offshore

Technical Skills:

Data Engineering, Big Data

Location:

Ho Chi Minh - Viet Nam

Working Policy:

Hybrid

Salary:

Negotiation

Job ID:

J01710

Status:

Active

Related Job:

Senior Deep Learning Algorithms Engineer

Ho Chi Minh, Ha Noi - Viet Nam


Product

  • Machine Learning
  • Algorithm

Analyze and optimize deep learning training and inference workloads on advanced hardware and software platforms. Work with researchers and engineers to enhance workload performance. Develop high-quality software for deep learning platforms. Create automated tools for workload analysis and optimization.

Negotiation

View details

Software Engineer

Ho Chi Minh - Viet Nam


Product

Create and develop the API Platform with a focus on reliability, performance, and providing a top-tier developer experience Deploy and enhance AI/ML models in scalable, production environments in collaboration with research and applied ML teams Manage and advance a contemporary, cloud-native infrastructure stack utilizing Kubernetes, Docker, and infrastructure-as-code (IaC) tools Ensure platform dependability by designing and implementing telemetry, monitoring, alerting, autoscaling, failover, and disaster recovery mechanisms Contribute to developer and operations workflows, encompassing CI/CD pipelines, release management, and on-call rotations Work collaboratively across teams to implement secure APIs with fine-grained access control, usage metering, and billing integration Continuously enhance platform performance, cost-efficiency, and observability to accommodate scaling and serve users globally.

Negotiation

View details

Product Manager (Data & Models)

Ho Chi Minh - Viet Nam


Product

  • Product Management
  • AI

Designing data strategy and model integration for creating efficient data pipelines, evaluation frameworks, and annotation systems to maintain high-performance LLMs. Responsible for ensuring data quality standards and implementing bias mitigation and privacy-preserving techniques. Defining the product's core model roadmaps, taking into account technical feasibility, user needs, and ethical considerations. Collaboration with researchers to incorporate experimental breakthroughs into deployable features. Partnering with Engineering and Research teams to ensure model development aligns with product goals and advocating for transparency in model decision-making to build user trust. Analyzing usage patterns from open-source communities (Discord, Reddit, GitHub) to refine model behavior and address real-world edge cases, contributing to community-driven model evolution. Setting performance benchmarks, cost efficiency, and resource utilization standards for model scalability and reliability.

Negotiation

View details