Senior/Lead Data Engineer

ABOUT CLIENT

Our client is a global technology company that specializes in providing innovative IT solutions for the financial services industry

JOB DESCRIPTION

Implement technical infrastructure for compliance initiatives such as SOC 2 and GDPR, including building systems for a future data catalog and managing data access.
Design and develop scalable data pipelines for batch and event-driven data ingestion to facilitate real-time analytics and machine learning feature capabilities.
Establish foundational patterns for data quality monitoring, including automated freshness and integrity checks for critical data assets to enhance trust and reliability in the data.
Proactively lead the migration of analytical workloads from a shared production cluster to a scalable cloud data warehouse, such as Snowflake.

JOB REQUIREMENT

At least 7 years of experience in data engineering, emphasizing the construction and management of core data platforms.
Specialization in Modern Data Warehousing: Proven expertise in designing, constructing, and maintaining robust and scalable data warehouses such as Snowflake, BigQuery, or Redshift.
Proficiency in Event-Driven Architectures: Hands-on experience with real-time data processing technologies and patterns (e.g., Kafka, Kinesis, Flink, Spark Streaming).
Extensive Knowledge of Database Operations: Strong comprehension of database performance tuning, monitoring, disaster recovery, and the operational considerations of large-scale data systems.
Experience with Data Governance & Compliance: Hands-on involvement in creating technical solutions to meet compliance requirements like SOC 2 or GDPR, including data access controls and cataloging.
Pragmatic Problem-Solving Skills: Demonstrated ability to select appropriate solutions without over-engineering, while ensuring robustness in a startup environment. Proficiency in SQL and Python.
AWS Experience: Familiarity with core AWS services used in a platform context (RDS, S3, IAM, Kinesis, etc.). Experience using dbt (Data Build Tool) for data transformations in a production environment is highly desirable.
Infrastructure as Code Experience: Familiarity with tools such as Terraform for managing data infrastructure.
Experience in a Startup Environment: Comfortable working in an ambiguous and fast-paced setting.

WHAT'S ON OFFER

Generous salary package
Additional month's salary
Performance-based bonuses
Access to professional English training
Comprehensive health insurance
Ample annual leave opportunities

CONTACT

PEGASI – IT Recruitment Consultancy | Email: recruit@pegasi.com.vn | Tel: +84 28 3622 8666
We are PEGASI – IT Recruitment Consultancy in Vietnam. If you are looking for new opportunity for your career path, kindly visit our website www.pegasi.com.vn for your reference. Thank you!

Job Summary

Company Type:

Outsource

Technical Skills:

Data Engineering, Python, AWS

Location:

Ho Chi Minh, Ha Noi - Viet Nam

Working Policy:

Salary:

Negotiation

Job ID:

J00685

Status:

Active

Related Job:

Senior Deep Learning Algorithms Engineer

Ho Chi Minh, Ha Noi - Viet Nam


Product

  • Machine Learning
  • Algorithm

Analyze and optimize deep learning training and inference workloads on advanced hardware and software platforms. Work with researchers and engineers to enhance workload performance. Develop high-quality software for deep learning platforms. Create automated tools for workload analysis and optimization.

Negotiation

View details

Software Engineer

Ho Chi Minh - Viet Nam


Product

Create and develop the API Platform with a focus on reliability, performance, and providing a top-tier developer experience Deploy and enhance AI/ML models in scalable, production environments in collaboration with research and applied ML teams Manage and advance a contemporary, cloud-native infrastructure stack utilizing Kubernetes, Docker, and infrastructure-as-code (IaC) tools Ensure platform dependability by designing and implementing telemetry, monitoring, alerting, autoscaling, failover, and disaster recovery mechanisms Contribute to developer and operations workflows, encompassing CI/CD pipelines, release management, and on-call rotations Work collaboratively across teams to implement secure APIs with fine-grained access control, usage metering, and billing integration Continuously enhance platform performance, cost-efficiency, and observability to accommodate scaling and serve users globally.

Negotiation

View details

Product Manager (Data & Models)

Ho Chi Minh - Viet Nam


Product

  • Product Management
  • AI

Designing data strategy and model integration for creating efficient data pipelines, evaluation frameworks, and annotation systems to maintain high-performance LLMs. Responsible for ensuring data quality standards and implementing bias mitigation and privacy-preserving techniques. Defining the product's core model roadmaps, taking into account technical feasibility, user needs, and ethical considerations. Collaboration with researchers to incorporate experimental breakthroughs into deployable features. Partnering with Engineering and Research teams to ensure model development aligns with product goals and advocating for transparency in model decision-making to build user trust. Analyzing usage patterns from open-source communities (Discord, Reddit, GitHub) to refine model behavior and address real-world edge cases, contributing to community-driven model evolution. Setting performance benchmarks, cost efficiency, and resource utilization standards for model scalability and reliability.

Negotiation

View details