Data Engineer

ABOUT CLIENT

Our client is a global technology company that specializes in providing innovative IT solutions for the financial services industry

JOB DESCRIPTION

Design, maintain, and enhance analytical and operational services, including data lakes, databases, pipelines, and metadata repositories, ensuring timely and accurate delivery of insights.
Work closely with data science teams to define and implement data schemas and models, integrate new data sources with product teams, and collaborate with data engineers to adopt emerging technologies in the data domain.
Develop and improve large-scale batch and real-time data processing systems to support business expansion and performance enhancement initiatives.
Leverage tools like Apache Airflow and AWS Batch for efficient workflow scheduling, monitoring, and data management.
Implement robust testing strategies to guarantee the accuracy, reliability, and usability of data processing systems.Stay up to date with emerging technologies and industry best practices, identifying and implementing strategic optimizations to improve development efficiency.

JOB REQUIREMENT

Strong understanding of the principles and architectural design of distributed computing.
Extensive experience in ETL/ELT pipeline development, ensuring efficient data transformation and processing.
Proficiency in Hadoop/Spark performance tuning to enhance system efficiency.
Strong Python proficiency, with expertise in PySpark, Pandas, NumPy/SciPy, and Polars for data processing.
Experience working with version control systems, such as Git.
Hands-on experience with big data and columnar databases, including Athena, Redshift, Vertica, and Hive/Hadoop.
Familiarity with AWS and other cloud technologies, including Glue, EMR, EC2, S3, Lambda, and more.
Expertise in Docker, ECS, Kubernetes, and container management strategies.
Knowledge of CI/CD tools, including Jenkins, CircleCI, and AWS CodePipeline.
Familiarity with Java or Scala (JVM-based languages).
Experience with RDBMS (MySQL, PostgreSQL) and NoSQL databases (DynamoDB, Redis).
Exposure to enterprise-level BI platforms, such as Tableau, Looker, and PowerBI.
Understanding of cloud-based data science platforms, including AWS SageMaker and Databricks.
Experience with log ingestion and monitoring tools like ELK Stack, Datadog, and similar frameworks.
Strong grasp of data security principles, compliance, and privacy best practices.
Familiarity with event-driven architectures and message queuing technologies.

WHAT'S ON OFFER

Generous salary package
Additional month's salary
Performance-based bonuses
Access to professional English training
Comprehensive health insurance
Ample annual leave opportunities

CONTACT

PEGASI – IT Recruitment Consultancy | Email: recruit@pegasi.com.vn | Tel: +84 28 3622 8666
We are PEGASI – IT Recruitment Consultancy in Vietnam. If you are looking for new opportunity for your career path, kindly visit our website www.pegasi.com.vn for your reference. Thank you!

Job Summary

Company Type:

Information Technology & Services

Technical Skills:

Data Engineering, Python, AWS

Location:

Ho Chi Minh, Ha Noi - Viet Nam

Salary:

Negotiation

Job ID:

J00685

Status:

Active

Related Job:

Tech Lead Software Developer (Delphi, Oracle PL-SQL)

Ho Chi Minh - Viet Nam


Global Software Delivery Centers

  • Delphi

Supervising development teams in a local management role, reporting to the Software Engineering Manager in Europe Setting targets and offering guidance to local teams Ensuring quality in team development Participating in sprint planning and retrospective meetings Assigning and delivering development tasks as per sprint planning Estimating complexity and workload Selecting the most suitable technical solution to meet user requirements Designing, developing, and implementing changes to the LIMS in line with customer and business user needs Collaborating with other team members to support the LIMS Working with other team members (Engineers/QA) to assure high-quality solutions Implementing and enforcing good practices and high-quality standards

Negotiation

View details

Senior DevOps (Data Platform)

Ho Chi Minh - Viet Nam


Digital Bank, Product

  • Devops
  • Spark

Managing workloads on EC2 clusters using DataBricks/EMR for efficient data processing Collaborating with stakeholders to implement a Data Mesh architecture for multiple closely related enterprise entities Utilizing Infrastructure as Code (IaC) tools for defining and managing data platform user access Implementing role-based access control (RBAC) mechanisms to enforce least privilege principles Collaborating with cross-functional teams to design, implement, and optimize data pipelines and workflows Utilizing distributed engines such as Spark for efficient data processing and analysis Establishing operational best practices for data warehousing tools Managing storage technologies to meet business requirements Troubleshooting and resolving platform-related issues Staying updated on emerging technologies and industry trends Documenting processes, configurations, and changes for comprehensive system documentation.

Negotiation

View details

Python Developer (Distributed Systems)

Ho Chi Minh - Viet Nam


Outsourcing

  • Python
  • Flask

Engage in architecture, design, and code reviews. Contribute to strategic project development, testing, and deployment. Tackling scalability and reliability challenges will lead to meaningful discussions on Distributed Systems. Collaborate within a high-impact, cross-functional team. Utilize technologies including Kafka, PostgreSQL, Spark, BigQuery, GitLab with integrated CI/CD, etc.

Negotiation

View details