Cloud Engineer (AWS Kafka)

ABOUT CLIENT

Our client is using new technology to develop products for the banking industry

JOB DESCRIPTION

Designing, implementing, and maintaining streaming solutions using AWS Managed Streaming for Apache Kafka (MSK).
Monitoring and managing Kafka clusters to ensure optimal performance, scalability, and uptime.
Configuring and fine-tuning MSK clusters, including partitioning strategies, replication, and retention policies.
Collaborating with engineering teams to design and implement event-driven systems and microservices architectures.
Developing and maintaining robust data pipelines for real-time data processing and streaming using Kafka.
Ensuring seamless integration between MSK/SQS/SNS and other AWS services such as Lambda, EventBridge Pipes, S3.
Analyzing and optimizing the performance of Kafka clusters and streaming pipelines to meet high-throughput and low-latency requirements.
Implementing best practices for Kafka topic design, consumer group management, and message serialization (e.g., Avro).
Implementing security best practices for MSK, including encryption, authentication, and access controls.
Ensuring compliance with industry standards and regulations related to data streaming and event processing.
Setting up comprehensive monitoring and alerting for Kafka clusters and streaming applications using AWS CloudWatch and Datadog.
Troubleshooting and resolving issues related to data loss, message lag, and streaming failures.
Designing and implementing data integration solutions to stream data between various sources and targets using MSK.
Leading data transformation and enrichment processes to ensure data quality and consistency in streaming applications.

JOB REQUIREMENT

Bachelor's or Master's degree in Computer Science, Information Technology, or related field.
Minimum 5 years of experience in event-driven architectures and streaming solutions.
Proficiency in Apache Kafka, with at least 2 years specifically in AWS MSK.
Design and implementation experience of high-throughput, low-latency streaming applications in AWS environments.
Strong understanding of Kafka internals and proficiency in programming languages such as Java, Python, or Scala.
Experience with AWS services like Lambda, Kinesis, S3, and IAM in conjunction with MSK.
Familiarity with CI/CD tools and IaC tools like CloudFormation, Terraform, or CDK.
Strong analytical and problem-solving skills with effective communication and collaboration abilities.
AWS Certified Solutions Architect, AWS Certified Developer, or similar AWS certification.
Strong analytical and problem-solving skills and effective communication and collaboration abilities.
Ability to manage multiple priorities and projects in a fast-paced environment.

WHAT'S ON OFFER

Company offers meal and parking benefits.
Full benefits and probationary salary provided.
Insurance coverage as per Vietnamese labor law and premium health care for employees and their families.
Work environment is values-driven, international, and agile in nature.
Opportunities for overseas travel related to training and work.
Participation in internal Hackathons and company events such as team building, coffee runs, and blue card activities.
Additional benefits include a 13th-month salary and performance bonuses.
Employees receive 15 days of annual leave and 3 days of sick leave per year.
Work-life balance with a 40-hour workweek from Monday to Friday.

CONTACT

PEGASI – IT Recruitment Consultancy | Email: recruit@pegasi.com.vn | Tel: +84 28 3622 8666
We are PEGASI – IT Recruitment Consultancy in Vietnam. If you are looking for new opportunity for your career path, kindly visit our website www.pegasi.com.vn for your reference. Thank you!

Job Summary

Company Type:

Digital Bank, Product

Technical Skills:

Kafka, AWS

Location:

Ho Chi Minh - Viet Nam

Working Policy:

Hybrid

Salary:

Negotiation

Job ID:

J01556

Status:

Close

Related Job:

Senior DevOps Engineer

Ha Noi - Viet Nam


Financial services, Crypto

  • Devops
  • AWS

#About the Role We are looking for a highly skilled Senior DevOps Engineer to join our team. You will play a key role in designing, implementing, and maintaining infrastructure solutions that ensure the stability, reliability, and scalability of our systems. The role requires a balance of strong technical expertise, problem-solving skills, and a passion for automation. You will collaborate with cross-functional teams to drive process improvements, enhance engineering productivity, and support business-critical applications. As a Senior DevOps Engineer, you will also be responsible for ensuring system security, leading operational automation efforts, and participating in a 24/7 on-call rotation to maintain uptime and business continuity. In line with our philosophy that development and operations are inseparable, you will leverage your coding expertise to intervene in application modules, develop new automation tools, and ensure seamless integration between code and infrastructure.#Key Responsibilities Develop, maintain, and manage tools to automate operational activities and improve engineering efficiency, including writing custom modules in Node.js or Golang for task management and system orchestration. Troubleshoot, diagnose, and resolve complex software and infrastructure issues, including debugging and modifying application code in Node.js and Golang environments. Update, track, and resolve technical issues in a timely manner. Recommend architectural enhancements and propose process improvements for scalability and reliability. Contribute to application development by intervening in existing modules or creating new ones to enhance system manageability, scalability, and performance Evaluate and implement new technologies, frameworks, and vendor products to support business goals. Apply best-in-class security practices to safeguard critical systems and data. Ensure stability, reliability, and performance of production and non-production environments. Collaborate with engineering, QA, and product teams to align infrastructure with development needs. Participate in a 24/7 on-call rotation to support high-availability systems.

Negotiation

View details

DevOps Engineer

Ho Chi Minh - Viet Nam


Product, Offshore

  • Devops
  • Java
  • Kubernetes

As a Mid-level DevOps Engineer, you will play a key role in building, maintaining, and automating the environments, CI/CD pipelines, and infrastructure that power the mission-critical solutions for our internal teams and external clients. Maintain and ensure the availability of development, staging, and production environments Manage access, runtime stability, and environment upgrades Design, build, and improve CI/CD pipelines using tools like Jenkins, Automate build, testing, and deployment processes Troubleshoot and resolve pipeline issues Develop and maintain automation scripts using Ansible Standardize infrastructure configurations and automate environment provisioning Configure and maintain monitoring, logging, and alerting systems (Grafana, Prometheus, Splunk) Enhance observability with proactive alerting and dashboards Respond to alerts and incidents within agreed SLAs Triage and resolve infrastructure and pipeline issues Document incidents and implement preventative measures Apply system hardening, vulnerability remediation, and patching Support audits and compliance checks Monitor system performance and resource usage Conduct tuning (e.g., JVM, database, message broker) and provide optimization recommendations

Negotiation

View details

Senior Machine Learning Engineer

Ho Chi Minh, Ha Noi - Viet Nam


Information Technology & Services

  • Machine Learning

Creating the V1 Evaluation Platform: You will be responsible for designing and building the core backend systems for our new LLM Evaluation Platform, using Arize Phoenix as the basis for traces, evaluations, and experiments. Implementing Production Observability: You will need to architect and implement the observability backbone for our AI services by integrating Phoenix with OpenTelemetry to establish a centralized system for logging, tracing, and evaluating LLM behavior in production. Standardizing LLM Deployment Pipeline: You will be in charge of designing and implementing the CI/CD framework for versioning, testing, and deploying prompt-based logic and LLM configurations, ensuring reproducible and auditable deployments across all AI features. Providing Practical Solutions: Your role will involve making pragmatic technical decisions that prioritize business value and speed of delivery, in line with our early-stage startup environment. Collaborating with Other Teams: You will work closely with the Data Science team to understand their workflow and ensure that the platform you build meets their core needs for experiment tracking and validation. Establishing Core Patterns: You will also help in establishing and documenting the initial technical patterns for MLOps and model evaluation that will serve as the foundation for future development.

Negotiation

View details