Senior Data Engineer
Tài chính & bảo hiểm
Công nghệ thông tin
Our client, one of biggest banks in Vietnam is looking for a senior data engineer:
· Lead the team to improve scalability, reliability, and cost-efficiency of the Data Platform.
· Design, build, and deploy data pipelines (batch & streaming) using Spark orchestrated via Airflow.
· Develop libraries and frameworks for data ingestion, transformation, and governance with clean architecture principles.
· Collaborate with Data Architects to design/review data models, enforce data contracts, and maintain schema governance.
· Optimize performance with partitioning, caching, Z-Ordering, and metadata management in lakehouse environments (Delta/Iceberg/Hudi).
· Ensure security and compliance: IAM, encryption, secrets management, and GDPR/CCPA adherence.
· Drive CI/CD for data workflows, IaC (Terraform), and container orchestration (Kubernetes).
· Monitor SLOs/SLAs, implement alerting, and lead incident responses and postmortems.
· Design and operate end-to-end ML/LLM pipelines: data prep, training, evaluation, and deployment.
· Build RAG architectures, vector search, and embedding pipelines for LLM-based applications.
Must have:
· Bachelor’s or master’s degree in computer science, Software Engineering, Information Technology, or a related technical field
· English is required
· Have 5+ years of experience as a Data Engineer or Software Engineer
· Have experience in Cloud (AWS/Azure/GCP)
· Extremely proficient in at least 1 programming language (Python/Scala/Java)
· Strong experience in systems architecture – particularly in complex, scalable, and fault-tolerant distributed systems
· Good at multi-threading, atomic operations, computation framework: Spark (DataFrame, SQL, ...), distributed storage, distributed computing
· Understand designs of resilience, fault-tolerance, high availability, and high scalability, ...
· Tools: CI/CD, Gitlab, ...
· Good at communication & team working
· Being open-minded, willing to learn new things
Nice to have:
· Experience with Databricks (Delta Lake, Unity Catalog, Delta Live Tables) or similar lakehouse technologies is a strong plus.
· Proven ability in performance tuning and optimization for Big Data workloads (Spark/Flink, partitioning, shuffle strategies, caching).
· Familiarity with modern data transformation frameworks (dbt).
· Knowledgeable in AI and LLM technologies is a plus, including prompt engineering, embeddings, and retrieval-augmented generation (RAG).
· Hands-on experience with vector databases (ChromaDB, Vector Search) and LLMOps practices.
Please contact our senior consultant, Phuong Linh via: + 84 94 335 6541 (Whatsapp/zalo) for further support.
Ref: JN-012026-186909