GCP Data Engineer

GeekSoft Consulting

Engaging, Enabling, Empowering.

About the Company

GeekSoft is a technology consulting firm dedicated to delivering high-quality IT-enabled services that help businesses overcome challenges and enhance performance. The company’s expertise lies in technology consulting, ERP integration, and CRM solutions built primarily on SAP and Salesforce platforms. GeekSoft’s team of young and dynamic professionals focuses on client success through a customer-oriented service model, offering tailored consulting, development, implementation, and support to create seamless and intelligent business experiences.

About the Role

The GCP Data Engineer will play a vital role in designing, building, and optimizing data pipelines and cloud-based platforms for clients. This position involves working with large-scale data systems, implementing advanced analytics solutions, and ensuring the reliability, scalability, and performance of data infrastructure. The role provides an opportunity to work in a collaborative and innovative environment with continuous learning and professional growth.

Responsibilities

  • Design, develop, and enhance data pipelines and ETL processes using Spark and Python.
  • Build and maintain Data Warehouse and Data Lake solutions for large-scale environments.
  • Research, propose, and implement new technologies aligned with industry best practices.
  • Ensure high availability, resiliency, and scalability of data products and platforms.
  • Work with distributed systems including Kafka, Apache Beam, and YARN.
  • Orchestrate data workflows using Airflow or Oozie.
  • Troubleshoot, debug, and optimize Hadoop, Spark, and Oozie jobs for performance improvements.
  • Collaborate with cross-functional teams to meet project goals and deadlines.

Required Skills

  • Proven hands-on experience in ETL job development using Spark and Python.
  • Strong programming proficiency in Python, Java, or Scala.
  • In-depth knowledge of Data Warehouse and Data Lake architectures.
  • Experience with Kafka, Apache Beam, and NoSQL databases such as HBase, Cassandra, or MongoDB.
  • Proficiency in RDBMS and PL/SQL (Oracle, MySQL, DB2).
  • Experience with Airflow, Oozie, Redis, and YARN.
  • Ability to identify and resolve performance and reliability issues in distributed environments.
  • Strong understanding of cloud platforms, especially Google Cloud Platform (GCP); AWS is a plus.
  • GCP or AWS certification is highly desirable.

Preferred Qualifications

  • Experience with big data frameworks and distributed computing systems.
  • Strong analytical and problem-solving skills with attention to detail.
  • Ability to collaborate in an agile, fast-paced environment.
  • Eagerness to learn and implement emerging data technologies.

Benefits

  • Opportunity to work in a challenging and innovative environment.
  • Continuous learning and skill development support.
  • Collaborative and dynamic team culture focused on growth and excellence.

Please refer to the official website below for a comprehensive job description:

Copyright © 2025 hadoop-jobs. All Rights Reserved.