About the Company
Prodapt is a leading AI-first strategic technology partner, specializing in telecom and tech industries. Recognized by Gartner as a Large, Telecom-Native Regional IT Service Provider, Prodapt delivers tailored solutions to tackle industry-specific challenges. With a focus on customer-centric partnerships, Prodapt drives business success by aligning with client goals. Operating in over 30 countries, Prodapt combines a 130-year legacy with cutting-edge innovation to shape the digital future.
About the Role
Prodapt is seeking an experienced GCP Hadoop Data Engineer to design, build, and manage scalable data architectures on Google Cloud Platform (GCP). This role will collaborate closely with business stakeholders to develop secure, high-performance data solutions that enable advanced analytics and machine learning initiatives. The ideal candidate will have extensive experience with GCP services, real-time data architectures, and a strong background in data engineering.
Responsibilities
- Solution Design & Architecture: Design scalable, reliable data pipelines for batch and real-time streaming use cases. Architect cloud-native solutions on GCP for high availability, performance, and cost efficiency. Develop embedded analytics platforms to deliver actionable insights within operational workflows.
- Data Stack Development: Build real-time data pipelines using GCP services with quality metrics. Optimize batch pipelines and streaming data architectures to handle high-volume, high-velocity data.
- Advanced Analytics & Data Science Enablement: Support ML and data science workflows, including model development, deployment, and monitoring. Facilitate predictive analytics to drive data-driven decision-making.
- Application Architecture Collaboration: Collaborate with software development teams to align applications with data and analytics architecture. Define and enforce architectural standards that meet business needs.
- Technology Leadership & Innovation: Stay updated with industry trends and technologies in data engineering, cloud, and analytics. Mentor teams on best practices in data architecture and cloud solutions.
Required Skills
- Experience: Minimum 3+ years in Data Architecture and 5+ years hands-on with Google Cloud Platform (GCP). Deep experience with GCP services such as BigQuery, Cloud Storage, Pub/Sub, Dataflow, Airflow, Spanner, Bigtable, and Cloud Composer.
- Technical Expertise: Proficient in building and managing streaming data pipelines with Apache Kafka and Google Cloud Dataflow. Strong programming skills in Python, Java, and SQL. Familiarity with Docker, Kubernetes, Terraform, and Google Cloud Deployment Manager. Expertise in ETL/ELT pipelines, data migration, and data warehousing. Knowledge of Hadoop and Spark frameworks is a plus.
- Soft Skills: Excellent problem-solving and analytical abilities. Strong communication skills to engage stakeholders and manage client relationships. Ability to work in a fast-paced environment with multiple priorities.
Preferred Qualifications
- Experience: 10+ years in solution architecture or related data roles. Familiarity with data governance, security practices, and compliance standards.
- Certifications: Google Professional Cloud Architect certification is required.
- Desirable Experience: Experience with embedded analytics platforms (Tableau Embedded, Power BI Embedded) and ML frameworks (TensorFlow, PyTorch).