Principal Apache Hadoop Engineer

BairesDev

BairesDev is an expert partner that works with clients to build high-quality software solutions.

About the Company

BairesDevยฎ has been at the forefront of technology innovation for over 15 years, delivering advanced digital solutions to global leaders such as Google and some of the most dynamic startups in Silicon Valley. With a team of over 4,000 professionals representing the Top 1% of global tech talent, the company drives transformative results across industries. BairesDev is dedicated to aligning individual passions and expertise with world-class projects, empowering professionals to grow, lead, and make a measurable impact worldwide.

About the Role

The Principal Apache Hadoop Engineer will take a leading role in designing, optimizing, and maintaining large-scale distributed data processing systems within the Apache Hadoop ecosystem. This position involves deep technical engagement with open-source projects, guiding architectural decisions, and mentoring engineering teams. The ideal candidate will be a hands-on technical leader passionate about building scalable, high-performance data platforms and contributing to the broader open-source community.

Responsibilities

  • Contribute as a core developer to Apache Hadoop and related open-source projects, dedicating approximately 20% of time to upstream development.
  • Architect and optimize big data ecosystem solutions for large-scale data processing.
  • Design and enhance HDFS components to ensure robust distributed storage capabilities.
  • Implement advanced MapReduce paradigms for scalable data workflows.
  • Optimize YARN resource management for improved cluster efficiency and throughput.
  • Provide technical leadership, mentorship, and best practice guidance to engineering teams.
  • Drive innovation and collaboration across multiple projects and teams.

Required Skills

  • 10+ years of experience in software development, with a focus on big data and distributed systems.
  • Expert-level understanding of HDFS architecture and distributed file system concepts.
  • Strong experience with MapReduce, YARN, and cluster orchestration.
  • Proven contributions to Apache open-source projects such as Hadoop, HDFS, or Spark.
  • Deep technical knowledge of parallel processing and large-scale data system design.
  • Exceptional problem-solving, analytical, and communication skills.
  • Advanced English proficiency.

Benefits

  • Competitive compensation in USD or local currency.
  • Paid parental leave, vacation, and national holidays.
  • Innovative, multicultural, and collaborative work environment.
  • Opportunities to work alongside the global Top 1% of engineering talent.
  • Comprehensive career growth support through mentorship, upskilling, and promotions.
  • Access to diverse global projects with cutting-edge technologies.

For additional information and the full job description, visit the link to our official website below:

Copyright © 2025 hadoop-jobs. All Rights Reserved.