Hadoop Developer

  • Full Time
  • Plano
Dexian

In the business for good.

About the Company

Dexian, launched in 2023 from the combination of DISYS, Signature Consultants, and others, is a leading staffing and IT workforce provider with 12,000 employees across 70 locations. Dexian connects talent and technology to drive transformative results for clients. Brands include Dexian DISYS, Signature Consultants, Government Solutions, Talent Development, and IT Solutions.

About the Role

Seeking an experienced developer with 3+ years designing ETL solutions using tools such as Informatica or SSIS. Strong background required in Hadoop ecosystem development with hands-on skills in Spark, Impala, Hive, and Python. Expertise in operating and troubleshooting Cloudera Hadoop components, including Hadoop FS, Hive, Impala, Spark, Kafka, Hue, Oozie, Yarn, Sqoop, and Flume. Experience with Autosys JIL scripting and proficiency in Unix shell and Perl scripting is essential. Experience handling large volumes of structured/unstructured data via MapReduce, SQL, and relational database extraction is required. Python or Scala programming skills are advantageous.

Key Responsibilities

  • Design, develop, and maintain ETL workflows and data pipelines using Hadoop ecosystem tools.

  • Build, optimize, and troubleshoot big data solutions for data ingestion, processing, and transformation.

  • Monitor and support ETL jobs and Hadoop cluster health, proactively addressing issues and performance bottlenecks.

  • Collaborate with cross-functional teams to deliver scalable data solutions aligned with business needs.

  • Automate repetitive tasks through scripting to improve efficiency and reduce manual errors.

  • Manage scheduling and execution of batch jobs using Autosys.

  • Ensure data integrity and quality across complex data environments.

Required Skills

  • Minimum 3 years of ETL development experience with tools such as Informatica or SSIS.

  • Strong hands-on experience with Hadoop ecosystem technologies: Spark, Impala, Hive, Kafka, Oozie, Yarn, Sqoop, and Flume.

  • Proficiency in Python programming and Unix shell & Perl scripting.

  • Experience with Autosys JIL scripting and job scheduling.

  • Solid SQL skills and familiarity with relational databases.

  • Ability to troubleshoot and resolve data-related and performance issues in large-scale distributed systems.

  • Knowledge of MapReduce and handling large volumes of structured and unstructured data.

  • Bonus: Experience with Scala programming.

Visit the official website below to access the full details of this vacancy:

Copyright © 2025 hadoop-jobs. All Rights Reserved.