Find Amazing Talent Find Your Dream Job

Hadoop Python Data Engineer

Contract: Washington, Washington DC, US

Salary Range: 64.00 - 65.00 | Per Hour

Job Code: 367605

End Date: 2026-03-19

Days Left: 28 days, 23 hours left

Job Title: Hadoop Data Engineer
Location: Chicago, IL/Denver, CO/Washington, DC
Work Arrangement: Fully Onsite
Client Industry: Enterprise BFSI client
Duration: 18+ Months Contract
Schedule: Monday to Friday, Standard working hours

About the Role:

Seeking a highly skilled Data Engineer (hadoop plus python) to join the client in Chicago or Denver or Washington. This is a 100% onsite role supporting critical data ingestion, pipeline development, and performance optimization initiatives in a secure banking environment.

Top Three skills:

 

 

  • Python-Based Data Pipeline Engineering

    • Building production-grade pipelines

    • Performance tuning, memory management, error handling

    • Batch & near-real-time processing

  • Hadoop Ecosystem Expertise

    • Strong understanding of HDFS, YARN, MapReduce

    • Distributed processing with Spark/Hive

    • Scalable and reliable data workflows

  • Custom Data Ingestion & Integration

    • Designing ingestion from APIs, queues, DBs, files

    • Kafka & message-driven architectures

    • PHP-based data services & integration endpoints

 

What We’re Looking For:

· Design, develop, and maintain data pipelines primarily using Python, writing production-grade code to ingest, process, and move large-scale datasets across distributed environments.

· Build and enhance data processing workflows on Hadoop-based platforms, leveraging a strong understanding of Hadoop architecture (HDFS, YARN, MapReduce) to ensure scalability, reliability, and performance.

· Develop custom data ingestion and transformation logic, rather than relying solely on prebuilt tools, to support batch and near-real-time processing requirements.

· Write, maintain, and optimize Python code for distributed data processing, including performance tuning, memory management, and error handling in large data environments.

· Contribute hands-on code in PHP to support data services, ingestion endpoints, or integration layers that interact with data pipelines or upstream/downstream system

Compensation:
Hourly Rate: $63 – $64 per hour
This range reflects base compensation and may vary based on location, market conditions, experience, and candidate qualifications.

Benefits:
The Company offers the following benefits for this position, subject to applicable eligibility requirements: medical insurance, dental insurance, vision insurance, 401(k) retirement plan, life insurance, long-term disability insurance, short-term disability insurance, paid parking/public transportation, (paid time , paid sick and safe time , hours of paid vacation time, weeks of paid parental leave, paid holidays annually - AS Applicable)

About Us:

At Collabera, we don’t just offer jobs—we build careers. As a global leader in talent solutions, we provide opportunities to work with top organizations, cutting-edge technologies, and dynamic teams. Our culture thrives on innovation, collaboration, and a commitment to excellence. With continuous learning, career growth, and a people-first approach, we empower you to achieve your full potential. Join us and be part of a company that values passion, integrity, and making an impact.

 

Ready to Apply?

Apply now on shivani.soni@collabera.com/973-841-2468

Note: we are looking for python programming with data integration rather than ETL Tooling.

Job Requirement
  • python programming
  • data integration
  • hadoop
  • hadoop python
  • Design develop maintain data pipelines primarily using Python
  • data processing workflows on Hadoop-based platforms
  • Develop custom data ingestion
  • Write maintain optimize Python code
  • Contribute hands-on code in PHP to support data services
  • Data Pipeline Engineering
Reach Out to a Recruiter
  • Recruiter
  • Email
  • Phone
  • Shivani Soni
  • shivani.soni@collabera.com
Apply Now
Apply Now
close-icon