We are hiring for a large enterprise healthcare organization to support the development and ongoing operations of distributed data platforms used across the business.
This role is focused on hands-on development of batch data processing workflows within a Hadoop ecosystem. It is not a reporting or BI position. The work centers on building, maintaining, and improving data pipelines that feed enterprise analytics and operational systems.
What You’ll Do
Design, build, and enhance data processing solutions on Hadoop platforms
Develop and support Spark-based batch workflows for large-scale data processing
Write and maintain production-grade Python and SQL code
Troubleshoot and resolve issues in existing data pipelines and jobs
Participate in code reviews and contribute to development standards
Create and maintain technical documentation and system specifications
Support applications through the full SDLC, including testing and release cycles
Ensure data integrity, availability, and performance across systems
Collaborate with developers, analysts, and project teams in a structured enterprise environment
Required Experience
3+ years of experience in software development or data engineering
Hands-on experience working with Hadoop ecosystems
Experience developing Spark batch processing jobs
Strong programming experience in Python
Solid SQL skills, including stored procedures and data manipulation
Experience supporting production environments and resolving issues
Nice to Have
Exposure to Scala or other JVM-based languages
Experience working in regulated or enterprise environments
Familiarity with change management and release processes
Technical Environment
Hadoop
Spark
Python
SQL / Stored Procedures
Additional Details
Location: Jacksonville, FL (Hybrid – onsite required)
Hours: Full-time, standard business hours
Travel: None
