Duties: Work as part of a scrum team and provide big data ETL solutions for within an agile methodology.
Participate in technical design discussions aiming for optimal performance and risk mitigation.
Ensure code adheres to enterprise standards and principles.
Leverage development operations techniques and practices like Continuous Integration, Continuous Deployment, Test Automation, Build Automation and Test-Driven Development to enable the rapid delivery of working code utilizing tools like GIT
Work independently and drive solutions end-to-end leveraging various big data technologies to solve data problems and develop innovative big data solutions
Coordination with the offshore delivery team on the day to day operational activities
Bachelor's Degree from an accredited college or university with a major in Computer Science, Information Systems, Engineering, Business, or other related scientific or technical discipline is required.
At least six (6) years of overall experience developing software solutions.
At least three (3) years of experience of Python + Spark in big data environments.
At least two (2) years of experience in working, defining, maintaining architectural frameworks, patterns & data structures.
Proven experience in building in data driven applications using a combination of Python and the Spark framework.
Experience working in AWS
Proven expertise w/leveraging big data components (including but not limited to Redshift, Hive, HBASE, Oozie, Kafka) to build large scale data processing systems.
Proven experience in two or more of the programming languages (Spark, Python).
Experience working in an agile environment.
Knowledge on Life Sciences domain will be an added advantage