Senior Lead Hadoop Developer

Texas  ‐ Onsite
This project has been archived and is not accepting more applications.
Browse open projects on our job board.

Description

Description:

  • The Senior/Lead Hadoop Developer is responsible for designing, developing, testing, tuning and building a large-scale data processing system, for Data Ingestion and Data products that allow the client to improve quality, velocity and monetization of our data assets for both Operational Applications and Analytical needs. This position supports this goal with strong experience in software engineering and development of solutions within the Hadoop Ecosystem.
  • This project entails finishing the build a new platform for cost analytics, prepaid claims, cost containing edits, etc. and moving all data over from the old Legacy system to the new built platform.
  • They will be migrating data analytics from the SQL Server platform and shifting to run into a Hadoop environment
  • Healthcare is very strongly preferred
  • Will help complete the new platform

Required Experience:

  • Must have big data experience and know the concepts
  • Data integration experience
  • 4 years of Hadoop
  • 5 years of enterprise big data solutions experience
  • Scala is strongly preferred but Java background is fine too
  • Exp with Big Data Concepts.
  • Scala Program Experience.
  • Experience with Hadoop, Scala, Java, or Apache Nifi.
  • 8 years' experience in managing data lineage and performing impact analyses.
  • 5 years' experience with ETL tool development
  • 4 years' experience with Hadoop Eco System
  • Experience working in Data Management projects.Experience working in Hive or related tools on Hadoop, Performance tuning, File Format, executing designing complex hive HQL's, data migration conversion.
  • Experience working with programing language like Java/Scala or python.
  • Experience working in Agile environment.
  • Experience working with Spark for data manipulation, preparation, cleansing.
  • Experience working with ETL Tools (Informatica/DS/SSIS) for data Integration.
  • Experience designing and developing automated analytic software, techniques, and algorithms
  • Ability to handle multiple tasks and adapt to a constantly changing environment
  • Self-starter with the ability to work independently and take initiative. Ability to translate ideas and business requirements into fully functioning ETL workflows.
  • Strong analytical and problem solving skills.
  • Excellent written and oral communication skills, with the ability to articulate and document processes and workflows for use by various individuals of varying technical abilities.
  • Excellent organizational skills.
  • Knowledge of HealthCare a plus.
  • Expert ability and hands on experience in SQL and Core Java a must.
  • Experience with Unix/Linux and Shell scripting.
  • Ability to demonstrate experience in distributed UNIX environments
  • Experience working in Agile, DevOps environment a plus.
  • Experience or knowledge of web architecture, (Javascript, SOAP/XML, Weblogic, Tomcat) is a plus.
  • Experience with an ORM framework, SOA Architecture, Microservices is a plus.
  • Experience with Middleware components (ESB, API Gateway) is a plus.

Responsibilities:

  • Responsible for design, development and delivery of data from operational systems and files into ODSs (operational data stores), downstream Data Marts and files.
  • Troubleshoot and develop on Hadoop technologies including HDFS, Hive, Pig, Flume, HBase, Spark, Impala and Hadoop ETL development via tools such as Informatica,
  • Knowledge of and experience with any Azure Data Platform components - Azure Data Lake, Data Factory, Data Management Gateway, Azure Storage Options, DocumentDB, Data Lake Analytics, Stream Analytics, EventHubs, Azure SQL
  • Translate, load and present disparate data sets in multiple formats and multiple sources including JSON, Avro, text files, Kafka queues, and log data.
  • Will implement quality logical and physical ETL designs that have been optimized to meet the operational performance requirements for our multiple solutions and products, this will include the implementation of sound architecture, design, and development standards.
  • Has the experience to design the optimal performance strategy, and manage the technical metadata across all ETL jobs.
  • Responsible for building solutions involving large data sets using SQL methodologies, Data Integration Tools like Informatica in any Database preferably in an MPP platform.
  • Has strong Core Java Programming experience to apply in Data Integration
  • Works with BA's, end users and architects to define and process requirements, build code efficiently and work in collaboration with the rest of the team for effective solutions.
  • Deliver projects on-time and to specification with quality.
Start date
September 2018
Duration
4 months +
(extension possible)
From
MCG - Midwest Consulting Group
Published at
18.08.2018
Project ID:
1617467
Contract type
Freelance
To apply to this project you must log in.
Register