Anshul ROY available

Anshul ROY

Sr. Consultant/Sr. Data Engineer/Data Scientist

Profileimage by Anshul ROY Sr. Consultant/Sr. Data Engineer/Data Scientist from FrankfurtamMain
  • 60435 Frankfurt am Main Freelancer in
  • Graduation: Post Graduate in Data Analytics from IIT Bangalore
  • Hourly-/Daily rates:
  • Languages: English (Limited professional) | Hindi (Native or Bilingual)
  • Last update: 20.08.2017
Profileimage by Anshul ROY Sr. Consultant/Sr. Data Engineer/Data Scientist from FrankfurtamMain
Data Science/Data Analytics - 3 Years
R/Python - 3 Years
Data warehousing - 10 Years
ETL (IBM Datastage) - 7.5 Years
Oracle/DB2/SQL Server SQL - 10 Years
Big Data (Hadoop, Hive, Spark, Scala, Sqoop) - 3 Years
Data Modelling - 1 Year
Tableau - 3 Years
Sr. Consultant/Data Engineer/Data Scientist  Dec 2015 - Present
Deloitte Consulting USI • Hyderabad  
o Leading a team of 20+ individuals.
o Architecting & designing of integrating 9 Source Systems with 300 tables having different data transformations using 1 Generic Spark & Scala Code in Hadoop.
o Implemented the creation of feature sets on the ingested data using Datasets in Spark & Scala.
o Architecting the Azure environment & Created Standards for Project.
o Predicted Telecom Customer Churn for a Telecom Company for 70000+ customer’s data. Identified best model out of KNN, Naïve Bayes, Logistic, and SVM. Final prediction accuracy was 82.2%.
o Predicted Engagement Analytics for a Telecom Company for 70000+ customer’s data. Identified best model out of Naïve Bayes, Logistic and Random Forest. Final prediction accuracy was 81.1%. Used Lift and Gain Chart to showcase the cost difference due to analytics to Business.
o ML on Fraud Detection Model: Created model to reduce fraud by 50% by using Random forest on credit history data of customers.
o Involved in Acquisition Analytics project for a banking firm. The project was to predict the response and find the optimal customers and increase the conversion rate and reduce the overall conversion cost. Used Logistic Regression, Random Forest and XGBoost with Gain and Lift chart for optimal cost.
o Created Data Models for Azure SQL DB Databases using Toad Data Modeler.
o Creating different reports using Tableau for client presentation.

Technical Specialist                         Jul 2011 - Dec 2015
Cognizant Technology Solutions India PVT Ltd • Hyderabad      
o Lead a team of 10 to complete project development on Hadoop & its ecosystem tools like Hive, Sqoop, and Spark.
o ML on Rail-Wagon project for Canadian National Railways to predict the possibility of rail engine failure based on available sensor data & secondly to predict the faulty component with an accuracy of 75%. Identified best model out of KNN, Naïve Bayes, Logistic, and SVM.
o Implemented BI Solutions using IBM Datastage on Hadoop. Creation of reusable generic Datastage Jobs & UNIX Scripts.
o Implemented complex business requirements, Hierarchy architecture & performance tuning.
o Involved in creating Tableau Reports on Salesforce Data using filters, actions, labels etc.
o Have done POC (Proof of Concept) on connectivity to Cloud based solutions (Salesforce, Azure & AWS) from IBM Datastage.

Software Engineer                           Dec 2009 - Jul 2011
HSBC GLT India • Hyderabad
o Participated in the requirement gathering and feasibility analysis process of the project, understanding the source data and integration of the source data
o Designed & testing of complex Datastage requirement for Account Opening project of HSBC. 
o Conversion of complex DB stored procedure to Datastage Jobs.
o Responsible for END to END of the Project Development & support.
o Responsible from interacting with Client, evaluating the data in different source system & integrating them & creating Technical requirement document.
o Responsible of Design & Development of Datastage Jobs, Scheduling, Performing Check Standard & Deployment of Jobs to Production.
o Involved in design of dimensional data model – Star schema and Snow Flake Schema
o Worked SCDs to populate Type I and Type II slowly changing dimension tables from several operational
source files
o Implemented multi-node declaration using configuration files APT_Config_file for performance
o Repartitioned job flow by determining DataStage PX best available resource consumption.
o Involved in the creation of a Number of Generic Jobs to increase re-usability.
o Created a number of UNIX & Batch scripts for Automation like CheckStandard Script, Load Netezza Script, Performance Tuning Script etc.
o Alone Responsible for Integration Testing using Generic Compare Jobs.
o Applied Optimization Techniques for the better performance of the reports.

Software engineer                           Oct 2007 - Dec 2009
HCL Technologies • Chennai 
  • Skilled with Datastage & worked alongside with Business team, bridging the gap between business and ETL team for successful deliverables.
  • Interacted with the BAs in order to prepare and finalize the Functional Specs for different sources system and countries & timely closure of SIT/UAT.
  • Created Datastage Jobs for B&CPR reconciliation.
  • As Software Engineer, he was responsible to understand the project detailed data warehouse architecture.
  • He was responsible to write SQL Stored Procedures for Medium to complex requirements
  • Interact with the business team for clarity on the requirement.
  • Planning & Strategy building of the requirement implementation in IBM Datastage
  • He was responsible for the design & development of Medium complexity ETL requirements using IBM Datastage.
  • He was responsible for detailed testing of designed ETL Jobs in IBM Datastage to ensure minimal defects.
  • Involve in deployment of ETL projects in different environments like UAT/PROD.
  • He follows the quality standard set and delivers the ETL jobs with the best quality possible.
  • Involved in POC on preparation of reports using OBIEE.
  • Creation of technical specification documents for the source systems to be used for future modification and maintenance.
  • Writing, updating and maintaining end user documentation and operational procedures
  • Performing code migration and tracking code version change using SCM.
  • Writing SQL queries to fetch and validate the converted data.