09/22/2024 updated

**** ******** ****
100 % available

Data Engineer | AWS, Snowflake, Pyspark, Sql, Python,ETL,Stored Proc

NOIDA, India BTech Computer Science
NOIDA, India BTech Computer Science

Profile attachments

Harsh_Latest_5.pdf

AirflowAmazon Web ServicesAmazon S3Business LogicAutomationBig DataBusiness DevelopmentBusiness ProcessesCloud ComputingContinuous IntegrationData ValidationInformation EngineeringExtract Transform Load (ETL)Data TransformationData MigrationData QualityData SecurityAmazon DynamoDBPython (Programming Language)Oracle ApplicationsRole-Based Access ControlSQL DatabasesStreamlineWorkflowsData ProcessingSnowflakeApache SparkPysparkBitbucketApi GatewayData PipelineJenkins
Cloud Data Engineering: Extensive experience with AWS services including Glue, Lambda, DynamoDB, S3, API Gateway, and Step Functions. I automate workloads and orchestrate complex data workflows to streamline business operations. Data Pipeline Optimization: Proven ability to optimize SQL code, even with massive datasets (up to 4.5 billion rows), and shift processes from monthly to weekly runs, enhancing performance and reducing processing time by up to 50%. Data Migration and Integration: Expert in migrating complex systems to the cloud using AWS Data Migration Service, Snowpipe, and Snowflake. I’ve successfully transitioned hundreds of tables and critical business modules, ensuring seamless data validation and integrity. Automation and CI/CD: Skilled in automating data workflows using Bitbucket, Jenkins, and Apache Airflow, including dynamic DAG creation for new business logic deployments. Big Data and Spark: Developed robust Spark applications and wrote complex business logic in PySpark to handle large-scale data processing tasks efficiently. Role-Based Access Control (RBAC): Created custom RBAC models in Snowflake for secure and compliant data access management tailored to client needs. Advanced SQL and Data Validation: Expertise in writing and validating SQL for data transformation, including CDC processes and ensuring data consistency across platforms like Oracle, Snowflake, and AWS.

Languages

EnglishFluent

Project history

Data Engineer

Lumiq

Banking & Financial Services

250-500 team member

I am a Data Engineer with expertise in converting SAS business logic to SQL, optimizing it for Apache Hive and Impala. I developed Spark applications, automated CI/CD pipelines using Bitbucket and Jenkins, and improved loan processing time by 50% for a major financial institution. Skilled in AWS services like Data Migration Service, Glue, Lambda, DynamoDB, API Gateway, and Step Functions for data loading and automation. I’ve built complex PySpark logic, established RBAC in Snowflake, migrated 100+ tables via Snowpipe, and transitioned modules to dbt-Snowflake with thorough validation. I optimized SQL handling 4.5 billion rows, automated Airflow DAG creation, and validated data with Snowpark. I also recommended using AWS SageMaker for customer insights. My technical skills include Python, SQL, PySpark, DBT, Airflow, AWS, and Snowflake, with strong teamwork, time management, and communication abilities.

Certificates

aws developer associate

Amazon

2023


Contact form

Log in to get in touch

You need to be logged in to use the contact form.

Sign upLog in