
Project Lead
September 2015 - February 2019
Worked at Fidelity Investments Chennai,India as Project Lead. Finance Data Hub Project is internal data ware house to consume enterprise wide assets and flow financial data build on mysql, oracle IBM Netezza as databases and Analytics is done on AWS Hadoop. It uses Oracle business intelligence, tableau & PoweBI as reporting tools. The initial data ETL is handled Informatica tool. All the procedures , packages & pyspark codes are executed using Unix shell scripts. These batch jobs are scheduled using control-M.
Responsibilities:
Coordinate with business customers and create Technical design documents, traceability matrix & Business Requirement documents for developers.
Conduct daily Scrum meetings, backlog grooming meeting, participate in epic meeting & track team delivery every sprint using JIRA board.
Leading the Team and managing the business enhancement request & effectively tack the delivery deadlines.
Desgin Data workflow and Data pipelines in NoSQL Big Data platforms using PYSpark, Scoop, NIFI, Kafka, Hive, Python, S3 and HDFS.
Working on data migration POC from from IBM Netezza to AWS Bigdata Architecture.
Worked in various AWS tooks like EMR, EC2, Lamda, S3, Cloud formation, Redshift and SES. Created several utility scripts in AWS.
Development of Python scripts , Java scripts and Unix Shell, spark & scoop Scripts.
Perform ETL based development tasks in various RDBMS data bases like Oracle.
PLSQL & SQL development in mysql, IBM Netezza , mysql & oracle databases.
Create ETL workflow using Informatica and SSIS on both OLTP & OLAP application.
Do performance tuning on long running SQL statements, Spark SQL & hive queries.
Development of scripts to store data in parquet & avero external tables in S3 & HDFS.
Perform Data Modelling using Star, snowflake or 3NF. Build Datalake and Datamarts to perform data injection from master data manager( MDM). Perform data quality checks using Informatica MMR.
Create source qualifiers, transformations, mapping and workflows in SSIS & Informatica 10.1.1. Create new jobs and job dependencies using control-M scheduler.
Monitor ETL workflows in production and tune long running SQL queries in production. Lead team in task in stories assignment and task delivery.
Work on procedures, functions and packages creations using PLSQL in oracle DB using sqlplus and PLSQL developer. Use NZloader to load data from oracle to netezza.
Develop simple reports for business requirements using Power BI and Tableau
Monitor the daily production jobs in control-M and work on critical production failures.
Coordinate Release Management with development teams Do code versioning in Jenkins and deployment using Concourse, IBM u-Deploy Devops tools.