A7F19791 – Harsha V A – BigData/Hadoop Developer

Resume posted by Glady in BigData Developer.
Desired Rate: $70.00/hr
Desired position type: C2C
Current Location: Bentonville Arkansas, United States

gcorreya@compunnel.com
Tel:
609-779-1361
Mobile:

Summary

• Possess 11+ years of overall IT experience with around 4+ years in Big Data/Hadoop/Spark technology.
• Have work experience in Pharma, Retail and Banking domain.
• In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and Map Reduce concepts and experience in working with Hive using Apache Hadoop for working with Big Data to analyze large data sets efficiently.
• Good hands on experience over Hadoop components – HIVE, HDFS, Sqoop, YARN, Spark
• Have better understanding over data models, referential relations and data integrity.
• Good knowledge in how to import and export data between RDBMS and HDFS using Sqoop
• Worked with Oozie that enabled to run workflow jobs with actions that run Hive
• Experienced in working with spark eco system using Spark SQL through PySpark shell on different formats like Text file, Avro, Parquet files
• Excellent understanding of Hadoop Architecture and underlying Hadoop framework.
• Extensive experience in SQL in query processing, execution, optimization, performance
• Experience in creating tables, views, functions, indexes in Oracle databases
• Experience with Agile Methodology, Scrum Methodology.
• Have regular interactions with client over understanding or clarifications on client specific requirements.
• Hands on experience in Functional testing, System testing, Regression testing and Database testing.
• Provide expertise and leadership for offshore transition planning, knowledge acquisition and steady state delivery, and set client/team expectations throughout the engagement
• Handled several techno-functional responsibilities including estimates, identifying functional and technical gaps, requirements gathering, designing solutions, development, developing documentation, and production support
• Lead execution of complex tests according to established plans and scripts, documents defects in tracking system and retests software corrections to ensure problems are resolved.
• Involved in time estimation for client specific projects.
• A good team player with excellent communication, presentation and interpersonal skills and ability to prioritize and coordinate work across different geographic locations

TECHNICAL SKILLS:

Programming Languages: C, SQL
Scripting Languages: UNIX Shell scripting
Operating Systems: Windows NT/2000, UNIX
Test Management Tool: Test Director 8.0, BBU, HP ALM12, JIRA v6.4.14
Data Warehousing Tool: Informatica 6.2, DataStage 8.5
Database Skills : MS SQL Server 2000, Oracle 10g, Teradata
Hadoop Eco System: Hadoop 2.5.0, HDFS, Hive 1.2.1, PySpark, Sqoop

Education

B.E. in Mechanical Engineering with 77% (First Class with Distinction) in the year 2002 from University B.D.T College of Engineering, Davanagere, Karnataka.

Experience

PROJECTS:
Wipro Technologies – Walmart, Bentonville January 2016 – Present
Lead Solutions Engineer
Project-1 : Customer Knowledge Platform (CKP)
Environment : Hadoop 2.5.0, HiveQL, UNIX
Test tools : JIRA v6.4.14
Description
Customer Knowledge Platform (CKP) is a data set product comprised of aggregated internal and external data, profile quality standardization, matching and enhancement to support Customer Analytics. This mainly deals with the group of detailed data sets created from integrating information from multiple data sources both internal and external. The information gathered will help –
• To create a clear view of retail performance, customer penetration, trends and predictors.
• To summarize and track performance over time at an overall company level or at granular level details such as individual customers, households, items, stores or markets.

My Role :
As Team member performed
• Acting as the lead solution engineer or technical lead responsible for the implementation of approved technologies that align with technical architecture, customer’s business plans and long-term strategy.
• Working closely with business on understanding the processing of financial transformation data requirements and provide big data solutions using HIVE, HADOOP, SPARK and tools like OOZIE, AUTOMIC and AORTA.
• Loading data from different databases and deciding on which file format is efficient for a task. The data will be loaded primarily from Teradata to Hadoop for data transformations, using Sqoop application for data imports.
• Automation of data quality checks to be done on the data that has been brought to Data Lake using Map Reduce framework in Unix shell scripts.
• Build distributed, reliable and scalable data pipelines to ingest and process data in real-time using Automic and Oozie.
• Managing Hadoop jobs using Automic tool and making sure that job flows get completed without any errors.
• Design and develop automation scripts using Unix shell scripting to trigger the HIVE queries present in HQL files to bring the data in the Raw zone and Cataloged zone of the Data Lake.
• Developing HIVE queries and automating them by creating workflows using Automic and Oozie tools.
• Defining Hadoop Job Flows in Automic tool and execute the job flows to bring the data to Walmart Data Lake.
• Identify the process gaps by analyzing data from Jira and Confluence tools to understand the gaps and dashboarding them and provide alternate ways of working to project team.
• Gathering business requirements from customer and suggest enhancements from end-user perspective and then sign off for development. This is done by analyzing requirements utilizing domain knowledge and current market/business trends.
• Delivered technical designs, test plan that meet business and product features.

Project-2 : Legacy Assortment Discipline
Environment : Hadoop 2.5.0, HiveQL, UNIX
Test tools : JIRA v6.4.14, HPALM 12

Description
Legacy Assortment Discipline project that allows the business to generate customer insights through analytics such as customer loyalty to products by region or brand, customer buying habits in terms of frequency or purchase cycle, substitution of products and much more. This also provides the customer data that assists the merchants in making a better-informed decision on what products/number of items should be placed on store shelves and made available to the customer.

My Role :
As Team member performed
• Analyze and understand source to target mapping requirements and be able to translate those requirements into written test plans, detailed test cases and complex Hive queries as part of test scenarios.
• Building HIVE queries to validate the data completeness and data quality against the aggregated data in the target table.
• Involved in creation of hive tables, loading the structured / unstructured data from the files using HDFS commands.
• Interacted with business users for any clarifications with respect to domain and incorporated the related scenarios in testing.
• Have led the offshore QA team in driving the testing activities from onsite and assisting the team in understanding on each User stories.
• Preparation of Test script comprising both positive and negative test scenarios.
• Involved in peer review of test deliverables within the team and has arranged review meetings with other teams such as Dev/UAT team etc.
• Validating the completeness, integrity and accuracy of data across various systems.
• Preparation of test reports and metrics based on execution and defect process and responsible for sending the daily and weekly status reports to the client and project managers within Wipro.
• Ensured that design and execution of test cases and defect management are maintained through JIRA.
• Ensured that the test cases are linked to the respective user stories in JIRA in compliance to Agile Methodology.

Wipro Technologies – Target, India, Bangalore March 2012 – November 2012
Specialist
Project-3 : PARS DBI DENTON
Environment : Teradata, DataStage 8.5, Unix
Test tools : ETL Testing, Quality Center 11.0
Description :
The PARS Program introduces new data warehouse technology and functionality to enable enterprise Business Intelligence and Analytics. As a result, PARS DBI and Denton will replace the current IBM database with Teradata as the enterprise data warehouse (EDW).
PARS Distribution BI project is mainly focused on Food DC’s and operations around it. Target would like to capture all the operations In & Out DC, which is from Inbound to Outbound, by knowing these operations, target will be able to optimize the Inventory maintenance and can monitor the equipment performance and able to make better decisions on the DC operations. Till now, Target doesn’t have any reporting system which will give greater visibility towards FDC operations and this project will fulfill the needs of reporting.

My Role :
As Team member performed
• Analyze and understand ETL mapping requirements and be able to translate those requirements into written test plans, detailed test cases and complex test SQL scripts.
• Have played the Lead role for a team size of 15 working @ Wipro ODC and has provided direction and support to QA team members with respect to testing activities coordinating with team lead @Client (Target India).
• Interacted with business users for any clarifications with respect to domain and incorporated the related scenarios in ETL testing.
• Have led the team in driving the assigned track and as a senior test analyst has involved in core test activities.
• Preparation of Traceability Matrix with respect to functional requirement.
• Preparation of Test script comprising both positive and negative test scenarios.
• Create and Execute Test cases for Functional and System Testing and maintain the test results in Quality Center.
• Involved in peer review of test deliverables within the team and has arranged review meetings with other teams such as domain, UAT team etc.
• Validating the completeness, integrity and accuracy of data across various systems.
• Preparation of test reports and metrics based on execution and defect process and responsible for sending the daily and weekly status reports to the client and project managers within Wipro.
• Have participated and ensure the participation of other team members in status meetings arranged by the client and has been responsible in providing detailed information.
• Ensure that defect logging and tracking is maintained through Quality Center.
• Ensure that testing activities are followed in accordance with project plan and customers quality process.
• Have supported the UAT team with respect to data load and data mockup activities.

Cegedim Software, Bangalore September September 2007 – March 2012
Technical Expert

Cegedim Dendrite is the Life Sciences industry’s leading provider of Customer Relationship Management (CRM) solutions. It provides vital support for several pharmaceutical companies with well designed unique software solutions such as Force Pharma/Mobile Intelligence, Nucleus Pharma and Organization Manager.
MOBILE INTELLIGENCE provides sales forces with a robust set of customer relationship management capabilities that promotes information sharing, field reporting, and analytics.
NUCLEUS Pharma® integrates with both internal and third-party data sources to successfully import, cleanse, standardize, match, and merge the customer information.
Project-1 : NP SFA EDW Data Interface
Client : Daiichi Sankyo
Environment : Oracle 10g, Unix
Test tools : Manual Testing, JIRA

Description :
Daiichi Sankyo (DSI) is implementing an Enterprise Data warehouse to replace existing ODS & Data Mart systems hosted in Cegedim Dendrite. DSI has requested to build data integration between existing Nucleus Customer Master and DSI EDW and between existing EDGE SFA system and DSI EDW.

 My Role :
As Team member performed
• Regular interaction with onsite developers in relation with understanding the requirement.
• Involved in providing the time estimation for the client specific project.
• Involved in the preparation of Test plan with Testing scope and approach.
• Preparation of Traceability Matrix with respect to functional requirement.
• Preparation of Test script comprising both positive and negative test scenarios.
• Executed Test cases for Functional and System Testing.
• Involved in validation of processing of data load to ODS, Data Mart and NP systems.
• Ensuring the data flow is accurate across various systems.
• Validating the completeness, integrity and accuracy of data in ODS and Data Mart systems.
• Bug allocation and tracking through Defect Tracking tool JIRA.
• Performed Performance and Regression testing.

Infosys Technologies, Bangalore June 2006 – September 2007
Technical Consultant
FINACLE (Universal Banking Solution)
Finacle is the Integrated, Online, Enterprise Banking system designed to provide the “e-platform” for the global banking industry. The complete web-enabled solution is a centralized, multi-currency, multi-lingual, CRM-enabled, functionally rich solution that addresses the core banking, e-banking, Treasury, wealth management and cash management requirements of Universal, retail, corporate, community and private banks worldwide.
Project-1
Client : URALSIB (Version 9.5)
Environment : C, Java 2.0, Oracle 10GR2, IBM Aix, HP, JSP, HTML
Test tools : Manual Testing, RADAR
Team size : 04
Description :
Enhancements are made to base product based on the Central Bank of Russia. The enhancements are incorporated in various modules such as Loans, Term Deposits, Trade finance, General Banking, Future Value dated transactions, Bills & payments.

Skills

  • Business Requirements, Data Integration, Data Mart, Data Quality, Data Sources, Data Warehouse, Databases, Hdfs, Html, Integration Testing, Jsp, Map Reduce, Ms Sql Server 2000, Ods, Oracle 10g, Project Plan, Pyspark, Shell Scripting, Shell Scripts, Sqoop, Team Lead, Technical Architecture, Teradata, Test Plan, Trade Finance, Uat, Unix Shell, Unstructured Data

Specialties

    Business Requirements, Data Integration, Data Mart, Data Quality, Data Sources, Data Warehouse, Databases, Hdfs, Html, Integration Testing, JSP, Map Reduce, Ms Sql Server 2000, Ods, Oracle 10g, Project Plan, PySpark, Shell Scripting, Shell Scripts, Sqoop, Team Lead, Technical Architecture, Teradata, Test Plan, Trade Finance, Uat, Unix Shell, Unstructured Data

Groups & Associations

    H1B

38 total views, 1 today