User

Snowflake Developer

BMT Score
86
86%
  • Remote

Available for

About Vamsi K

Having around 9.2 years of IT experience including 3.9yrs of Snowflake & Snap logic Developer and 5yrs of Bigdata Teradata Development, Integration and Deployment and DWH & ETL technologies.
Certified Snowpro Core
Very Good knowledge on Telicommunication , Health Care, Automobile & Finance Domains.
Good understanding and Experience in implementation of Snowlake Features of Storage Integrations, Internal and External stages, file formats, tables, snowpipes, streams and Tasks.
Experience in building complex data models and table designs as per the client's requirement.
Working with Snaplogic for ELT process
Developed pipelines and Tasks, API Integration pipelines in snaplogic.
Troubleshoot the issues in Snaplogic Ex Generic Pipelines while data integrate into snowflake.
Experience in building new Snaplogic generic pipe line to perform SQL Server, My SQL and API Integrations to Snowflake.
Experience in building library files Snaplogic to integrate SRC vs TGT(Snowflake) tables and cloumns .
Build the Logical and Physical data model for snowflake as per the changes required.
In-depth knowledge of Snowflake Database, Schema and Table structures.
Define virtual warehouse sizing for Snowflake for different type of workloads.
Very good Knowledge on DWH concepts and implementation.
Extensively developed the Dataware house and data lake projects in snowflake. 
Extensively developed the DDL’S,DML’S, Views & Stored Procedures in snowflake. 
Extensively worked Teradata to Snowflake and Hive to Teradata Migration Projects.
Extensively developed the Stored procedures to identify the run time errors and capture into the error log tables.
Extensively developed Java based stored Procedures to load Data vault (2.0) tables for IT Data Lake.
Extensively worked with structured(csv) and semi structured data (JSON) data.
Worked across multiple functional projects to understand data usage and implications for data migration.

Tech Stack Expertise

  • Tech Stack Expertise

    Microsoft .Net

    Stream

    0 Years
  • Tech Stack Expertise

    Python

    Python

    1 Years
  • Tech Stack Expertise

    AWS

    AWS

    1 Years
  • Tech Stack Expertise

    Big Data

    BigData

    0 Years

Work Experience

Images

Snowflake Developer

  • January 2015 - July 2023 - 8 Year
  • India

Projects

Images

GIS IT DATA LAKE

  • December 2022 - August 2023 - 9 Months
Technologies
Role & Responsibility

     Caterpillar Inc. is an American construction equipment manufacturer. The company is the world's largest manufacturer of construction equipment. In 2018, Caterpillar was ranked number 73 on the Fortune 500 list and number 265 on the Global Fortune 500 list
    Project: IT Data Lake is a large repository of data in Caterpillar, this consists of different domains of data integrated in one place. Domains are Applications, Global Directory, Peoples, Vulnerabilities, RADWare and Service Management. These data will load into Data vault model to access live data By Data scientists, Data developers, and Business analysts. 
    Responsibilities:

    Working with data architects to develop tables and data pipelines to load Data vault Data model.
    Extensively developed the VWH, Clusters & Tables based on requirement to improve performance.
    Implement the RBAC Controls and Data security for the data.
    Extensively developed Java based stored Procedures to load Data vault (2.0) tables for IT Data Lake.
    Extensivel Developed Data Vault Frame work for data lake project.
    Working with Snaplogic for ELT process
    Extensively worked identify the ETL process for each SRC system and design the pipe lines to integrate into snowflake.
    Developed new pipelines and Tasks, API Integrations in snaplogic.
    Troubleshoot the issues in Snaplogic Ex Generic Pipelines to load the live data integrate into snowflake.
    Enchance the pipeline to improve performance.
    Perform the DQ checks to validate the data and document for knowledge transfer.
    Designed the reusable pipelines in snaplogic
    Perform the peer reviews with data architecht on snaplogic pilpines and snowflake for approvals to migrate Higher env. 
    Design the technical document which will incorporate snaplogic pipelines and snowflake integration information for each source system data flow.
    Extensively Worked with the Files from AWS integrate into Snowflake.
    Desined Copy command to load bulk data from S3
    Created Internal,External stage and  External tables transformed data during load.
    Designed/Redesigned the Views in snowflake to increase the performance.
    Work across multiple functional projects to understand data usage and implications for data migration.
    Extensively implemented Snow Pipes, Streams &Tasks to automate the CDC. 
    Tune the queries in such a way that the code meets the performance standards of the snowflake.
    Validating the data from Different sources to Snowflake to make sure it has Apple to Apple match with DBeaver tool.
    Worked with BI team to clear the reporting issues while generating Power bi reports.
...see less
Images

PADL

  • February 2020 - December 2021 - 23 Months
Technologies
Role & Responsibility
    The Provider Analytical Data Layer is a centralized analytics platform that enables robust, accurate, and up-to-date provider profiling and comparability in the areas of quality, cost and resource wise. PADL is a combination of both BKBN & PCR modules. BKBN- Back Bone maintains the provider data demographics; whereas PCR is all about using this provider information along with Claims information and process the data which serves as a source for reporting layer to evaluate Providers performance. In addition to these, we have other modules like RHI PQP, RHIG2 and Data Quality Metrics.
     Responsibilities:
    Attending daily scrum meeting with the architects and data security teams to have possible impact of migrating data into the cloud.
    Extensively developer Hummer Frame work in order to export files from Teradata to S3 bucket.
    Extensively developed the Hummer framework with YML Scripts to configure the files Python and Pandas DF in order to Export on premise data into files format to AWS S3. 
    Worked on SnowSQL and Snowpipe.
    Created Snowpipe for continuous data load based on requirement.
    Created internal and external stage and transformed data during load.
    Redesigned the Views in snowflake to increase the performance.
    Unit tested the data between Teradata and Snowflake.
    Work across multiple functional projects to understand data usage and implications for data migration.
    Modifying the existing informatica mappings and logics, Teradata code into SnowSQL format and load data into the snowflake warehouse.
    Used COPY to bulk load the data.
    Tune the queries in such a way that the code meets the performance standards of the snowflake.
    Validating the data from Teradata to Snowflake to make sure it has Apple to Apple match.
    Consulting on Snowflake Data Platform Solution Architecture, Design, Development and deployment focused to bring the data driven culture across the enterprises.
    Create and trigger the Step functions to automatically trigger the Glue jobs which hold the snowsql code.
    Develop stored procedures/views in Snowflake and use  for loading Dimensions and Facts.
...see less
Images

Voyager Ninja

  • June 2019 - January 2020 - 8 Months
Technologies
Role & Responsibility
    the purpose of this project is to build the code from Hadoop Ecosystem to Teradata forESI to CVS data from CVS and internal sources (including Claims, Drug, Provider, Benefits and Membership, Rebates, Specialty, Formulary and Pharmacy Networks) into the CVS Landing Zone and make it available to WellPoint warehouses and downstream applications. In some cases additional elements to meet and/or improve our end user reporting needs will be provided. Data from the CVS Landing Zone will be fed to EDWard. Regional Data Warehouses will be provided data extracts as required to maintain their current functionality.


    Responsibilities:
    Work as a consultant and closely working with product owners and BAs to develop the application phases of the software development cycle
    Developed the Bteq scripts and Implement the code 
    Building & Loading data into APPZone Hive tables in Hadoop.
    Loading data into Teradata IRX Env.
    Experience in stream sets and Pipelines.
    Building & Implementing Business logics from IRX to EDW and load the data based o mapping sheet.
     very good knowledge on Ex: Membership & Claims  Subject areas
    Taking End to End  non GBD data responsibility 
    Develop the TDD document and implementing the code.
    Working with PO to approve the TDD document and Code
    Conduction code review calls to get approval from higher end
    Working on JIRA to track the sprint releases
    Promoting the Dev code to Sit and Prod every releases.
    Working on Control M, developing the CTM jobs 
    Develop and execute detailed ETL related functional, performance, integration and regression test cases, and documentation
    Extensively Developed the SQL queries based on the mapping documents/Approach documents.
    Extensively worked on Teradata database.
...see less
Images

EDWARD

  • March 2016 - June 2019 - 40 Months
Technologies
Role & Responsibility
    DWARD(Enterprise data warehouse and Research Depot).
    It consist of various subject areas like Claims, Member, Provider and Product which intern will have many Source Systems like Claims has Facets, Aces, Star (Small group), NASCO, WGS (Large group), CS90 etc. as source systems in it, LO Team will track the user stories with the Help of DIM Portal and analyze the data and will fix the issues raised by BA's by developing/Enhancing the code.
    Responsibilities:
    Work closely with software developers/product owners and BAs to develop and execute thorough test suites in all phases of the software development cycle
    Extensively worked on ICEDQ (Automation Tool)
    POC & very good knowledge on Ex: Membership& Claims Subject areas
    Leading offshore team and reaching SLA in time, due hard work received several client appreciations as well
    Taking End to End LO_ITT non GBD data responsibility 
    Develop Test strategy, test plan/design, execute test cases and defect management for the ETL & BI systems
    Develop and execute detailed ETL related functional, performance, integration and regression test cases, and documentation
    Extensively Developed the SQL queries based on the mapping documents/Approach documents.
    Executing Regression test cases through Automation tool (ICEDQ).
    Extensively worked on Teradata database.
    Very good knowledge to analyze the BTEQ scripts and develop the test cases
    Finding the issues and reporting to the Developers to fix the issues through JIRA.
    Raising/Closing the open defects in defect tracking tool JIRA.
    Following Agile Methodology, attending standup calls and retrospective meetings to track/Discuss the open issues.  
    Analyze and understand the ETL work flows developed.
    Quality Management - Knowledge of quality management methods, tools and techniques
    Communicate and manage expectations of the senior management team and during the planning and roll out of project releases.
    Involved in Development support during the deployment process and post production support as well.
...see less
Images

Unify Phase 2

  • May 2014 - May 2016 - 25 Months
Technologies
Role & Responsibility
    Vodafone Netherlands, part of the Vodafone Group is the second largest telecom provider in the Netherlands and was previously called Libertel. Telecommunication Industry is rapidly changing day by day and looking to this change, market requirements are also changing very fast. As new services are coming which require a complex BSS and looking to this market scenario Vodafone has decided to replace the current BSS landscape with the latest version of the Amdocs CES BSS suite, based upon an out-of-the-box (OOB) implementation.The Unify program exists to enable this.  In this project data from various online transaction processing applications and other sources is selectivity extracted related, transformed and loaded into Teradata Warehouse using BTEQ, MULTILOAD and FASTLOAD. 
    Responsibilities:
    Prepared mapping sheet for Script Level on individual data mart.
    Participated in Designing, Loading and maintaining OLAP Layer, basically used for reporting purpose.
    Created scripts (BTEQ, Fast Load, and MultiLoad) and written queries to move the data from source to destination.
    Data Quality Checks on the Source Systems
    Worked in creating Unit test cases, Technical specifications etc... According to the business specification from BRD’s (Business Required Documents and Mapping documents).
    Created appropriate indexes depend on the table situation and requirement.
    Involved in developing and update existing scripts as per project requirements.
    Did performance checks and tuning opportunities like partition on table level, compression, join indexes and collect statics etc.
    Tested the scripts in UT, IT and pre-prod (UAT) before getting successfully deployed in production.
    Involved in Development support during the deployment process and post production support as well.
    Experience in working with (HP QC) for finding defects and fixing the issues.
...see less

Industry Expertise

Education

Education

in Bachelors of Technology

Hyderabad University
  • June 2006 - June 2009

Our Suggestions