L

Lakshmi Sowjanya

About

Detail

Sr. Cloud Data Engineer at Centene Corporation//OPEN FOR CONTRACT on C2C
North Brunswick Township, New Jersey, United States

Contact Lakshmi regarding: 
work
Full-time jobs

Timeline


work
Job
school
Education

Résumé


Jobs verified_user 0% verified
  • Centene Corporation
    Sr. AWS Data Engineer
    Centene Corporation
    Oct 2022 - Current (3 years 8 months)
    • Migrate data from on-premises to AWS storage buckets. • Develop a Python script to transfer data from on-premises to AWS S3. • Develop a Python script to hit REST API’s and extract data to AWS S3. • Work on Ingesting data by going through cleansing and transformations and leveraging AWS Lambda, AWS Glue and Step Functions. • Develop a Python script to extract data from Netezza databases and transfer it to AWS S3. • Develop Lambda functions and assigned IAM roles to run python scripts along with various triggers (SQS, Event Bridge, SNS). • Create a Lambda Deployment function and configured it to receive events from S3 buckets. • Writing UNIX shell scripts to automate the jobs and scheduling cron jobs for job automation using commands with
  • Staples
    Cloud Big Data Engineer
    Staples
    Jan 2022 - Sep 2022 (9 months)
    • Developing ETL solutions using Spark SQL in Azure Data bricks for data extraction, transformation and aggregation from multiple file formats and data sources for analyzing & transforming the data to uncover insights into the customer usage patterns. • Integrated and automated data workloads to Snowflake Warehouse. • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios. • Designed and developed data pipelines in a variety of platforms such as Talend, Fivetran to Azure cloud data solutions. • Migrated an existing on-premises application to Azure • Developed Spark scripts by using Scala, Java as per the requirement. • Experience in developing customiz
  • Micron Technology
    Data Engineer
    Micron Technology
    Jul 2018 - Jul 2021 (3 years 1 month)
    • Migrated Tables and views from Oracle database to Azure SQL data warehouse using Azure Data Factory. The pipelines are parameterized, and values are provided using canvas-level variables on Azure Data Factory Triggers. • Implemented ETL pipelines using Azure Data bricks which read data from SFTP and Netezza databases. The raw data is transformed using PySpark code on Azure Data bricks. Versioning has been implemented using Azure DevOps GIT. • The Data Lake has been successfully migrated from HDFS to Azure Data Lake storage GEN1. Installed Azure Auto-resolve Integration runtime to authenticate tragedy between the client’s data center and Azure cloud. • Implemented log aggregation and dash boarding using Azure Log Analytics service. Integra
  • Micron
    Data Engineer
    Micron
    Jul 2018 - Jul 2021 (3 years 1 month)
    Responsibilities: • Migrated Tables and views from Oracle database to Azure SQL data warehouse using Azure Data Factory. The pipelines are parameterized, and values are provided using canvas-level variables on Azure Data Factory Triggers. • Implemented ETL pipelines using Azure Data bricks which read data from SFTP and Netezza databases. The raw data is transformed using PySpark code on Azure Data bricks. Versioning has been implemented using Azure DevOps GIT. • The Data Lake has been successfully migrated from HDFS to Azure Data Lake storage GEN1. Installed Azure Auto-resolve Integration runtime to authenticate tragedy between the client's data center and Azure cloud. • Implemented log aggregation and dash boarding using Azure Log Anal
  • Caterpillar Inc
    Data Developer
    Caterpillar Inc
    Apr 2015 - Jun 2018 (3 years 3 months)
    • Implemented DDL Curated Data Store logic using Spark Scala and Data frames concepts. Worked on and designed a Big Data analytics platform for processing customer interface preferences and comments using Hadoop, Hive and Pig, and Cloudera. • Importing and exporting data into HDFS and Hive using Sqoop from Oracle and vice versa. Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team. • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs and PySpark. Developed Python code to gather the data from HBase and designs the solution to implement using PySpark. • Enhanced the performance of queries and daily running spark jobs using the efficient
  • IBM
    Python Developer
    IBM
    May 2013 - Mar 2015 (1 year 11 months)
    • Creating web-based applications using Python on Django framework for data processing. • Implementing the preprocessing procedures along with deployment using the AWS services and creating virtual machines using EC2. • Good knowledge in exploratory data analysis and performed data wrangling and data visualization. • Validating the data to check for the proper conversion and identifying and cleaning unwanted data, data profiling for accuracy, completeness, consistency. • Preparing standard reports, charts, graphs, and tables from a structured data source by querying data repositories using Python and SQL. • Developed and produced a dashboard, key performance indicators and monitor organization performance. • Define data needs, evaluate data
Education verified_user 0% verified
  • University of New Haven
    Master of Science
    University of New Haven
    Nov 2016 - Jan 2023 (6 years 3 months)
  • J
    Bachelor of Science in Information Technology
    Jntuk
    Nov 2010 - Jan 2014 (3 years 3 months)