s

saimanikanta kukkadapu

About

Detail

Data Engineer
United States

Contact saimanikanta regarding: 
work
Full-time jobs

Timeline


work
Job

Résumé


Jobs verified_user 0% verified
  • Paychex
    Data Engineer
    Paychex
    Nov 2022 - Current (3 years 6 months)
    • Interacted with clients to gather business and system requirements which involved documentation of processes based on the user requirements. • Used Spark and Scala for developing machine learning algorithms which analyses click stream data. • Worked on spark core transformation and actions along with RDD and Dataset API. • Created PySpark frame to bring data from RDBMS to Amazon S3. • Designed and Developed Scala workflows for data pull from cloud based systems and applying transformations on it. • Developed various spark applications using Scala to perform various enrichment of these click stream data merged with user profile data. • Developed highly complex Python and Scala code, which is maintainable, easy to use, and satisfies
  • D
    Data Engineer
    Aug 2021 - Apr 2022 (9 months)
    • Participated in requirement gathering session with business users and sponsors to understand and document the business requirements. • Developed programs in Spark to use on application for faster data processing than standard MapReduce programs. • Used Spark for interactive queries, processing of streaming data and integration with NoSQL database for huge volume of data. • Designed and implemented Spark jobs to support distributed data processing. • Designed and Developed Scala workflows for data pull from cloud based systems and applying transformations on it. • Designed Data Quality Framework to perform schema validation and data profiling on Spark (PySpark). • Used Kafka producer to ingest the raw data into Kafka topics run the
  • GeBBS Healthcare Solutions
    Data Engineer
    GeBBS Healthcare Solutions
    Jul 2019 - Aug 2020 (1 year 2 months)
    • Gathering business requirements, business analysis and design various data products. • Developed Scala based Spark applications for performing data cleansing, event enrichment, data aggregation, de-normalization and data preparation needed for machine learning and reporting teams to consume. • Developed spark applications in python (PySpark) on distributed environment to load huge number of CSV files with different schema in to Hive ORC tables. • Developed Simple to complex MapReduce Jobs using Hive. • Wrote pre-processing queries in python for internal spark jobs • Reviewed basic SQL queries and edited inner, left, and right joins in Tableau Desktop by connecting live/dynamic and static datasets. • Performed Tableau type conversio
  • D
    Data Engineer
    Feb 2017 - Jun 2019 (2 years 5 months)
    • Performed Data analysis, Data Profiling and Requirement Analysis. • Developed Spark code using Scala and Spark-SQL/Streaming for faster testing and processing of data. • Performed Spark jobs with the Spark core, SparkSQL libraries for processing the data. • Worked on migrating MapReduce programs into Spark transformations using Scala. • Integrated data quality plans as a part of ETL processes. • Building data pipelines and complex ETL to process external client data using Python, Spark. • Performed Data cleaning and Preparation on XML files. • Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms • Developed various Python scripts to find vulnerabilities with SQL Queries by doing SQL injection,