Learn Data Skills
Beta
Harika Kellampalli

Harika Kellampalli

Data Engineer

NIKE

Technologies

My Portfolio Highlights

My New Project

Hypothesis Testing with Men's and Women's Soccer Matches

My New Course

Exploratory Data Analysis in Python

Numbers enthusiast, translating data into meaningful insights.

My Work

Take a look at my latest work.

project

Exploring NYC Public School Test Result Scores

project

Investigating Netflix Movies

PythonSQLTheory
course

Exploratory Data Analysis in Power BI

Power BI
DataLab

Project: Exploring NYC Public School Test Result Scores

1Upvotes

My Certifications

These are the industry credentials that I’ve earned.

Other Certificates

Damien Carey Oracle Database SQL Certified Associate

Jonathan Cornelissen Certified Power BI Data Analyst

Microsoft Microsoft Certified: Power BI Data Analyst

DataCamp Course Completion

Take a look at all the courses I’ve completed on DataCamp.

My Work Experience

Where I've interned and worked during my career.

Nike | Sep 2022 - Present

Cloud Databricks Engineer

• Built scalable Data Lakehouse architectures on Databricks, consolidating structured and unstructured data and integrating data from various sources for unified analytics. • Integrated the end-to-end data pipeline to take data from source systems to target data repositories ensuring the quality and consistency of data are maintained at all times • Worked on various file formats such as ORC, Avro, Parquet, and JSON. • Developed robust ETL/ELT pipelines using Databricks and Apache Spark, ensuring efficient data extraction, transformation, and loading. • Optimized data storage and query performance in the Data Lake using partitioning and columnar file formats like Parquet. • Developed ETL processes for data ingestion, transformation, and storage, ensuring high performance and reliability • Implemented real-time data processing solutions using Databricks Structured Streaming and Kafka, enabling immediate insights and decision-making. • Integrated data from disparate sources into a single data platform, enhancing data accessibility and usability. • Prepared detailed specifications and translated designs into functional software by developing, documenting, testing, and debugging code. • Performed data cleansing and preprocessing using SQL and Python, ensuring high data quality and reliability for analysis. • Implemented AWS security best practices, including IAM policies, VPC configurations, and encryption mechanisms. • Automated data ingestion from various sources, reducing manual intervention and errors. • Achieved significant improvements in data processing times and resource utilization. • Tuned Spark jobs and queries to optimize performance, reduce processing times, and efficiently utilize resources. • Maintained compliance with industry standards and regulations, protecting sensitive data • Implemented monitoring and alerting mechanisms to track pipeline performance and detect issues. • Conducted root cause analysis and resolved data pipeline failures, ensuring minimal downtime. • Working within an Agile delivery to deliver proof of concept and production implementation in iterative sprints. • Worked closely business stakeholders to understand requirements and deliver tailored data solutions.
Show More

Optum | Aug 2020 - Sep 2021

Cloud Engineer

• PySpark Developer using AWS services like Glue, Lambda, and Athena. • Involved in the Design and building of projects right from scratch. • Involved in converting Hive/SQL queries into Spark transformations using Spark RDD and Python. • Worked in an Agile development environment in sprint cycles of two weeks by dividing and organizing tasks. Participated in the daily scrum and other design-related meetings. • Collaborated with product teams, data analysts and data scientists to design and build data-forward solutions. • Created Python scripts using Boto3 to automate AWS infrastructure provisioning tasks. • Key contributor in building the complete workflows, triggers, PySpark jobs, crawlers, and Lambda functions. • Using AWS Glue to schedule and track job performance by enabling the job. • Using AWS Cloud watch to track all log activities. • ETL pipelines in and out of the data warehouse using a combination of Python and Snowflakes Snow SQL Writing SQL queries against Snowflake. • Using AWS Redshift as a database/data warehouse to perform all analytics. • Using AWS Lambda to perform rest API activities and dump data into S3. • Wrote various data normalization jobs for new data ingested into Redshift. • The Process involved Design, Development, Build, Testing, Implementation, and support till there were no further issues.

Cigna | Apr 2019 - Jul 2021

data analyst

 Designed Crystal Reports and wrote complex formulas at the summary level and group level to implement the business logic.  Created complex Crystal Reports Ad Hoc Reports, Frequency Reports, Summary Reports, Drill-down, Dynamic grouping, graphical, aging reports.  Created various PL/SQL stored procedures, views, functions and temporary tables for data input to Crystal Reports.  Involved in modifying the SSIS packages and optimizing the package to fetch better performance and results.  Modified the SSIS package based on the new requirement of customer.  Trained business users on how to use Google Looker to explore data and generate custom reports, empowering teams to become self-sufficient in accessing the insights they needed for decision-making.  Set up automated reporting in Google Looker, scheduling the delivery of daily, weekly, and monthly reports to key stakeholders, improving efficiency and reducing manual report generation efforts.  Created the package configuration and loggings for better handling.  SSIS- re-wrote the package to reduce the redundant SSIS package doing same task.  Reduced the manual effort by the run team daily.  Prepared a complete flow of the application, suggested, and successfully implemented the changes.  Scheduled the application and other automation for sending reports and emails.  Performance tuning of DTS/SSIS packages.  Developed dashboards and ad-hoc reports using MS Power BI and SSRS for senior management team for analysis.  Worked on generated various reports with drilldowns, calculated members, dropdowns from the cubes by connecting to Analysis server from SSRS.  Analyzed business data using Google Looker to identify trends, patterns, and insights that drove strategic initiatives, leading to increase in customer retention.  Deployed reports on the web using Crystal publishing wizard and scheduled reports using e-Portfolio and Crystal management console.  Responsible for completing maintenance and delivery of scheduled releases.  Performance Tuning - Memory, CPU, I/O, Application, Hardware configuration (Different RAID levels) etc.  Troubleshot the long running queries and stored procedures to increase the performance by adding the index or by changing the query logic on need basis.  Analyzed the executing plan of SQL Server queries or stored procedures.  Used various sources to pull data into Power BI such as Sql Server, SAP BW, Oracle, SQL Azure etc.  Utilized Power BI Dataflows to create and manage reusable data entities and entities, enabling centralized data storage and management for self-service analytics.  Developed and optimized LookML data models and interactive dashboards in Google Looker, providing business stakeholders with real-time insights and reducing the time spent on manual reporting.  Monitored the performance and usage of Power BI reports and dashboards using usage metrics and performance analyzer tools, identifying opportunities for optimization and improvement.  Scheduled and maintained nightly and weekly loads of data by creating the corresponding job tasks.  Development and testing of new packages to ensure the data were extracted from database and presented for the project requirements and data validation.  Created SSIS jobs, which run on daily, weekly, and monthly basis to update fields in reporting files.

Kotak Bank | Jun 2015 - Dec 2018

data analyst

• Developed story telling dashboards in Tableau Desktop and published them on to Tableau Server, which allowed end users to understand the data on the fly with the usage of quick filters for on demand needed information. • Developed and created classes with Dimension, Detail & Measure objects and developed Custom hierarchies to support drill down reports. • Responsible for gathering requirements from Business Analysts and identifying the data sources required for the requests. • Wrote SAS Programs to convert Excel data into Teradata tables. • Actively reviewed over 208 unique variables and 4,700 rows of data using Excel and Python. • Used Python's multiple data science packages like Pandas, NumPy, SciPy, Scikit-learn and NLTK. • Involved in code-review sessions on existing SSRS report’s performance and Dataset query tuning. • Used MDX scripting for querying OLAP cubes. • Perform detailed data analysis (i.e. determine the structure, content, and quality of the data through examination of source systems and data samples) using SQL and Python. • Created drill down and drill up in worksheets in tableau. • Developed Teradata SQL scripts using OLAP functions like rank and rank () Over to improve the query performance while pulling the data from large tables. • Worked on importing/exporting large amounts of data from files to Teradata and vice versa. • Created Derived tables in the BO Designer to enhance the capabilities and performance of the universe. • Created report schedules on Tableau server. • Used Inner Join, Outer join while creating tables from multiple tables. • Designed and developed weekly, monthly reports related to the Logistics and manufacturing departments using Teradata SQL.

My Education

Take a look at my formal education

Bachelor's in ECETirumala Engineering College | 2014

About Me

Harika Kellampalli

As a passionate and motivated individual, I am actively seeking opportunities that will allow me to further develop my skills and gain valuable experience in my chosen field. Whether through an internship, a project collaboration, or new learning exp

Powered by

  • Work
  • Certifications
  • Courses
  • Experience
  • Education
  • About Me
  • Create Your Data Portfolio for Free