Loading visitor count...

About Me


Let Me Introduce Myself.

I'm a data engineer focused on building scalable, production-ready pipelines.

Hello! I'm Andrea, a certified data engineer with a background in both software engineering and business leadership. I specialize in building high-performance, cloud-native data pipelines using Databricks, PySpark, Azure Data Factory, and AWS Glue.

My core data engineering skills include:

  • Batch and streaming pipeline development with a focus on scalability
  • Production-grade ETL design with 100+ pipelines deployed
  • Extensive work with Change Data Capture (CDC) architectures
  • JSON and XML transformation for structured and semi-structured data
  • Dynamic schema handling for flexible, evolving data models
  • Automation of data quality checks, validations, and monitoring
  • Debugging Spark performance issues and enabling anomaly detection in streaming data
  • Scalable infrastructure supporting analytics and enterprise reporting

I thrive at the intersection of clean data engineering and business impact - translating complex technical needs into scalable systems that support analytics, ML, and decision-making. If you're looking for someone who can own pipelines, drive quality, and collaborate across teams, let’s connect!

📬 Reach me at andreahayes.dev@gmail.com or LinkedIn.

My Skills

SQL

Azure Data Factory

Databricks

PySpark

Delta Lake

Name: Andrea Hayes

LinkedIn: Andrea_Hayes_MSML

GitHub: NikkLuna

"Patience, persistence and perspiration make an unbeatable combination for success."

-Napoleon Hill

My Experience

My Work History

10/2024
until
Present
Tekletics
Data Engineer
  • Designed and deployed 100 + end-to-end CDC pipelines using Azure Synapse, ADF, and SQL Server CDC for scalable, reliable data ingestion.
  • Built metadata-driven ETL processes to parse 100+ nested XML/JSON files daily, transforming them into clean, normalized tables.
  • Engineered dynamic ADF workflows with validation logic, audit logging, row-level tracking, and concurrency tuning for high-throughput workloads.
  • Authored a 30-page CDC manual and troubleshooting guide to standardize practices and onboard new engineers.

ACHIEVEMENTS:

  • Created a dynamic ingestion framework that automatically scanned source tables and loaded only those with updated records, reducing unnecessary compute costs and improving pipeline efficiency.
9/2022
until
6/2024
Amazon
Amazon Robotics Floor Monitor ICQA
  • Obtained FAST (Floor Access Safety Training) certification to operate on the AR robotics floor.
  • Responded to incidents by clearly guiding associates through initial troubleshooting steps; escalated tickets to AR Technician if unresolved.
  • Communicated error codes, asset IDs, issue descriptions, severity level, urgency, and expected resolution time to relevant teams.

ACHIEVEMENTS:

  • Created guidebook for the Quarterback / Team Lead position that outlined higher-level troubleshooting steps, protocols for escalation, and methods for obtaining team performance metrics which were presented to the management team.

My Education

May
2024
Western Governors University
Bachelors Science, Software Engineering

RELEVANT COURSEWORK:

  • Learned best practices in systems and services, networking and security, scripting and programming, data management, and the business of IT.

ACHIEVEMENTS:

  • Excellence Award for Exemplary User Interface Design
November 2018
Western Governors University 
Masters Degree, Business Management and Leadership 

RELEVANT COURSEWORK:

  • Developed the comprehensive managerial and leadership skill set needed to execute organizational change, drive strategic initiatives, and develop high-performance teams.

ACHIEVEMENTS:

  • Successfully completed a rigorous program in Business Management and Leadership encompassing data-driven decision making, business acumen, and strategic change management.

"When I have fully decided that a result is worth getting I go ahead of it and make trial after trial until it comes."

-Thomas A. Edison

Technical Skills

See my highlighted technical skills below for related skillsets to the required skills for the position I'm applying for.

SQL

Experience building production queries and data models

Databricks

Experience designing and tuning notebooks and jobs

PySpark

Used for data transformation, joins, windowing, and performance tuning

Delta Lake

Experience with upserts (MERGE), schema evolution, and partitioning

Python

Used for scripting, validation, and PySpark jobs

Git / GitHub

Experience with version control, collaboration, and pipeline CI/CD

Airflow

Designed DAGs to automate ETL tasks and manage task dependencies

Azure Data Factory

Used ADF to build dynamic pipelines with parameters and conditional logic