Loading visitor count...

About Me


Let Me Introduce Myself.

I’m a Remote Data Engineer with extensive production pipeline delivery experience (100+ pipelines deployed).

I design scalable pipelines in Azure Synapse, ADF, and SQL Server, specializing in CDC ingestion, dimensional modeling, and automated data quality.

My core data engineering skills include:

  • Production-grade ETL design with 100+ pipelines deployed
  • Spark + PySpark (Databricks + AWS Glue)
  • Medallion architecture, Delta Lake, schema evolution
  • ADF orchestration, metadata-driven ingestion, JSON parsing
  • Automation of data quality checks, validations, and monitoring
  • Technical documentation and onboarding
  • Scalable infrastructure supporting analytics and enterprise reporting

I combine platform-level thinking with strong execution and high autonomy. I own pipelines end to end - built to move fast, stay clean, and remain reliable as they scale.

If you're looking for someone who can drive quality and collaborate across teams, I'd love to connect!

📬 Reach me at andreahayes.dev@gmail.com or LinkedIn.

My Skills

SQL

Azure Data Factory

Databricks

PySpark

Delta Lake

Name: Andrea Hayes

LinkedIn: Andrea_Hayes_MSML

GitHub: NikkLuna

"Patience, persistence and perspiration make an unbeatable combination for success."

-Napoleon Hill

My Experience

My Work History

10/2024
until
Present
Tekletics
Data Engineer
  • Designed and deployed 100+ end-to-end CDC pipelines using Azure Synapse, ADF, and SQL Server CDC to support reliable, scalable incremental ingestion.
  • Built metadata-driven ETL processes to parse and normalize 100+ nested XML/JSON files daily into analytics-ready relational tables.
  • Engineered dynamic ADF workflows with embedded validation, audit logging, row-level tracking, and concurrency tuning for high-throughput workloads.
  • Authored a 30-page CDC implementation manual and troubleshooting guide to standardize ingestion patterns and accelerate engineer onboarding.

ACHIEVEMENTS:

  • Delivered 12 enterprise Core tables by independently designing data models, reconciliation logic, and validation workflows across fragmented legacy systems, enabling reliable downstream reporting and analytics.
9/2022
until
6/2024
Amazon
Amazon Robotics Floor Monitor ICQA
  • Obtained FAST (Floor Access Safety Training) certification to operate on the AR robotics floor.
  • Responded to incidents by clearly guiding associates through initial troubleshooting steps; escalated tickets to AR Technician if unresolved.
  • Communicated error codes, asset IDs, issue descriptions, severity level, urgency, and expected resolution time to relevant teams.

ACHIEVEMENTS:

  • Created guidebook for the Quarterback / Team Lead position that outlined higher-level troubleshooting steps, protocols for escalation, and methods for obtaining team performance metrics which were presented to the management team.

My Education

May
2024
Western Governors University
Bachelors Science, Software Engineering

RELEVANT COURSEWORK:

  • Learned best practices in systems and services, networking and security, scripting and programming, data management, and the business of IT.

ACHIEVEMENTS:

  • Excellence Award for Exemplary User Interface Design
November 2018
Western Governors University 
Masters Degree, Business Management and Leadership 

RELEVANT COURSEWORK:

  • Developed the comprehensive managerial and leadership skill set needed to execute organizational change, drive strategic initiatives, and develop high-performance teams.

ACHIEVEMENTS:

  • Successfully completed a rigorous program in Business Management and Leadership encompassing data-driven decision making, business acumen, and strategic change management.

"When I have fully decided that a result is worth getting I go ahead of it and make trial after trial until it comes."

-Thomas A. Edison

Technical Skills

See my highlighted technical skills below for related skillsets to the required skills for the position I'm applying for.

SQL

Experience building production queries and data models

Databricks

Experience designing and tuning notebooks and jobs

PySpark

Used for data transformation, joins, windowing, and performance tuning

Delta Lake

Experience with upserts (MERGE), schema evolution, and partitioning

Python

Used for scripting, validation, and PySpark jobs

Git / GitHub

Experience with version control, collaboration, and pipeline CI/CD

Airflow

Designed DAGs to automate ETL tasks and manage task dependencies

Azure Data Factory

Used ADF to build dynamic pipelines with parameters and conditional logic