A highly skilled and results-driven
Senior Data Architect with
over 19 years of experience in designing, developing,
and optimizing database systems,
ETL/ELT processes, Data Governance, Master Data Management (MDM),
Slowly Changing Dimensions (SCD), and providing business intelligence solutions on
Data Warehouse, Data Lake, Delta Lake, and Data Lakehouse using
Star Schema, Snowflake Schema, and Data Vault. Proven expertise in leading offshore
teams, managing end-to-end data projects, and delivering scalable
solutions in industries such as
banking, insurance, healthcare, and
financial services. Adept at working with a range of
technologies, including
SQL Server, SSIS, Informatica, PySpark, Pandas, PostgreSQL, Oracle,
MySQL, Snowflake, Hadoop, SSRS, Apache Airflow, Apache Hive, Apache
Iceberg, Apache Ranger, Apache Atlas, Apache Superset, Apache Spark,
Tableau,
and Power BI. Committed to delivering
high-performance, secure, and accurate data solutions that drive
business insights and operational efficiency.
Project Highlights
Data Warehouse & Data Lake Architecture: Designed
and implemented a robust data architecture for a leading healthcare
client, enabling efficient data storage,
retrieval, and analytics.
Data Pipeline Optimization: Led the optimization of
data pipelines using
Apache Airflow and PySpark,
resulting in a 30% reduction in processing time and
improved data quality.
Business Intelligence Solutions: Developed
interactive dashboards using
Tableau and Power BI, providing
real-time insights to senior management and enhancing
decision-making processes.
Data Governance Implementation: Implemented
data governance frameworks using
Apache Ranger and Apache Atlas to
ensure data security, compliance,
and lineage tracking across the organization.
Master Data Management (MDM): Led
MDM initiatives to ensure
data consistency and
accuracy across multiple systems, improving
data integrity and
reporting capabilities.
Modern Big Data Architecture Transformation: This
project involved rewriting legacy data models to
modern big data architecture, transitioning from
Data Staging to a Data Lake.
Replaced SSIS with
Apache Airflow and PySpark for
ETL processes and used
Apache Iceberg for new
table storage. The
data model shifted from
Star Schema to Data Vault 2.0 to
support scalability and
flexibility. The solution integrates both
structured and unstructured data,
leveraging tools like Hive,
Apache Parquet, and JSON formats.
Data is migrated via AWS S3 and
Minio S3A, processed through
Airflow, and reported in
Tableau using Snowflake as the
centralized data platform.
Trizetto Facets Data Import: Led the
data integration process for importing
healthcare data, ensuring smooth
ETL operations and compliance with
healthcare regulations.
Business Intelligence Dashboards: Developed and
deployed real-time interactive dashboards using
SSRS, Power BI, and
Tableau, enabling
decision-makers to visualize key
business metrics.
Data Migration & Reporting Modernization:
Successfully migrated and modernized
reporting systems from
Crystal Reports to SSRS, improving
reporting accuracy and
performance.
Experience
πSenior Data
Architect @ Connect Media Networks
Jan 2025 - Present
Pipeline Architect – Design, develop, and
maintain ETL processes using SQL Server Integration Services
(SSIS) & Python (PySpark) to automate data workflows and improve
efficiency.
Data & Delta Lake Developer – Develop and
optimize Data & Delta Lake Tables, catalogs, queries, and scripts
to support data loading performance and data integrity.
Stored Procedure Development– Write and enhance
complex SQL stored procedures for data manipulation, retrieval,
and optimization in SQL Server.
Data Modeler – Create and refine logical and
physical data models using Lucid Chart and io draw, ensuring
structured data representation for business processes.
Report Developer – Design and develop interactive
reports using SQL Server Reporting Services (SSRS) to provide
data-driven insights for decision-making.
πData Architect @
EVIDEN (Atos)
Nov 2022 - Oct 2024
Developed and managed a Data Warehouse and
Data Lake architecture, ensuring effective
storage and retrieval of structured and unstructured data to
support business intelligence and analytics.
Led offshore teams in the design, development, and optimization of
data pipelines using
PySpark, Pandas, Apache Airflow, and
Trino to handle large-scale data integration
tasks.
Utilized modeling tools such as Draw.io,
Lucid Chart, ER/Studio, and
ER/Win for conceptual, logical, and physical data
modeling, ensuring effective representation and alignment with
organizational data needs.
Developed Data Vault data models to enable
scalable, flexible, and high-performing data architecture.
Implemented data governance using
Apache Ranger and
data lineage using
Apache Atlas to ensure data security, quality,
and compliance.
Led business intelligence development, building
interactive dashboards using Tableau and
Apache Superset for real-time reporting.
Created and maintained robust data models for complex data
systems, focusing on
Master Data Management (MDM) and
Slowly Changing Dimensions (SCD Type 4 for Data Lake)
to ensure data integrity and consistency across systems.
Led the adoption of industry-standard tools for
data modeling and
database management, ensuring that the team
stayed up-to-date with the latest technology trends.
Led and managed an offshore team to deliver complex data
integration projects, including
Trizetto Facets Data Import for healthcare
clients.
Designed and implemented data models using
Star Schema, Snowflake Schema,
and Data Vault 2.0 to support business
intelligence and reporting needs.
Developed complex ETL pipelines using SQL Server,
SSIS, and SSRS for data
extraction, transformation, and loading.
Created custom Jasper Reports and
Power BI visualizations, delivering interactive
dashboards and reports to senior stakeholders.
Led initiatives around
Master Data Management (MDM), ensuring that
clients had accurate, consistent, and reliable master data across
their systems.
Designed and implemented
Slowly Changing Dimensions (SCD) to track data
changes and improve data accuracy and historical analysis.
Evaluated and implemented new database technologies and tools,
including different database platforms (SQL Server, Oracle, MySQL,
PostgreSQL), and identified opportunities for
further automation and efficiency gains.
πProject Lead (Data
Solutions) @ Syntel LLC
Jan 2011 - Jan 2015
Led and managed team developers in the design and implementation
of data pipelines and
ETL processes using SSIS for
various clients.
Designed and implemented data models using
Star Schema, Snowflake Schema,
and OLAP to support business intelligence and
reporting needs.
Developed complex ETL pipelines using
SQL Server, SSIS, and
SSRS for data extraction, transformation, and
loading.
Created custom Reporting Services Model and
SSRS visualizations, delivering interactive
Adhoc Reporting and reports to senior
stakeholders.
Optimized and wrote SQL queries and
stored procedures for
SQL Server, SSIS, and
SSRS.
Worked on database optimization, writing complex
queries and
stored procedures across
SQL Server, Oracle, and
MySQL.
Designed Star Schema and
Snowflake Schema models and implemented
Slowly Changing Dimensions (SCD) strategies to
ensure accurate historical data.
Developed SSRS reports and
SSAS cubes to enable multidimensional analysis
for business users.
Collaborated with business stakeholders to
understand reporting needs and translated them
into technical solutions, ensuring alignment with business goals
and delivering high-quality data products. Acted
as the primary technical point of contact for clients in the
phone insurance,
Medicare/Medicaid insurance, and
stock market domains, ensuring that database
solutions were tailored to meet their specific business
requirements.
Designed and implemented data pipelines using
SSIS and DTS to efficiently
extract, transform, and load (ETL) large datasets
across multiple systems, ensuring
data consistency and availability for reporting
and analysis.
Developed and optimized complex SQL queries and
stored procedures for databases such as
SQL Server, Oracle, and
MySQL, ensuring efficient
data extraction and transformation to meet
business requirements.
Led report conversion and
rewriting projects to migrate reports from
Crystal Reports to SSRS,
improving the quality and performance of client reports with more
interactive features.
Created and managed interactive
Cognos dashboards for
real-time reporting and
business intelligence insights.
Developed SSAS cubes using
MDX queries for multidimensional analysis.
Conducted performance tuning for
SQL queries, SSIS packages, and
reporting services to improve the speed and
efficiency of data processing and
report generation.
Developed and optimized T-SQL queries and
stored procedures for
SQL Server and Oracle, improving
data retrieval and processing.
Managed and optimized
DTS (Data Transformation Services) and
DM Express for ETL processes,
ensuring smooth data integration.
Led the development of Crystal Reports and
Report 2.0 to provide clear, insightful, and
accurate business reports for stakeholders.
Implemented
Slowly Changing Dimensions (SCD) strategies,
particularly Type 1 and
Type 2 SCDs, to track historical changes in
critical business data while ensuring
data consistency and accuracy.
Worked closely with business users and
stakeholders at
Accordia Life and Annuity Company to understand
reporting requirements, ensuring that
database solutions met their specific needs.
No comments:
Post a Comment