0% found this document useful (0 votes)
108 views3 pages

ANSAR HAYAT BigData Architect

This document summarizes the skills and experience of Ansar Hayat. He has over 7 years of experience as a Microsoft and Databricks Certified Data Engineer, specializing in building end-to-end data pipelines in cloud environments like AWS, Azure, and Databricks. He has extensive hands-on experience with big data technologies and platforms including AWS services, Azure services, Spark, Hadoop, Kafka and databases like SQL, MongoDB and PostgreSQL.

Uploaded by

Sana Ali
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
108 views3 pages

ANSAR HAYAT BigData Architect

This document summarizes the skills and experience of Ansar Hayat. He has over 7 years of experience as a Microsoft and Databricks Certified Data Engineer, specializing in building end-to-end data pipelines in cloud environments like AWS, Azure, and Databricks. He has extensive hands-on experience with big data technologies and platforms including AWS services, Azure services, Spark, Hadoop, Kafka and databases like SQL, MongoDB and PostgreSQL.

Uploaded by

Sana Ali
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

Ansar Hayat

Big Data & Data Engineer


Mob: +923013200515
Email:[email protected]
Skype: ansarh_3

Summary

A Microsoft and Databricks Certified Data Engineer Professional with almost 7 + years of experience in the
creation, orchestration, and deployments of end-to-end batch and stream data pipelines and managed big data
platforms in cloud-native environments such as AWS, Azure, and Databricks. Experienced in processing large
datasets of different forms including structured, semi-structured, and unstructured data. Hands-on experience with
AWS Big Data Services (EMR, EMR Studio, Glue, Glue Studio, Glue catalog, Athena, S3, RDS, Databricks, Lake-
house) as well as on Azure Data Services (Synapse Analytics, Azure Data factory, ADLS Gen2, Azure Databricks,
Lake-house).

Professional History

Sep 2023 – Present


(EXP: 2 Month) Working at The Contour Software Lahore in different capacities.
since Sep 2023 as Big Data Engineer || Senior Data Engineer
(https://contour-software.com/)
Work Summary:

• Architect and developed a data lake architecture that allowed for efficient storage and retrieval of large
datasets.
• Loaded and transformed large sets of structured, and semi-structured data from different sources.
• Creation and monitoring end to end-to-end data pipeline by using Pyspak and glue code.
• Saving data in s3 with parquet format and segregated data as a bronze, silver, and gold layer.
• Built a data catalog to store metadata and enable data governance.
• Querying and data verification were done with Athena.
• A Final Aggregated data set with move-in dataware house to make fact and dim tables.
• The dataware house is connected with Power BI for different dashboards.
• Automated data ingestion processes to reduce manual effort.
• Implemented a data security framework to ensure compliance with data privacy regulations.

November 2021 – August 2023


(EXP: 1.9 Year) Working at The Systems Limited. Lahore in different capacities
since October 2021 as Big Data Engineer || Senior Data Consultant
(https://www.systemsltd.com/)
Work Summary:

• In Systems worked with different clients to provide big data solutions by using AWS data platform
services.
• By using Data Lake, we Ingest data from SharePoint, SFTP, and cloud storage using Apache NiFi and
Pyspark job.
• Post-ingestion was done with the Pyspark script, deployed scripts on Jenkins, and scheduled DAG on
airflow and job run on the EMR cluster.
• we with the Databricks team to Implementation using AWS Databricks E2.
• Creation of s3 root bucket for audit log and terraform State files via terraform scripts.
• Create CI/CD pipelines and integrate with Bitbucket.
• Creation of new Databricks Workspace and Polices.
• Configuration of Unravel with Databricks for Hive and S3 role-based access control management.
• Worked on Informatica cloud by moving data to Redshift.
• Implemented Kafka producer and consumer application on Kafka Cluster setup with the help of
Zookeeper.
• Use Kafka API calls to process the messages smoothly on the Kafka cluster setup.
• Used Kafka Connect Source and Sink Connector of MYSQL, PostgreSQL, Mongo, and MYSQL.
• Use Alteryx for different project ETL pipelines. Created functional diagrams and documented data flow
process on confluence.
July 2019 – October 2021
(EXP: 2 Year) Worked at The Entertainer Pakistan. Lahore in different capacities
since July 2019 as Data Engineer
(https://www.theentertainerme.com/)
Work Summary:

• Big Data architect Implementation using Databricks services Data Lake, Delta Lake, and control
data quality.
• Implement the big data solution by using Databricks services Data Lake, Delta Lake, and Synapse
analytics·
• Spark Batch stream by using Pyspark and spark SQL to get data from Mongo to move into the data
lake.
• Azure Synapse Analytics pipeline was created, scheduled, and monitored by using Azure data
factory.
• Spark Real-time streaming by using Azure event hub and move data into Delta Lake.
• Tableau Desktop and Server for data visualization and scheduling different reports.
• Extensively used the advanced features of MongoDB to get different analytics-related ad hoc
reports.
• GDPR data removal from different data sources like Mongo, MySQL, and DWH.
• B2C and B2B app analytics flow control and investigating data analytics flow issues.

October 2016 – June 2019


(EXP: 3 Year) Worked at NETSOL Technologies Ltd. Lahore in different capacities
since October 2016 as Software Engineer
(http://www.netsoltech.com)

Work Summary:
• Implement business logic of finance and leasing using Stored Procedures, Functions, views, Cursors,
CTE in MS SQL Server 2014 and Oracle.
• Database modeling and modification for different business domains.
• Conducted deep business analysis with respect to client’s actual business requirements, current
practices, and procedure of work, on prioritized risk issues
• Implementing ETL tools from data Migration, by using Visual studio data tools for SQL.
• Telerik Reporting development to show different client data of different dates.
• Account data developments by using different business logic.
• Participation in Gap Analysis and Requirement gathering with different clients.
• Analyzing Client’s Business operational procedures, methods and mapping them with current business
processes and practices.
• Suggesting business processes related recommendations on proposed solutions.
• Conducted deep business analysis with respect to client’s actual business requirements, current.
practices, and procedure of work, on prioritized risk issues.

Development Skills

• Languages: Python, Java


• Big Data: Apache Spark, Glue, Glue Catalog, AWS Athena, Apache NiFi, Databricks, Apache
Airflow, HDFS, YARN, Amazon EMR, EMR Studio, Hive, Zookeeper, Kafka, Event Hub, Azure
Synapse Analytics, Snowflake, Jenkins, Docker
• Cloud: Azure, Aws, Databricks, Hortonworks, Cloudera
• Business Intelligence Tolls: Power BI, Tableau
• ETL: Azure Data Factory, SSIS, Pyspark, Spark SQL, Alteryx, Informatica
• SQL DBs: SQL, My SQL, SQLite and PostgreSQL, Azure SQL Databases
• NoSQL DBs: MongoDB, Cosmos DB
• Query languages: SQL, PL/SQL
• Web Technologies: JSON, HTML, CSS, Java Script, Telerik Reporting
• Version Control: Tortoise SVN, Bitbucket, GitHub
• Issue Tracking: JIRA
• Operating Systems: Microsoft Windows (XP, Vista, Windows 7 and Windows 10), Linux
Education
2016 BS (Hons) Computing (Specialization: Software Engineering &
Information Technology)
GC University

Awards and Achievements

❖ Microsoft Certified: Azure Data Engineer Associate


o Credential ID: 16346316
o Credential Link:
https://www.credly.com/badges/7dfdb623-dc8c-4ecd-8e04-
b54c3b2a2cdb/public_url

❖ Databricks Certified: Databricks Certified Data Engineer Associate


o Credential ID: ed161c08-1cc6-48d8-a45d-8237119be9da
o Credential Link:
https://credentials.databricks.com/ed161c08-1cc6-48d8-
a45d-8237119be9da

❖ Databricks Certified: Academy Accreditation - Platform Administrator


o Credential ID: 73c3bf57-470d-4a4e-99be-ae114e434b9f
o Credential Link:
https://credentials.databricks.com/73c3bf57-470d-4a4e-
99be-ae114e434b9f

❖ Databricks Certified: Academy Accreditation - Databricks Lakehouse Fundamentals


o Credential ID: 8a339206-6aac-4b66-a9b9-1b17e0fc7206
o Credential Link: https://credentials.databricks.com/8a339206-6aac-4b66-
a9b9-1b17e0fc7206

❖ Microsoft Certified: Azure Administrator Associate


o Credential ID: 16346316
o Credential Link:
https://www.credly.com/badges/7dfdb623-dc8c-4ecd-8e04-
b54c3b2a2cdb/public_url

❖ Microsoft Certified: Azure Data Fundamentals


o Credential ID: 16346316
o Credential Link: https://www.credly.com/badges/40aa6b83-96f2-46d3-8d0b-
df0c0934377f/public_url

❖ Microsoft Certified: Azure Fundamentals


o Credential ID: 16346316
o Credential Link:
https://www.credly.com/badges/514f30d0-c5e1-4b80-a9ae-
a347d6d2d44c/public_url

References

https://www.linkedin.com/in/ansarhayat/

You might also like