MLops 12 Draft
MLops 12 Draft
PROFESSIONAL SUMMARY
Accomplished Machine Learning Engineer and Data Scientist with over 12 years of diverse IT
experience, including more than 8 years in designing, developing, deploying, and monitoring advanced
ML/AI solutions.
Adept in architecting end-to-end machine learning systems from data collection and preprocessing to
training, deployment, and production monitoring using best practices in MLOps and DevOps.
Proficient in building and maintaining automated ML pipelines using Airflow, MLflow, SageMaker
Pipelines, TFX, and Kubeflow to enable reproducibility, versioning, and CI/CD for ML models.
Deep experience implementing model monitoring and retraining workflows using tools like
Prometheus, Grafana, EvidentlyAI, and AWS CloudWatch, ensuring model performance over time.
Hands-on expertise with cloud platforms: AWS (SageMaker, Lambda, ECS, S3, Redshift), Azure (ML
Studio, Synapse), and GCP (Vertex AI, BigQuery) – including hybrid and multi-cloud architectures.
Strong background in supervised, unsupervised, and semi-supervised learning, with practical
application of algorithms like Random Forests, XGBoost, LightGBM, Gradient Boosting, SVM, and
KNN.
Proficient in deep learning techniques, using PyTorch, TensorFlow, and Keras to build models like
CNNs for image classification, LSTM/GRU/RNN for sequence modeling, and Transformers (BERT,
T5) for NLP tasks.
Developed and deployed multiple Natural Language Processing (NLP) solutions, including sentiment
analysis, topic modeling, text summarization, chatbot systems, and information extraction pipelines.
Designed and productionized computer vision solutions for image segmentation, defect detection,
object tracking, OCR, and image-based recommendation using OpenCV, YOLOv5, Bytetrack, and
Detectron2.
Built and scaled recommendation systems (collaborative filtering, content-based, and hybrid), fraud
detection systems, and churn prediction models for Fortune 500 clients across sectors.
Experience in stream processing frameworks such as Apache Kafka, Apache Flink, and Spark
Streaming to support real-time ML inference and low-latency decision systems.
Created data lakes and lakehouses on AWS S3/Glue and Azure Data Lake with Delta Lake to store
petabyte-scale raw, processed, and feature-engineered data efficiently.
Led the design of feature stores to standardize feature engineering and sharing across teams using
Feast, Tecton, and custom-built in-house platforms.
Skilled in data wrangling, transformation, and ETL/ELT workflows using Spark (PySpark/Scala),
Hive, Airflow, and Databricks, working with structured and semi-structured data (JSON, Parquet,
Avro).
Conducted A/B testing and statistical analysis of ML models and business experiments using Bayesian
methods, t-tests, ANOVA, and uplift modeling.
Created comprehensive dashboards and visualizations using Power BI, Tableau, and Plotly Dash to
track model metrics, business KPIs, and data quality across various pipelines.
Strong expertise in model interpretability and fairness using SHAP, LIME, Captum, and Fairlearn to
support ethical AI practices and transparency for regulated industries.
Developed containerized ML services with Docker and Kubernetes, enabling scalable and resilient
deployment of inference endpoints and batch scoring jobs.
Experience building APIs with FastAPI, Flask, and gRPC to serve machine learning models with
latency and throughput constraints in production environments.
Practiced in MLOps lifecycle, maintaining complete traceability of data, code, parameters, and models
using DVC, Git, MLflow, Weights & Biases, and other versioning tools.
Familiar with Data Governance, PII handling, GDPR, HIPAA, and responsible AI practices, especially
in healthcare and finance sectors.
Excellent communication, documentation, and client engagement skills; have worked closely with
stakeholders to translate vague business goals into measurable ML deliverables.
Actively mentor junior data scientists and engineers, contribute to code reviews, and participate in
internal AI/ML communities of practice to promote knowledge sharing.
TECHINALS SKILLS
Framework Flask, FastAPI, Django (basic), Apache Airflow, Luigi, TFX, Spark MLlib.
MLOps Tools MLflow, DVC, Weights & Biases, TensorBoard, Neptune.ai, Feast, Tecton, Kubeflow
(basic), Metaflow, Dagshub.
Version Control Git, GitHub, GitLab, Bitbucket
CI/CD Jenkins, GitHub Actions, GitLab CI/CD, Argo Workflows, CodePipeline, Airflow
CI/CD
Monitoring & Logging Prometheus, Grafana, ELK Stack (Elasticsearch, Logstash, Kibana), AWS
CloudWatch
Databases PostgreSQL, MySQL, Oracle, SQL Server, MongoDB, DynamoDB, Snowflake,
Redshift
IDE & Notebooks JupyterLab, Jupyter Notebook, RStudio, PyCharm, VS Code, Notepad++.
Statistical Analysis R, Excel, Statsmodels, Bayesian Inference, A/B Testing, ANOVA, Uplift Modeling
PROFESSIONAL EXPERIENCE