SlideShare a Scribd company logo
The Age of Language Models in NLP
Tuesday | 23rd June, 2020
LIVE WEBINAR
Presented by
AGENDA
1 . About Tyrone
 World’s high performing AI platform system – A100
 Get Development, Training, Inference in one
 Era of Modern Mixed Workloads
 Tyrone Kubyts™
2. Word Embeddings
 How Word embeddings create a context based relationship
 How to create Word Embeddings
3. Sequence Modelling
 Introduction of Deep learning in NLP
 Overview on the model architecture to use
4. Advanced Language Models
 Overview of the Language models
 How are they created
 Transformers
 BERT , GP2 etc
5. NLP Attention Mechanism
 Overview on attention Mechanism
6. Case Studies
Tyrone Systems at a Glance
NVIDIA HGX A100 PERFORMANCE
New Tensor Core for AI & HPC
New Multi-instance GPU
New Hardware Engines
Increase in GPU
interconnect
bandwidth
Increase in GPU
memory
Increase in
memory
bandwidth
Speedup in
AI performance
54 Billion
XTORS
3rd Gen
Tensor cores
Sparsity
Acceleration
Multi
Instance GPU
3rd GEN NVLINK
& NVSwitch
NVIDIA A100
Greatest Generational Leap – 20X Volta
54B XTOR | 826mm2 | TSMC 7N | 40GB Samsung HBM2 | 600 GB/s NVLink
Peak Vs Volta
FP32 TRAINING 312 TFLOPS 20X
INT8 INFERENCE 1,248 TOPS 20X
FP64 HPC 19.5 TFLOPS 2.5X
MULTI INSTANCE GPU 7X GPUs
New tf32 tensor cores on A100
20X Higher FLOPS for AI, Zero Code Change
20X Faster than Volta FP32 | Works like FP32 for AI with Range of FP32 and Precision of FP16
No Code Change Required for End Users | Supported on PyTorch, TensorFlow and MXNet Frameworks Containers
Most flexible ai platform with MULTI-INSTANCE GPU (MIG)
Optimize GPU Utilization, Expand Access to More Users with Guaranteed Quality of Service
Up To 7 GPU Instances In a Single A100:
Simultaneous Workload Execution With
Guaranteed Quality Of Service:
All MIG instances run in parallel with predictable
throughput & latency
Flexibility to run any type of workload on a MIG
instance
Right Sized GPU Allocation:
Different sized MIG instances based on target
workloads
Amber
GPU Mem
GPU
GPU Mem
GPU
GPU Mem
GPU
GPU Mem
GPU
GPU Mem
GPU
GPU Mem
GPU
GPU Mem
GPU
ONE SYSTEM FOR ALL ai INFRASTRUCTURE
AI Infrastructure Re-Imagined, Optimized, and Ready for Enterprise AI-at-Scale
any job | any size | any node | anytime
Analytics  Training  Inference
Flexible AI infrastructure that adapts to the
pace of enterprise
• One universal building block for the AI data
center
• Uniform, consistent performance across the
data center
• Any workload on any node - any time
• Limitless capacity planning with predictably
great performance with scale
Game-changing performance for innovators
9x Mellanox ConnectX-6 200Gb/s Network Interface
450GB/sec Peak Bi-directional Bandwidth
Dual 64-core AMD Rome CPUs and 1TB RAM
3.2X More Cores to Power the Most Intensive AI Jobs
8x NVIDIA A100 GPUs with 320GB Total GPU Memory
12 NVLinks/GPU 600GB/sec GPU-to-GPU Bi-directional Bandwidth
15TB Gen4 NVME SSD
4.8TB/sec Bi-directional Bandwidth
2X More than Previous Generation NVSwitch
6x NVIDIA NVSwitches
25GB/sec Peak Bandwidth
2X Faster than Gen3 NVME SSDs
2U GPU server up to 4 NVIDIA HGX ™ A100 GPU
Camarero DAS7TGVQ-24RT
Tyrone NVIDIA A100 based SERVERS
• Supports 4x A100 40GB SXM4 GPUs
• Supports CPU TDP up to 280W
• Dual AMD EPYC™ 7002 Series Processors in up to 128 Cores
• Flexible storage with 4 hotswap for SAS, SATA or NVMe
• PCI-E Gen 4 NVLink for fast GPU-GPU connections
• 32 DIMM Slots that allow up to 8TB of 3200Mhz DDR4 memory
• 4 Hot-swap heavy duty fans
• 2x 2200W Redundant Power Supplies, Titanium Level
PCI-E Gen 4
NEW LAUNCH
NVIDIA NVLink
4U GPU server up to 8 NVIDIA HGX ™ A100 GPU
Tyrone NVIDIA A100 based SERVERS
NVIDIA NVLink & NVSwitch
NEW LAUNCH
• Supports up to 8 double-width GPUs,
• Supports CPU TDP up to 280W
• Dual AMD EPYC™ 7002 Series Processors in up to 128 Cores
• Flexible storage with 4 hotswap for SAS, SATA or NVMe
• PCI-E Gen 4 NVLink for fast GPU-GPU connections
• 32 DIMM Slots that allow up to 8TB of 3200Mhz DDR4 memory
• 4 Hot-swap heavy duty fans
• 2x 2000W Redundant Power Supplies, Titanium Level
4U GPU server up to 8 NVIDIA HGX ™ A100 GPU
Tyrone NVIDIA A100 based SERVERS
NVIDIA NVLink
COMING SOON
• Supports Intel Xeon
• Supports NVLink
• 8 x NVIDIA Tesla A100 SXM4
Delivers 4XFASTER TRAINING
than other GPU-based systems
Your Personal AI Supercomputer
Power-on to Deep Learning in Minutes
Pre-installed with Powerful
Deep Learning
Software
Extend workloads from your
Desk-to-Cloud in Minutes
Mixed Workloads Convergence of
AI|HPC| Cloud | Containers
The Era of Modern Mixed Workload
F L E X I B L E Is the usage going to be constant?
O P T I M I Z A T I O N Is optimal utilization required?
R E S I L I E N C E Do we need the application to run all the time.
E A S E Is ‘ease of maintenance’ key?
S C A L A B I L I T Y & S P E E D Do we have one size that fits all?
Connectivity and usage
Virtual Desktop
Laptop
Tyrone Cloud
Manager
Tyrone Cloud
Manager
Laptop
Run Multiple Applications
simultaneously
Flow Architecture Revolutionizing Deep Learning CPU-GPU Environment
10X20X30X40X50X60X70X
SPEED
WITH TYRONE KUBYTS™ CLIENT
Compatible Workstations
has a repository of :
50 containerized applications
100s of Containers
CLOUD
Tyrone KUBITS : Revolutionizing Deep Learning CPU-GPU Environment
Run different
applications
simultaneously
Check for Tyrone
KUBITS Compatible
Workstations
Get access to over
100+ Containers on
Tyrone KUBITS Cloud.
High scalability
Affordable price
Has both GPU &
CPU Optimized
Containers
Design a simple Workstation
or Large Clusters with KUBITS
technology.
Talk to our experts & build
the right workstation within
your budget.
KUBITS
CLOUDCOMPATIBLE
AGENDA
1 . About Tyrone
 World’s high performing AI platform system – A100
 Get Development, Training, Inference in one
 Era of Modern Mixed Workloads
 Tyrone Kubyts™
2. Word Embeddings
 How Word embeddings create a context based relationship
 How to create Word Embeddings
3. Sequence Modelling
 Introduction of Deep learning in NLP
 Overview on the model architecture to use
4. Advanced Language Models
 Overview of the Language models
 How are they created
 Transformers
 BERT , GP2 etc
5. NLP Attention Mechanism
 Overview on attention Mechanism
6. Case Studies
Word Embedding
• Word Embedding is a language modeling technique used
for mapping words to vectors of real numbers.
• It represents words or phrases in vector space with
several dimensions.
• Word embeddings can be generated using various
methods like neural networks, co-occurrence matrix,
probabilistic models, etc.
• Word2Vec consists of models for generating word
embedding. These models are shallow two layer neural
networks having one input layer, one hidden layer and
one output layer. Word2Vec utilizes two architectures
CBOW – Continuous Bag of Words
• CBOW model predicts the
current word given context
words within specific window.
The input layer contains the
context words and the output
layer contains the current word.
• The hidden layer contains the
number of dimensions in which
we want to represent current
word present at the output layer.
Skip Gram – Word Embeddings
• Skip gram predicts the surrounding
context words within specific
window given current word. The
input layer contains the current
word and the output layer contains
the context words.
• The hidden layer contains the
number of dimensions in which we
want to represent current word
present at the input layer.
Advanced Language Models and Transformers
ELMo ULMFit
BERT Transformer
Transformer
BERT Architecture
• Transformer is an attention-based architecture for NLP
• Transformer composed of two parts: Encoding component and
Decoding component
• BERT is a multi-layer bidirectional Transformer encoder
Attention Mechanism
BERT vs GPT
Q&A Session
Hirdey Vikram
Hirdey.vikram@netwebindia.com
India (North)
Niraj
niraj@netwebindia.com
India (South)
Vivek
vivek@netwebindia.com
India (East)
Navin
navin@netwebindia.com
India (West)
Anupriya
anupriya@netwebtech.com
Singapore
Arun
arun@netwebtech.com
UAE
Agam
agam@netwebtech.com
Indonesia
Contact our team if you have any further questions after this webinar
ai@netwebtech.comTalk to our AI Experts

More Related Content

What's hot (15)

PDF
2016-JAN-28 -- High Performance Production Databases on Ceph
Ceph Community
 
PPTX
Ceph Day Seoul - AFCeph: SKT Scale Out Storage Ceph
Ceph Community
 
PPTX
Rack Cluster Deployment for SDSC Supercomputer
Rebekah Rodriguez
 
PPTX
Which Hypervisor is Best?
Kyle Bader
 
PDF
SPACK: A Package Manager for Supercomputers, Linux, and MacOS
inside-BigData.com
 
PPTX
Manta Unleashed BigDataSG talk 2 July 2013
Christopher Hogue
 
PDF
IMCSummit 2015 - Day 2 IT Business Track - 4 Myths about In-Memory Databases ...
In-Memory Computing Summit
 
PPTX
FPGAs in the cloud? (October 2017)
Julien SIMON
 
PPTX
Orchestrate Your AI Workload with Cisco Hyperflex, Powered by NVIDIA GPUs
Renee Yao
 
PPTX
Walk Through a Software Defined Everything PoC
Ceph Community
 
PDF
Ucx an open source framework for hpc network ap is and beyond
inside-BigData.com
 
PPTX
MySQL Head to Head Performance
Kyle Bader
 
PPTX
Red Hat Ceph Storage Acceleration Utilizing Flash Technology
Red_Hat_Storage
 
PPTX
Ceph Day Melabourne - Community Update
Ceph Community
 
PPTX
OpenStack and Ceph case study at the University of Alabama
Kamesh Pemmaraju
 
2016-JAN-28 -- High Performance Production Databases on Ceph
Ceph Community
 
Ceph Day Seoul - AFCeph: SKT Scale Out Storage Ceph
Ceph Community
 
Rack Cluster Deployment for SDSC Supercomputer
Rebekah Rodriguez
 
Which Hypervisor is Best?
Kyle Bader
 
SPACK: A Package Manager for Supercomputers, Linux, and MacOS
inside-BigData.com
 
Manta Unleashed BigDataSG talk 2 July 2013
Christopher Hogue
 
IMCSummit 2015 - Day 2 IT Business Track - 4 Myths about In-Memory Databases ...
In-Memory Computing Summit
 
FPGAs in the cloud? (October 2017)
Julien SIMON
 
Orchestrate Your AI Workload with Cisco Hyperflex, Powered by NVIDIA GPUs
Renee Yao
 
Walk Through a Software Defined Everything PoC
Ceph Community
 
Ucx an open source framework for hpc network ap is and beyond
inside-BigData.com
 
MySQL Head to Head Performance
Kyle Bader
 
Red Hat Ceph Storage Acceleration Utilizing Flash Technology
Red_Hat_Storage
 
Ceph Day Melabourne - Community Update
Ceph Community
 
OpenStack and Ceph case study at the University of Alabama
Kamesh Pemmaraju
 

Similar to Age of Language Models in NLP (20)

PDF
Distributed deep learning reference architecture v3.2l
Ganesan Narayanasamy
 
PPTX
An Introduction to Natural Language Processing
Tyrone Systems
 
PPT
Enabling a hardware accelerated deep learning data science experience for Apa...
DataWorks Summit
 
PDF
Fascinating Tales of a Strange Tomorrow
Julien SIMON
 
PDF
The Convergence of HPC and Deep Learning
inside-BigData.com
 
PDF
GIST AI-X Computing Cluster
Jax Jargalsaikhan
 
PDF
infoShare AI Roadshow 2018 - Tomasz Kopacz (Microsoft) - jakie możliwości daj...
Infoshare
 
PDF
Power AI introduction
Snowy Chen
 
PDF
Exploring AI as tools in your career.pdf
videongamesrfun
 
PDF
FPGA Hardware Accelerator for Machine Learning
Dr. Swaminathan Kathirvel
 
PPTX
IBM COE AI Lab at your University
Ganesan Narayanasamy
 
PDF
OpenPOWER Boot camp in Zurich
Ganesan Narayanasamy
 
PPTX
2018 bsc power9 and power ai
Ganesan Narayanasamy
 
PPTX
PowerAI Deep dive
Ganesan Narayanasamy
 
PPTX
AI on the Edge
Jared Rhodes
 
PDF
NVIDIA DGX User Group 1st Meet Up_30 Apr 2021.pdf
MuhammadAbdullah311866
 
PPTX
Introduction to PowerAI - The Enterprise AI Platform
Indrajit Poddar
 
PPTX
[DSC Europe 24] Thomas Kitzler - Building the Future – Unpacking the Essentia...
DataScienceConferenc1
 
PDF
AI/Cloud Technology access
Ganesan Narayanasamy
 
PDF
2018-11-05 Intro to AI
Jeff Boleman, M.S.C.S.
 
Distributed deep learning reference architecture v3.2l
Ganesan Narayanasamy
 
An Introduction to Natural Language Processing
Tyrone Systems
 
Enabling a hardware accelerated deep learning data science experience for Apa...
DataWorks Summit
 
Fascinating Tales of a Strange Tomorrow
Julien SIMON
 
The Convergence of HPC and Deep Learning
inside-BigData.com
 
GIST AI-X Computing Cluster
Jax Jargalsaikhan
 
infoShare AI Roadshow 2018 - Tomasz Kopacz (Microsoft) - jakie możliwości daj...
Infoshare
 
Power AI introduction
Snowy Chen
 
Exploring AI as tools in your career.pdf
videongamesrfun
 
FPGA Hardware Accelerator for Machine Learning
Dr. Swaminathan Kathirvel
 
IBM COE AI Lab at your University
Ganesan Narayanasamy
 
OpenPOWER Boot camp in Zurich
Ganesan Narayanasamy
 
2018 bsc power9 and power ai
Ganesan Narayanasamy
 
PowerAI Deep dive
Ganesan Narayanasamy
 
AI on the Edge
Jared Rhodes
 
NVIDIA DGX User Group 1st Meet Up_30 Apr 2021.pdf
MuhammadAbdullah311866
 
Introduction to PowerAI - The Enterprise AI Platform
Indrajit Poddar
 
[DSC Europe 24] Thomas Kitzler - Building the Future – Unpacking the Essentia...
DataScienceConferenc1
 
AI/Cloud Technology access
Ganesan Narayanasamy
 
2018-11-05 Intro to AI
Jeff Boleman, M.S.C.S.
 
Ad

More from Tyrone Systems (20)

PDF
Kubernetes in The Enterprise
Tyrone Systems
 
PDF
Why minio wins the hybrid cloud?
Tyrone Systems
 
PDF
why min io wins the hybrid cloud
Tyrone Systems
 
PDF
5 ways hci (hyper-converged infrastructure) powering today’s modern learning ...
Tyrone Systems
 
PDF
5 current and near-future use cases of ai in broadcast and media.
Tyrone Systems
 
PDF
How hci is driving digital transformation in the insurance firms to enable pr...
Tyrone Systems
 
PDF
How blockchain is revolutionising healthcare industry’s challenges of genomic...
Tyrone Systems
 
PDF
5 ways hpc can provides cost savings and flexibility to meet the technology i...
Tyrone Systems
 
PDF
How Emerging Technologies are Enabling The Banking Industry
Tyrone Systems
 
PDF
Five Exciting Ways HCI can accelerates digital transformation for Media and E...
Tyrone Systems
 
PPTX
Design and Optimize your code for high-performance with Intel® Advisor and I...
Tyrone Systems
 
PDF
Fast-Track Your Digital Transformation with Intelligent Automation
Tyrone Systems
 
PDF
Top Five benefits of Hyper-Converged Infrastructure
Tyrone Systems
 
PDF
An Effective Approach to Cloud Migration for Small and Medium Enterprises (SMEs)
Tyrone Systems
 
PDF
How can Artificial Intelligence improve software development process?
Tyrone Systems
 
PDF
3 Ways Machine Learning Facilitates Fraud Detection
Tyrone Systems
 
PDF
Four ways to digitally transform with HPC in the cloud
Tyrone Systems
 
PDF
How to Secure Containerized Environments?
Tyrone Systems
 
PPTX
OneAPI Series 2 Webinar - 9th, Dec-20
Tyrone Systems
 
PPTX
OneAPI dpc++ Virtual Workshop 9th Dec-20
Tyrone Systems
 
Kubernetes in The Enterprise
Tyrone Systems
 
Why minio wins the hybrid cloud?
Tyrone Systems
 
why min io wins the hybrid cloud
Tyrone Systems
 
5 ways hci (hyper-converged infrastructure) powering today’s modern learning ...
Tyrone Systems
 
5 current and near-future use cases of ai in broadcast and media.
Tyrone Systems
 
How hci is driving digital transformation in the insurance firms to enable pr...
Tyrone Systems
 
How blockchain is revolutionising healthcare industry’s challenges of genomic...
Tyrone Systems
 
5 ways hpc can provides cost savings and flexibility to meet the technology i...
Tyrone Systems
 
How Emerging Technologies are Enabling The Banking Industry
Tyrone Systems
 
Five Exciting Ways HCI can accelerates digital transformation for Media and E...
Tyrone Systems
 
Design and Optimize your code for high-performance with Intel® Advisor and I...
Tyrone Systems
 
Fast-Track Your Digital Transformation with Intelligent Automation
Tyrone Systems
 
Top Five benefits of Hyper-Converged Infrastructure
Tyrone Systems
 
An Effective Approach to Cloud Migration for Small and Medium Enterprises (SMEs)
Tyrone Systems
 
How can Artificial Intelligence improve software development process?
Tyrone Systems
 
3 Ways Machine Learning Facilitates Fraud Detection
Tyrone Systems
 
Four ways to digitally transform with HPC in the cloud
Tyrone Systems
 
How to Secure Containerized Environments?
Tyrone Systems
 
OneAPI Series 2 Webinar - 9th, Dec-20
Tyrone Systems
 
OneAPI dpc++ Virtual Workshop 9th Dec-20
Tyrone Systems
 
Ad

Recently uploaded (20)

PDF
CIFDAQ Weekly Market Wrap for 11th July 2025
CIFDAQ
 
PDF
Chris Elwell Woburn, MA - Passionate About IT Innovation
Chris Elwell Woburn, MA
 
PDF
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
PDF
How Startups Are Growing Faster with App Developers in Australia.pdf
India App Developer
 
PPTX
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
PDF
From Code to Challenge: Crafting Skill-Based Games That Engage and Reward
aiyshauae
 
PDF
Using FME to Develop Self-Service CAD Applications for a Major UK Police Force
Safe Software
 
PPTX
AI Penetration Testing Essentials: A Cybersecurity Guide for 2025
defencerabbit Team
 
PDF
SWEBOK Guide and Software Services Engineering Education
Hironori Washizaki
 
PPTX
AUTOMATION AND ROBOTICS IN PHARMA INDUSTRY.pptx
sameeraaabegumm
 
PDF
[Newgen] NewgenONE Marvin Brochure 1.pdf
darshakparmar
 
PDF
Fl Studio 24.2.2 Build 4597 Crack for Windows Free Download 2025
faizk77g
 
PDF
Empower Inclusion Through Accessible Java Applications
Ana-Maria Mihalceanu
 
PDF
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
PDF
"AI Transformation: Directions and Challenges", Pavlo Shaternik
Fwdays
 
PDF
Newgen 2022-Forrester Newgen TEI_13 05 2022-The-Total-Economic-Impact-Newgen-...
darshakparmar
 
PDF
Exolore The Essential AI Tools in 2025.pdf
Srinivasan M
 
PDF
Log-Based Anomaly Detection: Enhancing System Reliability with Machine Learning
Mohammed BEKKOUCHE
 
PDF
HCIP-Data Center Facility Deployment V2.0 Training Material (Without Remarks ...
mcastillo49
 
PDF
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 
CIFDAQ Weekly Market Wrap for 11th July 2025
CIFDAQ
 
Chris Elwell Woburn, MA - Passionate About IT Innovation
Chris Elwell Woburn, MA
 
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
How Startups Are Growing Faster with App Developers in Australia.pdf
India App Developer
 
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
From Code to Challenge: Crafting Skill-Based Games That Engage and Reward
aiyshauae
 
Using FME to Develop Self-Service CAD Applications for a Major UK Police Force
Safe Software
 
AI Penetration Testing Essentials: A Cybersecurity Guide for 2025
defencerabbit Team
 
SWEBOK Guide and Software Services Engineering Education
Hironori Washizaki
 
AUTOMATION AND ROBOTICS IN PHARMA INDUSTRY.pptx
sameeraaabegumm
 
[Newgen] NewgenONE Marvin Brochure 1.pdf
darshakparmar
 
Fl Studio 24.2.2 Build 4597 Crack for Windows Free Download 2025
faizk77g
 
Empower Inclusion Through Accessible Java Applications
Ana-Maria Mihalceanu
 
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
"AI Transformation: Directions and Challenges", Pavlo Shaternik
Fwdays
 
Newgen 2022-Forrester Newgen TEI_13 05 2022-The-Total-Economic-Impact-Newgen-...
darshakparmar
 
Exolore The Essential AI Tools in 2025.pdf
Srinivasan M
 
Log-Based Anomaly Detection: Enhancing System Reliability with Machine Learning
Mohammed BEKKOUCHE
 
HCIP-Data Center Facility Deployment V2.0 Training Material (Without Remarks ...
mcastillo49
 
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 

Age of Language Models in NLP

  • 1. The Age of Language Models in NLP Tuesday | 23rd June, 2020 LIVE WEBINAR Presented by
  • 2. AGENDA 1 . About Tyrone  World’s high performing AI platform system – A100  Get Development, Training, Inference in one  Era of Modern Mixed Workloads  Tyrone Kubyts™ 2. Word Embeddings  How Word embeddings create a context based relationship  How to create Word Embeddings 3. Sequence Modelling  Introduction of Deep learning in NLP  Overview on the model architecture to use 4. Advanced Language Models  Overview of the Language models  How are they created  Transformers  BERT , GP2 etc 5. NLP Attention Mechanism  Overview on attention Mechanism 6. Case Studies
  • 3. Tyrone Systems at a Glance
  • 4. NVIDIA HGX A100 PERFORMANCE New Tensor Core for AI & HPC New Multi-instance GPU New Hardware Engines Increase in GPU interconnect bandwidth Increase in GPU memory Increase in memory bandwidth Speedup in AI performance
  • 5. 54 Billion XTORS 3rd Gen Tensor cores Sparsity Acceleration Multi Instance GPU 3rd GEN NVLINK & NVSwitch
  • 6. NVIDIA A100 Greatest Generational Leap – 20X Volta 54B XTOR | 826mm2 | TSMC 7N | 40GB Samsung HBM2 | 600 GB/s NVLink Peak Vs Volta FP32 TRAINING 312 TFLOPS 20X INT8 INFERENCE 1,248 TOPS 20X FP64 HPC 19.5 TFLOPS 2.5X MULTI INSTANCE GPU 7X GPUs
  • 7. New tf32 tensor cores on A100 20X Higher FLOPS for AI, Zero Code Change 20X Faster than Volta FP32 | Works like FP32 for AI with Range of FP32 and Precision of FP16 No Code Change Required for End Users | Supported on PyTorch, TensorFlow and MXNet Frameworks Containers
  • 8. Most flexible ai platform with MULTI-INSTANCE GPU (MIG) Optimize GPU Utilization, Expand Access to More Users with Guaranteed Quality of Service Up To 7 GPU Instances In a Single A100: Simultaneous Workload Execution With Guaranteed Quality Of Service: All MIG instances run in parallel with predictable throughput & latency Flexibility to run any type of workload on a MIG instance Right Sized GPU Allocation: Different sized MIG instances based on target workloads Amber GPU Mem GPU GPU Mem GPU GPU Mem GPU GPU Mem GPU GPU Mem GPU GPU Mem GPU GPU Mem GPU
  • 9. ONE SYSTEM FOR ALL ai INFRASTRUCTURE AI Infrastructure Re-Imagined, Optimized, and Ready for Enterprise AI-at-Scale any job | any size | any node | anytime Analytics  Training  Inference Flexible AI infrastructure that adapts to the pace of enterprise • One universal building block for the AI data center • Uniform, consistent performance across the data center • Any workload on any node - any time • Limitless capacity planning with predictably great performance with scale
  • 10. Game-changing performance for innovators 9x Mellanox ConnectX-6 200Gb/s Network Interface 450GB/sec Peak Bi-directional Bandwidth Dual 64-core AMD Rome CPUs and 1TB RAM 3.2X More Cores to Power the Most Intensive AI Jobs 8x NVIDIA A100 GPUs with 320GB Total GPU Memory 12 NVLinks/GPU 600GB/sec GPU-to-GPU Bi-directional Bandwidth 15TB Gen4 NVME SSD 4.8TB/sec Bi-directional Bandwidth 2X More than Previous Generation NVSwitch 6x NVIDIA NVSwitches 25GB/sec Peak Bandwidth 2X Faster than Gen3 NVME SSDs
  • 11. 2U GPU server up to 4 NVIDIA HGX ™ A100 GPU Camarero DAS7TGVQ-24RT Tyrone NVIDIA A100 based SERVERS • Supports 4x A100 40GB SXM4 GPUs • Supports CPU TDP up to 280W • Dual AMD EPYC™ 7002 Series Processors in up to 128 Cores • Flexible storage with 4 hotswap for SAS, SATA or NVMe • PCI-E Gen 4 NVLink for fast GPU-GPU connections • 32 DIMM Slots that allow up to 8TB of 3200Mhz DDR4 memory • 4 Hot-swap heavy duty fans • 2x 2200W Redundant Power Supplies, Titanium Level PCI-E Gen 4 NEW LAUNCH NVIDIA NVLink
  • 12. 4U GPU server up to 8 NVIDIA HGX ™ A100 GPU Tyrone NVIDIA A100 based SERVERS NVIDIA NVLink & NVSwitch NEW LAUNCH • Supports up to 8 double-width GPUs, • Supports CPU TDP up to 280W • Dual AMD EPYC™ 7002 Series Processors in up to 128 Cores • Flexible storage with 4 hotswap for SAS, SATA or NVMe • PCI-E Gen 4 NVLink for fast GPU-GPU connections • 32 DIMM Slots that allow up to 8TB of 3200Mhz DDR4 memory • 4 Hot-swap heavy duty fans • 2x 2000W Redundant Power Supplies, Titanium Level
  • 13. 4U GPU server up to 8 NVIDIA HGX ™ A100 GPU Tyrone NVIDIA A100 based SERVERS NVIDIA NVLink COMING SOON • Supports Intel Xeon • Supports NVLink • 8 x NVIDIA Tesla A100 SXM4
  • 14. Delivers 4XFASTER TRAINING than other GPU-based systems Your Personal AI Supercomputer Power-on to Deep Learning in Minutes Pre-installed with Powerful Deep Learning Software Extend workloads from your Desk-to-Cloud in Minutes
  • 15. Mixed Workloads Convergence of AI|HPC| Cloud | Containers
  • 16. The Era of Modern Mixed Workload F L E X I B L E Is the usage going to be constant? O P T I M I Z A T I O N Is optimal utilization required? R E S I L I E N C E Do we need the application to run all the time. E A S E Is ‘ease of maintenance’ key? S C A L A B I L I T Y & S P E E D Do we have one size that fits all?
  • 17. Connectivity and usage Virtual Desktop Laptop Tyrone Cloud Manager Tyrone Cloud Manager Laptop
  • 18. Run Multiple Applications simultaneously Flow Architecture Revolutionizing Deep Learning CPU-GPU Environment 10X20X30X40X50X60X70X SPEED WITH TYRONE KUBYTS™ CLIENT Compatible Workstations has a repository of : 50 containerized applications 100s of Containers CLOUD
  • 19. Tyrone KUBITS : Revolutionizing Deep Learning CPU-GPU Environment Run different applications simultaneously Check for Tyrone KUBITS Compatible Workstations Get access to over 100+ Containers on Tyrone KUBITS Cloud. High scalability Affordable price Has both GPU & CPU Optimized Containers Design a simple Workstation or Large Clusters with KUBITS technology. Talk to our experts & build the right workstation within your budget. KUBITS CLOUDCOMPATIBLE
  • 20. AGENDA 1 . About Tyrone  World’s high performing AI platform system – A100  Get Development, Training, Inference in one  Era of Modern Mixed Workloads  Tyrone Kubyts™ 2. Word Embeddings  How Word embeddings create a context based relationship  How to create Word Embeddings 3. Sequence Modelling  Introduction of Deep learning in NLP  Overview on the model architecture to use 4. Advanced Language Models  Overview of the Language models  How are they created  Transformers  BERT , GP2 etc 5. NLP Attention Mechanism  Overview on attention Mechanism 6. Case Studies
  • 21. Word Embedding • Word Embedding is a language modeling technique used for mapping words to vectors of real numbers. • It represents words or phrases in vector space with several dimensions. • Word embeddings can be generated using various methods like neural networks, co-occurrence matrix, probabilistic models, etc. • Word2Vec consists of models for generating word embedding. These models are shallow two layer neural networks having one input layer, one hidden layer and one output layer. Word2Vec utilizes two architectures
  • 22. CBOW – Continuous Bag of Words • CBOW model predicts the current word given context words within specific window. The input layer contains the context words and the output layer contains the current word. • The hidden layer contains the number of dimensions in which we want to represent current word present at the output layer.
  • 23. Skip Gram – Word Embeddings • Skip gram predicts the surrounding context words within specific window given current word. The input layer contains the current word and the output layer contains the context words. • The hidden layer contains the number of dimensions in which we want to represent current word present at the input layer.
  • 24. Advanced Language Models and Transformers ELMo ULMFit BERT Transformer
  • 26. BERT Architecture • Transformer is an attention-based architecture for NLP • Transformer composed of two parts: Encoding component and Decoding component • BERT is a multi-layer bidirectional Transformer encoder
  • 29. Q&A Session Hirdey Vikram [email protected] India (North) Niraj [email protected] India (South) Vivek [email protected] India (East) Navin [email protected] India (West) Anupriya [email protected] Singapore Arun [email protected] UAE Agam [email protected] Indonesia Contact our team if you have any further questions after this webinar [email protected] to our AI Experts