Megatron Unleashed: NVIDIA’s NLP Model “Megatron-LM” is the Largest Transformer Ever Trained
Megatron-LM: Entering the Frontiers of NLP If you’ve followed the latest advancements in Natural Language Processing (NLP), you’ll know that Transformer Models are all the latest craze. These language models…
A Summary of DeepMind’s Protein Folding Upset at CASP13
At last year’s Critical Assessment of protein Structure Prediction competition (CASP13), researchers from DeepMind made headlines by taking the top position in the free modeling category by a considerable margin,…
Examining the Transformer Architecture – Part 2: A Brief Description of How Transformers Work
This post was written by Nityesh Agarwal in collaboration with and sponsorship of Exxact (@Exxactcorp). As we learned in Part 1, The GPT-2 is based on the Transformer, which is being…
What You Missed: Exxact at GTC 2019
Major Annoucement Announced during the GTC keynote by NVIDIA co-founder, Jensen Huang, Exxact will be offering workstations optimized for data-science, along with being one of the partners. What We Showcased…
Is Docker Ideal for Running TensorFlow GPU? Let’s measure using the RTX 2080 Ti
Note: This blog compares only the performance of TensorFlow for the training deep deep neural networks. TensorFlow can be compiled for many different use cases, as with TensorFlow GPU Docker…