Category Archives: TensorFlow

Faster Training of Large Language Models with Parallelization

Implementing Faster Training of Large Language Models with Parallelization. Learn about key parallelism techniques for GPUs and TPUs to accelerate training.

Image Synthesis using Pixel CNN based Autoregressive Generative Model

Image Synthesis using Pixel CNN based Autoregressive Generative Models Recent advances in the field of deep learning have led to the development of complex generative models that are capable of generating high quality content in the form of text, audio, pictures, videos and so on. Generative models that make use of deep learning architectures to tackle the task… Read More »

Convolutional Denoising Autoencoders for image noise reduction

Autoencoders are unsupervised Deep Learning techniques that are extensively used for dimensionality reduction, latent feature learning (Learning Representations), and also as generative models (Generative Adversarial Networks: GANs). Denoising Autoencoders are slight modifications to the vanilla autoencoders that can be used for reducing noise from real-world noisy datasets. In this tutorial, we will investigate Convolutional Denoising Autoencoders to reduce… Read More »

Autoencoders in Keras and Deep Learning

We all are well aware of the Supervised Machine Learning algorithms where ML algorithm tries to understand the relationship between input features and labels from the training data and is expected to automatically generate a similar kind of relationship between test data features and set of possible output labels. Although we solve lots of problems with supervised learning… Read More »

Sentiment Classification with Deep Learning: RNN, LSTM, and CNN

Sentiment classification is a common task in Natural Language Processing(NLP). There are various ways to do sentiment classification in Machine Learning (ML). In this article, we talk about how to perform sentiment classification with Deep Learning (Artificial Neural Networks). In my previous two articles, We have already talked about how to perform sentiment analysis using different traditional machine… Read More »

Optimizers explained for training Neural Networks

Overview Training a Deep Learning model (or any machine learning model in fact) is all about bringing the model predictions (model output) close to the real output(Ground truth) for a given set of input-output pairs. Once the model’s results are close to the real results our job is done. To understand how close model predictions are with respect… Read More »

Optimizing TensorFlow models with Quantization Techniques

Deep Learning models are great at solving extremely complex tasks efficiently but this superpower comes at a cost. Due to a large number of parameters, these models are typically big in size(memory footprint) and also slow in the inference (during predictions). Slow and heavy models are not much appreciated when it comes to the deployment part. As we… Read More »