Articles About Machine Learning

9 Books on Generative Adversarial Networks (GANs)

Last Updated on August 21, 2019 Generative Adversarial Networks, or GANs for short, were first described in the 2014 paper by Ian Goodfellow, et al. titled “Generative Adversarial Networks.” Since then, GANs have seen a lot of attention given that they are perhaps one of the most effective techniques for generating large, high-quality synthetic images. As such, a number of books have been written about GANs, mostly focusing on how to develop and use the models in practice. In this […]

Read more

A Gentle Introduction to BigGAN the Big Generative Adversarial Network

Generative Adversarial Networks, or GANs, are perhaps the most effective generative model for image synthesis. Nevertheless, they are typically restricted to generating small images and the training process remains fragile, dependent upon specific augmentations and hyperparameters in order to achieve good results. The BigGAN is an approach to pull together a suite of recent best practices in training class-conditional images and scaling up the batch size and number of model parameters. The result is the routine generation of both high-resolution […]

Read more

How to Evaluate Generative Adversarial Networks

Generative adversarial networks, or GANs for short, are an effective deep learning approach for developing generative models. Unlike other deep learning neural network models that are trained with a loss function until convergence, a GAN generator model is trained using a second model called a discriminator that learns to classify images as real or generated. Both the generator and discriminator model are trained together to maintain an equilibrium. As such, there is no objective loss function used to train the […]

Read more

How to Implement the Inception Score (IS) for Evaluating GANs

Last Updated on October 11, 2019 Generative Adversarial Networks, or GANs for short, is a deep learning neural network architecture for training a generator model for generating synthetic images. A problem with generative models is that there is no objective way to evaluate the quality of the generated images. As such, it is common to periodically generate and save images during the model training process and use subjective human evaluation of the generated images in order to both evaluate the […]

Read more

How to Implement the Frechet Inception Distance (FID) for Evaluating GANs

Last Updated on October 11, 2019 The Frechet Inception Distance score, or FID for short, is a metric that calculates the distance between feature vectors calculated for real and generated images. The score summarizes how similar the two groups are in terms of statistics on computer vision features of the raw images calculated using the inception v3 model used for image classification. Lower scores indicate the two groups of images are more similar, or have more similar statistics, with a […]

Read more

A Gentle Introduction to Generative Adversarial Network Loss Functions

The generative adversarial network, or GAN for short, is a deep learning architecture for training a generative model for image synthesis. The GAN architecture is relatively straightforward, although one aspect that remains challenging for beginners is the topic of GAN loss functions. The main reason is that the architecture involves the simultaneous training of two models: the generator and the discriminator. The discriminator model is updated like any other deep learning neural network, although the generator uses the discriminator as […]

Read more

A Gentle Introduction to Jensen’s Inequality

Last Updated on July 31, 2020 It is common in statistics and machine learning to create a linear transform or mapping of a variable. An example is a linear scaling of a feature variable. We have the natural intuition that the mean of the scaled values is the same as the scaled value of the mean raw variable values. This makes sense. Unfortunately, we bring this intuition with us when using nonlinear transformations of variables where this relationship no longer […]

Read more

How to Develop and Evaluate Naive Classifier Strategies Using Probability

Last Updated on September 25, 2019 A Naive Classifier is a simple classification model that assumes little to nothing about the problem and the performance of which provides a baseline by which all other models evaluated on a dataset can be compared. There are different strategies that can be used for a naive classifier, and some are better than others, depending on the dataset and the choice of performance measures. The most common performance measure is classification accuracy and common […]

Read more

Resources for Getting Started With Probability in Machine Learning

Last Updated on September 25, 2019 Machine Learning is a field of computer science concerned with developing systems that can learn from data. Like statistics and linear algebra, probability is another foundational field that supports machine learning. Probability is a field of mathematics concerned with quantifying uncertainty. Many aspects of machine learning are uncertain, including, most critically, observations from the problem domain and the relationships learned by models from that data. As such, some understanding of probability and tools and […]

Read more

5 Reasons to Learn Probability for Machine Learning

Last Updated on November 8, 2019 Probability is a field of mathematics that quantifies uncertainty. It is undeniably a pillar of the field of machine learning, and many recommend it as a prerequisite subject to study prior to getting started. This is misleading advice, as probability makes more sense to a practitioner once they have the context of the applied machine learning process in which to interpret it. In this post, you will discover why machine learning practitioners should study […]

Read more
1 198 199 200 201 202 226