# The CIFAR-10 Dataset

Learn about the CIFAR-10 dataset.

## We'll cover the following

## About this chapter

At this point in the course, we are comfortable with the structure of a neural network. All the networks that we’ve seen so far share the same architecture: **a sequence of dense layers**, whereby “**dense**” means that each node in a layer is connected to all the nodes in the neighboring layers. That blueprint is also called a **fully connected neural network**, and it comes with a drastic limitation. Whether it’s classifying images, parsing a product review, or predicting traffic congestion, a fully connected network treats all data the same, *as an indistinct sequence of bytes*. That generalistic approach can only go so far and fails when dealing with complex datasets.

Deep learning is not just about making neural networks deeper; it’s also about making them **smarter**, and *adopting different architectures to deal with different kinds of data*. Deep learning researchers came up with quite a few variations to the basic fully connected network, and more are probably getting invented as we read this lesson. Each of those architectures would deserve its own course, and some add up to an entire field of study.

We certainly would not be able to describe or even mention all those different blueprints in this chapter. However, we can take a brief look at one of them. In the context of the rest of this course, we’ll focus on an architecture that’s customized to process images called a **convolutional neural network**, or **CNN** for short. Where fully connected networks multiply each layer by a set of weights, CNNs are based on a more complicated operation called **convolution**, which makes them particularly good at dealing with spatial data such as images.

Get hands-on with 1200+ tech skills courses.