The freeCodeCamp.org PyTorch deep learning tutorial is over 25 hours (1537 minutes) long. How long can I endure it?
Here are my notes from learning the subject during spring 2025.
Deep learning is a subset of machine learning, which finds patterns in data. This requires the idealized output (labels; supervised learning) from the inputs (features) so the algorithm can check how good it is. This is better-suited than traditional programming for processes with patterns that are difficult for humans to understand.
Deep learning (DL) is a subset of machine learning (ML), which finds patterns in data. This requires the idealized output (labels; supervised learning) from the inputs (features) so the algorithm can check how good it is. This is better-suited than traditional programming for processes with patterns that are difficult for humans to understand (too many rules/variations/too much data).
Deep learning is not explainable, much like our conscious brains must estimate the thought process that occurred in our subconscious. This can make it a bit difficult to predict or fix errors. DL is also rather data-hungry, similarly to the necessitation of Meta's mass piracy of literature to feed their AI models.
Traditional machine learning, such as a gradient-boosted machine, is good for structured data. Deep learning, such as neural networks (NNs), is better for unstructured data such as observations in the real world.
Traditional ML methods ("shallow methods"):
DL methods ("deep methods"):
The above NN types are the original fundamentals.
NNs are a biomimetic structure of neuron-like nodes that are connected by synapse-like edges that send signals with weights that are altered by the nodes before it. This does a bunch of transformations on the inputs that leads to a certain conclusion in whatever output space. They can be trained in various ways to do this most effectively (to get the best "weights").
This is used for unstructured data, but, because we are working with computers, they must be numbers encoded in bits.
The architecture of a NN has an input layer, various numbers of hidden layers, and an output layer. I theorize that, if hidden layers can indeed be represented as matrix transformations, then maybe the hidden layers can just be combined to one layer.
Supervised learning is the traditional method, wherein you have a lot of inputs and give the ML algorithm all of the idealized outputs so it knows how it's doing. It's able to grade and "supervise" itself.
Unsupervised learning finds patterns between the data, but it doesn't really have any concept of what is right or wrong or what anything means to us humans.
Transfer learning adapts weights to our own model so that our model can get a head start on finding the right weights.
Reinforcement learning (which is not being covered in this tutorial) operates on carrot-and-stick rewards and punishments to incentivize the model to succeed.
DL is different from ML due to the large amount of layers and sheer complexity of the data. This sort of avoids the Google ML rule of avoiding ML for things that can just be traditionally programmed and really stretches the capabilities of ML for the profound-yet-unexplainable observations that we expect oft-hyped omniscient models to produce.
I will be using PyTorch v2.6.0 on Ubuntu 24 LTS. The second most-used machine learning program is TensorFlow, but the market share is lesser by far.
Input data is converted into -dimensional arrangements of numbers called tensors. We preform transformations upon these tensors using the hidden layers (many of them--deep), harvest the output tensors (perhaps in a different space), and convert that output to something that humans can understand.
Since I am now getting into ML development on the same Ubuntu install where I have a lot of other projects going on, I was looking for a method to isolate these environments without having to worry about dependency conflicts and installing multiple versions of the same thing and having to choose which one to run (e.g. ROS1 Noetic and ROS2 Humble).
Docker is an industry-standard tool for doing this, wherein you create various Dockerfiles that set up an isolated environment. You have to describe the packages that must be installed in sequence. In this way, you maintain consistency and also get self-documentation capabilities for the installation process. Follow the installation instructions for Ubuntu here.
You can also apparently piggyback off of existing Docker images to make your script similar. This is my first Dockerfile:
FROM pytorch/pytorch:2.2.0-cuda12.1-cudnn8-runtime
WORKDIR /workspace
RUN apt-get update && apt-get install -y git wget unzip
RUN pip install torch torchvision torchaudio numpy matplotlib opencv-python pandas matplotlib scikit-learn tqdm requests jupyter seaborn
CMD ["bash"]
I have no idea what the CMD ["bash"]
does, but I'll roll with it.
We can then run the Docker container and clone freeCodeCamp's tutorial repo after:
docker build -t fcc_pytorch_dl
docker run -it --rm --gpus all -v $(pwd):/workspace -p 8888:8888 fcc_pytorch_dl
# After env is good
## Clone tutorial repo
git clone https://github.com/mrdbourke/pytorch-deep-learning#course-materialsoutline
## Locally start the JUpyter notebook on port 8888
jupyter notebook --ip=0.0.0.0 --port=8888 --allow-root
To list the current containers, I can use docker ps -a
.
sudo docker build -t fcc_pytorch_dl .
sudo docker run -it fcc_pytorch_dl bash
I then cloned the tutorial repository to the /workspace
directory. To be able to access the jupyter notebook server from your browser outside of the container:
docker run -it -p 8888:8888 fcc_pytorch_dl jupyter notebook --ip=0.0.0.0 --allow-root
Then go to localhost:8888
and paste in the token you get embedded in a URL from the output of the above command.
There's an error here--when I access the Jupyter server from outside of the container, there are no files present.
Let's worry about how to fix Docker later. Now back to the tutorial.
For an -dimensional tensor, where are each of the dimensions.
A tensor of -dim can be created with torch.zeros([d1, d2, d3, ...])
. Inputting the values manually can be done with torch.tensor()
.
.ndim
finds the number of dimensions .item()
gets the number for a one-element tensor