Skip to content
This repository was archived by the owner on Apr 28, 2025. It is now read-only.

Commit ca42700

Browse files
Initial commit
0 parents  commit ca42700

File tree

5 files changed

+300
-0
lines changed

5 files changed

+300
-0
lines changed

.github/workflows/docker.yml

Lines changed: 41 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,41 @@
1+
name: Build notebook
2+
on:
3+
schedule:
4+
- cron: '0 0 * * 0'
5+
push:
6+
7+
env:
8+
REGISTRY: ghcr.io
9+
IMAGE_NAME: ${{ github.repository }}
10+
11+
jobs:
12+
build-and-push-image:
13+
runs-on: ubuntu-latest
14+
permissions:
15+
contents: read
16+
packages: write
17+
18+
steps:
19+
- name: Checkout notebook
20+
uses: actions/checkout@v3
21+
22+
- name: Log in to registry
23+
uses: docker/login-action@v2
24+
with:
25+
registry: ${{ env.REGISTRY }}
26+
username: ${{ github.actor }}
27+
password: ${{ secrets.GITHUB_TOKEN }}
28+
29+
- name: Generate tags and labels
30+
id: meta
31+
uses: docker/metadata-action@v4
32+
with:
33+
images: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
34+
35+
- name: Build and push notebook
36+
uses: docker/build-push-action@v4
37+
with:
38+
context: .
39+
push: true
40+
tags: ${{ steps.meta.outputs.tags }}
41+
labels: ${{ steps.meta.outputs.labels }}

Dockerfile

Lines changed: 26 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,26 @@
1+
# 1) choose base container
2+
# generally use the most recent tag
3+
4+
# base notebook, contains Jupyter and relevant tools
5+
# See https://github.com/ucsd-ets/datahub-docker-stack/wiki/Stable-Tag
6+
# for a list of the most current containers we maintain
7+
ARG BASE_CONTAINER=ghcr.io/ucsd-ets/datascience-notebook:2023.2-stable
8+
9+
FROM $BASE_CONTAINER
10+
11+
LABEL maintainer="UC San Diego ITS/ETS <[email protected]>"
12+
13+
# 2) change to root to install packages
14+
USER root
15+
16+
RUN apt-get -y install htop
17+
18+
# 3) install packages using notebook user
19+
USER jovyan
20+
21+
# RUN conda install -y scikit-learn
22+
23+
RUN pip install --no-cache-dir networkx scipy
24+
25+
# Override command to disable running jupyter notebook at launch
26+
# CMD ["/bin/bash"]

README.md

Lines changed: 161 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,161 @@
1+
# Instructions on Building a Custom Image for DataHub/DSMLP
2+
3+
This guide is for advanced DSMLP users (both students and instructors) who want to add or modify applications on their working environment using a custom Docker container.
4+
5+
For CUDA-enabled images, checkout [cuda.md](cuda.md).
6+
7+
## Introduction
8+
9+
A Docker **image** is a snapshot of packaged applications, dependencies and the underlying operating system. Users can use the same Docker image anywhere on any machine running the Docker platform while having the same software functionality and behavior. **Github Container Registry (GHCR)** is a public container registry you can download ("pull") and upload ("push") Docker images located under the **[Packages](https://github.com/orgs/ucsd-ets/packages?repo_name=datahub-example-notebook)** section of this Github repo. In this guide, we will build a custom Docker image by modifying a **Dockerfile**, building the image on a desired platform, and publishing it on GHCR.
10+
11+
**Docker Hub** is also a public container registry you can download ("pull") and upload ("push") Docker images. In other words, Docker Hub hosts and distributes Docker images just like GHCR.
12+
13+
Building and maintaining a Docker image follows three essential steps: build, share and deploy/test. It's likely for you to go through these steps several times until it achieves what you want. You can find an official tutorial from [docs.docker.com](https://docs.docker.com/get-started/) that demonstrates a general case, but this document is tailored specifically for DSMLP users.
14+
15+
## Step 0: Prerequisites
16+
17+
- A new **public** GitHub git repo using this as a template. Click "Use this template" at upper-right corner. You can also use an existing **public** repo by adding a `Dockerfile` at the repo's root level. The public visibility is to stay on the free plan that GitHub offers, which comes in to play later.
18+
- An **optional** Docker Hub account if following [Option 1, Push](#option-1:-use-docker-desktop/docker-engine)
19+
20+
## Step 1: Customize the Dockerfile
21+
22+
### Step 1.1 Base Image
23+
24+
- Choose the base container by uncommenting the corresponding line that sets the `BASE_CONTAINER` argument
25+
- [An overview of standard Datahub/DSMLP containers maintained by UCSD EdTech Services](https://github.com/ucsd-ets/datahub-docker-stack/wiki/Stable-Tag)
26+
- The `datahub-base-notebook` image contains Jupyter and common data science tools from Python and R. Derived from [`jupyter/datascience-notebook`](https://jupyter-docker-stacks.readthedocs.io/en/latest/using/selecting.html#jupyter-datascience-notebook).
27+
- The `datascience-notebook` image has a few more packages installed using pip.
28+
- The `scipy-ml` image has a wider range of packages including tensorflow, pytorch, including CUDA 11 support, generally used for GPU-accelerated workflows.
29+
30+
- **Note**: Although `scipy-ml` has more features, building an image on top of it will take longer. It's better to apply a minimal set of required tools to the `base-notebook` than for saving image size and reduce image build time.
31+
32+
### Step 1.2 System Packages
33+
34+
- Use `USER root` to gain root privileges for installing system packages. This line is already typed out for you.
35+
36+
- Install system-level packages using `apt-get`
37+
- The example at line 19 installs a system utility called `htop`.
38+
- Specify more packages as extra arguments: `apt-get -y install htop ping`
39+
40+
### Step 1.3a Python Libraries
41+
42+
- Note: It is recommended to use `pip` instead of `conda` as much as possible. `pip` is more forgiving in resolving package conflicts, and generally much faster.
43+
44+
- Install conda packages
45+
- Use `RUN conda install --yes <package1> <package2>` to install all required conda packages in one go
46+
- (Optional) Use `RUN conda clean -tipy` to reduce image size
47+
48+
- Install pip packages
49+
- Only use pip after conda.
50+
- Use `pip install --no-cache-dir <package>`
51+
- Alternatively, you can provide a `requirements.txt` file in the project root and use `pip install --no-cache-dir -r requirements.txt` to reference it. List each package as a single line.
52+
53+
- Leave the rest of the Dockerfile as is
54+
55+
### Step 1.3b R Libraries (If Needed)
56+
57+
- Note: It is recommended to use the "install.packages" command in your Dockerfile instead of Conda whenver possible. This is because installing Conda packages increases the build time exponentially. Example:
58+
59+
```
60+
RUN R -e "install.packages( \
61+
c('ggplot2', \
62+
'dplyr'), repos='http://cran.rstudio.com/')"
63+
```
64+
If you cannot use cran to install your libraries, try using conda-forge in your Dockerfile instead. Example:
65+
```
66+
RUN conda install -c conda-forge r-magick r-tesseract r-pdftools
67+
```
68+
69+
## Step 2: Build the Image
70+
71+
In this step you will build the image using the Dockerfile you created. Here you have two options:
72+
73+
1. Install the Docker Client and build the image locally. This will require you have [Docker Desktop](https://www.docker.com/products/docker-desktop) installed on your local Windows PC/Mac or [Docker Engine](https://docs.docker.com/engine/install/) on Linux. You can also use a remote Linux server with Docker installed for building Docker Images and testing them. The commands will be the same, but if you don't have docker locally, you cannot develop locally, only on DSMLP. For development on Windows, I strongly recommend you to install the [WSL 2 engine](https://docs.microsoft.com/en-us/windows/wsl/install) for a better experience.
74+
75+
2. Setup a [GitHub Actions](https://github.com/features/actions) configuration file inside your project. A template is provided and require minimum modification.
76+
77+
It is recommended to use both routes for easier debugging and shorter turnaround time on successful builds. Option 2 is much easier but to understand the basics, please go through Option 1 beforehand. Installing Docker on your local setup is sometimes a headache, however you can take advantage of the $100 DigitalOcean credit from the [GitHub Student Developer Pack](https://education.github.com/pack) and launch a Docker Droplet there. Or use any remote Linux box you can find for free. Check out the resources in the Appendix.
78+
79+
### Option 1: Use Docker Desktop/Docker Engine
80+
81+
After Docker is installed, launch a terminal and navigate to the your git directory containing the Dockerfile.
82+
83+
Make sure to double check the name of your GitHub/DockerHub account and the names of the GitHub repo and Docker Hub repo. The full image name, for when you pull the image from somewhere else, will be `<dockerhub-username>/<dockerhub-repo-name>`, with an optional `:<tag>` that immediately follows it. When the tag isn't supplied, it will use `:latest` as default. We will reference this full name as `<image-fullname>` for the commands that follow.
84+
85+
#### Step 2.1.1 Build
86+
87+
Type "`docker build -t <image-fullname> .`" and hit \<Enter\>, notice the "period/dot" at the end of the command, which denotes the current directory. Docker will then build the image in the current directory's context. The resulting image will be labeled `<image-fullname>`. Monitor the build process for errors.
88+
89+
#### Step 2.1.2 Debug
90+
91+
If the build fails, take note of the last command Docker that was run and start debugging from there. Run the build command again after editing the Dockerfile. Sometimes, it is better to launch the intermediate docker image that follows a step and launch the image from there and try a few commands. To do this, use the command in Step 2.1.3 Test Run.
92+
93+
This will help you find the name of the intermediate image. If `Step 4` fails, look through the output and finds the image from `Step 3`, in the following example output, we will use the image `51a4d2ec5e16` to debug.
94+
95+
```
96+
Step 3/14 : USER root # Step count and command
97+
---> Running in 4e32937f1e93 # temporary container
98+
---> 51a4d2ec5e16 # result image (use this to debug)
99+
```
100+
101+
Some commmon errors/mistakes here include:
102+
103+
1. not supplying the default yes `-y` option to the install command, causing it to timeout.
104+
105+
2. If the error message says it finds `/r/n` in one of your files, change the end of line sequence to from CRLF to LF. You can do this in an editor or use the utility `dos2unix`. This typically happens on Windows machines.
106+
107+
#### Step 2.1.3 Test Run
108+
109+
- Once the image is successfully built, use `docker run --rm -it <image-fullname> /bin/bash` to enter the image. `-it` denotes interactive mode, and `--rm` tells docker to remove the container after exit. Check if it has all the functionality you want. Use `exit` to exit from the container. If something is wrong, go back to the build step and start over.
110+
111+
#### Step 2.1.4 Push
112+
113+
- Log in to Docker Hub on the Docker Client. This can be done using the GUI or by using the command `docker login <username>`.
114+
115+
- push the image `docker push <image-fullname>`, the tag will default to `latest`. If you want to push a different tag, add `:<tag>` at the end of the full name.
116+
117+
- If you are using a remote pay-as-you-go Linux VM such as DigitalOcean, don't forget to remove the instance to save cost!
118+
119+
### Option 2: Use Github Actions
120+
121+
After going through the previous option, you should be familiar with the entire workflow of building, testing, and pushing the Docker image. Now we can use Github Actions to automatically do these things for us.
122+
123+
1. Follow the file at `.github/workflows/docker.yml`. Notice here the `.github` will be a hidden directory, which can be hidden graphically on Windows if you don't have that setting enabled. This workflow uses a [standard action](https://docs.github.com/en/packages/managing-github-packages-using-github-actions-workflows/publishing-and-installing-a-package-with-github-actions) for building and pushing Docker images to GHCR
124+
125+
*Note: this action will create a Docker tag based off the Git branch name, e.g., the "main" branch will create a Docker image:tag as `ghcr.io/ucsd-ets/datahub-example-notebook:main`*
126+
127+
2. Leave the rest of the content as is. You will find that it contains all the necessary steps in Option 1. If you are feeling confident, add more steps to augment the workflow.
128+
129+
3. Commit and push the changes to GitHub. In the "Actions" tab, there will be a new workflow and under there, you can check the progress and output.
130+
131+
4. If your Github action was successful, you'll see a newly created package under [Packages](https://github.com/orgs/ucsd-ets/packages?repo_name=datahub-example-notebook)
132+
133+
For more information, check out the syntax for Github Actions and relevant documentation, [here](https://docs.github.com/en/actions/learn-github-actions/workflow-syntax-for-github-actions).
134+
135+
## Step 3: Launch a Pod on DSMLP
136+
137+
- SSH to `dsmlp-login.ucsd.edu`
138+
139+
- RUN `launch-scipy-ml.sh -i <image-fullname> -P Always` . The `-P Always` flag will force the docker host to sync, as it pulls the latest version of the image manifest. You will not need the `-P Always` flag once you are done with development. Note: a docker image name follows the format `<user>/<image>:<tag>`. The `:<tag>` part will be assumed to be `:latest` if you don't supply it to the launch script. Use tags like `v1` or `test` in the build step to have control over different versions of the same docker image.
140+
141+
- Wait for the node to download the image. Download time depends on the image size.
142+
143+
- If it timeout/fails to launch, check `kubectl logs <pod-name>` or contact ETS service desk for help.
144+
145+
# Tips
146+
147+
- If you are repeatedly using the pod or sharing the custom image among a few other people within a day, use the same node to reduce spawn time (without download). You can do this by adding a `-n <node-number>` at the end of the launch command.
148+
149+
- To disable launching jupyter notebook upon entry, override the default executable by adding `CMD ["/bin/bash"]` as the last layer (as last line in `Dockerfile`). You can always launch the notebook again and manually port-forward on dsmlp-login. `kubectl port-forward pods/<POD_NAME> <DSMLP_PORT>:8888`
150+
151+
# Resources/Further Reading
152+
153+
- [**DSMLP Knowledge Base Articles**](https://support.ucsd.edu/its?id=kb_category&kb_category=7defd803db49fb08bd30f6e9af961979&kb_id=e343172edb3c1f40bd30f6e9af961996)
154+
- [Best practices for writing Dockerfiles](https://docs.docker.com/develop/develop-images/dockerfile_best-practices/)
155+
- [Docker Cheat Sheet (pdf)](https://www.docker.com/sites/default/files/d8/2019-09/docker-cheat-sheet.pdf)
156+
- [Original version of this guide](https://docs.google.com/document/d/1LPfqHvk2Itm_ckafrxRVxXQdr5BSozjsv_TURQDj9x8/edit)
157+
158+
# Appendix
159+
160+
- [DigitalOcean Docker Droplet](https://marketplace.digitalocean.com/apps/docker), this will give you a Linux VM with Docker installed ready to go. Start with any configuration with at least 8GB of memory and up it if working with a large image.
161+
- [GitHub Student Developer Pack](https://education.github.com/pack)

additional-information.md

Lines changed: 40 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,40 @@
1+
# Additional Information
2+
3+
## Dockerfile Best Practices
4+
5+
### Additional Kernels
6+
7+
To install a new kernel that can be selected within a jupyter notebook, you can look into creating a second conda environment and use [nb_conda_kernels](https://github.com/Anaconda-Platform/nb_conda_kernels) to add it in.
8+
9+
### Keep your images small
10+
11+
Concatentate `RUN` steps into a single `RUN` block
12+
13+
```bash
14+
# Creates a separate image layer for every RUN block
15+
RUN apt-get -y update
16+
RUN apt-get install -y python
17+
18+
# Single RUN block results in a smaller image
19+
RUN apt-get -y update && \
20+
apt-get install -y python
21+
```
22+
23+
If you've converted all your `RUN` statements and find that your container is still prohibitively large, try breaking your container into a multi-stage build or have multiple containers for different purposes.
24+
25+
### Use Mamba instead of Conda when Conda is slow
26+
27+
Sometimes you may need to install a conda package. If you include a statement like below:
28+
29+
```bash
30+
RUN conda install <package> -y
31+
```
32+
33+
and you find that it an in-ordinate amount of time to install, you can try using [mamba](https://github.com/mamba-org/mamba). It works faster.
34+
35+
If that still doesn't work, you can try to manually install your software
36+
37+
### Remove unused files and software
38+
39+
If you need to reduce the size of your image, try removing files and/or software you may not need. You can also try starting from our smallest image `ucsd-ets/datahub-base-notebook` as your starting point [An overview of standard Datahub/DSMLP containers maintained by UCSD EdTech Services](https://github.com/ucsd-ets/datahub-docker-stack/wiki/Stable-Tag)
40+

cuda.md

Lines changed: 32 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,32 @@
1+
# Building CUDA-enabled Images for DataHub/DSMLP
2+
3+
In this branch we will cover the starting steps of creating a GPU accelerated Docker Image for DSMLP. It's recommended to follow the steps in the "master" branch before continuing.
4+
5+
**Note**: our standard [scipy-ml-notebook](https://github.com/ucsd-ets/datahub-docker-stack/wiki/Stable-Tag) has CUDA installed with a compatible PyTorch and Tensorflow for use on DSMLP. Please use that image if that's all you require since installing a working CUDA environment is time consuming.
6+
7+
## DSMLP: Available GPUs
8+
9+
As of Fall 2020, there are 15 GPU nodes on the DSMLP cluster available for classroom use, and each node has 8 NVIDIA GPUs installed. These GPUs are dynamically assigned to a container on start-up when requested and will stay attached until that container is deleted, meaning a GPU will remain occupied even if it's actually not running anything.
10+
11+
12+
| GPU Model | VRAM | Amount | Node |
13+
|----------------|------|--------|---------------------------------|
14+
| NVIDIA 1080 Ti | 11GB | 80 | n01 through n12 except n09, n10 |
15+
| NVIDIA 2080 Ti | 11GB | 32 | n18, n21, n22, n24 |
16+
| NVIDIA 1070 Ti | 8GB | 7 | n10 |
17+
18+
19+
## Install CUDA, PyTorch, and TensorFlow
20+
21+
*Note: The Datahub/DSMLP cannot guarantee that versions of Python libraries that use particular CUDA versions will work on the platform as CUDA versions must be synced to specific [Linux x86_64 Driver Versions](https://docs.nvidia.com/deploy/cuda-compatibility/index.html#binary-compatibility__table-toolkit-driver)*
22+
23+
The graphics driver will be installed automatically on the container on start-up. It can be challenging to get CUDA compatibility working in a custom environment. You can find the latest CUDA Toolkit supported on DSMLP by checking how we install it within our [scipy-ml-notebook](https://github.com/ucsd-ets/datahub-docker-stack/blob/main/images/scipy-ml-notebook/Dockerfile) container. The version of CUDA in this Dockerfile is the latest version of CUDA the DSMLP supports, and versions > what's defined may not work on the platform.
24+
25+
Please see the [scipy-ml-notebook](https://github.com/ucsd-ets/datahub-docker-stack/blob/main/images/scipy-ml-notebook/Dockerfile) Dockerfile for supported TensorFlow and Pytorch versions.
26+
27+
<!-- ## Dockerfile: Write Access to /opt/conda -->
28+
29+
# Resources/Further Reading
30+
- [**DSMLP Knowledge Base Articles**](https://support.ucsd.edu/its?id=kb_category&kb_category=7defd803db49fb08bd30f6e9af961979&kb_id=e343172edb3c1f40bd30f6e9af961996)
31+
- [CUDA Compatibility Table](https://docs.nvidia.com/deploy/cuda-compatibility/index.html#binary-compatibility__table-toolkit-driver)
32+
- [cuDNN Support Matrix](https://docs.nvidia.com/deeplearning/cudnn/support-matrix/index.html)

0 commit comments

Comments
 (0)