fbpx

Docker Certified Associate Exam Series (Part – 4): Image Creation, Management and Registry

We use images in Docker to package, share and transport applications and preconfigured server environments. An image is, in essence, a template that contains the set of instructions we use to create and run containers on the Docker platform. This section tackles the operations involved in creating, running, and managing images in Docker.

The Docker Image Registry and Operations

In Docker, an image registry is a central repository where you can store and retrieve images. The default repository service for storing and retrieving images is known as the Docker Hub. Here, you will find three kinds of images. Official Images that are published and maintained by Docker, and these include Operating Systems, Databases, and other running software. Verified Images are created and published by third-party vendors such as Oracle and MySQL. Finally, User Images are created and published by developers and users globally.

Within an organization, you can create a repository to publish and share images between colleagues using the Docker Trusted Registry Service. Public cloud providers also offer registry services for individuals and organizations. These include: Google Container Registry, AWS Elastic Container Registry, and Azure Container Registry.

Image Tags help you to categorize and describe an image, ensuring it conveys important information. This way, users and developers can differentiate the various available images in the registry. If you don’t specify an image tag, it is given the latest tag by default.

To search for a public image, it is easy to use the search bar on Docker Hub’s Graphical User Interface.

You can also download and deploy images using the docker run command:

$ docker run ubuntu

You can also search for an image on the Docker CLI. To search for the image httpd, use the command:

$ docker search httpd

To limit your search results to the top 2, use the command:

$ docker search httpd   --limit 2

To limit your search results to images with a 10-star rating, use the command:

$ docker search --filter stars=10 httpd

To limit your search results to verified images with a 10-star rating, use the command:

$ docker search --filter stars=10 --filter is-official=true httpd

To list the images running in a host, run the command:

$ docker image ls

To download an image without running it:

$ docker image pull httpd

Image Addressing Conventions

  • Image names are typically made up of slash-separated elements. The image name is typically prefixed by a hostname that complies with standard DNS naming rules, but need not include underscores. 
  • Any image name that lacks a hostname is assumed to belong to the default public registry docker.io.
  • Other public registry addresses include: gcr.io for google, public.ecr.aws for Amazon, and azurecr.io for Microsoft Azure.

When specifying a name for your image, you will need to give the User/Account name of your organization on the repository, and the name of the image and repository to which it belongs, as seen in the screenshot below:

Authenticating to Registries

When pulling an image from a public registry, you don’t need to authenticate. For instance, to download and run the image for Ubuntu OS from Docker Hub, you only need to run the command:

$ docker run ubuntu

To pull an image from a private repository, you will need to provide authentication, otherwise, you will get an access denied error. You will also need to provide authentication to push (publish) an image into a public or private registry. First, you will log in to the registry. 

To log into a public registry (Docker Hub), run the command:

$ docker login docker.io

To log into a private registry on Google Cloud, run the command:

$ docker login gcr.io

You will then be asked for your username and password. For the former, you can use Docker Hub credentials. If you don’t have a Docker Hub account, you can create one here.

For the Google Cloud account, use the credentials set up in your organization’s repository.

You can then push the image into the registry by running the command:

$ docker image push httpd

Retagging an Image Locally

We retag an image so we can create a soft link to an existing image, providing a replica without having to actually create a copy. The following command will retag an image already in the repository:

$ docker image tag httpd:alpine httpd:customv1

This will create a new image tagged customv1 but with the same ImageID as the one tagged alpine. You can check this by listing images in your registry using the command:

$ docker image list

You may also want to tag this new image into the private registry. To accomplish this use the command:

$ docker image tag httpd:alpine gcr.io/company/httpd:customv1

You can now publish this image privately using the command:

$ docker image push gcr.io/company/httpd:customv1

To display the size of all objects running in Docker, type the command:

$ docker system df

Removing a Docker Image

You should delete Docker Images that are not in use anymore as this helps reclaim space and other resources. To remove an image that you no longer use, run the command:

$ docker image rm

To remove one version of an image, specify its tag in the command, as shown below:

$ docker image rm httpd:customv1

Note that this command only gets rid of an image version, but preserves the image layers as they are being used by other versions. When you remove the last version of an image, all these layers are deleted to reclaim space. 

You can use the prune command to delete all unused images in the Docker environment.

$ docker image prune -a

Inspecting a Docker Image

Various commands allow you to check information on containers straight from the Docker CLI. The docker image list command lets you view all images running in Docker.

To list all layers/versions linked to an image, use the image history command, as follows:

$ docker image history ubuntu

To check on the configuration details in the Dockerfile used to create an image, run the image inspect command:

$ docker image inspect httpd

You can use a JSON expression to filter specific configuration details you need from the Dockerfile. For instance, to get the Operating System the image runs on, you’ll use the expression:

$ docker image inspect httpd -f '{{.Os}}'

To get information on the architecture supporting the image, run the command:

$ docker image inspect httpd -f '{{.Architecture}}'

To get both the architecture and operating system, run the command:

$ docker image inspect httpd -f '{{.Os}}{{.Architecture}}'

To get nested information, separate the components using a period as shown:

$ docker image inspect httpd -f '{{.ContainerConfig.ExposedPorts}}'

Save and Load Image

You may need to get images to restricted environments with no access to the internet. To do this, you convert an image file into a .tar file using the save command:

$ docker image save alpine:latest -o alpine.tar

This creates a .tar bundle that can be transferred to any restricted environment using a network folder or pen drive. In this environment, you can extract the contents of this bundle using the load command:

$ docker image load -i alpine.tar

You can also convert containers into.tar image bundles for transfer to other environments. To achieve this use the export command:

$ docker export <container_name> > testcontainer.tar

You can now import this file as an image on any system using the import command:

$ docker image import testcontainer.tar newimage:latest

You can use the list command to list all the images and check whether your new image has been created.

Building a Custom Image

In some cases, you may not find a component or service in the repository that offers the functionality you need. In such cases, you can create your own image. This is also useful when you are containerizing your application for ease of shipping and deployment. 

In this demo, you will containerize an application made using Python’s Flask framework.

The general process will involve the following steps:

  1. Specify the Operating System (Ubuntu)
  2. Update Repositories using the apt command
  3. Install Python dependencies using the pip command
  4. Copying source code to the /opt folder
  5. Run the webserver using the flask command

Here’s a rundown of the actual demo:

  1. Create a Dockerfile and save it in your project directory as Dockerfile

The file should have the following attributes:

FROM ubuntu
RUN    apt-get update && apt-get -y install python python3-pip
RUN    pip3 install flask flask-mysql
COPY ./opt/source code
ENTRYPOINT FLASK_APP=/opt/source-code/app.py flask run

  • The FROM attribute specifies the base OS or image on which the application will run.
  • RUN command specifies dependencies that interact with the application.
  • COPY command copies files from the local folder to the registry.
  • While ENTRYPOINT will specify the command that will initiate the container once the image has been created.

2. Following the above, build this image using the command:

$ docker build -t account-name/my-custom-app .

3. Following the above, publish this image to the registry using the command:

$ docker push account-name/my-custom-app

You can check information on your image using the history command:

$ docker image history my-custom-app

HTTPD Image

In this demo section, you will build a custom httpd image in Docker and publish it in Docker Hub.

First, begin by creating a new project directory, using the commands:

$ mkdir firstimage

$ cd firstimage

Then, create a Dockerfile using the command:

$ vi Dockerfile

Then use the cat Dockerfile command to check that this file has been created.

You will then create this Dockerfile with the instructions and arguments as shown:

FROM centos:7
RUN         yum -y update
RUN         yum -y install httpd
COPY        ./index.html /var/www/index.html
EXPOSE      80
CMD         ["httpd","-D","FOREGROUND"]

Now create a file named index.html, which will be a simple web application built to display the words, Welcome to KodeKloud Again. You will create it using the command:

$ vi index.html

This will be written in html format, as shown:

<h1> Welcome to KodeKloud Again </h1>

Now, build an image using this file by running the command:

$ docker image build -t user-account/imagename:v1

To verify that the image has been created, run the docker image list command:

$ docker image ls

To check on this image’s specific layers, use the history command:

$ docker image history

To perform a quick test on the health and presence of this image, use the command:

$ docker container run -itd --name=testimage -p 82:80 
user-account/imagename:v1

You can now publish this image to Docker Hub using the command:

$ docker push user-account/imagename:v1

Tomcat Image

This section demonstrates how you can build and publish a Tomcat image. This is a practical Dockerfile with new instructions. To run a Tomcat image, you will need to first clone the Tomcat repository using the command:

$ git clone https://github.com/user-account/dockertomcat.git

Once you have successfully cloned the repository, create a Dockerfile using the cat command, with the properties as follows:

FROM        centos:7
ARG         tomcat version=8.5.6
RUN         yum -y epel-release java-1.8.0-openjdk.x86_64
RUN         groupadd tomcat && mkdir /opt/tomcat
RUN         useradd -s /bin/nologin -g tomcat -d/opt/tomcat tomcat
WORKDIR /
RUN         wget  https://archive.apache.org/dist/tomcat/tomcat-8/v$tomcat_version/bin/apache-tomcat-$tomcat_version.tar.gz
RUN         tar -zxvf apache-tomcat-$tomcat_version.tar.gz -C/opt/tomcat  --strip-components=1
RUN         cd/opt/tomcat && chgrp -R tomcat conf
RUN         chmod g+rwx /opt/tomcat/conf && chmod g+r /opt/tomcat/conf/*
RUN         chown -R tomcat /opt/tomcat/logs/      /opt/tomcat/temp/opt/tomcat/webapps/opt/tomcat/work
RUN         chgrp -R tomcat /opt/tomcat/bin && chgrp -R tomcat /opt/tomcat/lib && chgrp -R tomcat /opt/tomcat/lib && chmod g+rwx /opt/tomcat/bin && chmod g+r /opt/tomcat/bin/*
WORKDIR     /opt/tomcat/webapps
RUN         wget https://tomcat.apache.org/tomcat-7.0-doc/appdev/sample/sample.war
EXPOSE      8080
CMD         ["/opt/tomcat/bin/catalina.sh","run"]

You can now build an image on this Dockerfile using the command:

$ docker image build -t user-account/tomcatone:v1

Create a test container to check on this image by running the command:

$ docker container run -itd --name=imagetesting -p 84:8080  
user-account/tomcatone

You can now test this application by inputting its IP address on your browser. To deploy a newer version, use the command:

$ docker image build -t user-account/tomcatone:v2--build-arg 
tomcat_version=8.5.8

The Docker Commit Method

We can use the Docker Commit method to build images using existing containers. To do this, we’ll first create a container with the image httpd using the command:

$ docker run -d --name httpd httpd

We’ll now build a custom image version based on the one created in the previous step:

$ docker exec -it httpd bash

You can also commit a container as it is from an image using the command:

$ docker  container commit -a "author-name" httpd customhttpd

Creating Images using Docker Commit

In this demo lesson, let us build a custom image using the Docker commit method.

  1. First, you will pull an image for the Operating System using the command:
$ docker image pull centos:7

2. Then, create a container named test:

$ docker container create -it --name=test centos:7

3. To confirm that you have successfully created this container, run the command:

$ docker container ls -l 

4. To start working inside this container, initiate it with the command:

$ docker container start test

5. Update packages and dependencies using the commands:

$ docker container exec -it test /bin/bash
$ yum -y update

6. You can now customize your web page by typing in the HTML code as shown:

$ echo '<h1> "Hello from KodeKloud" </h1>'

To ascertain that you have properly configured your dependencies and packages, you can do so by running the Docker container list command.

7. To stop this test container and exit using the command:

$ docker container stop test

8. Let us now go through the various commit options in Docker Hub to familiarize yourself with the method. To commit a version to the current container, run the command:

$ docker container commit -a "author-name" -c 'CMD 
["httpd","-D","FOREGROUND"]' test webtest:v1

9. Create a container to test whether the new image version is working by running the command:

$ docker container run -itd --name=webtesting -p 80:80 webtest:v1

10. Test access of this webpage from your browser, If successful, it means the container is healthy and has retained your customizations. You can now easily tag this image and push it to Docker Hub. The following command can be used to tag your image:

$ docker image tag webtest:v1 author-name/image-name

11. To push this image to Docker Hub, you’ll need to log in to the Docker account using the command:

$ docker login

12. After providing your details, you can now push your image to the Docker Registry:

$ docker push author-name/image-name

Build Contexts

In Docker, a Build Context specifies a path to the directory that contains all the files needed to run your containers. Consider the application we tried to run earlier:

FROM ubuntu
RUN    apt-get update $$ apt-get -y install python python-pip3
RUN    pip3 install flask flask-mysql
COPY ./opt/source code
ENTRYPOINT FLASK_APP=/opt/source-code/app.py flask run

The Docker CLI stores the files in a folder named:opt/source code

The Docker Daemon that runs the containers usually creates a similar directory into which it transfers the files temporarily – /var/lib/docker/tmp/docker-builderxxxxx

Special Note: You should only store the files you need to run your application in the build context, as any more would slow the build process. You can also create the .dockerignore  file to tell Docker Daemon which files to ignore during the build process.

You can also specify the URL to a repository to build an image from a remote source. This clones the repository within your Docker environment then builds images from it. The command for this would be:

$ docker build https://github.com/my-account/my-app

To specify the branches that your daemon will check out while building the image using the extension:

$ docker build https://github.com/my-account/my-app#<branch>

You can also point the builder to a specific folder within the repository if required:

$ docker build https://github.com/my-account/my-app:<folder>

Often you may need to create a separate file with instructions on how to build the image. In such cases you build a Dockerfile together with the application image, as shown:

$ docker build -f Dockerfile.dev https://github.com/my-account/my-app

Build Cache

When creating an image, each instruction creates a new layer, which then adds some storage to the image based on the output of instructions. Docker builds the layers progressively from first to last, and each layer is cached so that if one fails, Docker will not have to rebuild them from scratch.

When rebuilding an application, Docker uses various criteria to determine which layers can be rebuilt from the cache, and which ones have to be rebuilt from scratch. It first checks instructions in the Dockerfile. If an instruction has changed, the cache from that and the remaining layers are deemed invalid. Docker also compares the checksums of files specified in the COPY and ADD instructions, and if there is a difference, these caches are deemed invalid.

One cache best practice is known as Cache Busting, where we combine the update and install arguments into one, forcing an update before any installation. We keep the elements in different lines, separated by a slash, and in alphabetic order to make the files easier to locate. Version Pinning is another best practice where we point the build to the latest version of the image by specifying it in the build command. The final best practice for Build Cache is listing the most frequently modified instructions at the bottom of the Dockerfile. This means fewer caches are invalidated during frequent rebuilds.

COPY vs ADD

  • The COPY and ADD commands have sometimes been used interchangeably since they both transfer files from local repositories to the image’s file system, but they have fundamental differences. 
  • The  ADD command can access a local .tar file, extract its files and build an image out of it. 
  • As a best practice, always use the COPY instruction whenever possible. This instruction easily copies files locally and is straightforward so anyone can understand the file paths. All arguments in the COPY instruction can be run on a single line, reducing the number of layers in your image, making it small and lean enough.
  • As an illustration, here are two code snippets of images created with each command:For the COPY command:
FROM centos:7
RUN         curl http://app.tar.xz\|tar -CJ/testdir/file.tar.xz\&&yarn build \ && rm /testdir/file.tar.xz
FROM centos:7
ADD          http://app.tar.xz/testdir
RUN          tar-xJF/testdir/app.tar.xz -C /tmp/app
RUN          make -C /tmp/app

The snippet above shows the Dockerfile configurations for the ADD command.

CMD vs ENTRYPOINT

Containers are ephemeral, meaning they only stay active when they are running a process. When you run the command:

$ docker run ubuntu

With this, the container created is exited immediately. This is because containers cannot host Operating Systems. Instead, they are designed to run specific tasks including application servers, databases, and web servers. Every Dockerfile used to build an image typically comes with a CMD instruction in the end. This command usually decides the process that runs on a container when it is initialized. A quick point to note here is that Ubuntu uses bash as its running command. Bash is not a process, rather it is a shell that listens to requests from a terminal. With no terminal running in the environment, this container always exits immediately.

When you run the command:

$ docker ps

You won’t see the Ubuntu container as it is not listed among the running containers. Instead, use the command:

$ docker ps -a

The above command will show you all containers, including the Ubuntu file which was just created and exited. You can keep this container running for some time by appending the run command to specify the time it takes before executing the bash command, as shown in the code:

$ docker run ubuntu sleep 5

In this case, the Ubuntu image will initiate, run for 5 seconds, then execute the bash command, at which point it will exit the container.

We can make this change permanent by enforcing the command in a Dockerfile as shown:

FROM ubuntu
CMD  sleep 5

You can now build a newer version of the application by running the command:

$ docker build -t ubuntu-sleeper

Initiate this application by running the command:

$ docker run ubuntu-sleeper

When you launch this application, the command at startup will be sleep 5

To change the waiting time, you could enforce it in the command line as follows:

$ docker run ubuntu-sleeper sleep 10

When you launch this application, the command at startup will be sleep 10

An ENTRYPOINT helps specify how the program runs when the container starts so you won’t have to specify the sleep command every time by changing this number. This means that all changes to be appended in the command line are written in the container.

FROM ubuntu
ENTRYPOINT ["sleep"]
CMD ["5"]

To overwrite an existing ENTRYPOINT, adjust it within the command line as follows:

$ docker run --entrypoint sleep2.0 ubuntu-sleeper 10

The command at startup will be sleep 2.0 10

Base vs Parent Image

There are various types of images and dependencies in Docker. A Base Image is one that has been built from scratch. Examples of base images are Ubuntu and debian:buster-slim.

A parent image lends its dependencies to other images, and these include httpd and custom-php. 
Scratch is the top-most level of image abstraction, and most scratch images are available of the Docker Registry. For instance, the Dockerfile for the debian:buster-slim is:

FROM scratch
ADD  rootfs.tar.xz/
CMD [“bash]

Multi-Stage Builds

When creating an application in your development server, you will always work with multiple files and dependencies. Considering a Node.js application, developing a containerized version typically involves two steps:

  1. Building the application using the command:
$ npm run build

We then containerize this application with the command:

FROM nginx
COPY dist/usr/share/nginx/html
CMD [“nginx”,”-g”,”daemon off;”]

2. Then use the build command to run this container:

$ docker build -t my-app

In most production environments, you will need to create multiple versions of applications that share dependencies, and you’ll need to ensure that all development servers are well matched. 

In this case, we use a builder Dockerfile to guide the build process, using the specifications below:

$ dockerfile.builder

FROM node
COPY . .
RUN  npm install

RUN npm run build

Now build this Dockerfile by using the command:

$ docker build -t builder

This gives you a clean, consistent build. You can then copy the distribution file path from the builder using the format:

$ copy-dist-from-builer.sh

$ docker container create --name builder builder
$ docker container cp builder: dist./dist
$ docker container -f builder

3. This is a third stage in the container building process that helps extract distribution files. You can create a single, multi-stage build by combining the building, extraction, and containerization process into a single Dockerfile. A multi-stage build contains more than one FROM instruction, and dependencies are shared by adding a pointer to the COPY instruction as shown.

FROM node As builder
COPY . .
RUN  npm install
RUN  npm run build

FROM nginx
COPY --from=builder dist/usr/share/nginx/html
CMD [“nginx”,”-g”,”daemon off;”]

The As builder argument creates and names the application as builder. The COPY command on the second part then instructs the Docker Daemon to get information from this application using the argument --from=builder.

Multi-stage builds help us:

  • Optimize Dockerfiles by making them easy to read and maintain.
  • Keep image size small
  • Avoid having to keep different container files for development and production
  • Eliminate the need for intermediate images

Dockerfile – Best Practices

Here are a few tips to make the most of your Docker images:

  1. Do not build images that run multiple containers. Each image should be running a specific task. These images can then communicate with each other and form a single application that runs multiple services. This makes it easy to scale each component independently.
  2. Do not store data or state in containers since containers are ephemeral. Always use external volumes or a caching service (e.g Redis) to store persistent information.
  3. Try to keep your images as lean and minimal as possible. You can find existing minimal images for Operating Systems, Databases, and other services. Only install the packages necessary to run your application. 
  4. Maintain different images for different environments. For instance, images running in the development environment may require tools like monitoring and debugging. Keep these out of your production environment.
  5. Use multi-stage builds so you won’t need to create many images.
  6. Do not send unnecessary files to the build context.

These building patterns are helpful for anyone looking to build applications in Docker. It is also advised that you keep yourself updated with Docker documentation and blogs on recent developments to improve Docker image builds, and integrate these into your practice as you go. As you keep practicing with Docker builds, you will also discover workflows and patterns to make your applications lean and minimal.

Research Questions & Conclusion

This concludes the Image Creation, Management and Registry chapter of the DCA certification exam. To test your knowledge, it is strongly recommended that you access research questions of all core concepts covered in the coursework and a test to prepare you for the exam. You can also send your feedback to the course developers, whether you have feedback or would like something changed within the course. 

Here is a quick quiz to help you assess your knowledge. Leave your answers in the comments below and tag us back

Quick Tip – Questions below may include a mix of DOMC and MCQ types.

1. What is a recommended best practice for installing packages and libraries using the apt-get package manager while building an image?

[A] Download packages on the host and use ADD instructions to add them to the image.

[B] Use the ADD instruction to provide a URL to the package on the remote host.

[C] Use the RUN instruction and have the apt-get update and apt-get install commands on the same instruction

[D] Use the RUN instruction and have the apt-get update and apt-get install commands as separate instructions

2. An application you are developing requires an httpd server as frontend, a python application as the backend API server, a MongoDB database and a worker developed in Python. What is the recommended approach in building images for these containers

[A] Build httpd, python API server, MongoDB database and Python worker into a single image to allow ease of deployment

[B] Build httpd into an image, MongoDB database to another and Python API and worker together into a single image 

[C] Build separate images for each component of the application

3. Which of the following commands may be used to list all images matching the com.example.version label?

[A] docker images --label=”com.example.version”

[B] docker images --filter "com.example.version"

[C] docker images --filter "label=com.example.version"

[D] docker images --format "label=com.example.version"

4. What is the command to change the tag of httpd:latest to httpd:v1

[A] docker container image retag httpd:latest httpd:v1

[B] docker container image tag httpd:latest httpd:v1

[C] docker image retag httpd:latest httpd:v1

[D] docker image tag httpd:latest httpd:v1

5. When you log in to a registry using the docker login command, the credentials are stored locally at…

[A] $HOME/.docker/config.json

[B] /etc/docker/.docker/config.json

[C] /var/lib/docker/.docker/config.json

[D] /var/lib/docker/containers/.docker/config.json

6. Which of the following is the correct docker image address to be used to access an image named payapp hosted under the organization payroll at a private registry registry.company.io?

[A] registry.company.io/payapp/payroll

[B] payroll/registry.company.io/payapp

[C] payapp/registry.company.io/payroll

[D] registry.company.io/payroll/payapp

7. What is the command used to find images with a name containing postgres, with at least 12 stars?

[A] docker find  --filter=stars=12 postgres

[B] docker search --filter=stars=12 postgres

[C] docker find  --limit=12 postgres

[D] docker search --limit=12 postgres

8. After building the below code with an image named webapp, What will happen when you run docker run webapp sleep 1000?

FROM ubuntu:18.04
COPY . /app
RUN make /app
CMD python /app/app.py

  [A] docker overrides the ENTRYPOINT instruction with sleep 1000

  [B] docker overrides the CMD instruction with sleep 1000

  [C] docker override ENTRYPOINT instruction with sleep and CMD instruction with 1000

9. Choose the correct instruction to add the command instruction with the command `echo “Hello World”` in the Dockerfile

[A] CMD [echo "Hello World"]

[B] CMD ["echo", "Hello World"]

[C] CMD ["Hello World"]

[D] CMD Hello World

10.  What is the right instruction to download a file from https://file.tar.xz and auto-extract it into /testdir in the image

[A] ADD https://file.tar.xz  /testdir

[B] COPY https://file.tar.xz  /testdir

[C] RUN https://file.tar.xz  /testdir

[D] Volume https://file.tar.xz  /testdir

11. Which of the below can help minimize the image size?

[A] Only install necessary packages within the image

[B] Avoid sending unwanted files to the build context using .dockerignore

[C] Combine multiple dependent instructions into a single instruction and cleanup temporary files 

[D] Move the instructions that are likely to change most frequently to the bottom of the Dockerfile

[E] Use multi-stage builds

12. Which of the following statements best describe Docker Trusted Registry (DTR)?

[A] Docker Trusted Registry (DTR) is Mirantis’s enterprise-grade image storage solution. 

[B] Installed behind the firewall only on-premises 

[C] DTR provides a secure environment from which users can store and manage Docker images.

[D] Using DTR in Docker EE we can control who can access and make changes to your cluster and applications

13.  A government facility runs a secure data center with no internet connectivity. A new application requires access to docker images hosted on docker hub. What is the best approach to solve this?

[A] Get the Dockerfile of the image and build a local version from within the restricted environment.

[B] Establish a secure link between the host in the restricted environment and docker hub

[C] Pull docker images from a host with access to docker hub, convert to a tarball using docker image save command, and copy to the restricted environment and extract the tarball

[D] Pull docker images from a host with access to docker hub, then push to a registry hosted within the restricted environment.

14. Which of the following is the correct format for CMD instruction?

[A] CMD [“executable”,”param1″,”param2″]

[B] CMD [“param1″,”param2”]

[C] CMD command param1 param2

[D] CMD param1 param2

15. Print the value of ‘Architecture’ and ‘Os’ of an image named `webapp`

[A] `docker image inspect webapp -f ‘{{.Os}}’ -f ‘{{.Architecture}}’`

[B] `docker image inspect webapp -f ‘{{.Os}} {{.Architecture}}’`

[C] `docker image inspect webapp -f ‘{{.Os}}’, -f ‘{{.Architecture}}’`

[D] `docker image inspect webapp -f ‘{{.Os .Architecture}}’`

16. While building a docker image from code stored in a remote URL, which command will be used to build from a directory called `docker` in the branch `dev`?

[A] docker build https://github.com/kk/dca.git#dev:docker

[B] docker build https://github.com/kk/dca.git#docker:dev

[C] docker build https://github.com/kk/dca.git:dev

[D] docker build https://github.com/kk/dca.gitdev:#docker

17. What are the features of docker trusted registry (DTR)?

[A] Built-in Access Control

[B] Image and Job Management

[C] Security Scanning

[D] Auto scaling applications

[E] Image Signing

18. Docker Trusted Registry (DTR) is a containerized application that runs on a Docker Universal Control Plane cluster.

[A] True 

[B] False

19. What is the command to push the image to a docker private registry?

[A] docker push <docker-registry-address>/username/docke-repo-name

[B] docker push <docker-registry-address>/docker-repo-name

[C] docker upload <docker-registry-address>/username/docker-repo-name

[D] docker upload <docker-registry-address>/docker-repo-name

20. Inspect the following output and answer the following questions:

REPOSITORY          TAG                 IMAGE ID            CREATED             SIZE
nginx               latest              c39a868aad02        4 days ago          150MB
redis               latest              4760dc956b2d        2 years ago         100MB
ubuntu              trusty              f975c5035748        2 years ago         100MB
webapp              latest              c39a868aad02        4 days ago          150MB
alpine              latest              3fd9065eaf02        2 years ago           5MB

1. What command might have generated the above output?

[A] docker container ps

[B] docker ps

[C] docker image ps

[D] docker image list

2. What is the total space consumed by images on this system?

[A] 355 MB

[B] 505 MB

[C] 405 MB

[D] 455 MB

By properly following this study guide till this part of the series, you have prepared yourself to handle all Image Creation, Management and Registry tasks of Docker – and are of course a step closer to pass the DCA certification test.

On KodeKloud, you also get a learning path with recommendations, sample questions and tips for clearing the DCA exam. Once you are done with this section, you can head on over to the research questions and practice test sections to examine your understanding of Docker Image Creation, Management and Registry.

Related Articles

Docker Certified Associate Exam Series (Part-3): Kubernetes

Introduction to Kubernetes Kubernetes is the most popular container orchestration tool that integrates seamlessly with Docker. The DCA certification test emphasizes specifically on services, architecture, and commands used in Kubernetes, so it is crucial to develop a firm grasp of the framework. The Kubernetes section of the DCA exam preparation tutorials combines the Kubernetes for…

Docker Certified Associate Exam Series (Part 6): Networking

Introduction to Docker Networking Default Networks When you install Docker on a host machine, it automatically creates 3 types of networks: none, bridge, and host networks. The Bridge network is the default network that Docker attaches to containers once they are created. This is a private internal network, whose IP address is typically 172.17.0.1, and…

Responses

Your email address will not be published. Required fields are marked *

Secured By miniOrange