What is a container ?

A container is just a way to achieve process isolation. Unlike virtual machines, they don’t achieve isolation by simulating hardware, but by using existing Linux kernel features.

In a typical Unix/Linux OS all processes share the same user space, but with the introduction of new features in Linux 2.6+, you can create a process that has its own particular set of isolated contexts like file tree, threads, etc. These features in combination with other kernel technologies are the magic behind containers.

In this article I will introduce basic Docker commands and concepts. After you finish reading, you will be able to adopt some Docker capabilities to accelerate and simplify your day-to-day workflow.

Installing Docker

Installation of Docker is a simple task in OS X/Windows using the installation wizard You can find the installer for your OS on the Docker community page, On Linux, Docker is usually available in the distribution package manager.

Installing Docker in Fedora:

sudo dnf install docker

To start the process:

sudo systemctl start docker

To make the Docker process start and boot time:

sudo systemctl enable docker

The steps should be similar in others Linux distributions that use Systemd.

Getting Started

Hello World

Once you finished the installation we should try a small Hello World! for containers:

#sudo is only for Linux
sudo docker run --name hello -it busybox echo "Hello World!" # Hello World!

Using sudo is only necessary if you are running some Linux distributions, but keep in mind that Docker require admin rights to create the containment, in OSX and Windows at the moment of writing this article they use some Linux based virtual machine behind the scene, so Docker command can be run without privilege user in those systems.

How it works

sudo docker run <options> <image> <command to run inside the container>

The run option creates and run a container, one of the properties is that Docker binds the life of the container to the running process (in this case the Linux command echo), which means that when process finish the container will be terminated.

  • name we set the name of the container, if you don't choose anything Docker will choose one at random.

  • it this mean interactive, it will connect our terminal to the output of the container virtual TTY, allowing interact with the running process.

  • busybox This the base image to create the container, think of it as zip file with the files and folders necessary to run the desired application. There is a whole community base images available in Docker Hub, I use busybox because is very light just 715 KB compressed.

echo, As we mentioned earlier, is the command we try to execute that is included inside busybox image.

If you want check the commands available in busybox just do this:

sudo docker run -it busybox ls /bin/

Listing Cached Images

When you execute the Docker command for the first time the image is downloaded and is cached to speed up things. You can check local images by using this command:

sudo docker images

Running in Daemon mode

In some cases we don't want to interact directly with some applications like servers, in this case we want to spawn the process and get back our terminal to continue doing some work, Docker provide us a way to execute the process in daemon mode using the -d parameter like this:

sudo docker run -d --name snooze busybox sleep 15

This process will run for 15 seconds in the background and then exit.

Listing background running containers

Once the container is running in the background, you can check its status with ps:

sudo docker ps

Killing running containers

To stop a container is simple just:

sudo docker stop [name of your container]

#example
sudo docker stop snooze

This command will stop the running container, but the docker service will keep the container you created including its associated command cached in disk, in the case you want to replay the exact command again just need to do docker start and pass the container name.

sudo docker start snooze

If you want to change the configuration and reuse the container name then you need to stop and delete the container, let say we want to change our snooze container, to sleep for 10 seconds instead of 15:

# stop & clean
sudo docker stop snooze
sudo docker rm snooze

# re-create
sudo docker run -d --name snooze busybox sleep 10

Mounting

The v parameter will allow us to mount/map a folder from the host (our computer) to a folder inside the container.

Example:

Let's create an file

echo 'Hello World' > hello

Now we want to open the file using an isolated text editor available in busybox:

# Opening the file with our contained VIM
sudo docker run -it busybox vi hello

Nothing happens, this is because the vi process we are calling is isolated and is unable to access the file outside the contained area, to solve this, we need to mount the folder so our editor is able to find the file.

#the :z in /app:z -> is for SELinux, non-Linux can ignore this.
sudo docker run -it -v "$(pwd)":/app:z busybox vi app/text

This will mount the actual folder $(pwd), to the folder /app inside the container, if the folder doesn't exist inside it will be created, then, if we use vi and pass the location of file of the mounted folder vi app/hello.

Some observations:

  • The v will overwrite any previous folder inside the container, if exist it will be replaced with the provided folder.
  • This command literally mount the folder, so every change made by the container to this folder will be persisted once the container has been killed, this can be a good idea if you want a DB to persist its data beyond container lifecycle.
  • The container will have access to your system resource (the shared folder) so be careful.

Networking

Option p allow us to expose an isolated port and pass it through a specific Host port.

To illustrate how networking works with containers, first let start by writing a simple Javascript script to startup a server, we are going to do this in our local machine, so let's write some code.

require('http')
  .createServer((req, res) => { res.end('Hello World!') }).listen(8080)

We will call this file index.js it basically create a server that waits for connection in port 8080, when somebody connects it will send a Hello World!.

Next step is to run our script inside a container, we can do this by writing the following command:

sudo docker run -it -v "$(pwd)":/app:z --name myserver mhart/alpine-node node app/index.js

The new stuff here is the base image mhart/alpine-node it will pull a Node.JS container, then will just mount using -v the folder as we did before and then execute the isolated node app/index.js process.

Let see if our server is working:

sudo docker exec -it myserver wget -qO- localhost:8080
# Hello World%

This command tests that our server is working from within the container and we should get back Hello World%, Now let's try to connect from our Host, open a new terminal, and write:

curl http://localhost:8080
#curl: (7) Failed to connect to localhost port 8080: Connection refused

We are unable to connect because the container network is isolated, we need to forward the port:

# stopping our container
sudo docker stop myserver  
sudo docker rm myserver

sudo docker run -it -v "$(pwd)":/app:z -p 8080:8080 --name myserver \
 mhart/alpine-node node app/index.js

Now try to open http://locahost:8080 in your browser and you should see a hello world!.

Congrats!, you have wroted a nice contained NodeJS application. One of the greatest advantage is that you can do this without installing NodeJS and you can use this capability to install other kind of software like DB's, other microservices, etc.

Some quick tips

In my day to day I always need to integrate with MongoDB and Redis, but installing those are usually a painful process, I have solved this problem by creating some bash scripts in my .zshrc.

function new_mongo {
  docker run -d --name mongodb -p  27017:27017 mongo
}

# the : here means image tag, usually if the image is done correctly
# like in this case tag version match the Redis version
function new_redis {
  docker run -d  --name redis  -p 6379:6379 redis:3.2
}

function stop_mongo {
  docker stop mongodb
  docker rm mongodb
}

function stop_redis {
  docker stop redis
  docker rm redis
}

After adding this line to the bottom of your .bashrc || .zshrc , then just execute source ~/.bashrc || source ~/.zshrc, then you should be able to do this.

new_mongo # it will spin up a new mongodb instance.
new_redis # it will spin up a new redis instance.

# to stop this containers
stop_mongo
stop_redis

Now you will be able to deploy a local MongoDB or Redis instance on-demand with zero configuration, and one advantage (at least in my view) is that the data in these instances are ephemeral, meaning that when you terminate the container it will reset the database as well, releasing the occupied space.

Linux security

If you execute to mount a folder by using v parameter in Fedora you may get this error:

sudo docker run -it -v "$(pwd)":/app busybox ls app/text
#ls: can't open '.': Permission denied

This is because SELinux default policy will protect any read/write in the Host, in case an attacker can get out of the container, SELinux will hold your back by enforcing security rules at Kernel level.

To mount the folder in a SELinux aware machine you need to pass the z parameter, this will change the SELinux context and will allow the container to perform the mounting.

# "$(pwd)" will get the actual directory, is equivalent to do pwd
docker run -it -v "$(pwd)":/app:z busybox /bin/sh

Other way (but not recommended) to do this is to temporarily disable this protection with:

su -c "setenforce 0"

To enable it again just do:

su -c "setenforce 1"