Changes

From Amahi Wiki
Jump to: navigation, search
13,927 bytes added ,  22:16, 27 August 2017
no edit summary
Containers!
==Overview==
Amahi supports one click install for different kind of apps. To add any app to amahi they have to be packaged properly. Each app might require a completely different environment to run and it might not be possible to provide that environment to every app because their can be dependnecy conflicts. For example some app might require php5 and some might require php7. We need a mechanism to address this concern such that we can add any app to Amahi irrespective of the technology stack that it might require or without version conflicts. Using containers is one way of addressing this problem.
 
==Important Terms==
Before going further in this document, understanding the following terms might be useful:
 
* [https://www.docker.com/what-container What is a Container?]
* [http://blog.codesupport.info/docker-images-vs-containers/ Image vs Container]
* amahi.net -> Domain name of the amahi hda for all examples here
 
==Architecture Overview==
 
// TODO - Add image here
 
The idea of implementation is very similar to what is showing in the image above. Each app will run as a container. Each container will expose their own port/ports as can be seen as `exposed port` in the diagram above. We can map a host system port (`mapped port` in the diagram) to the `exposed port` in the container. And then using reverse proxy we can connect different subdomains to different container apps.
 
This has been explained in detail in the Design Document.
 
==How the app installation works?==
 
This section presents an overview of how installation works.
 
Assuming we have official container for an app available we can easily integrate them to amahi. If the official image is not available then we might have to build one of our own like I did for osticket and coppermine.
 
===Building Images===
Building images can be tricky and does require some knowledge of the apps as well (For example which php libraries to install, etc). There's a well defined procedure for building images for node and rails apps as well.
* [Dockerizing Node Apps](https://nodejs.org/en/docs/guides/nodejs-docker-webapp/)
* [Dockerizing Rails Apps](https://semaphoreci.com/community/tutorials/dockerizing-a-ruby-on-rails-application)
 
Building images can be tricky and the image size is a very major issue. To reduce the image size I would suggest the readers to look up the following articles:
* [Reduce Docker Image Size](https://blog.codeship.com/reduce-docker-image-size/)
* [Alpine Docker](https://hub.docker.com/_/alpine/)
 
===Install Script===
Once the image is available we can write an install script. You can see below a sample install script for gitlab.
 
```sh
cat > docker-compose.yml << 'EOF'
gitlab-container:
image: 'gitlab/gitlab-ce:latest'
container_name: "APP_IDENTIFIER"
restart: unless-stopped
hostname: 'APP_HOSTNAME'
environment:
GITLAB_OMNIBUS_CONFIG: |
external_url 'http://APP_HOSTNAME:HOST_PORT'
gitlab_rails['gitlab_shell_ssh_port'] = 2224
ports:
- 'HOST_PORT:HOST_PORT'
- '2224:22'
volumes:
- './srv/gitlab/config:/etc/gitlab'
- './srv/gitlab/logs:/var/log/gitlab'
- './srv/gitlab/data:/var/opt/gitlab'
EOF
docker-compose up -d
```
This script creates a docker-compose.yml file and then runs `docker-compose up -d` command which essentially creates and runs the container.
 
The `yml` file can have different parameters. For that we might have to refer to docker and docker-compose documentation. `restart : unless-stopped` is used to handle failovers of containers. If a container crashes for some reaosn then it will restart automatically.
 
===Understanding the script===
Container script for each app will have these parameters for sure:
```
container-name: "APP_IDENTIFIER"
restart : unless-stopped
ports:
- 'HOST_PORT:xyz' # Not required for apps which are not webapp
 
# xyz = any port inside the container
# HOST_PORT is port on the host machine. Two different containers cannot have same value for
# HOST_PORT but can have the same value for xyz (Just to clear up the confusion about
# HOST_PORT)
```
 
Most of the configuration related to the container will be written in the docker-compose.yml file but some data we have to extract during run time like the HOST_PORT using which will reverse proxy to the container. HOST_PORT is 35000+app_id and has to be derived during app installation. Similarly APP_IDENTIFIER and APP_HOSTNAME are derived during runtime. The code below is used to put that data into install script during runtime.
 
```rb
#app/models/app.rb Line 322 to 326
install_script = installer.install_script
install_script = install_script.gsub(/HOST_PORT/, (BASE_PORT+self.id).to_s)
install_script = install_script.gsub(/WEBAPP_PATH/, webapp_path)
install_script = install_script.gsub(/APP_IDENTIFIER/, identifier)
install_script = install_script.gsub(/APP_HOSTNAME/, app_host)
```
 
Sample uninstall script
```sh
docker-compose stop
docker-compose rm -f
# Not removing the image. Just stopping the container.
```
 
For most containers the above uninstallation script will work fine. This stops the running container and removes it. Please note that this doesn't delete any of the volumes attached (persistent storage. Please refer to docker documentatio for more details regarding volumes) with the container so if you add a volume during installation (as we have done in the gitlab example above) then we have to remove them here during uninstallation. For example if we were to remove gitlab completely along with all files that were added by gitlab container then the uninstall script would look something like this:
 
```sh
docker-compose stop
docker-compose rm -f
rm -rf srv # Removing the srv folder which holds the persistent files for gitlab container
```
 
This behaviour might not be intended for all applications. Right now I haven't removed static files for any apps that I have added.
 
===Reverse Proxy===
====Why is it needed?====
Each app will be running insidea container and they will attach to some port/ports on the host machine. For example, imagine if you are running an app like gitlab in container. Gitlab might require different ports for different things. Inside the container different services can run on any port as containers run in an isolated environment and have their own network stack. But ultimately the services running on these containers have to be reachable from the outside world. To make that possible we have to bind a port inside the container with a host port.
 
Gitlab App - Inside Container - Web Server Running on port 80
We have to bind this with a port on host machine (our HDA). Let's assume we bind it to port 35001. Now typing amahi.net:35001/ in the browser will open gitlab website.
 
OSticket app - Inside Container - Web Server Running on port 9000
Let's bind this to port 35002 on host machine. Now typing amahi.net:35002/ in the browser will open osticket website.
 
As we can see that each of these apps web interface is accessible on different ports but we want to achieve a behaviour like:
 
osticket.amahi.net -> opens osticket which is running on amahi.net:35002/
gitlab.amahi.net -> opens gitlab which is running on amahi.net:35001/
 
But the url end point should be the same, i.e osticket.amahi.net and gitlab.amahi.net . To attain this behaviour we have to use Reverse Proxy.
 
 
For reverse proxy I have added a new app-container.conf file which can be seen below. The `APP_PORT` is changed during runtime.
```sh
<VirtualHost *:80>
 
ServerName HDA_APP_NAME
ServerAlias HDA_APP_NAME.HDA_DOMAIN
 
APP_ALIASES
 
APP_CUSTOM_OPTIONS
 
ProxyPreserveHost On
 
 
ProxyPass / http://localhost:APP_PORT/
ProxyPassReverse / http://localhost:APP_PORT/
 
ErrorLog APP_ROOT_DIR/logs/error_log
CustomLog APP_ROOT_DIR/logs/access_log combined env=!dontlog
</VirtualHost>
 
```
APP_PORT part is derived from the app id. After installation the app will have some id in the database base. The APP_PORT will be 35000+app_id
 
Please note that we can run any kind of app in container. It might be a headless app and it might be a webapp. In case of web applications we need to define an external port (mapped port on host - Refer to Architecture Overview section) through which the app will be bind. Then to reach that app we have to reverse proxy. "APP_PORT" is essentially that. For apps which don't require a web interface we might not use this file at all.
 
 
## How to add a new app?
Taking example of [Hydra](https://hub.docker.com/r/linuxserver/hydra/)
 
See the usage as mentioned by the maintainer:
```sh
docker create --name=hydra \
-v <path to data>:/config \
-v <nzb download>:/downloads \
-e PGID=<gid> -e PUID=<uid> \
-e TZ=<timezone> \
-p 5075:5075 linuxserver/hydra
```
 
Convert the above to a docker-compose file. Ignore the `-e PGID=<gid> -e PUID=<uid>`, even though it's relevant, it is out of the scope of this discussion.
 
```
hydra-container:
image: 'docker.io/linuxserver/hydra'
container_name: "hydra"
restart: unless-stopped
ports:
- '5075:5075'
volumes:
- './config:/config'
- './downloads:/downloads'
- '/etc/localtime:/etc/localtime:ro'
 
 
# Understanding the volume mounts:
# ./config:/config -> As seen in the docker create commnad the -v command mentions the volumes.
# Path to data that we are providing is a relative path. Every installed app has a path in which the
# install script runs. So "config" and "downloads" folder will be created there in that path.
 
# /etc/localtime:/etc/localtime:ro -> This is to make sure that the container uses the same time as used
# by the host system. To avoid this mount we can also use
# environment:
# - TZ=<timezone>
# in the docker compose file
```
 
NOTE: Please note that adding apps might require knowledge about docker and docker-compose and discussing those is out of the scope of this documentation though the links mentioned below might be useful.
* [Container and Layers](https://docs.docker.com/engine/userguide/storagedriver/imagesandcontainers/#container-and-layers)
* [Docker Compose](https://docs.docker.com/compose/overview/)
* [Restart Policies](https://docs.docker.com/engine/admin/start-containers-automatically/). Apps run as containers which are managed by docker. If docker daemon is shut down or stopped then the app will also stop. If the container crashes for some reason then it has to be restarted. Using restart policies we can manage this.
 
Now once we are done with making a docker-compose file we can test it on our local system to see if it is working properly or not. Once that's done, we can go ahead and add this image to amahi.org
 
For adding to amahi.org some modifications have to be done. The final changes can be seen below. Notice the `APP_IDENTIFIER` and `HOST_PORT` (For more info on this refer to "Understandin the script" section)
```
hydra-container:
image: 'docker.io/linuxserver/hydra'
container_name: "APP_IDENTIFIER"
restart: unless-stopped
ports:
- 'HOST_PORT:5075'
volumes:
- './config:/config'
- './downloads:/downloads'
- '/etc/localtime:/etc/localtime:ro'
```
 
The final install and uninstall scripts to be added on amahi.org will be
 
Install Script
```sh
cat > docker-compose.yml << 'EOF'
hydra-container:
image: 'docker.io/linuxserver/hydra'
container_name: "APP_IDENTIFIER"
restart: unless-stopped
ports:
- 'HOST_PORT:5075'
volumes:
- './config:/config'
- './downloads:/downloads'
- '/etc/localtime:/etc/localtime:ro'
EOF
docker-compose up -d
```
 
Uninstall Script
```sh
docker-compose stop
docker-comose rm -f
rm -rf config # Use this if you want all files to be removed after uninstall
rm -rf downloads # Use this if you want all files to be removed after uninstall
```
## NOTE
 
You should be careful about the following:
 
## Future Work
 
### Cleanup - too many images.
Presently when a containerised app is uninstalled the container is stopped and removed but the image used to run the container stays (Note the difference between container and image). I am not removing the image since images are generally large and what if users decides to reinstall the app? This has to be thought of more. Adding the feature of deleting image after installation is essentially 1 line of code. Going forward we might have to come up with a mechanism for clearing images which haven't been used for a long time.
 
### Reducing Download Size
We can't build and maintain containers for all apps. Vendors provide official containers but some of them are huge in size. To fix this problem we can reduce the download to that of a normal installation by using an apt-cache, gem server or npm server on hda itself.
 
The idea is following: Instead of we building images on our server we can push the Dockerfile to the client and the client can build the docker image. While building the image they will download the required packages, gem files, node dependencies or whatever. If we have a mechanism to cache this download so that all the subsequent builds can use this data then we can save a lot of Internet usage. One possible way of doing this was running apt-cache, gem server and npm server on the client itself.
 
### Updating apps
With containers updates can be really easy. We can support single click update of apps. It's just a matter of supporting them now.
 
### Support for advance configuration.
With containers we can limit the cpu/memory/disk usage of each app. We just have to modify the docker-compose.yml file for the app for these changes without doing any modification to the source code. If required, this can be used.
 
### Collecting Logs
It's a new feature, we will need to collect a lot of metrics from the users to understand how this feature is working and how it can be improved. Some of those features include
* The CPU info of systems which are running amahi.
* RAM and Storage information.
* Logs of containerised apps to debug errors.
15

edits