You build and start MIM+db with two commands:
sudo ./start_miquido_mim.sh build <path_to_mim_root>
sudo ./start_miquido_mim.sh run <path_to_mim_root>
It loads ejabberd.cfg from ./config/ejabberd.cfg and uses *.pem
files form there.
MongooseIM is Erlang Solutions' robust and efficient XMPP server aimed at large installations. Specifically designed for enterprise purposes, it is fault-tolerant, can utilize resources of multiple clustered machines and easily scale in need of more capacity (by just adding a box/VM). Its home at GitHub is http://github.com/esl/MongooseIM.
This work, though it's not reflected in git history, was bootstrapped from Paweł Pikuła's (@ppikula) great https://github.com/ppikula/MongooseIM-docker/.
If you need vanila MongooseIM as found on https://github.com/esl/MongooseIM please use docker images from https://hub.docker.com/r/mongooseim/mongooseim/
If customised images are needed, following documentation may be useful.
In order to build MongooseIM tarball a builder image and container need to be created. You can create an image by running the following command:
docker build -f Dockerfile.builder -t mongooseim-builder .
After that you can run the builder container.
It's important to mount the container's /builds
directory as a volume because MongooseIM tarball will be placed there after the build.
For simplicity it's assumed that env var VOLUMES
is exported and set to an existing absolute path, f.e: pwd
docker run -d --name mongooseim-builder -h mongooseim-builder \
-v ${VOLUMES}/builds:/builds mongooseim/mongooseim-builder
You can modify which Erlang/OTP version is used by MongooseIM when creating a builder image by providing OTP_VSN
build argument:
docker build --build-arg OTP_VSN=19.3.6 -f Dockerfile.builder -t mongooseim-builder:otp19.3.6 .
By default the builder will use Erlang/OTP 20.3.
Now building MongooseIM tarball is as simple as runing following command:
docker exec -i mongooseim-builder /build.sh
This command will by default build MongooseIM's master branch from: https://github.com/esl/MongooseIM.
This can be changed by specificing parameter to the build.sh
command:
/build.sh project_name repo commit
-
project_name
- friendly name for the build -
commit
- commit or branch or tag - what to checkout? -
repo
- where to checkout from
In order to build a specific commit, following command can be used:
docker exec -i mongooseim-builder /build.sh MongooseIM https://github.com/esl/MongooseIM a37c196
A log file of the build is available at /builds/build.log
,
so it's accessible from the host system at ${VOLUMES}/builds/build.log
.
Finally, a tarball you get after a successful build will land
at ${VOLUMES}/builds/mongooseim-myproject-3414588-2015-11-20_095715.tar.gz
(it's mongooseim-${PROJECT}-${COMMIT}-${TIMESTAMP}.tar.gz
).
Provided a tarball was produced by mongooseim-builder a small image with only
MongooseIM can be build now from Dockerfile.member
. In order to build the image
the MongooseIM tarball has to be copied to members
directory.
The image can now be build with this command:
docker build -f Dockerfile.member -t mongooseim .
First, we need to setup some volumes:
${VOLUMES}/
├── myproject-mongooseim-1
│ ├── ejabberd.cfg
│ ├── hosts
│ ├── mongooseim
│ └── mongooseim.tar.gz
└── myproject-mongooseim-2
├── ejabberd.cfg
├── hosts
├── mongooseim
└── mongooseim.tar.gz
We're preparing a 2 node cluster hence two directories (myproject-mongooseim-X
).
The only file we need to place there is ejabberd.cfg
(a predefined config file).
The rest is actually created when we build our cluster member containers.
The member container can be created with the following command
docker run -t -d -h mongooseim-1 --name mongooseim-1 mongooseim
After docker logs mongooseim-1
shows something similar to:
MongooseIM cluster primary node mongooseim@myproject-mongooseim-1
Clustered mongooseim@myproject-mongooseim-1 with mongooseim@myproject-mongooseim-1
Exec: /member/mongooseim/erts-6.3/bin/erlexec -boot /member/mongooseim/releases//mongooseim -embedded -config /member/mongooseim/etc/app.config -args_file /member/mongooseim/etc/vm.args -- live --noshell -noinput +Bd -mnesia dir "/member/mongooseim/Mnesia.mongooseim@myproject-mongooseim-1"
Root: /member/mongooseim
2015-11-20 10:42:35.903 [info] <0.7.0> Application lager started on node 'mongooseim@myproject-mongooseim-1'
...
2015-11-20 10:42:36.420 [info] <0.145.0>@ejabberd_app:do_notify_fips_mode:270 Used Erlang/OTP does not support FIPS mode
2015-11-20 10:42:36.453 [info] <0.7.0> Application mnesia exited with reason: stopped
2015-11-20 10:42:36.535 [info] <0.7.0> Application mnesia started on node 'mongooseim@myproject-mongooseim-1'
2015-11-20 10:42:36.571 [info] <0.7.0> Application p1_cache_tab started on node 'mongooseim@myproject-mongooseim-1'
We can health-check the MongooseIM node with telnet
.
Supply the IP based on your setup - Docker Machine or localhost - and port
which translates to the container's 5222:
$ telnet $BOOT2DOCKER_IP 32822
Trying 192.168.99.100...
Connected to 192.168.99.100.
Escape character is '^]'.
<?xml version='1.0'?><stream:stream xmlns='jabber:client' xmlns:stream='http://etherx.jabber.org/streams' id='1996592071' from='localhost' version='1.0'><stream:error><xml-not-well-formed xmlns='urn:ietf:params:xml:ns:xmpp-streams'/></stream:error></stream:stream>Connection closed by foreign host.
Success! MongooseIM is accepting XMPP connections.
Let's start another cluster member:
docker run -t -d -h mongooseim-2 --name mongooseim-2 mongooseim
Redo the docker logs
and telnet
checks, but this time against mongooseim-2
.
The nodes should already form a cluster.
Let's check it:
$ docker exec -it myproject-mongooseim-1 /member/mongooseim/bin/mongooseimctl mnesia running_db_nodes
['mongooseim@myproject-mongooseim-2','mongooseim@myproject-mongooseim-1']
$ docker exec -it myproject-mongooseim-2 /member/mongooseim/bin/mongooseimctl mnesia running_db_nodes
['mongooseim@myproject-mongooseim-1','mongooseim@myproject-mongooseim-2']
Tadaa! There you have a brand new shiny cluster running.
There are plenty of ready to use Docker images with databases
or external services you might want to integrate with the cluster.
For example, I'm running a stock postgres:9.6.1
container.
docker run -d --name mongooseim-postgres \
-e POSTGRES_PASSWORD=mongooseim -e POSTGRES_USER=mongooseim \
-v ${PATH_TO_MONGOOSEIM_PGSQL_FILE}:/docker-entrypoint-initdb.d/pgsql.sql:ro \
-p 5432:5432 postgres:9.6.1
Where ${PATH_TO_MONGOOSEIM_PGSQL_FILE}
is an absolute path to pgsql.sql file
which can be found in MongooseIM's repo in apps/ejabberd/priv/pgsql.sql
Don't forget to tweak your ejabberd.cfg
to connect with the services you set up!
For example, like this in case of the PostgreSQL container mentioned above:
{odbc_server, {pgsql, "mongooseim-postgres", "mongooseim", "mongooseim", "mongooseim"}}.
There is another Dockerfile in this repo Dockerfile.multistage
, which allows
to build MongooseIM Docker image from current source code, instead of pulling
it from GitHub, like builder
does.
It may be used with multistage_build.sh
script, which is overlay over
Dockerfile. Usage is simple:
./multistage_build.sh <path_to_moongooseim_root_dir>
It will copy required files into MongooseIM directory and start build process.
Files to be copied: member/start.sh
and Dockerfile.multistage
. They
will be deleted once build process is finished or canceled.
It is possible to pass custom image tag with environmental variable:
IMAGE_TAG=my_mongooseim_docker_build \
./multistage_build.sh <path_to_moongooseim_root_dir>
By default multistage_build.sh
script uses git branch and git reference
as an image tag, e.g.: new_feature-c36dce4fd
. Branch names may contain
characters, which are invalid tags. For instance /
cannot be present
in a tag name. In such a case, image's custom tag has to be explicitly set
using aforementioned IMAGE_TAG
variable. Please visit Docker docs
to read about valid Docker tag names.