DockerFlow creates the necessary Docker containers (webserver, database, php, mail, redis, elasticsearch, couchdb)
to run your Flow Framework or Neos CMS project. The package provides a wrapper script in bin/dockerflow
which simplifies the handling of docker and does all the configuration necessary.
We created this package to make development on Flow Framework and Neos CMS projects easier and to create a simple reusable package which can easily be maintained and serves well for the standard project.
Development will continue further as the package is already reused in several projects. Contributions and feedback are very welcome.
https://docs.docker.com/installation/ (tested with docker v1.9)
We use docker-compose to do all the automatic configuration:
http://docs.docker.com/compose/install/ (tested with docker-compose v1.5)
The repository contains a Dockerfile which will automatically be built in the docker hub after each change and used by docker-compose to build the necessary containers.
http://boot2docker.io
You will not have a performance as good as on linux but it's workable. Checkout the boot2docker support branch of dockerflow for a Vagrantfile which builds a boot2docker instance with working NFS shared. This makes it much faster (still not as fast as on linux).
Add shel/dockerflow
as dev dependency in your composer, using the latest stable release is highly recommended.
Example:
composer require --dev shel/dockerflow 3.0.*
bin/dockerflow up -d
The command will echo the url with which you can access your project. Add the hostname then to your /etc/hosts
and set the ip to your docker host (default for linux is 0.0.0.0) or your boot2docker ip. You can also use any
subdomain with *.hostname
and it will point to the same server. What you need to do is to add exact subdomain name
to your /etc/hosts
.
The parameter -d
will keep it running in the background until you run:
bin/dockerflow stop
The default database configuration for your Settings.yaml
is:
TYPO3:
Flow:
persistence:
backendOptions:
dbname: dockerflow
user: root
password: root
host: db
driver: pdo_mysql
Also note that there is a second database dockerflow_test
available for your testing context. The testing context url
would be test.hostname
and this hostname should be added to your /etc/hosts
too.
bin/dockerflow ps
This will show the running containers. The data
container can be inactive to do it's work.
FLOW_CONTEXT=Production bin/dockerflow up -d
Dockerflow also setup a sub-context for testing depends on the current context you are running. In the above example,
it would be Production/Testing
. Anyway, you can only use the parent context with the bin/dockerflow
command. So when
there is a need to execute command for the testing context, you need to first get into app
container and then call the
command prefixed by the context variable.
FLOW_CONTEXT=Production bin/dockerflow up -d
bin/dockerflow run app /bin/bash
FLOW_CONTEXT=Production/Testing ./flow doctrine:migrate
bin/dockerflow run app ./flow help
FLOW_CONTEXT=Production bin/dockerflow run app ./flow flow:cache:flush --force
Add this configuration to your Settings.yaml
in Flow:
TYPO3:
Flow:
utility:
environment:
temporaryDirectoryBase: /tmp/dockerflow/Temporary/
For caches that has tags, Neos becomes slow with lots of content. Add the following to your Caches.yaml
to
store those mentioned caches in Redis instead:
TYPO3_TypoScript_Content:
backend: TYPO3\Flow\Cache\Backend\RedisBackend
backendOptions:
hostname: 'redis'
port: '6379'
database: 0
Flow_Mvc_Routing_Resolve:
backend: TYPO3\Flow\Cache\Backend\RedisBackend
backendOptions:
hostname: 'redis'
port: '6379'
database: 0
Flow_Mvc_Routing_Route:
backend: TYPO3\Flow\Cache\Backend\RedisBackend
backendOptions:
hostname: 'redis'
port: '6379'
database: 0
Add this configuration to yourSettings.yaml
:
TYPO3:
SwiftMailer:
transport:
type: 'Swift_SmtpTransport'
options:
host: 'mail'
port: 1025
And open MyNeosProject:8025
in your browser (use your own hostname) to see your mails.
Send emails from your Flow app and have fun.
Add this configuration to yourSettings.yaml
:
Flowpack:
ElasticSearch:
clients:
default:
-
host: elasticsearch
port: 9200
DockerFlow installs by the default xdebug with the following config on the server:
xdebug.remote_enable = On
xdebug.remote_host = 'dockerhost'
xdebug.remote_port = '9001'
xdebug.max_nesting_level = 500
So you can do remote debugging from your host to the container through port 9001. From your IDE, you need to configure the port accordingly. If you are using PHPStorm, this link may be useful for you to configure your IDE properly.
bin/dockerflow run SERVICE /bin/bash
SERVICE can currently be app
, web
, data
, db
, redis
, elasticsearch
or couchdb
.
As of current docker doesn't support bi-directional link, you cannot access web container from app container.
But in some case you will need this connection. For example in behat tests without selenium, you need the url of
your site in Testing
context while running the tests has to be done inside the app
container.
Dockerflow adds additional script after starting all containers to fetch the IP address of web container and
append it to /etc/hosts
inside app container as below:
WEB_CONTAINER_IP project-url
WEB_CONTAINER_IP test.project-url
You need to define the default test suite url in your behat.yml
to use http://test.project-url:8080
and then you can
run the behat tests without having to connect external selenium server
bin/dockerflow run app bin/behat -c Path/To/Your/Package/Tests/Behaviour/behat.yml
While you can easily login to shell of the db
container with bin/dockerflow run db /bin/bash
and execute your mysql commands, there are some cases that you want to run mysql commands directly
from your host without having to login to the db
container first. One of the best use cases,
for example, is to access the databases inside the container from MySQL Workbench tool.
To be able to do that, we have mapped database port inside the container (which is 3306
) to your
host machine through 3307
port.
DockerFlow installs by default sqlite
in the base image so that functional tests can be run out-of-the-box.
Example below is for running all functional tests of Flow Framework package in one-off command:
bin/dockerflow run app /var/www/bin/phpunit -c /var/www/Build/BuildEssentials/PhpUnit/FunctionalTests.xml /var/www/Packages/Framework/TYPO3.Flow/Tests/Functional/
Make sure you run composer install with --dev
mode when setting up your Flow project
and adjust the path to the test directory of your own package.
From your host machine, you can access couchdb from web interface or command line:
Web: http://0.0.0.0:5984/_utils/
Cli: curl -X GET http://0.0.0.0:5984/_all_dbs
From inside your app
container, you can also access couchdb through the command line:
bin/dockerflow run app /bin/bash
curl -X GET http://couchdb:5984/_all_dbs
Run bin/dockerflow ps
and copy the container's name that you want to attach to.
Run docker exec -it <containername> /bin/bash
with the name you just copied.
With this you can work in a running container instead of creating a new one.
bin/dockerflow run SERVICE netstat --listen