mirror of
https://github.com/docker/compose.git
synced 2026-02-12 03:29:27 +08:00
Compare commits
177 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
0d4edbbd19 | ||
|
|
fe42107f41 | ||
|
|
30fcb72cf3 | ||
|
|
db94eb0c26 | ||
|
|
c9527747eb | ||
|
|
58dc3ad3c1 | ||
|
|
9f2fd6f358 | ||
|
|
b750e5bb4c | ||
|
|
6eb51d3b5b | ||
|
|
f35db7f31b | ||
|
|
cd97ac0fc1 | ||
|
|
2011bc3cea | ||
|
|
7c82d8ad2c | ||
|
|
d06e15323a | ||
|
|
fa7a7e2cb5 | ||
|
|
3e797d6544 | ||
|
|
60856228ea | ||
|
|
4a51af09d6 | ||
|
|
318181b21e | ||
|
|
4d5aab95d1 | ||
|
|
53237b9f9c | ||
|
|
2a485a02de | ||
|
|
e815126c68 | ||
|
|
872e7025a0 | ||
|
|
e6db77b073 | ||
|
|
cc97ce94b5 | ||
|
|
addfb6c2b2 | ||
|
|
ce8e6d0a5f | ||
|
|
e5358f273b | ||
|
|
a091f468d6 | ||
|
|
acb91db8a6 | ||
|
|
dfe472118d | ||
|
|
947953143c | ||
|
|
2791d31cfb | ||
|
|
a99377720c | ||
|
|
b8f292bbed | ||
|
|
73f8cf9132 | ||
|
|
53add5f372 | ||
|
|
203576363e | ||
|
|
0eaf57466d | ||
|
|
96a8e3b3c5 | ||
|
|
aca6edd631 | ||
|
|
39ab3aee50 | ||
|
|
39c2d42a0e | ||
|
|
397ca0764c | ||
|
|
aa45dedb3d | ||
|
|
afb36f236b | ||
|
|
f01e535cf4 | ||
|
|
853efbfb54 | ||
|
|
6ea24001fa | ||
|
|
c3d8e558a2 | ||
|
|
8f9f1111f3 | ||
|
|
1b42ecba14 | ||
|
|
adae403b27 | ||
|
|
0bea52b18d | ||
|
|
2d0c9366ff | ||
|
|
8a6dc88f9e | ||
|
|
d79745a2cd | ||
|
|
1250bb7481 | ||
|
|
3ba41b98d3 | ||
|
|
7f7370b811 | ||
|
|
a9e8ae190f | ||
|
|
2c6f4b7028 | ||
|
|
3eee3e093a | ||
|
|
ad770b272c | ||
|
|
e5cab3ced5 | ||
|
|
7984767db2 | ||
|
|
c81046aac0 | ||
|
|
84c816e887 | ||
|
|
5b6c3e8389 | ||
|
|
1607674374 | ||
|
|
683fac0dbf | ||
|
|
ddee2958ec | ||
|
|
1ab1cd202b | ||
|
|
1752927dcd | ||
|
|
c2ddd71e5f | ||
|
|
72bbd9c3a6 | ||
|
|
3c9ee678e7 | ||
|
|
92fefbc9cc | ||
|
|
e496c64127 | ||
|
|
84afa518e8 | ||
|
|
0f6a55e036 | ||
|
|
8ce5e235e4 | ||
|
|
c1dddbe608 | ||
|
|
981b0cd641 | ||
|
|
5ec8af582c | ||
|
|
f5342b600c | ||
|
|
4a26d95de4 | ||
|
|
5b7851f55b | ||
|
|
eaa22df151 | ||
|
|
551f680751 | ||
|
|
3e071ec8d9 | ||
|
|
858ff26731 | ||
|
|
2a7c06a050 | ||
|
|
d0b7bc3110 | ||
|
|
fe4f16e448 | ||
|
|
1da4301650 | ||
|
|
c594cb3fc3 | ||
|
|
89ad637d50 | ||
|
|
6ca2aed7ec | ||
|
|
fc744a0cc9 | ||
|
|
245ede1d75 | ||
|
|
72f7b086d7 | ||
|
|
2f48b6f5e9 | ||
|
|
e6fcde422c | ||
|
|
75b2d7905f | ||
|
|
efa5969086 | ||
|
|
2a4aca7f54 | ||
|
|
9c8f5a5705 | ||
|
|
62bbc5cfe2 | ||
|
|
66375c2871 | ||
|
|
c760600a65 | ||
|
|
4daad056c4 | ||
|
|
74c09cac66 | ||
|
|
36e470d640 | ||
|
|
d28d717884 | ||
|
|
42c2cfd7a6 | ||
|
|
5b983ac653 | ||
|
|
93425218eb | ||
|
|
49d0ee2de5 | ||
|
|
a92c6d7e17 | ||
|
|
b8800db52e | ||
|
|
ccabfde353 | ||
|
|
3297bb50bb | ||
|
|
e688006444 | ||
|
|
e4a83c15ff | ||
|
|
824b9f138e | ||
|
|
8654eb2ea3 | ||
|
|
9407ee65e5 | ||
|
|
66c6d2757a | ||
|
|
17daa93edf | ||
|
|
9795e39d0c | ||
|
|
393abc5b33 | ||
|
|
d0866c8c18 | ||
|
|
546133c977 | ||
|
|
9a2f94713e | ||
|
|
b88f635514 | ||
|
|
31002aeacd | ||
|
|
6209baccf3 | ||
|
|
28f8b8549d | ||
|
|
76a19ec8c5 | ||
|
|
bba8cd0322 | ||
|
|
f2ec6a2176 | ||
|
|
7f7f1607de | ||
|
|
4990a7f935 | ||
|
|
72f8551466 | ||
|
|
487779960c | ||
|
|
99b6776fd2 | ||
|
|
4e382b9c28 | ||
|
|
862107a32a | ||
|
|
6a3af5b707 | ||
|
|
205d520805 | ||
|
|
8f2bb66e73 | ||
|
|
af4eaae006 | ||
|
|
1c547b270e | ||
|
|
1c499bb2eb | ||
|
|
4fa72a066a | ||
|
|
b9249168bd | ||
|
|
e36ac32120 | ||
|
|
5be6bde76c | ||
|
|
c380604a9e | ||
|
|
369eb3220a | ||
|
|
2e273c5029 | ||
|
|
21e196f20a | ||
|
|
b9d86f4b51 | ||
|
|
1b5278f977 | ||
|
|
affb0d504d | ||
|
|
8034bc3bd6 | ||
|
|
89fcfc5499 | ||
|
|
40a4ec1624 | ||
|
|
6c55ef6a5d | ||
|
|
3f46dc1d76 | ||
|
|
f2bc89a876 | ||
|
|
fee4756e33 | ||
|
|
030b347673 | ||
|
|
4139d701f3 | ||
|
|
3ebfa4b089 |
15
.github/ISSUE_TEMPLATE/bug_report.md
vendored
15
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@@ -7,6 +7,16 @@ assignees: ''
|
||||
|
||||
---
|
||||
|
||||
<!--
|
||||
**DEPRECATION NOTICE:**
|
||||
|
||||
Compose V1 is end-of-life, and as such only issues relating to security vulnerabilities will be considered.
|
||||
|
||||
Please do not submit issues regarding bugs or improvements.
|
||||
|
||||
For a more up-to-date compose, check v2: https://github.com/docker/compose/tree/v2/
|
||||
-->
|
||||
|
||||
<!--
|
||||
Welcome to the docker-compose issue tracker! Before creating an issue, please heed the following:
|
||||
|
||||
@@ -25,7 +35,10 @@ Welcome to the docker-compose issue tracker! Before creating an issue, please he
|
||||
|
||||
## Context information (for bug reports)
|
||||
|
||||
**Output of `docker-compose version`**
|
||||
- [ ] Using Compose V2 `docker compose ...`
|
||||
- [ ] Using Compose V1 `docker-compose ...`
|
||||
|
||||
**Output of `docker(-)compose version`**
|
||||
```
|
||||
(paste here)
|
||||
```
|
||||
|
||||
13
.github/ISSUE_TEMPLATE/feature_request.md
vendored
13
.github/ISSUE_TEMPLATE/feature_request.md
vendored
@@ -7,6 +7,16 @@ assignees: ''
|
||||
|
||||
---
|
||||
|
||||
<!--
|
||||
**DEPRECATION NOTICE:**
|
||||
|
||||
Compose V1 is end-of-life, and as such only issues relating to security vulnerabilities will be considered.
|
||||
|
||||
Please do not submit issues regarding bugs or improvements.
|
||||
|
||||
For a more up-to-date compose, check v2: https://github.com/docker/compose/tree/v2/
|
||||
-->
|
||||
|
||||
<!--
|
||||
Welcome to the docker-compose issue tracker! Before creating an issue, please heed the following:
|
||||
|
||||
@@ -19,6 +29,9 @@ Welcome to the docker-compose issue tracker! Before creating an issue, please he
|
||||
the original discussion.
|
||||
-->
|
||||
|
||||
/!\ If your request is about evolving the compose file format, please report on the [Compose Specification](https://github.com/compose-spec/compose-spec)
|
||||
|
||||
|
||||
**Is your feature request related to a problem? Please describe.**
|
||||
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
|
||||
|
||||
|
||||
@@ -7,6 +7,16 @@ assignees: ''
|
||||
|
||||
---
|
||||
|
||||
<!--
|
||||
**DEPRECATION NOTICE:**
|
||||
|
||||
Compose V1 is end-of-life, and as such only issues relating to security vulnerabilities will be considered.
|
||||
|
||||
Please do not submit issues regarding bugs or improvements.
|
||||
|
||||
For a more up-to-date compose, check v2: https://github.com/docker/compose/tree/v2/
|
||||
-->
|
||||
|
||||
Please post on our forums: https://forums.docker.com for questions about using `docker-compose`.
|
||||
|
||||
Posts that are not a bug report or a feature/enhancement request will not be addressed on this issue tracker.
|
||||
|
||||
28
.github/dependabot.yml
vendored
Normal file
28
.github/dependabot.yml
vendored
Normal file
@@ -0,0 +1,28 @@
|
||||
version: 2
|
||||
updates:
|
||||
- package-ecosystem: pip
|
||||
directory: "/"
|
||||
schedule:
|
||||
interval: weekly
|
||||
time: "14:00"
|
||||
timezone: America/Los_Angeles
|
||||
open-pull-requests-limit: 10
|
||||
ignore:
|
||||
- dependency-name: python-dotenv
|
||||
versions:
|
||||
- 0.15.0
|
||||
- 0.16.0
|
||||
- dependency-name: urllib3
|
||||
versions:
|
||||
- 1.26.2
|
||||
- 1.26.3
|
||||
- dependency-name: coverage
|
||||
versions:
|
||||
- 5.3.1
|
||||
- "5.4"
|
||||
- dependency-name: packaging
|
||||
versions:
|
||||
- "20.8"
|
||||
- dependency-name: cached-property
|
||||
versions:
|
||||
- 1.5.2
|
||||
58
.github/workflows/artifacts.yml
vendored
Normal file
58
.github/workflows/artifacts.yml
vendored
Normal file
@@ -0,0 +1,58 @@
|
||||
name: Publish Artifacts
|
||||
on:
|
||||
issue_comment:
|
||||
types: [created]
|
||||
jobs:
|
||||
publish-artifacts:
|
||||
if: github.event.issue.pull_request != '' && contains(github.event.comment.body, '/generate-artifacts')
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Set up Go 1.16
|
||||
uses: actions/setup-go@v2
|
||||
with:
|
||||
go-version: 1.16
|
||||
id: go
|
||||
|
||||
- name: Checkout code into the Go module directory
|
||||
uses: actions/checkout@v2
|
||||
|
||||
- uses: actions/cache@v2
|
||||
with:
|
||||
path: ~/go/pkg/mod
|
||||
key: go-${{ hashFiles('**/go.sum') }}
|
||||
|
||||
- name: Build cross platform compose-plugin binaries
|
||||
run: make -f builder.Makefile cross
|
||||
|
||||
- name: Upload macos-amd64 binary
|
||||
uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: docker-compose-darwin-amd64
|
||||
path: ${{ github.workspace }}/bin/docker-compose-darwin-amd64
|
||||
|
||||
- name: Upload macos-arm64 binary
|
||||
uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: docker-compose-darwin-arm64
|
||||
path: ${{ github.workspace }}/bin/docker-compose-darwin-arm64
|
||||
|
||||
- name: Upload linux-amd64 binary
|
||||
uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: docker-compose-linux-amd64
|
||||
path: ${{ github.workspace }}/bin/docker-compose-linux-amd64
|
||||
|
||||
- name: Upload windows-amd64 binary
|
||||
uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: docker-compose-windows-amd64.exe
|
||||
path: ${{ github.workspace }}/bin/docker-compose-windows-amd64.exe
|
||||
|
||||
- name: Update comment
|
||||
uses: peter-evans/create-or-update-comment@v1
|
||||
with:
|
||||
comment-id: ${{ github.event.comment.id }}
|
||||
body: |
|
||||
This PR can be tested using [binaries](https://github.com/docker/compose-cli/actions/runs/${{ github.run_id }}).
|
||||
reactions: eyes
|
||||
|
||||
104
.github/workflows/ci.yml
vendored
Normal file
104
.github/workflows/ci.yml
vendored
Normal file
@@ -0,0 +1,104 @@
|
||||
name: Continuous integration
|
||||
|
||||
on:
|
||||
push:
|
||||
branches:
|
||||
- v2
|
||||
pull_request:
|
||||
branches:
|
||||
- v2
|
||||
|
||||
jobs:
|
||||
lint:
|
||||
name: Lint
|
||||
runs-on: ubuntu-latest
|
||||
env:
|
||||
GO111MODULE: "on"
|
||||
steps:
|
||||
- name: Set up Go 1.16
|
||||
uses: actions/setup-go@v2
|
||||
with:
|
||||
go-version: 1.16
|
||||
id: go
|
||||
|
||||
- name: Checkout code into the Go module directory
|
||||
uses: actions/checkout@v2
|
||||
|
||||
- name: Validate go-mod is up-to-date and license headers
|
||||
run: make validate
|
||||
|
||||
- name: Run golangci-lint
|
||||
env:
|
||||
BUILD_TAGS: e2e
|
||||
run: |
|
||||
curl -sSfL https://raw.githubusercontent.com/golangci/golangci-lint/master/install.sh | sudo sh -s -- -b /usr/bin/ v1.39.0
|
||||
make -f builder.Makefile lint
|
||||
|
||||
# only on main branch, costs too much for the gain on every PR
|
||||
validate-cross-build:
|
||||
name: Validate cross build
|
||||
runs-on: ubuntu-latest
|
||||
if: github.ref == 'refs/heads/main'
|
||||
env:
|
||||
GO111MODULE: "on"
|
||||
steps:
|
||||
- name: Set up Go 1.16
|
||||
uses: actions/setup-go@v2
|
||||
with:
|
||||
go-version: 1.16
|
||||
id: go
|
||||
|
||||
- name: Checkout code into the Go module directory
|
||||
uses: actions/checkout@v2
|
||||
|
||||
- uses: actions/cache@v2
|
||||
with:
|
||||
path: ~/go/pkg/mod
|
||||
key: go-${{ hashFiles('**/go.sum') }}
|
||||
|
||||
# Ensure we don't discover cross platform build issues at release time.
|
||||
# Time used to build linux here is gained back in the build for local E2E step
|
||||
- name: Build packages
|
||||
run: make -f builder.Makefile cross
|
||||
|
||||
build:
|
||||
name: Build
|
||||
runs-on: ubuntu-latest
|
||||
env:
|
||||
GO111MODULE: "on"
|
||||
steps:
|
||||
- name: Set up Go 1.16
|
||||
uses: actions/setup-go@v2
|
||||
with:
|
||||
go-version: 1.16
|
||||
id: go
|
||||
|
||||
- name: Set up gosum
|
||||
run: |
|
||||
go get -u gotest.tools/gotestsum
|
||||
|
||||
- name: Setup docker CLI
|
||||
run: |
|
||||
curl https://download.docker.com/linux/static/stable/x86_64/docker-20.10.3.tgz | tar xz
|
||||
sudo cp ./docker/docker /usr/bin/ && rm -rf docker && docker version
|
||||
|
||||
- name: Checkout code into the Go module directory
|
||||
uses: actions/checkout@v2
|
||||
|
||||
- uses: actions/cache@v2
|
||||
with:
|
||||
path: ~/go/pkg/mod
|
||||
key: go-${{ hashFiles('**/go.sum') }}
|
||||
|
||||
- name: Test
|
||||
env:
|
||||
BUILD_TAGS: kube
|
||||
run: make -f builder.Makefile test
|
||||
|
||||
- name: Build for local E2E
|
||||
env:
|
||||
BUILD_TAGS: e2e
|
||||
run: make -f builder.Makefile compose-plugin
|
||||
|
||||
- name: E2E Test
|
||||
run: make e2e-compose
|
||||
11
.github/workflows/pr-closed.yml
vendored
Normal file
11
.github/workflows/pr-closed.yml
vendored
Normal file
@@ -0,0 +1,11 @@
|
||||
name: PR cleanup
|
||||
on:
|
||||
pull_request:
|
||||
types: [closed]
|
||||
jobs:
|
||||
delete_pr_artifacts:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: stefanluptak/delete-old-pr-artifacts@v1
|
||||
with:
|
||||
workflow_filename: ci.yaml
|
||||
19
.github/workflows/rebase.yml
vendored
Normal file
19
.github/workflows/rebase.yml
vendored
Normal file
@@ -0,0 +1,19 @@
|
||||
name: Automatic Rebase
|
||||
on:
|
||||
issue_comment:
|
||||
types: [created]
|
||||
jobs:
|
||||
rebase:
|
||||
name: Rebase
|
||||
if: github.event.issue.pull_request != '' && contains(github.event.comment.body, '/rebase')
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Checkout the latest code
|
||||
uses: actions/checkout@v2
|
||||
with:
|
||||
token: ${{ secrets.GITHUB_TOKEN }}
|
||||
fetch-depth: 0 # otherwise, you will fail to push refs to dest repo
|
||||
- name: Automatic Rebase
|
||||
uses: cirrus-actions/rebase@1.4
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
49
.github/workflows/release.yaml
vendored
Normal file
49
.github/workflows/release.yaml
vendored
Normal file
@@ -0,0 +1,49 @@
|
||||
name: Releaser
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
tag:
|
||||
description: 'Release Tag'
|
||||
required: true
|
||||
dry-run:
|
||||
description: 'Dry run'
|
||||
required: false
|
||||
default: 'true'
|
||||
|
||||
jobs:
|
||||
upload-release:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Set up Go 1.16
|
||||
uses: actions/setup-go@v2
|
||||
with:
|
||||
go-version: 1.16
|
||||
id: go
|
||||
|
||||
- name: Setup docker CLI
|
||||
run: |
|
||||
curl https://download.docker.com/linux/static/stable/x86_64/docker-20.10.3.tgz | tar xz
|
||||
sudo cp ./docker/docker /usr/bin/ && rm -rf docker && docker version
|
||||
|
||||
- name: Checkout code into the Go module directory
|
||||
uses: actions/checkout@v2
|
||||
|
||||
- uses: actions/cache@v2
|
||||
with:
|
||||
path: ~/go/pkg/mod
|
||||
key: ${{ runner.os }}-go-${{ hashFiles('**/go.sum') }}
|
||||
restore-keys: |
|
||||
${{ runner.os }}-go-
|
||||
|
||||
- name: Build
|
||||
run: make -f builder.Makefile cross-compose-plugin
|
||||
|
||||
- name: License
|
||||
run: cp packaging/* bin/
|
||||
|
||||
- uses: ncipollo/release-action@v1
|
||||
with:
|
||||
artifacts: "bin/*"
|
||||
prerelease: true
|
||||
token: ${{ secrets.GITHUB_TOKEN }}
|
||||
@@ -1,26 +1,28 @@
|
||||
- repo: git://github.com/pre-commit/pre-commit-hooks
|
||||
sha: 'v0.9.1'
|
||||
hooks:
|
||||
- id: check-added-large-files
|
||||
- id: check-docstring-first
|
||||
- id: check-merge-conflict
|
||||
- id: check-yaml
|
||||
- id: check-json
|
||||
- id: debug-statements
|
||||
- id: end-of-file-fixer
|
||||
- id: flake8
|
||||
- id: name-tests-test
|
||||
exclude: 'tests/(integration/testcases\.py|helpers\.py)'
|
||||
- id: requirements-txt-fixer
|
||||
- id: trailing-whitespace
|
||||
- repo: git://github.com/asottile/reorder_python_imports
|
||||
sha: v1.3.4
|
||||
hooks:
|
||||
- id: reorder-python-imports
|
||||
language_version: 'python3.9'
|
||||
args:
|
||||
- --py3-plus
|
||||
- repo: https://github.com/asottile/pyupgrade
|
||||
exclude: .github/
|
||||
repos:
|
||||
- repo: git://github.com/pre-commit/pre-commit-hooks
|
||||
sha: 'v0.9.1'
|
||||
hooks:
|
||||
- id: check-added-large-files
|
||||
- id: check-docstring-first
|
||||
- id: check-merge-conflict
|
||||
- id: check-yaml
|
||||
- id: check-json
|
||||
- id: debug-statements
|
||||
- id: end-of-file-fixer
|
||||
- id: flake8
|
||||
- id: name-tests-test
|
||||
exclude: 'tests/(integration/testcases\.py|helpers\.py)'
|
||||
- id: requirements-txt-fixer
|
||||
- id: trailing-whitespace
|
||||
- repo: git://github.com/asottile/reorder_python_imports
|
||||
sha: v1.3.4
|
||||
hooks:
|
||||
- id: reorder-python-imports
|
||||
language_version: 'python3.7'
|
||||
args:
|
||||
- --py3-plus
|
||||
- repo: https://github.com/asottile/pyupgrade
|
||||
rev: v2.1.0
|
||||
hooks:
|
||||
- id: pyupgrade
|
||||
|
||||
197
CHANGELOG.md
197
CHANGELOG.md
@@ -1,6 +1,203 @@
|
||||
Change log
|
||||
==========
|
||||
|
||||
1.29.2 (2021-05-10)
|
||||
-------------------
|
||||
|
||||
[List of PRs / issues for this release](https://github.com/docker/compose/milestone/59?closed=1)
|
||||
|
||||
### Miscellaneous
|
||||
|
||||
- Remove prompt to use `docker compose` in the `up` command
|
||||
|
||||
- Bump `py` to `1.10.0` in `requirements-indirect.txt`
|
||||
|
||||
1.29.1 (2021-04-13)
|
||||
-------------------
|
||||
|
||||
[List of PRs / issues for this release](https://github.com/docker/compose/milestone/58?closed=1)
|
||||
|
||||
### Bugs
|
||||
|
||||
- Fix for invalid handler warning on Windows builds
|
||||
|
||||
- Fix config hash to trigger container recreation on IPC mode updates
|
||||
|
||||
- Fix conversion map for `placement.max_replicas_per_node`
|
||||
|
||||
- Remove extra scan suggestion on build
|
||||
|
||||
1.29.0 (2021-04-06)
|
||||
-------------------
|
||||
|
||||
[List of PRs / issues for this release](https://github.com/docker/compose/milestone/56?closed=1)
|
||||
|
||||
### Features
|
||||
|
||||
- Add profile filter to `docker-compose config`
|
||||
|
||||
- Add a `depends_on` condition to wait for successful service completion
|
||||
|
||||
### Miscellaneous
|
||||
|
||||
- Add image scan message on build
|
||||
|
||||
- Update warning message for `--no-ansi` to mention `--ansi never` as alternative
|
||||
|
||||
- Bump docker-py to 5.0.0
|
||||
|
||||
- Bump PyYAML to 5.4.1
|
||||
|
||||
- Bump python-dotenv to 0.17.0
|
||||
|
||||
1.28.6 (2021-03-23)
|
||||
-------------------
|
||||
|
||||
[List of PRs / issues for this release](https://github.com/docker/compose/milestone/57?closed=1)
|
||||
|
||||
### Bugs
|
||||
|
||||
- Make `--env-file` relative to the current working directory and error out for invalid paths. Environment file paths set with `--env-file` are relative to the current working directory while the default `.env` file is located in the project directory which by default is the base directory of the Compose file.
|
||||
|
||||
- Fix missing service property `storage_opt` by updating the compose schema
|
||||
|
||||
- Fix build `extra_hosts` list format
|
||||
|
||||
- Remove extra error message on `exec`
|
||||
|
||||
### Miscellaneous
|
||||
|
||||
- Add `compose.yml` and `compose.yaml` to default filename list
|
||||
|
||||
1.28.5 (2021-02-25)
|
||||
-------------------
|
||||
|
||||
[List of PRs / issues for this release](https://github.com/docker/compose/milestone/55?closed=1)
|
||||
|
||||
### Bugs
|
||||
|
||||
- Fix OpenSSL version mismatch error when shelling out to the ssh client (via bump to docker-py 4.4.4 which contains the fix)
|
||||
|
||||
- Add missing build flags to the native builder: `platform`, `isolation` and `extra_hosts`
|
||||
|
||||
- Remove info message on native build
|
||||
|
||||
- Avoid fetching logs when service logging driver is set to 'none'
|
||||
|
||||
1.28.4 (2021-02-18)
|
||||
-------------------
|
||||
|
||||
[List of PRs / issues for this release](https://github.com/docker/compose/milestone/54?closed=1)
|
||||
|
||||
### Bugs
|
||||
|
||||
- Fix SSH port parsing by bumping docker-py to 4.4.3
|
||||
|
||||
### Miscellaneous
|
||||
|
||||
- Bump Python to 3.7.10
|
||||
|
||||
1.28.3 (2021-02-17)
|
||||
-------------------
|
||||
|
||||
[List of PRs / issues for this release](https://github.com/docker/compose/milestone/53?closed=1)
|
||||
|
||||
### Bugs
|
||||
|
||||
- Fix SSH hostname parsing when it contains leading s/h, and remove the quiet option that was hiding the error (via docker-py bump to 4.4.2)
|
||||
|
||||
- Fix key error for '--no-log-prefix' option
|
||||
|
||||
- Fix incorrect CLI environment variable name for service profiles: `COMPOSE_PROFILES` instead of `COMPOSE_PROFILE`
|
||||
|
||||
- Fix fish completion
|
||||
|
||||
### Miscellaneous
|
||||
|
||||
- Bump cryptography to 3.3.2
|
||||
|
||||
- Remove log driver filter
|
||||
|
||||
1.28.2 (2021-01-26)
|
||||
-------------------
|
||||
|
||||
### Miscellaneous
|
||||
|
||||
- CI setup update
|
||||
|
||||
1.28.1 (2021-01-25)
|
||||
-------------------
|
||||
|
||||
### Bugs
|
||||
|
||||
- Revert to Python 3.7 bump for Linux static builds
|
||||
|
||||
- Add bash completion for `docker-compose logs|up --no-log-prefix`
|
||||
|
||||
1.28.0 (2021-01-20)
|
||||
-------------------
|
||||
|
||||
### Features
|
||||
|
||||
- Support for Nvidia GPUs via device requests
|
||||
|
||||
- Support for service profiles
|
||||
|
||||
- Change the SSH connection approach to the Docker CLI's via shellout to the local SSH client (old behaviour enabled by setting `COMPOSE_PARAMIKO_SSH` environment variable)
|
||||
|
||||
- Add flag to disable log prefix
|
||||
|
||||
- Add flag for ansi output control
|
||||
|
||||
### Bugs
|
||||
|
||||
- Make `parallel_pull=True` by default
|
||||
|
||||
- Bring back warning for configs in non-swarm mode
|
||||
|
||||
- Take `--file` in account when defining `project_dir`
|
||||
|
||||
- On `compose up`, attach only to services we read logs from
|
||||
|
||||
### Miscellaneous
|
||||
|
||||
- Make COMPOSE_DOCKER_CLI_BUILD=1 the default
|
||||
|
||||
- Add usage metrics
|
||||
|
||||
- Sync schema with COMPOSE specification
|
||||
|
||||
- Improve failure report for missing mandatory environment variables
|
||||
|
||||
- Bump attrs to 20.3.0
|
||||
|
||||
- Bump more_itertools to 8.6.0
|
||||
|
||||
- Bump cryptograhy to 3.2.1
|
||||
|
||||
- Bump cffi to 1.14.4
|
||||
|
||||
- Bump virtualenv to 20.2.2
|
||||
|
||||
- Bump bcrypt to 3.2.0
|
||||
|
||||
- Bump gitpython to 3.1.11
|
||||
|
||||
- Bump docker-py to 4.4.1
|
||||
|
||||
- Bump Python to 3.9
|
||||
|
||||
- Linux: bump Debian base image from stretch to buster (required for Python 3.9)
|
||||
|
||||
- macOS: OpenSSL 1.1.1g to 1.1.1h, Python 3.7.7 to 3.9.0
|
||||
|
||||
- Bump pyinstaller 4.1
|
||||
|
||||
- Loosen restriction on base images to latest minor
|
||||
|
||||
- Updates of READMEs
|
||||
|
||||
|
||||
1.27.4 (2020-09-24)
|
||||
-------------------
|
||||
|
||||
|
||||
14
Dockerfile
14
Dockerfile
@@ -1,13 +1,13 @@
|
||||
ARG DOCKER_VERSION=19.03
|
||||
ARG PYTHON_VERSION=3.9.0
|
||||
ARG PYTHON_VERSION=3.7.10
|
||||
|
||||
ARG BUILD_ALPINE_VERSION=3.12
|
||||
ARG BUILD_CENTOS_VERSION=7
|
||||
ARG BUILD_DEBIAN_VERSION=slim-buster
|
||||
ARG BUILD_DEBIAN_VERSION=slim-stretch
|
||||
|
||||
ARG RUNTIME_ALPINE_VERSION=3.12
|
||||
ARG RUNTIME_CENTOS_VERSION=7
|
||||
ARG RUNTIME_DEBIAN_VERSION=buster-slim
|
||||
ARG RUNTIME_DEBIAN_VERSION=stretch-slim
|
||||
|
||||
ARG DISTRO=alpine
|
||||
|
||||
@@ -38,7 +38,7 @@ RUN apt-get update && apt-get install --no-install-recommends -y \
|
||||
git \
|
||||
libc-dev \
|
||||
libffi-dev \
|
||||
libgcc-8-dev \
|
||||
libgcc-6-dev \
|
||||
libssl-dev \
|
||||
make \
|
||||
openssl \
|
||||
@@ -68,8 +68,8 @@ WORKDIR /code/
|
||||
COPY docker-compose-entrypoint.sh /usr/local/bin/
|
||||
COPY --from=docker-cli /usr/local/bin/docker /usr/local/bin/docker
|
||||
RUN pip install \
|
||||
virtualenv==20.2.1 \
|
||||
tox==3.20.1
|
||||
virtualenv==20.4.0 \
|
||||
tox==3.21.2
|
||||
COPY requirements-dev.txt .
|
||||
COPY requirements-indirect.txt .
|
||||
COPY requirements.txt .
|
||||
@@ -79,7 +79,7 @@ COPY tox.ini .
|
||||
COPY setup.py .
|
||||
COPY README.md .
|
||||
COPY compose compose/
|
||||
RUN tox --notest
|
||||
RUN tox -e py37 --notest
|
||||
COPY . .
|
||||
ARG GIT_COMMIT=unknown
|
||||
ENV DOCKER_COMPOSE_GITSHA=$GIT_COMMIT
|
||||
|
||||
136
INSTALL.md
Normal file
136
INSTALL.md
Normal file
@@ -0,0 +1,136 @@
|
||||
# Install Docker Compose
|
||||
|
||||
This page contains information on how to install Docker Compose. You can run Compose on macOS, Windows, and 64-bit Linux.
|
||||
|
||||
> ⚠️ The installation instructions on this page will help you to install Compose v1 which is a deprecated version. We recommend that you use the [latest version of Docker Compose](https://docs.docker.com/compose/install/).
|
||||
|
||||
## Prerequisites
|
||||
|
||||
Docker Compose relies on Docker Engine for any meaningful work, so make sure you
|
||||
have Docker Engine installed either locally or remote, depending on your setup.
|
||||
|
||||
|
||||
- Install
|
||||
[Docker Engine](https://docs.docker.com/engine/install/#server)
|
||||
for your OS and then come back here for
|
||||
instructions on installing the Python version of Compose.
|
||||
|
||||
- To run Compose as a non-root user, see [Manage Docker as a non-root user](https://docs.docker.com/engine/install/linux-postinstall/).
|
||||
|
||||
## Install Compose
|
||||
|
||||
|
||||
Follow the instructions below to install Compose using the `pip`
|
||||
Python package manager or to install Compose as a container.
|
||||
|
||||
> Install a different version
|
||||
>
|
||||
> The instructions below outline installation of the current stable release
|
||||
> (**v1.29.2**) of Compose. To install a different version of
|
||||
> Compose, replace the given release number with the one that you want. For instructions to install Compose 2.x.x on Linux, see [Install Compose 2.x.x on Linux](https://docs.docker.com/compose/install/#install-compose-on-linux-systems).
|
||||
>
|
||||
> Compose releases are also listed and available for direct download on the
|
||||
> [Compose repository release page on GitHub](https://github.com/docker/compose/releases).
|
||||
> To install a **pre-release** of Compose, refer to the [install pre-release builds](#install-pre-release-builds)
|
||||
> section.
|
||||
|
||||
- [Install using pip](#install-using-pip)
|
||||
- [Install as a container](#install-as-a-container)
|
||||
|
||||
#### Install using pip
|
||||
|
||||
> For `alpine`, the following dependency packages are needed:
|
||||
> `py-pip`, `python3-dev`, `libffi-dev`, `openssl-dev`, `gcc`, `libc-dev`, `rust`, `cargo`, and `make`.
|
||||
{: .important}
|
||||
|
||||
You can install Compose from
|
||||
[pypi](https://pypi.python.org/pypi/docker-compose) using `pip`. If you install
|
||||
using `pip`, we recommend that you use a
|
||||
[virtualenv](https://virtualenv.pypa.io/en/latest/) because many operating
|
||||
systems have python system packages that conflict with docker-compose
|
||||
dependencies. See the [virtualenv
|
||||
tutorial](https://docs.python-guide.org/dev/virtualenvs/) to get
|
||||
started.
|
||||
|
||||
```console
|
||||
$ pip3 install docker-compose
|
||||
```
|
||||
|
||||
If you are not using virtualenv,
|
||||
|
||||
```console
|
||||
$ sudo pip install docker-compose
|
||||
```
|
||||
|
||||
> pip version 6.0 or greater is required.
|
||||
|
||||
#### Install as a container
|
||||
|
||||
You can also run Compose inside a container, from a small bash script wrapper. To
|
||||
install Compose as a container run this command:
|
||||
|
||||
```console
|
||||
$ sudo curl -L --fail https://github.com/docker/compose/releases/download/1.29.2/run.sh -o /usr/local/bin/docker-compose
|
||||
$ sudo chmod +x /usr/local/bin/docker-compose
|
||||
```
|
||||
|
||||
|
||||
### Install pre-release builds
|
||||
|
||||
If you're interested in trying out a pre-release build, you can download release
|
||||
candidates from the [Compose repository release page on GitHub](https://github.com/docker/compose/releases).
|
||||
Follow the instructions from the link, which involves running the `curl` command
|
||||
in your terminal to download the binaries.
|
||||
|
||||
Pre-releases built from the "master" branch are also available for download at
|
||||
[https://dl.bintray.com/docker-compose/master/](https://dl.bintray.com/docker-compose/master/).
|
||||
|
||||
> Pre-release builds allow you to try out new features before they are released,
|
||||
> but may be less stable.
|
||||
|
||||
----
|
||||
|
||||
## Upgrading
|
||||
|
||||
If you're upgrading from Compose 1.2 or earlier, remove or
|
||||
migrate your existing containers after upgrading Compose. This is because, as of
|
||||
version 1.3, Compose uses Docker labels to keep track of containers, and your
|
||||
containers need to be recreated to add the labels.
|
||||
|
||||
If Compose detects containers that were created without labels, it refuses
|
||||
to run, so that you don't end up with two sets of them. If you want to keep using
|
||||
your existing containers (for example, because they have data volumes you want
|
||||
to preserve), you can use Compose 1.5.x to migrate them with the following
|
||||
command:
|
||||
|
||||
```console
|
||||
$ docker-compose migrate-to-labels
|
||||
```
|
||||
|
||||
Alternatively, if you're not worried about keeping them, you can remove them.
|
||||
Compose just creates new ones.
|
||||
|
||||
```console
|
||||
$ docker container rm -f -v myapp_web_1 myapp_db_1 ...
|
||||
```
|
||||
|
||||
## Uninstall
|
||||
|
||||
To uninstall Docker Compose if you installed using `curl`:
|
||||
|
||||
```console
|
||||
$ sudo rm /usr/local/bin/docker-compose
|
||||
```
|
||||
|
||||
To uninstall Docker Compose if you installed using `pip`:
|
||||
|
||||
```console
|
||||
$ pip uninstall docker-compose
|
||||
```
|
||||
|
||||
> Got a "Permission denied" error?
|
||||
>
|
||||
> If you get a "Permission denied" error using either of the above
|
||||
> methods, you probably do not have the proper permissions to remove
|
||||
> `docker-compose`. To force the removal, prepend `sudo` to either of the above
|
||||
> commands and run again.
|
||||
14
Jenkinsfile
vendored
14
Jenkinsfile
vendored
@@ -2,7 +2,7 @@
|
||||
|
||||
def dockerVersions = ['19.03.13']
|
||||
def baseImages = ['alpine', 'debian']
|
||||
def pythonVersions = ['py39']
|
||||
def pythonVersions = ['py37']
|
||||
|
||||
pipeline {
|
||||
agent none
|
||||
@@ -23,7 +23,7 @@ pipeline {
|
||||
parallel {
|
||||
stage('alpine') {
|
||||
agent {
|
||||
label 'ubuntu && amd64 && !zfs'
|
||||
label 'ubuntu-2004 && amd64 && !zfs && cgroup1'
|
||||
}
|
||||
steps {
|
||||
buildImage('alpine')
|
||||
@@ -31,7 +31,7 @@ pipeline {
|
||||
}
|
||||
stage('debian') {
|
||||
agent {
|
||||
label 'ubuntu && amd64 && !zfs'
|
||||
label 'ubuntu-2004 && amd64 && !zfs && cgroup1'
|
||||
}
|
||||
steps {
|
||||
buildImage('debian')
|
||||
@@ -62,7 +62,7 @@ pipeline {
|
||||
|
||||
def buildImage(baseImage) {
|
||||
def scmvar = checkout(scm)
|
||||
def imageName = "dockerbuildbot/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
def imageName = "dockerpinata/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
image = docker.image(imageName)
|
||||
|
||||
withDockerRegistry(credentialsId:'dockerbuildbot-index.docker.io') {
|
||||
@@ -87,9 +87,9 @@ def buildImage(baseImage) {
|
||||
def runTests(dockerVersion, pythonVersion, baseImage) {
|
||||
return {
|
||||
stage("python=${pythonVersion} docker=${dockerVersion} ${baseImage}") {
|
||||
node("ubuntu && amd64 && !zfs") {
|
||||
node("ubuntu-2004 && amd64 && !zfs && cgroup1") {
|
||||
def scmvar = checkout(scm)
|
||||
def imageName = "dockerbuildbot/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
def imageName = "dockerpinata/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
def storageDriver = sh(script: "docker info -f \'{{.Driver}}\'", returnStdout: true).trim()
|
||||
echo "Using local system's storage driver: ${storageDriver}"
|
||||
withDockerRegistry(credentialsId:'dockerbuildbot-index.docker.io') {
|
||||
@@ -99,6 +99,8 @@ def runTests(dockerVersion, pythonVersion, baseImage) {
|
||||
--privileged \\
|
||||
--volume="\$(pwd)/.git:/code/.git" \\
|
||||
--volume="/var/run/docker.sock:/var/run/docker.sock" \\
|
||||
--volume="\${DOCKER_CONFIG}/config.json:/root/.docker/config.json" \\
|
||||
-e "DOCKER_TLS_CERTDIR=" \\
|
||||
-e "TAG=${imageName}" \\
|
||||
-e "STORAGE_DRIVER=${storageDriver}" \\
|
||||
-e "DOCKER_VERSIONS=${dockerVersion}" \\
|
||||
|
||||
56
README.md
56
README.md
@@ -4,6 +4,40 @@ Docker Compose
|
||||
|
||||

|
||||
|
||||
# :warning: *Compose V1 is DEPRECATED* :warning:
|
||||
Since [Compose V2 is now GA](https://www.docker.com/blog/announcing-compose-v2-general-availability/), Compose V1 is officially **End of Life**. This means that:
|
||||
- Active development and new features will only be added to the V2 codebase
|
||||
- Only security-related issues will be considered for V1
|
||||
|
||||
Check out the [V2 branch here](https://github.com/docker/compose/tree/v2/)!!
|
||||
|
||||
---------------------------------------------
|
||||
|
||||
** Compose V2 is **Generally Available**! :star_struck: **
|
||||
---------------------------------------------
|
||||
|
||||
Check it out [here](https://github.com/docker/compose/tree/v2/)!
|
||||
|
||||
Read more on the [GA announcement here](https://www.docker.com/blog/announcing-compose-v2-general-availability/)
|
||||
|
||||
|
||||
---------------------------------------------
|
||||
|
||||
V1 vs V2 transition :hourglass_flowing_sand:
|
||||
--------------------------------------------
|
||||
|
||||
"Generally Available" will mean:
|
||||
- New features and bug fixes will only be considered in the V2 codebase
|
||||
- Users on Mac/Windows will be defaulted into Docker Compose V2, but can still opt out through the UI and the CLI. This means when running `docker-compose` you will actually be running `docker compose`
|
||||
- Our current goal is for users on Linux to receive Compose v2 with the latest version of the docker CLI, but is pending some technical discussion. Users will be able to use [compose switch](https://github.com/docker/compose-switch) to enable redirection of `docker-compose` to `docker compose`
|
||||
- Docker Compose V1 will continue to be maintained regarding security issues
|
||||
- [v2 branch](https://github.com/docker/compose/tree/v2) will become the default one at that time
|
||||
|
||||
:lock_with_ink_pen: Depending on the feedback we receive from the community of GA and the adoption on Linux, we will come up with a plan to deprecate v1, but as of right now there is no concrete timeline as we want the transition to be as smooth as possible for all users. It is important to note that we have no plans of removing any aliasing of `docker-compose` to `docker compose`. We want to make it as easy as possible to switch and not break any ones scripts. We will follow up with a blog post in the next few months with more information of an exact timeline of V1 being marked as deprecated and end of support for security issues. We’d love to hear your feedback! You can provide it [here](https://github.com/docker/roadmap/issues/257).
|
||||
|
||||
About
|
||||
-----
|
||||
|
||||
Docker Compose is a tool for running multi-container applications on Docker
|
||||
defined using the [Compose file format](https://compose-spec.io).
|
||||
A Compose file is used to define how the one or more containers that make up
|
||||
@@ -21,26 +55,10 @@ the Docker CLI. You can read more about how to do this:
|
||||
Where to get Docker Compose
|
||||
----------------------------
|
||||
|
||||
### Windows and macOS
|
||||
All the instructions to install the Python version of Docker Compose, aka `v1`,
|
||||
are described in the [installation guide](./INSTALL.md).
|
||||
|
||||
Docker Compose is included in
|
||||
[Docker Desktop](https://www.docker.com/products/docker-desktop)
|
||||
for Windows and macOS.
|
||||
|
||||
### Linux
|
||||
|
||||
You can download Docker Compose binaries from the
|
||||
[release page](https://github.com/docker/compose/releases) on this repository.
|
||||
|
||||
### Using pip
|
||||
|
||||
If your platform is not supported, you can download Docker Compose using `pip`:
|
||||
|
||||
```console
|
||||
pip install docker-compose
|
||||
```
|
||||
|
||||
> **Note:** Docker Compose requires Python 3.6 or later.
|
||||
> ⚠️ This version is a deprecated version of Compose. We recommend that you use the [latest version of Docker Compose](https://docs.docker.com/compose/install/).
|
||||
|
||||
Quick Start
|
||||
-----------
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
|
||||
def dockerVersions = ['19.03.13', '18.09.9']
|
||||
def baseImages = ['alpine', 'debian']
|
||||
def pythonVersions = ['py39']
|
||||
def pythonVersions = ['py37']
|
||||
|
||||
pipeline {
|
||||
agent none
|
||||
@@ -23,7 +23,7 @@ pipeline {
|
||||
parallel {
|
||||
stage('alpine') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
buildImage('alpine')
|
||||
@@ -31,7 +31,7 @@ pipeline {
|
||||
}
|
||||
stage('debian') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
buildImage('debian')
|
||||
@@ -41,7 +41,7 @@ pipeline {
|
||||
}
|
||||
stage('Test') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
// TODO use declarative 1.5.0 `matrix` once available on CI
|
||||
@@ -61,7 +61,7 @@ pipeline {
|
||||
}
|
||||
stage('Generate Changelog') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
checkout scm
|
||||
@@ -98,7 +98,7 @@ pipeline {
|
||||
}
|
||||
stage('linux binary') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
checkout scm
|
||||
@@ -134,7 +134,7 @@ pipeline {
|
||||
}
|
||||
stage('alpine image') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
buildRuntimeImage('alpine')
|
||||
@@ -142,7 +142,7 @@ pipeline {
|
||||
}
|
||||
stage('debian image') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
buildRuntimeImage('debian')
|
||||
@@ -157,7 +157,7 @@ pipeline {
|
||||
parallel {
|
||||
stage('Pushing images') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
steps {
|
||||
pushRuntimeImage('alpine')
|
||||
@@ -166,7 +166,7 @@ pipeline {
|
||||
}
|
||||
stage('Creating Github Release') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
environment {
|
||||
GITHUB_TOKEN = credentials('github-release-token')
|
||||
@@ -198,7 +198,7 @@ pipeline {
|
||||
}
|
||||
stage('Publishing Python packages') {
|
||||
agent {
|
||||
label 'linux && docker && ubuntu-2004'
|
||||
label 'linux && docker && ubuntu-2004 && amd64 && cgroup1'
|
||||
}
|
||||
environment {
|
||||
PYPIRC = credentials('pypirc-docker-dsg-cibot')
|
||||
@@ -222,7 +222,7 @@ pipeline {
|
||||
|
||||
def buildImage(baseImage) {
|
||||
def scmvar = checkout(scm)
|
||||
def imageName = "dockerbuildbot/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
def imageName = "dockerpinata/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
image = docker.image(imageName)
|
||||
|
||||
withDockerRegistry(credentialsId:'dockerbuildbot-index.docker.io') {
|
||||
@@ -247,9 +247,9 @@ def buildImage(baseImage) {
|
||||
def runTests(dockerVersion, pythonVersion, baseImage) {
|
||||
return {
|
||||
stage("python=${pythonVersion} docker=${dockerVersion} ${baseImage}") {
|
||||
node("linux && docker && ubuntu-2004") {
|
||||
node("linux && docker && ubuntu-2004 && amd64 && cgroup1") {
|
||||
def scmvar = checkout(scm)
|
||||
def imageName = "dockerbuildbot/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
def imageName = "dockerpinata/compose:${baseImage}-${scmvar.GIT_COMMIT}"
|
||||
def storageDriver = sh(script: "docker info -f \'{{.Driver}}\'", returnStdout: true).trim()
|
||||
echo "Using local system's storage driver: ${storageDriver}"
|
||||
withDockerRegistry(credentialsId:'dockerbuildbot-index.docker.io') {
|
||||
@@ -259,6 +259,8 @@ def runTests(dockerVersion, pythonVersion, baseImage) {
|
||||
--privileged \\
|
||||
--volume="\$(pwd)/.git:/code/.git" \\
|
||||
--volume="/var/run/docker.sock:/var/run/docker.sock" \\
|
||||
--volume="\${DOCKER_CONFIG}/config.json:/root/.docker/config.json" \\
|
||||
-e "DOCKER_TLS_CERTDIR=" \\
|
||||
-e "TAG=${imageName}" \\
|
||||
-e "STORAGE_DRIVER=${storageDriver}" \\
|
||||
-e "DOCKER_VERSIONS=${dockerVersion}" \\
|
||||
|
||||
@@ -1 +1 @@
|
||||
__version__ = '1.28.0dev'
|
||||
__version__ = '1.30.0dev'
|
||||
|
||||
@@ -1,3 +1,6 @@
|
||||
import enum
|
||||
import os
|
||||
|
||||
from ..const import IS_WINDOWS_PLATFORM
|
||||
|
||||
NAMES = [
|
||||
@@ -12,6 +15,21 @@ NAMES = [
|
||||
]
|
||||
|
||||
|
||||
@enum.unique
|
||||
class AnsiMode(enum.Enum):
|
||||
"""Enumeration for when to output ANSI colors."""
|
||||
NEVER = "never"
|
||||
ALWAYS = "always"
|
||||
AUTO = "auto"
|
||||
|
||||
def use_ansi_codes(self, stream):
|
||||
if self is AnsiMode.ALWAYS:
|
||||
return True
|
||||
if self is AnsiMode.NEVER or os.environ.get('CLICOLOR') == '0':
|
||||
return False
|
||||
return stream.isatty()
|
||||
|
||||
|
||||
def get_pairs():
|
||||
for i, name in enumerate(NAMES):
|
||||
yield (name, str(30 + i))
|
||||
|
||||
@@ -129,7 +129,7 @@ def get_profiles_from_options(options, environment):
|
||||
if profile_option:
|
||||
return profile_option
|
||||
|
||||
profiles = environment.get('COMPOSE_PROFILE')
|
||||
profiles = environment.get('COMPOSE_PROFILES')
|
||||
if profiles:
|
||||
return profiles.split(',')
|
||||
|
||||
|
||||
@@ -17,10 +17,16 @@ class DocoptDispatcher:
|
||||
self.command_class = command_class
|
||||
self.options = options
|
||||
|
||||
@classmethod
|
||||
def get_command_and_options(cls, doc_entity, argv, options):
|
||||
command_help = getdoc(doc_entity)
|
||||
opt = docopt_full_help(command_help, argv, **options)
|
||||
command = opt['COMMAND']
|
||||
return command_help, opt, command
|
||||
|
||||
def parse(self, argv):
|
||||
command_help = getdoc(self.command_class)
|
||||
options = docopt_full_help(command_help, argv, **self.options)
|
||||
command = options['COMMAND']
|
||||
command_help, options, command = DocoptDispatcher.get_command_and_options(
|
||||
self.command_class, argv, self.options)
|
||||
|
||||
if command is None:
|
||||
raise SystemExit(command_help)
|
||||
|
||||
@@ -158,10 +158,8 @@ class QueueItem(namedtuple('_QueueItem', 'item is_stop exc')):
|
||||
|
||||
|
||||
def tail_container_logs(container, presenter, queue, log_args):
|
||||
generator = get_log_generator(container)
|
||||
|
||||
try:
|
||||
for item in generator(container, log_args):
|
||||
for item in build_log_generator(container, log_args):
|
||||
queue.put(QueueItem.new(presenter.present(container, item)))
|
||||
except Exception as e:
|
||||
queue.put(QueueItem.exception(e))
|
||||
@@ -171,20 +169,6 @@ def tail_container_logs(container, presenter, queue, log_args):
|
||||
queue.put(QueueItem.stop(container.name))
|
||||
|
||||
|
||||
def get_log_generator(container):
|
||||
if container.has_api_logs:
|
||||
return build_log_generator
|
||||
return build_no_log_generator
|
||||
|
||||
|
||||
def build_no_log_generator(container, log_args):
|
||||
"""Return a generator that prints a warning about logs and waits for
|
||||
container to exit.
|
||||
"""
|
||||
yield "WARNING: no logs are available with the '{}' log driver\n".format(
|
||||
container.log_driver)
|
||||
|
||||
|
||||
def build_log_generator(container, log_args):
|
||||
# if the container doesn't have a log_stream we need to attach to container
|
||||
# before log printer starts running
|
||||
|
||||
@@ -2,7 +2,6 @@ import contextlib
|
||||
import functools
|
||||
import json
|
||||
import logging
|
||||
import os
|
||||
import pipes
|
||||
import re
|
||||
import subprocess
|
||||
@@ -24,8 +23,11 @@ from ..config import resolve_build_args
|
||||
from ..config.environment import Environment
|
||||
from ..config.serialize import serialize_config
|
||||
from ..config.types import VolumeSpec
|
||||
from ..const import IS_LINUX_PLATFORM
|
||||
from ..const import IS_WINDOWS_PLATFORM
|
||||
from ..errors import StreamParseError
|
||||
from ..metrics.decorator import metrics
|
||||
from ..parallel import ParallelStreamWriter
|
||||
from ..progress_stream import StreamOutputError
|
||||
from ..project import get_image_digests
|
||||
from ..project import MissingDigests
|
||||
@@ -38,6 +40,8 @@ from ..service import ConvergenceStrategy
|
||||
from ..service import ImageType
|
||||
from ..service import NeedsBuildError
|
||||
from ..service import OperationFailedError
|
||||
from ..utils import filter_attached_for_up
|
||||
from .colors import AnsiMode
|
||||
from .command import get_config_from_options
|
||||
from .command import get_project_dir
|
||||
from .command import project_from_options
|
||||
@@ -52,60 +56,132 @@ from .log_printer import LogPrinter
|
||||
from .utils import get_version_info
|
||||
from .utils import human_readable_file_size
|
||||
from .utils import yesno
|
||||
from compose.metrics.client import MetricsCommand
|
||||
from compose.metrics.client import Status
|
||||
|
||||
|
||||
if not IS_WINDOWS_PLATFORM:
|
||||
from dockerpty.pty import PseudoTerminal, RunOperation, ExecOperation
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
console_handler = logging.StreamHandler(sys.stderr)
|
||||
|
||||
|
||||
def main():
|
||||
def main(): # noqa: C901
|
||||
signals.ignore_sigpipe()
|
||||
command = None
|
||||
try:
|
||||
command = dispatch()
|
||||
command()
|
||||
_, opts, command = DocoptDispatcher.get_command_and_options(
|
||||
TopLevelCommand,
|
||||
get_filtered_args(sys.argv[1:]),
|
||||
{'options_first': True, 'version': get_version_info('compose')})
|
||||
except Exception:
|
||||
pass
|
||||
try:
|
||||
command_func = dispatch()
|
||||
command_func()
|
||||
if not IS_LINUX_PLATFORM and command == 'help':
|
||||
print("\nDocker Compose is now in the Docker CLI, try `docker compose` help")
|
||||
except (KeyboardInterrupt, signals.ShutdownException):
|
||||
log.error("Aborting.")
|
||||
sys.exit(1)
|
||||
exit_with_metrics(command, "Aborting.", status=Status.CANCELED)
|
||||
except (UserError, NoSuchService, ConfigurationError,
|
||||
ProjectError, OperationFailedError) as e:
|
||||
log.error(e.msg)
|
||||
sys.exit(1)
|
||||
exit_with_metrics(command, e.msg, status=Status.FAILURE)
|
||||
except BuildError as e:
|
||||
reason = ""
|
||||
if e.reason:
|
||||
reason = " : " + e.reason
|
||||
log.error("Service '{}' failed to build{}".format(e.service.name, reason))
|
||||
sys.exit(1)
|
||||
exit_with_metrics(command,
|
||||
"Service '{}' failed to build{}".format(e.service.name, reason),
|
||||
status=Status.FAILURE)
|
||||
except StreamOutputError as e:
|
||||
log.error(e)
|
||||
sys.exit(1)
|
||||
exit_with_metrics(command, e, status=Status.FAILURE)
|
||||
except NeedsBuildError as e:
|
||||
log.error("Service '{}' needs to be built, but --no-build was passed.".format(e.service.name))
|
||||
sys.exit(1)
|
||||
exit_with_metrics(command,
|
||||
"Service '{}' needs to be built, but --no-build was passed.".format(
|
||||
e.service.name), status=Status.FAILURE)
|
||||
except NoSuchCommand as e:
|
||||
commands = "\n".join(parse_doc_section("commands:", getdoc(e.supercommand)))
|
||||
log.error("No such command: %s\n\n%s", e.command, commands)
|
||||
sys.exit(1)
|
||||
if not IS_LINUX_PLATFORM:
|
||||
commands += "\n\nDocker Compose is now in the Docker CLI, try `docker compose`"
|
||||
exit_with_metrics("", log_msg="No such command: {}\n\n{}".format(
|
||||
e.command, commands), status=Status.FAILURE)
|
||||
except (errors.ConnectionError, StreamParseError):
|
||||
sys.exit(1)
|
||||
exit_with_metrics(command, status=Status.FAILURE)
|
||||
except SystemExit as e:
|
||||
status = Status.SUCCESS
|
||||
if len(sys.argv) > 1 and '--help' not in sys.argv:
|
||||
status = Status.FAILURE
|
||||
|
||||
if command and len(sys.argv) >= 3 and sys.argv[2] == '--help':
|
||||
command = '--help ' + command
|
||||
|
||||
if not command and len(sys.argv) >= 2 and sys.argv[1] == '--help':
|
||||
command = '--help'
|
||||
|
||||
msg = e.args[0] if len(e.args) else ""
|
||||
code = 0
|
||||
if isinstance(e.code, int):
|
||||
code = e.code
|
||||
|
||||
if not IS_LINUX_PLATFORM and not command:
|
||||
msg += "\n\nDocker Compose is now in the Docker CLI, try `docker compose`"
|
||||
|
||||
exit_with_metrics(command, log_msg=msg, status=status,
|
||||
exit_code=code)
|
||||
|
||||
|
||||
def get_filtered_args(args):
|
||||
if args[0] in ('-h', '--help'):
|
||||
return []
|
||||
if args[0] == '--version':
|
||||
return ['version']
|
||||
|
||||
|
||||
def exit_with_metrics(command, log_msg=None, status=Status.SUCCESS, exit_code=1):
|
||||
if log_msg and command != 'exec':
|
||||
if not exit_code:
|
||||
log.info(log_msg)
|
||||
else:
|
||||
log.error(log_msg)
|
||||
|
||||
MetricsCommand(command, status=status).send_metrics()
|
||||
sys.exit(exit_code)
|
||||
|
||||
|
||||
def dispatch():
|
||||
setup_logging()
|
||||
console_stream = sys.stderr
|
||||
console_handler = logging.StreamHandler(console_stream)
|
||||
setup_logging(console_handler)
|
||||
dispatcher = DocoptDispatcher(
|
||||
TopLevelCommand,
|
||||
{'options_first': True, 'version': get_version_info('compose')})
|
||||
|
||||
options, handler, command_options = dispatcher.parse(sys.argv[1:])
|
||||
|
||||
ansi_mode = AnsiMode.AUTO
|
||||
try:
|
||||
if options.get("--ansi"):
|
||||
ansi_mode = AnsiMode(options.get("--ansi"))
|
||||
except ValueError:
|
||||
raise UserError(
|
||||
'Invalid value for --ansi: {}. Expected one of {}.'.format(
|
||||
options.get("--ansi"),
|
||||
', '.join(m.value for m in AnsiMode)
|
||||
)
|
||||
)
|
||||
if options.get("--no-ansi"):
|
||||
if options.get("--ansi"):
|
||||
raise UserError("--no-ansi and --ansi cannot be combined.")
|
||||
log.warning('--no-ansi option is deprecated and will be removed in future versions. '
|
||||
'Use `--ansi never` instead.')
|
||||
ansi_mode = AnsiMode.NEVER
|
||||
|
||||
setup_console_handler(console_handler,
|
||||
options.get('--verbose'),
|
||||
set_no_color_if_clicolor(options.get('--no-ansi')),
|
||||
ansi_mode.use_ansi_codes(console_handler.stream),
|
||||
options.get("--log-level"))
|
||||
setup_parallel_logger(set_no_color_if_clicolor(options.get('--no-ansi')))
|
||||
if options.get('--no-ansi'):
|
||||
setup_parallel_logger(ansi_mode)
|
||||
if ansi_mode is AnsiMode.NEVER:
|
||||
command_options['--no-color'] = True
|
||||
return functools.partial(perform_command, options, handler, command_options)
|
||||
|
||||
@@ -127,23 +203,23 @@ def perform_command(options, handler, command_options):
|
||||
handler(command, command_options)
|
||||
|
||||
|
||||
def setup_logging():
|
||||
def setup_logging(console_handler):
|
||||
root_logger = logging.getLogger()
|
||||
root_logger.addHandler(console_handler)
|
||||
root_logger.setLevel(logging.DEBUG)
|
||||
|
||||
# Disable requests logging
|
||||
# Disable requests and docker-py logging
|
||||
logging.getLogger("urllib3").propagate = False
|
||||
logging.getLogger("requests").propagate = False
|
||||
logging.getLogger("docker").propagate = False
|
||||
|
||||
|
||||
def setup_parallel_logger(noansi):
|
||||
if noansi:
|
||||
import compose.parallel
|
||||
compose.parallel.ParallelStreamWriter.set_noansi()
|
||||
def setup_parallel_logger(ansi_mode):
|
||||
ParallelStreamWriter.set_default_ansi_mode(ansi_mode)
|
||||
|
||||
|
||||
def setup_console_handler(handler, verbose, noansi=False, level=None):
|
||||
if handler.stream.isatty() and noansi is False:
|
||||
def setup_console_handler(handler, verbose, use_console_formatter=True, level=None):
|
||||
if use_console_formatter:
|
||||
format_class = ConsoleWarningFormatter
|
||||
else:
|
||||
format_class = logging.Formatter
|
||||
@@ -195,7 +271,8 @@ class TopLevelCommand:
|
||||
-c, --context NAME Specify a context name
|
||||
--verbose Show more output
|
||||
--log-level LEVEL Set log level (DEBUG, INFO, WARNING, ERROR, CRITICAL)
|
||||
--no-ansi Do not print ANSI control characters
|
||||
--ansi (never|always|auto) Control when to print ANSI control characters
|
||||
--no-ansi Do not print ANSI control characters (DEPRECATED)
|
||||
-v, --version Print version and exit
|
||||
-H, --host HOST Daemon socket to connect to
|
||||
|
||||
@@ -253,6 +330,7 @@ class TopLevelCommand:
|
||||
environment_file = self.toplevel_options.get('--env-file')
|
||||
return Environment.from_env_file(self.project_dir, environment_file)
|
||||
|
||||
@metrics()
|
||||
def build(self, options):
|
||||
"""
|
||||
Build or rebuild services.
|
||||
@@ -272,8 +350,6 @@ class TopLevelCommand:
|
||||
--no-rm Do not remove intermediate containers after a successful build.
|
||||
--parallel Build images in parallel.
|
||||
--progress string Set type of progress output (auto, plain, tty).
|
||||
EXPERIMENTAL flag for native builder.
|
||||
To enable, run with COMPOSE_DOCKER_CLI_BUILD=1)
|
||||
--pull Always attempt to pull a newer version of the image.
|
||||
-q, --quiet Don't print anything to STDOUT
|
||||
"""
|
||||
@@ -287,7 +363,7 @@ class TopLevelCommand:
|
||||
)
|
||||
build_args = resolve_build_args(build_args, self.toplevel_environment)
|
||||
|
||||
native_builder = self.toplevel_environment.get_boolean('COMPOSE_DOCKER_CLI_BUILD')
|
||||
native_builder = self.toplevel_environment.get_boolean('COMPOSE_DOCKER_CLI_BUILD', True)
|
||||
|
||||
self.project.build(
|
||||
service_names=options['SERVICE'],
|
||||
@@ -304,6 +380,7 @@ class TopLevelCommand:
|
||||
progress=options.get('--progress'),
|
||||
)
|
||||
|
||||
@metrics()
|
||||
def config(self, options):
|
||||
"""
|
||||
Validate and view the Compose file.
|
||||
@@ -315,6 +392,7 @@ class TopLevelCommand:
|
||||
--no-interpolate Don't interpolate environment variables.
|
||||
-q, --quiet Only validate the configuration, don't print
|
||||
anything.
|
||||
--profiles Print the profile names, one per line.
|
||||
--services Print the service names, one per line.
|
||||
--volumes Print the volume names, one per line.
|
||||
--hash="*" Print the service config hash, one per line.
|
||||
@@ -334,6 +412,15 @@ class TopLevelCommand:
|
||||
if options['--quiet']:
|
||||
return
|
||||
|
||||
if options['--profiles']:
|
||||
profiles = set()
|
||||
for service in compose_config.services:
|
||||
if 'profiles' in service:
|
||||
for profile in service['profiles']:
|
||||
profiles.add(profile)
|
||||
print('\n'.join(sorted(profiles)))
|
||||
return
|
||||
|
||||
if options['--services']:
|
||||
print('\n'.join(service['name'] for service in compose_config.services))
|
||||
return
|
||||
@@ -353,6 +440,7 @@ class TopLevelCommand:
|
||||
|
||||
print(serialize_config(compose_config, image_digests, not options['--no-interpolate']))
|
||||
|
||||
@metrics()
|
||||
def create(self, options):
|
||||
"""
|
||||
Creates containers for a service.
|
||||
@@ -381,6 +469,7 @@ class TopLevelCommand:
|
||||
do_build=build_action_from_opts(options),
|
||||
)
|
||||
|
||||
@metrics()
|
||||
def down(self, options):
|
||||
"""
|
||||
Stops containers and removes containers, networks, volumes, and images
|
||||
@@ -449,6 +538,7 @@ class TopLevelCommand:
|
||||
print(formatter(event))
|
||||
sys.stdout.flush()
|
||||
|
||||
@metrics("exec")
|
||||
def exec_command(self, options):
|
||||
"""
|
||||
Execute a command in a running container
|
||||
@@ -525,6 +615,7 @@ class TopLevelCommand:
|
||||
sys.exit(exit_code)
|
||||
|
||||
@classmethod
|
||||
@metrics()
|
||||
def help(cls, options):
|
||||
"""
|
||||
Get help on a command.
|
||||
@@ -538,6 +629,7 @@ class TopLevelCommand:
|
||||
|
||||
print(getdoc(subject))
|
||||
|
||||
@metrics()
|
||||
def images(self, options):
|
||||
"""
|
||||
List images used by the created containers.
|
||||
@@ -592,6 +684,7 @@ class TopLevelCommand:
|
||||
])
|
||||
print(Formatter.table(headers, rows))
|
||||
|
||||
@metrics()
|
||||
def kill(self, options):
|
||||
"""
|
||||
Force stop service containers.
|
||||
@@ -606,6 +699,7 @@ class TopLevelCommand:
|
||||
|
||||
self.project.kill(service_names=options['SERVICE'], signal=signal)
|
||||
|
||||
@metrics()
|
||||
def logs(self, options):
|
||||
"""
|
||||
View output from containers.
|
||||
@@ -618,7 +712,7 @@ class TopLevelCommand:
|
||||
-t, --timestamps Show timestamps.
|
||||
--tail="all" Number of lines to show from the end of the logs
|
||||
for each container.
|
||||
--no-log-prefix Don't print prefix in logs.
|
||||
--no-log-prefix Don't print prefix in logs.
|
||||
"""
|
||||
containers = self.project.containers(service_names=options['SERVICE'], stopped=True)
|
||||
|
||||
@@ -637,11 +731,12 @@ class TopLevelCommand:
|
||||
log_printer_from_project(
|
||||
self.project,
|
||||
containers,
|
||||
set_no_color_if_clicolor(options['--no-color']),
|
||||
options['--no-color'],
|
||||
log_args,
|
||||
event_stream=self.project.events(service_names=options['SERVICE']),
|
||||
keep_prefix=not options['--no-log-prefix']).run()
|
||||
|
||||
@metrics()
|
||||
def pause(self, options):
|
||||
"""
|
||||
Pause services.
|
||||
@@ -651,6 +746,7 @@ class TopLevelCommand:
|
||||
containers = self.project.pause(service_names=options['SERVICE'])
|
||||
exit_if(not containers, 'No containers to pause', 1)
|
||||
|
||||
@metrics()
|
||||
def port(self, options):
|
||||
"""
|
||||
Print the public port for a port binding.
|
||||
@@ -672,6 +768,7 @@ class TopLevelCommand:
|
||||
options['PRIVATE_PORT'],
|
||||
protocol=options.get('--protocol') or 'tcp') or '')
|
||||
|
||||
@metrics()
|
||||
def ps(self, options):
|
||||
"""
|
||||
List containers.
|
||||
@@ -681,7 +778,9 @@ class TopLevelCommand:
|
||||
Options:
|
||||
-q, --quiet Only display IDs
|
||||
--services Display services
|
||||
--filter KEY=VAL Filter services by a property
|
||||
--filter KEY=VAL Filter services by a property. KEY is either:
|
||||
1. `source` with values `image`, or `build`;
|
||||
2. `status` with values `running`, `stopped`, `paused`, or `restarted`.
|
||||
-a, --all Show all stopped containers (including those created by the run command)
|
||||
"""
|
||||
if options['--quiet'] and options['--services']:
|
||||
@@ -728,6 +827,7 @@ class TopLevelCommand:
|
||||
])
|
||||
print(Formatter.table(headers, rows))
|
||||
|
||||
@metrics()
|
||||
def pull(self, options):
|
||||
"""
|
||||
Pulls images for services defined in a Compose file, but does not start the containers.
|
||||
@@ -751,6 +851,7 @@ class TopLevelCommand:
|
||||
include_deps=options.get('--include-deps'),
|
||||
)
|
||||
|
||||
@metrics()
|
||||
def push(self, options):
|
||||
"""
|
||||
Pushes images for services.
|
||||
@@ -765,6 +866,7 @@ class TopLevelCommand:
|
||||
ignore_push_failures=options.get('--ignore-push-failures')
|
||||
)
|
||||
|
||||
@metrics()
|
||||
def rm(self, options):
|
||||
"""
|
||||
Removes stopped service containers.
|
||||
@@ -809,6 +911,7 @@ class TopLevelCommand:
|
||||
else:
|
||||
print("No stopped containers")
|
||||
|
||||
@metrics()
|
||||
def run(self, options):
|
||||
"""
|
||||
Run a one-off command on a service.
|
||||
@@ -869,6 +972,7 @@ class TopLevelCommand:
|
||||
self.toplevel_options, self.toplevel_environment
|
||||
)
|
||||
|
||||
@metrics()
|
||||
def scale(self, options):
|
||||
"""
|
||||
Set number of containers to run for a service.
|
||||
@@ -897,6 +1001,7 @@ class TopLevelCommand:
|
||||
for service_name, num in parse_scale_args(options['SERVICE=NUM']).items():
|
||||
self.project.get_service(service_name).scale(num, timeout=timeout)
|
||||
|
||||
@metrics()
|
||||
def start(self, options):
|
||||
"""
|
||||
Start existing containers.
|
||||
@@ -906,6 +1011,7 @@ class TopLevelCommand:
|
||||
containers = self.project.start(service_names=options['SERVICE'])
|
||||
exit_if(not containers, 'No containers to start', 1)
|
||||
|
||||
@metrics()
|
||||
def stop(self, options):
|
||||
"""
|
||||
Stop running containers without removing them.
|
||||
@@ -921,6 +1027,7 @@ class TopLevelCommand:
|
||||
timeout = timeout_from_opts(options)
|
||||
self.project.stop(service_names=options['SERVICE'], timeout=timeout)
|
||||
|
||||
@metrics()
|
||||
def restart(self, options):
|
||||
"""
|
||||
Restart running containers.
|
||||
@@ -935,6 +1042,7 @@ class TopLevelCommand:
|
||||
containers = self.project.restart(service_names=options['SERVICE'], timeout=timeout)
|
||||
exit_if(not containers, 'No containers to restart', 1)
|
||||
|
||||
@metrics()
|
||||
def top(self, options):
|
||||
"""
|
||||
Display the running processes
|
||||
@@ -962,6 +1070,7 @@ class TopLevelCommand:
|
||||
print(container.name)
|
||||
print(Formatter.table(headers, rows))
|
||||
|
||||
@metrics()
|
||||
def unpause(self, options):
|
||||
"""
|
||||
Unpause services.
|
||||
@@ -971,6 +1080,7 @@ class TopLevelCommand:
|
||||
containers = self.project.unpause(service_names=options['SERVICE'])
|
||||
exit_if(not containers, 'No containers to unpause', 1)
|
||||
|
||||
@metrics()
|
||||
def up(self, options):
|
||||
"""
|
||||
Builds, (re)creates, starts, and attaches to containers for a service.
|
||||
@@ -1022,7 +1132,7 @@ class TopLevelCommand:
|
||||
container. Implies --abort-on-container-exit.
|
||||
--scale SERVICE=NUM Scale SERVICE to NUM instances. Overrides the
|
||||
`scale` setting in the Compose file if present.
|
||||
--no-log-prefix Don't print prefix in logs.
|
||||
--no-log-prefix Don't print prefix in logs.
|
||||
"""
|
||||
start_deps = not options['--no-deps']
|
||||
always_recreate_deps = options['--always-recreate-deps']
|
||||
@@ -1034,7 +1144,7 @@ class TopLevelCommand:
|
||||
detached = options.get('--detach')
|
||||
no_start = options.get('--no-start')
|
||||
attach_dependencies = options.get('--attach-dependencies')
|
||||
keep_prefix = not options['--no-log-prefix']
|
||||
keep_prefix = not options.get('--no-log-prefix')
|
||||
|
||||
if detached and (cascade_stop or exit_value_from or attach_dependencies):
|
||||
raise UserError(
|
||||
@@ -1049,7 +1159,7 @@ class TopLevelCommand:
|
||||
for excluded in [x for x in opts if options.get(x) and no_start]:
|
||||
raise UserError('--no-start and {} cannot be combined.'.format(excluded))
|
||||
|
||||
native_builder = self.toplevel_environment.get_boolean('COMPOSE_DOCKER_CLI_BUILD')
|
||||
native_builder = self.toplevel_environment.get_boolean('COMPOSE_DOCKER_CLI_BUILD', True)
|
||||
|
||||
with up_shutdown_context(self.project, service_names, timeout, detached):
|
||||
warn_for_swarm_mode(self.project.client)
|
||||
@@ -1071,6 +1181,7 @@ class TopLevelCommand:
|
||||
renew_anonymous_volumes=options.get('--renew-anon-volumes'),
|
||||
silent=options.get('--quiet-pull'),
|
||||
cli=native_builder,
|
||||
attach_dependencies=attach_dependencies,
|
||||
)
|
||||
|
||||
try:
|
||||
@@ -1098,7 +1209,7 @@ class TopLevelCommand:
|
||||
log_printer = log_printer_from_project(
|
||||
self.project,
|
||||
attached_containers,
|
||||
set_no_color_if_clicolor(options['--no-color']),
|
||||
options['--no-color'],
|
||||
{'follow': True},
|
||||
cascade_stop,
|
||||
event_stream=self.project.events(service_names=service_names),
|
||||
@@ -1120,6 +1231,7 @@ class TopLevelCommand:
|
||||
sys.exit(exit_code)
|
||||
|
||||
@classmethod
|
||||
@metrics()
|
||||
def version(cls, options):
|
||||
"""
|
||||
Show version information and quit.
|
||||
@@ -1393,7 +1505,7 @@ def log_printer_from_project(
|
||||
keep_prefix=True,
|
||||
):
|
||||
return LogPrinter(
|
||||
containers,
|
||||
[c for c in containers if c.log_driver not in (None, 'none')],
|
||||
build_log_presenters(project.service_names, monochrome, keep_prefix),
|
||||
event_stream or project.events(),
|
||||
cascade_stop=cascade_stop,
|
||||
@@ -1401,13 +1513,11 @@ def log_printer_from_project(
|
||||
|
||||
|
||||
def filter_attached_containers(containers, service_names, attach_dependencies=False):
|
||||
if attach_dependencies or not service_names:
|
||||
return containers
|
||||
|
||||
return [
|
||||
container
|
||||
for container in containers if container.service in service_names
|
||||
]
|
||||
return filter_attached_for_up(
|
||||
containers,
|
||||
service_names,
|
||||
attach_dependencies,
|
||||
lambda container: container.service)
|
||||
|
||||
|
||||
@contextlib.contextmanager
|
||||
@@ -1583,7 +1693,3 @@ def warn_for_swarm_mode(client):
|
||||
"To deploy your application across the swarm, "
|
||||
"use `docker stack deploy`.\n"
|
||||
)
|
||||
|
||||
|
||||
def set_no_color_if_clicolor(no_color_flag):
|
||||
return no_color_flag or os.environ.get('CLICOLOR') == "0"
|
||||
|
||||
@@ -87,7 +87,7 @@
|
||||
"dockerfile": {"type": "string"},
|
||||
"args": {"$ref": "#/definitions/list_or_dict"},
|
||||
"labels": {"$ref": "#/definitions/list_or_dict"},
|
||||
"cache_from": {"$ref": "#/definitions/list_of_strings"},
|
||||
"cache_from": {"type": "array", "items": {"type": "string"}},
|
||||
"network": {"type": "string"},
|
||||
"target": {"type": "string"},
|
||||
"shm_size": {"type": ["integer", "string"]},
|
||||
@@ -188,7 +188,7 @@
|
||||
"properties": {
|
||||
"condition": {
|
||||
"type": "string",
|
||||
"enum": ["service_started", "service_healthy"]
|
||||
"enum": ["service_started", "service_healthy", "service_completed_successfully"]
|
||||
}
|
||||
},
|
||||
"required": ["condition"]
|
||||
@@ -330,12 +330,11 @@
|
||||
"privileged": {"type": "boolean"},
|
||||
"profiles": {"$ref": "#/definitions/list_of_strings"},
|
||||
"pull_policy": {"type": "string", "enum": [
|
||||
"always", "never", "if_not_present"
|
||||
"always", "never", "if_not_present", "build"
|
||||
]},
|
||||
"read_only": {"type": "boolean"},
|
||||
"restart": {"type": "string"},
|
||||
"runtime": {
|
||||
"deprecated": true,
|
||||
"type": "string"
|
||||
},
|
||||
"scale": {
|
||||
@@ -367,6 +366,7 @@
|
||||
"stdin_open": {"type": "boolean"},
|
||||
"stop_grace_period": {"type": "string", "format": "duration"},
|
||||
"stop_signal": {"type": "string"},
|
||||
"storage_opt": {"type": "object"},
|
||||
"tmpfs": {"$ref": "#/definitions/string_or_list"},
|
||||
"tty": {"type": "boolean"},
|
||||
"ulimits": {
|
||||
|
||||
@@ -10,7 +10,11 @@ from operator import attrgetter
|
||||
from operator import itemgetter
|
||||
|
||||
import yaml
|
||||
from cached_property import cached_property
|
||||
|
||||
try:
|
||||
from functools import cached_property
|
||||
except ImportError:
|
||||
from cached_property import cached_property
|
||||
|
||||
from . import types
|
||||
from ..const import COMPOSE_SPEC as VERSION
|
||||
@@ -149,9 +153,14 @@ DOCKER_VALID_URL_PREFIXES = (
|
||||
SUPPORTED_FILENAMES = [
|
||||
'docker-compose.yml',
|
||||
'docker-compose.yaml',
|
||||
'compose.yml',
|
||||
'compose.yaml',
|
||||
]
|
||||
|
||||
DEFAULT_OVERRIDE_FILENAMES = ('docker-compose.override.yml', 'docker-compose.override.yaml')
|
||||
DEFAULT_OVERRIDE_FILENAMES = ('docker-compose.override.yml',
|
||||
'docker-compose.override.yaml',
|
||||
'compose.override.yml',
|
||||
'compose.override.yaml')
|
||||
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
@@ -304,7 +313,16 @@ def find(base_dir, filenames, environment, override_dir=None):
|
||||
if filenames:
|
||||
filenames = [os.path.join(base_dir, f) for f in filenames]
|
||||
else:
|
||||
# search for compose files in the base dir and its parents
|
||||
filenames = get_default_config_files(base_dir)
|
||||
if not filenames and not override_dir:
|
||||
# none found in base_dir and no override_dir defined
|
||||
raise ComposeFileNotFound(SUPPORTED_FILENAMES)
|
||||
if not filenames:
|
||||
# search for compose files in the project directory and its parents
|
||||
filenames = get_default_config_files(override_dir)
|
||||
if not filenames:
|
||||
raise ComposeFileNotFound(SUPPORTED_FILENAMES)
|
||||
|
||||
log.debug("Using configuration files: {}".format(",".join(filenames)))
|
||||
return ConfigDetails(
|
||||
@@ -335,7 +353,7 @@ def get_default_config_files(base_dir):
|
||||
(candidates, path) = find_candidates_in_parent_dirs(SUPPORTED_FILENAMES, base_dir)
|
||||
|
||||
if not candidates:
|
||||
raise ComposeFileNotFound(SUPPORTED_FILENAMES)
|
||||
return None
|
||||
|
||||
winner = candidates[0]
|
||||
|
||||
@@ -556,8 +574,7 @@ def process_config_section(config_file, config, section, environment, interpolat
|
||||
config_file.version,
|
||||
config,
|
||||
section,
|
||||
environment
|
||||
)
|
||||
environment)
|
||||
else:
|
||||
return config
|
||||
|
||||
|
||||
@@ -54,9 +54,10 @@ class Environment(dict):
|
||||
if base_dir is None:
|
||||
return result
|
||||
if env_file:
|
||||
env_file_path = os.path.join(base_dir, env_file)
|
||||
else:
|
||||
env_file_path = os.path.join(base_dir, '.env')
|
||||
env_file_path = os.path.join(os.getcwd(), env_file)
|
||||
return cls(env_vars_from_file(env_file_path))
|
||||
|
||||
env_file_path = os.path.join(base_dir, '.env')
|
||||
try:
|
||||
return cls(env_vars_from_file(env_file_path))
|
||||
except EnvFileNotFound:
|
||||
@@ -113,13 +114,13 @@ class Environment(dict):
|
||||
)
|
||||
return super().get(key, *args, **kwargs)
|
||||
|
||||
def get_boolean(self, key):
|
||||
def get_boolean(self, key, default=False):
|
||||
# Convert a value to a boolean using "common sense" rules.
|
||||
# Unset, empty, "0" and "false" (i-case) yield False.
|
||||
# All other values yield True.
|
||||
value = self.get(key)
|
||||
if not value:
|
||||
return False
|
||||
return default
|
||||
if value.lower() in ['0', 'false']:
|
||||
return False
|
||||
return True
|
||||
|
||||
@@ -243,6 +243,7 @@ class ConversionMap:
|
||||
service_path('healthcheck', 'disable'): to_boolean,
|
||||
service_path('deploy', 'labels', PATH_JOKER): to_str,
|
||||
service_path('deploy', 'replicas'): to_int,
|
||||
service_path('deploy', 'placement', 'max_replicas_per_node'): to_int,
|
||||
service_path('deploy', 'resources', 'limits', "cpus"): to_float,
|
||||
service_path('deploy', 'update_config', 'parallelism'): to_int,
|
||||
service_path('deploy', 'update_config', 'max_failure_ratio'): to_float,
|
||||
|
||||
@@ -5,6 +5,7 @@ from .version import ComposeVersion
|
||||
DEFAULT_TIMEOUT = 10
|
||||
HTTP_TIMEOUT = 60
|
||||
IS_WINDOWS_PLATFORM = (sys.platform == "win32")
|
||||
IS_LINUX_PLATFORM = (sys.platform == "linux")
|
||||
LABEL_CONTAINER_NUMBER = 'com.docker.compose.container-number'
|
||||
LABEL_ONE_OFF = 'com.docker.compose.oneoff'
|
||||
LABEL_PROJECT = 'com.docker.compose.project'
|
||||
|
||||
@@ -186,11 +186,6 @@ class Container:
|
||||
def log_driver(self):
|
||||
return self.get('HostConfig.LogConfig.Type')
|
||||
|
||||
@property
|
||||
def has_api_logs(self):
|
||||
log_type = self.log_driver
|
||||
return not log_type or log_type in ('json-file', 'journald', 'local')
|
||||
|
||||
@property
|
||||
def human_readable_health_status(self):
|
||||
""" Generate UP status string with up time and health
|
||||
@@ -204,11 +199,7 @@ class Container:
|
||||
return status_string
|
||||
|
||||
def attach_log_stream(self):
|
||||
"""A log stream can only be attached if the container uses a
|
||||
json-file, journald or local log driver.
|
||||
"""
|
||||
if self.has_api_logs:
|
||||
self.log_stream = self.attach(stdout=True, stderr=True, stream=True)
|
||||
self.log_stream = self.attach(stdout=True, stderr=True, stream=True)
|
||||
|
||||
def get(self, key):
|
||||
"""Return a value from the container or None if the value is not set.
|
||||
|
||||
@@ -27,3 +27,8 @@ class NoHealthCheckConfigured(HealthCheckException):
|
||||
service_name
|
||||
)
|
||||
)
|
||||
|
||||
|
||||
class CompletedUnsuccessfully(Exception):
|
||||
def __init__(self, container_id, exit_code):
|
||||
self.msg = 'Container "{}" exited with code {}.'.format(container_id, exit_code)
|
||||
|
||||
0
compose/metrics/__init__.py
Normal file
0
compose/metrics/__init__.py
Normal file
64
compose/metrics/client.py
Normal file
64
compose/metrics/client.py
Normal file
@@ -0,0 +1,64 @@
|
||||
import os
|
||||
from enum import Enum
|
||||
|
||||
import requests
|
||||
from docker import ContextAPI
|
||||
from docker.transport import UnixHTTPAdapter
|
||||
|
||||
from compose.const import IS_WINDOWS_PLATFORM
|
||||
|
||||
if IS_WINDOWS_PLATFORM:
|
||||
from docker.transport import NpipeHTTPAdapter
|
||||
|
||||
|
||||
class Status(Enum):
|
||||
SUCCESS = "success"
|
||||
FAILURE = "failure"
|
||||
CANCELED = "canceled"
|
||||
|
||||
|
||||
class MetricsSource:
|
||||
CLI = "docker-compose"
|
||||
|
||||
|
||||
if IS_WINDOWS_PLATFORM:
|
||||
METRICS_SOCKET_FILE = 'npipe://\\\\.\\pipe\\docker_cli'
|
||||
else:
|
||||
METRICS_SOCKET_FILE = 'http+unix:///var/run/docker-cli.sock'
|
||||
|
||||
|
||||
class MetricsCommand(requests.Session):
|
||||
"""
|
||||
Representation of a command in the metrics.
|
||||
"""
|
||||
|
||||
def __init__(self, command,
|
||||
context_type=None, status=Status.SUCCESS,
|
||||
source=MetricsSource.CLI, uri=None):
|
||||
super().__init__()
|
||||
self.command = ("compose " + command).strip() if command else "compose --help"
|
||||
self.context = context_type or ContextAPI.get_current_context().context_type or 'moby'
|
||||
self.source = source
|
||||
self.status = status.value
|
||||
self.uri = uri or os.environ.get("METRICS_SOCKET_FILE", METRICS_SOCKET_FILE)
|
||||
if IS_WINDOWS_PLATFORM:
|
||||
self.mount("http+unix://", NpipeHTTPAdapter(self.uri))
|
||||
else:
|
||||
self.mount("http+unix://", UnixHTTPAdapter(self.uri))
|
||||
|
||||
def send_metrics(self):
|
||||
try:
|
||||
return self.post("http+unix://localhost/usage",
|
||||
json=self.to_map(),
|
||||
timeout=.05,
|
||||
headers={'Content-Type': 'application/json'})
|
||||
except Exception as e:
|
||||
return e
|
||||
|
||||
def to_map(self):
|
||||
return {
|
||||
'command': self.command,
|
||||
'context': self.context,
|
||||
'source': self.source,
|
||||
'status': self.status,
|
||||
}
|
||||
21
compose/metrics/decorator.py
Normal file
21
compose/metrics/decorator.py
Normal file
@@ -0,0 +1,21 @@
|
||||
import functools
|
||||
|
||||
from compose.metrics.client import MetricsCommand
|
||||
from compose.metrics.client import Status
|
||||
|
||||
|
||||
class metrics:
|
||||
def __init__(self, command_name=None):
|
||||
self.command_name = command_name
|
||||
|
||||
def __call__(self, fn):
|
||||
@functools.wraps(fn,
|
||||
assigned=functools.WRAPPER_ASSIGNMENTS,
|
||||
updated=functools.WRAPPER_UPDATES)
|
||||
def wrapper(*args, **kwargs):
|
||||
if not self.command_name:
|
||||
self.command_name = fn.__name__
|
||||
result = fn(*args, **kwargs)
|
||||
MetricsCommand(self.command_name, status=Status.SUCCESS).send_metrics()
|
||||
return result
|
||||
return wrapper
|
||||
@@ -11,10 +11,12 @@ from threading import Thread
|
||||
from docker.errors import APIError
|
||||
from docker.errors import ImageNotFound
|
||||
|
||||
from compose.cli.colors import AnsiMode
|
||||
from compose.cli.colors import green
|
||||
from compose.cli.colors import red
|
||||
from compose.cli.signals import ShutdownException
|
||||
from compose.const import PARALLEL_LIMIT
|
||||
from compose.errors import CompletedUnsuccessfully
|
||||
from compose.errors import HealthCheckFailed
|
||||
from compose.errors import NoHealthCheckConfigured
|
||||
from compose.errors import OperationFailedError
|
||||
@@ -60,7 +62,8 @@ def parallel_execute_watch(events, writer, errors, results, msg, get_name, fail_
|
||||
elif isinstance(exception, APIError):
|
||||
errors[get_name(obj)] = exception.explanation
|
||||
writer.write(msg, get_name(obj), 'error', red)
|
||||
elif isinstance(exception, (OperationFailedError, HealthCheckFailed, NoHealthCheckConfigured)):
|
||||
elif isinstance(exception, (OperationFailedError, HealthCheckFailed, NoHealthCheckConfigured,
|
||||
CompletedUnsuccessfully)):
|
||||
errors[get_name(obj)] = exception.msg
|
||||
writer.write(msg, get_name(obj), 'error', red)
|
||||
elif isinstance(exception, UpstreamError):
|
||||
@@ -83,10 +86,7 @@ def parallel_execute(objects, func, get_name, msg, get_deps=None, limit=None, fa
|
||||
objects = list(objects)
|
||||
stream = sys.stderr
|
||||
|
||||
if ParallelStreamWriter.instance:
|
||||
writer = ParallelStreamWriter.instance
|
||||
else:
|
||||
writer = ParallelStreamWriter(stream)
|
||||
writer = ParallelStreamWriter.get_or_assign_instance(ParallelStreamWriter(stream))
|
||||
|
||||
for obj in objects:
|
||||
writer.add_object(msg, get_name(obj))
|
||||
@@ -243,6 +243,12 @@ def feed_queue(objects, func, get_deps, results, state, limiter):
|
||||
'not processing'.format(obj)
|
||||
)
|
||||
results.put((obj, None, e))
|
||||
except CompletedUnsuccessfully as e:
|
||||
log.debug(
|
||||
'Service(s) upstream of {} did not completed successfully - '
|
||||
'not processing'.format(obj)
|
||||
)
|
||||
results.put((obj, None, e))
|
||||
|
||||
if state.is_done():
|
||||
results.put(STOP)
|
||||
@@ -259,19 +265,37 @@ class ParallelStreamWriter:
|
||||
to jump to the correct line, and write over the line.
|
||||
"""
|
||||
|
||||
noansi = False
|
||||
lock = Lock()
|
||||
default_ansi_mode = AnsiMode.AUTO
|
||||
write_lock = Lock()
|
||||
|
||||
instance = None
|
||||
instance_lock = Lock()
|
||||
|
||||
@classmethod
|
||||
def set_noansi(cls, value=True):
|
||||
cls.noansi = value
|
||||
def get_instance(cls):
|
||||
return cls.instance
|
||||
|
||||
def __init__(self, stream):
|
||||
@classmethod
|
||||
def get_or_assign_instance(cls, writer):
|
||||
cls.instance_lock.acquire()
|
||||
try:
|
||||
if cls.instance is None:
|
||||
cls.instance = writer
|
||||
return cls.instance
|
||||
finally:
|
||||
cls.instance_lock.release()
|
||||
|
||||
@classmethod
|
||||
def set_default_ansi_mode(cls, ansi_mode):
|
||||
cls.default_ansi_mode = ansi_mode
|
||||
|
||||
def __init__(self, stream, ansi_mode=None):
|
||||
if ansi_mode is None:
|
||||
ansi_mode = self.default_ansi_mode
|
||||
self.stream = stream
|
||||
self.use_ansi_codes = ansi_mode.use_ansi_codes(stream)
|
||||
self.lines = []
|
||||
self.width = 0
|
||||
ParallelStreamWriter.instance = self
|
||||
|
||||
def add_object(self, msg, obj_index):
|
||||
if msg is None:
|
||||
@@ -285,7 +309,7 @@ class ParallelStreamWriter:
|
||||
return self._write_noansi(msg, obj_index, '')
|
||||
|
||||
def _write_ansi(self, msg, obj_index, status):
|
||||
self.lock.acquire()
|
||||
self.write_lock.acquire()
|
||||
position = self.lines.index(msg + obj_index)
|
||||
diff = len(self.lines) - position
|
||||
# move up
|
||||
@@ -297,7 +321,7 @@ class ParallelStreamWriter:
|
||||
# move back down
|
||||
self.stream.write("%c[%dB" % (27, diff))
|
||||
self.stream.flush()
|
||||
self.lock.release()
|
||||
self.write_lock.release()
|
||||
|
||||
def _write_noansi(self, msg, obj_index, status):
|
||||
self.stream.write(
|
||||
@@ -310,17 +334,10 @@ class ParallelStreamWriter:
|
||||
def write(self, msg, obj_index, status, color_func):
|
||||
if msg is None:
|
||||
return
|
||||
if self.noansi:
|
||||
self._write_noansi(msg, obj_index, status)
|
||||
else:
|
||||
if self.use_ansi_codes:
|
||||
self._write_ansi(msg, obj_index, color_func(status))
|
||||
|
||||
|
||||
def get_stream_writer():
|
||||
instance = ParallelStreamWriter.instance
|
||||
if instance is None:
|
||||
raise RuntimeError('ParallelStreamWriter has not yet been instantiated')
|
||||
return instance
|
||||
else:
|
||||
self._write_noansi(msg, obj_index, status)
|
||||
|
||||
|
||||
def parallel_operation(containers, operation, options, message):
|
||||
|
||||
@@ -39,6 +39,7 @@ from .service import Service
|
||||
from .service import ServiceIpcMode
|
||||
from .service import ServiceNetworkMode
|
||||
from .service import ServicePidMode
|
||||
from .utils import filter_attached_for_up
|
||||
from .utils import microseconds_from_time_nano
|
||||
from .utils import truncate_string
|
||||
from .volume import ProjectVolumes
|
||||
@@ -489,7 +490,6 @@ class Project:
|
||||
log.info('%s uses an image, skipping' % service.name)
|
||||
|
||||
if cli:
|
||||
log.warning("Native build is an experimental feature and could change at any time")
|
||||
if parallel_build:
|
||||
log.warning("Flag '--parallel' is ignored when building with "
|
||||
"COMPOSE_DOCKER_CLI_BUILD=1")
|
||||
@@ -645,12 +645,10 @@ class Project:
|
||||
silent=False,
|
||||
cli=False,
|
||||
one_off=False,
|
||||
attach_dependencies=False,
|
||||
override_options=None,
|
||||
):
|
||||
|
||||
if cli:
|
||||
log.warning("Native build is an experimental feature and could change at any time")
|
||||
|
||||
self.initialize()
|
||||
if not ignore_orphans:
|
||||
self.find_orphan_containers(remove_orphans)
|
||||
@@ -671,12 +669,17 @@ class Project:
|
||||
one_off=service_names if one_off else [],
|
||||
)
|
||||
|
||||
def do(service):
|
||||
services_to_attach = filter_attached_for_up(
|
||||
services,
|
||||
service_names,
|
||||
attach_dependencies,
|
||||
lambda service: service.name)
|
||||
|
||||
def do(service):
|
||||
return service.execute_convergence_plan(
|
||||
plans[service.name],
|
||||
timeout=timeout,
|
||||
detached=detached,
|
||||
detached=detached or (service not in services_to_attach),
|
||||
scale_override=scale_override.get(service.name),
|
||||
rescale=rescale,
|
||||
start=start,
|
||||
@@ -780,7 +783,9 @@ class Project:
|
||||
return
|
||||
|
||||
try:
|
||||
writer = parallel.get_stream_writer()
|
||||
writer = parallel.ParallelStreamWriter.get_instance()
|
||||
if writer is None:
|
||||
raise RuntimeError('ParallelStreamWriter has not yet been instantiated')
|
||||
for event in strm:
|
||||
if 'status' not in event:
|
||||
continue
|
||||
|
||||
@@ -1,6 +1,5 @@
|
||||
import enum
|
||||
import itertools
|
||||
import json
|
||||
import logging
|
||||
import os
|
||||
import re
|
||||
@@ -45,6 +44,7 @@ from .const import LABEL_VERSION
|
||||
from .const import NANOCPUS_SCALE
|
||||
from .const import WINDOWS_LONGPATH_PREFIX
|
||||
from .container import Container
|
||||
from .errors import CompletedUnsuccessfully
|
||||
from .errors import HealthCheckFailed
|
||||
from .errors import NoHealthCheckConfigured
|
||||
from .errors import OperationFailedError
|
||||
@@ -112,6 +112,7 @@ HOST_CONFIG_KEYS = [
|
||||
|
||||
CONDITION_STARTED = 'service_started'
|
||||
CONDITION_HEALTHY = 'service_healthy'
|
||||
CONDITION_COMPLETED_SUCCESSFULLY = 'service_completed_successfully'
|
||||
|
||||
|
||||
class BuildError(Exception):
|
||||
@@ -712,6 +713,7 @@ class Service:
|
||||
'image_id': image_id(),
|
||||
'links': self.get_link_names(),
|
||||
'net': self.network_mode.id,
|
||||
'ipc_mode': self.ipc_mode.mode,
|
||||
'networks': self.networks,
|
||||
'secrets': self.secrets,
|
||||
'volumes_from': [
|
||||
@@ -753,6 +755,8 @@ class Service:
|
||||
configs[svc] = lambda s: True
|
||||
elif config['condition'] == CONDITION_HEALTHY:
|
||||
configs[svc] = lambda s: s.is_healthy()
|
||||
elif config['condition'] == CONDITION_COMPLETED_SUCCESSFULLY:
|
||||
configs[svc] = lambda s: s.is_completed_successfully()
|
||||
else:
|
||||
# The config schema already prevents this, but it might be
|
||||
# bypassed if Compose is called programmatically.
|
||||
@@ -1103,8 +1107,9 @@ class Service:
|
||||
'Impossible to perform platform-targeted builds for API version < 1.35'
|
||||
)
|
||||
|
||||
builder = self.client if not cli else _CLIBuilder(progress)
|
||||
build_output = builder.build(
|
||||
builder = _ClientBuilder(self.client) if not cli else _CLIBuilder(progress)
|
||||
return builder.build(
|
||||
service=self,
|
||||
path=path,
|
||||
tag=self.image_name,
|
||||
rm=rm,
|
||||
@@ -1125,30 +1130,7 @@ class Service:
|
||||
gzip=gzip,
|
||||
isolation=build_opts.get('isolation', self.options.get('isolation', None)),
|
||||
platform=self.platform,
|
||||
)
|
||||
|
||||
try:
|
||||
all_events = list(stream_output(build_output, output_stream))
|
||||
except StreamOutputError as e:
|
||||
raise BuildError(self, str(e))
|
||||
|
||||
# Ensure the HTTP connection is not reused for another
|
||||
# streaming command, as the Docker daemon can sometimes
|
||||
# complain about it
|
||||
self.client.close()
|
||||
|
||||
image_id = None
|
||||
|
||||
for event in all_events:
|
||||
if 'stream' in event:
|
||||
match = re.search(r'Successfully built ([0-9a-f]+)', event.get('stream', ''))
|
||||
if match:
|
||||
image_id = match.group(1)
|
||||
|
||||
if image_id is None:
|
||||
raise BuildError(self, event if all_events else 'Unknown')
|
||||
|
||||
return image_id
|
||||
output_stream=output_stream)
|
||||
|
||||
def get_cache_from(self, build_opts):
|
||||
cache_from = build_opts.get('cache_from', None)
|
||||
@@ -1304,6 +1286,21 @@ class Service:
|
||||
raise HealthCheckFailed(ctnr.short_id)
|
||||
return result
|
||||
|
||||
def is_completed_successfully(self):
|
||||
""" Check that all containers for this service has completed successfully
|
||||
Returns false if at least one container does not exited and
|
||||
raises CompletedUnsuccessfully exception if at least one container
|
||||
exited with non-zero exit code.
|
||||
"""
|
||||
result = True
|
||||
for ctnr in self.containers(stopped=True):
|
||||
ctnr.inspect()
|
||||
if ctnr.get('State.Status') != 'exited':
|
||||
result = False
|
||||
elif ctnr.exit_code != 0:
|
||||
raise CompletedUnsuccessfully(ctnr.short_id, ctnr.exit_code)
|
||||
return result
|
||||
|
||||
def _parse_proxy_config(self):
|
||||
client = self.client
|
||||
if 'proxies' not in client._general_configs:
|
||||
@@ -1790,20 +1787,77 @@ def rewrite_build_path(path):
|
||||
return path
|
||||
|
||||
|
||||
class _CLIBuilder:
|
||||
def __init__(self, progress):
|
||||
self._progress = progress
|
||||
class _ClientBuilder:
|
||||
def __init__(self, client):
|
||||
self.client = client
|
||||
|
||||
def build(self, path, tag=None, quiet=False, fileobj=None,
|
||||
def build(self, service, path, tag=None, quiet=False, fileobj=None,
|
||||
nocache=False, rm=False, timeout=None,
|
||||
custom_context=False, encoding=None, pull=False,
|
||||
forcerm=False, dockerfile=None, container_limits=None,
|
||||
decode=False, buildargs=None, gzip=False, shmsize=None,
|
||||
labels=None, cache_from=None, target=None, network_mode=None,
|
||||
squash=None, extra_hosts=None, platform=None, isolation=None,
|
||||
use_config_proxy=True):
|
||||
use_config_proxy=True, output_stream=sys.stdout):
|
||||
build_output = self.client.build(
|
||||
path=path,
|
||||
tag=tag,
|
||||
nocache=nocache,
|
||||
rm=rm,
|
||||
pull=pull,
|
||||
forcerm=forcerm,
|
||||
dockerfile=dockerfile,
|
||||
labels=labels,
|
||||
cache_from=cache_from,
|
||||
buildargs=buildargs,
|
||||
network_mode=network_mode,
|
||||
target=target,
|
||||
shmsize=shmsize,
|
||||
extra_hosts=extra_hosts,
|
||||
container_limits=container_limits,
|
||||
gzip=gzip,
|
||||
isolation=isolation,
|
||||
platform=platform)
|
||||
|
||||
try:
|
||||
all_events = list(stream_output(build_output, output_stream))
|
||||
except StreamOutputError as e:
|
||||
raise BuildError(service, str(e))
|
||||
|
||||
# Ensure the HTTP connection is not reused for another
|
||||
# streaming command, as the Docker daemon can sometimes
|
||||
# complain about it
|
||||
self.client.close()
|
||||
|
||||
image_id = None
|
||||
|
||||
for event in all_events:
|
||||
if 'stream' in event:
|
||||
match = re.search(r'Successfully built ([0-9a-f]+)', event.get('stream', ''))
|
||||
if match:
|
||||
image_id = match.group(1)
|
||||
|
||||
if image_id is None:
|
||||
raise BuildError(service, event if all_events else 'Unknown')
|
||||
|
||||
return image_id
|
||||
|
||||
|
||||
class _CLIBuilder:
|
||||
def __init__(self, progress):
|
||||
self._progress = progress
|
||||
|
||||
def build(self, service, path, tag=None, quiet=False, fileobj=None,
|
||||
nocache=False, rm=False, timeout=None,
|
||||
custom_context=False, encoding=None, pull=False,
|
||||
forcerm=False, dockerfile=None, container_limits=None,
|
||||
decode=False, buildargs=None, gzip=False, shmsize=None,
|
||||
labels=None, cache_from=None, target=None, network_mode=None,
|
||||
squash=None, extra_hosts=None, platform=None, isolation=None,
|
||||
use_config_proxy=True, output_stream=sys.stdout):
|
||||
"""
|
||||
Args:
|
||||
service (str): Service to be built
|
||||
path (str): Path to the directory containing the Dockerfile
|
||||
buildargs (dict): A dictionary of build arguments
|
||||
cache_from (:py:class:`list`): A list of images used for build
|
||||
@@ -1852,10 +1906,11 @@ class _CLIBuilder:
|
||||
configuration file (``~/.docker/config.json`` by default)
|
||||
contains a proxy configuration, the corresponding environment
|
||||
variables will be set in the container being built.
|
||||
output_stream (writer): stream to use for build logs
|
||||
Returns:
|
||||
A generator for the build output.
|
||||
"""
|
||||
if dockerfile:
|
||||
if dockerfile and os.path.isdir(path):
|
||||
dockerfile = os.path.join(path, dockerfile)
|
||||
iidfile = tempfile.mktemp()
|
||||
|
||||
@@ -1873,35 +1928,29 @@ class _CLIBuilder:
|
||||
command_builder.add_arg("--tag", tag)
|
||||
command_builder.add_arg("--target", target)
|
||||
command_builder.add_arg("--iidfile", iidfile)
|
||||
command_builder.add_arg("--platform", platform)
|
||||
command_builder.add_arg("--isolation", isolation)
|
||||
|
||||
if extra_hosts:
|
||||
if isinstance(extra_hosts, dict):
|
||||
extra_hosts = ["{}:{}".format(host, ip) for host, ip in extra_hosts.items()]
|
||||
for host in extra_hosts:
|
||||
command_builder.add_arg("--add-host", "{}".format(host))
|
||||
|
||||
args = command_builder.build([path])
|
||||
|
||||
magic_word = "Successfully built "
|
||||
appear = False
|
||||
with subprocess.Popen(args, stdout=subprocess.PIPE,
|
||||
with subprocess.Popen(args, stdout=output_stream, stderr=sys.stderr,
|
||||
universal_newlines=True) as p:
|
||||
while True:
|
||||
line = p.stdout.readline()
|
||||
if not line:
|
||||
break
|
||||
if line.startswith(magic_word):
|
||||
appear = True
|
||||
yield json.dumps({"stream": line})
|
||||
|
||||
p.communicate()
|
||||
if p.returncode != 0:
|
||||
raise StreamOutputError()
|
||||
raise BuildError(service, "Build failed")
|
||||
|
||||
with open(iidfile) as f:
|
||||
line = f.readline()
|
||||
image_id = line.split(":")[1].strip()
|
||||
os.remove(iidfile)
|
||||
|
||||
# In case of `DOCKER_BUILDKIT=1`
|
||||
# there is no success message already present in the output.
|
||||
# Since that's the way `Service::build` gets the `image_id`
|
||||
# it has to be added `manually`
|
||||
if not appear:
|
||||
yield json.dumps({"stream": "{}{}\n".format(magic_word, image_id)})
|
||||
return image_id
|
||||
|
||||
|
||||
class _CommandBuilder:
|
||||
|
||||
@@ -174,3 +174,18 @@ def truncate_string(s, max_chars=35):
|
||||
if len(s) > max_chars:
|
||||
return s[:max_chars - 2] + '...'
|
||||
return s
|
||||
|
||||
|
||||
def filter_attached_for_up(items, service_names, attach_dependencies=False,
|
||||
item_to_service_name=lambda x: x):
|
||||
"""This function contains the logic of choosing which services to
|
||||
attach when doing docker-compose up. It may be used both with containers
|
||||
and services, and any other entities that map to service names -
|
||||
this mapping is provided by item_to_service_name."""
|
||||
if attach_dependencies or not service_names:
|
||||
return items
|
||||
|
||||
return [
|
||||
item
|
||||
for item in items if item_to_service_name(item) in service_names
|
||||
]
|
||||
|
||||
@@ -138,7 +138,7 @@ _docker_compose_config() {
|
||||
;;
|
||||
esac
|
||||
|
||||
COMPREPLY=( $( compgen -W "--hash --help --no-interpolate --quiet -q --resolve-image-digests --services --volumes" -- "$cur" ) )
|
||||
COMPREPLY=( $( compgen -W "--hash --help --no-interpolate --profiles --quiet -q --resolve-image-digests --services --volumes" -- "$cur" ) )
|
||||
}
|
||||
|
||||
|
||||
@@ -164,10 +164,18 @@ _docker_compose_docker_compose() {
|
||||
_filedir "y?(a)ml"
|
||||
return
|
||||
;;
|
||||
--ansi)
|
||||
COMPREPLY=( $( compgen -W "never always auto" -- "$cur" ) )
|
||||
return
|
||||
;;
|
||||
--log-level)
|
||||
COMPREPLY=( $( compgen -W "debug info warning error critical" -- "$cur" ) )
|
||||
return
|
||||
;;
|
||||
--profile)
|
||||
COMPREPLY=( $( compgen -W "$(__docker_compose_q config --profiles)" -- "$cur" ) )
|
||||
return
|
||||
;;
|
||||
--project-directory)
|
||||
_filedir -d
|
||||
return
|
||||
@@ -290,7 +298,7 @@ _docker_compose_logs() {
|
||||
|
||||
case "$cur" in
|
||||
-*)
|
||||
COMPREPLY=( $( compgen -W "--follow -f --help --no-color --tail --timestamps -t" -- "$cur" ) )
|
||||
COMPREPLY=( $( compgen -W "--follow -f --help --no-color --no-log-prefix --tail --timestamps -t" -- "$cur" ) )
|
||||
;;
|
||||
*)
|
||||
__docker_compose_complete_services
|
||||
@@ -545,7 +553,7 @@ _docker_compose_up() {
|
||||
|
||||
case "$cur" in
|
||||
-*)
|
||||
COMPREPLY=( $( compgen -W "--abort-on-container-exit --always-recreate-deps --attach-dependencies --build -d --detach --exit-code-from --force-recreate --help --no-build --no-color --no-deps --no-recreate --no-start --renew-anon-volumes -V --remove-orphans --scale --timeout -t" -- "$cur" ) )
|
||||
COMPREPLY=( $( compgen -W "--abort-on-container-exit --always-recreate-deps --attach-dependencies --build -d --detach --exit-code-from --force-recreate --help --no-build --no-color --no-deps --no-log-prefix --no-recreate --no-start --renew-anon-volumes -V --remove-orphans --scale --timeout -t" -- "$cur" ) )
|
||||
;;
|
||||
*)
|
||||
__docker_compose_complete_services
|
||||
@@ -614,9 +622,11 @@ _docker_compose() {
|
||||
--tlskey
|
||||
"
|
||||
|
||||
# These options are require special treatment when searching the command.
|
||||
# These options require special treatment when searching the command.
|
||||
local top_level_options_with_args="
|
||||
--ansi
|
||||
--log-level
|
||||
--profile
|
||||
"
|
||||
|
||||
COMPREPLY=()
|
||||
|
||||
@@ -21,5 +21,7 @@ complete -c docker-compose -l tlscert -r -d 'Path to TLS certif
|
||||
complete -c docker-compose -l tlskey -r -d 'Path to TLS key file'
|
||||
complete -c docker-compose -l tlsverify -d 'Use TLS and verify the remote'
|
||||
complete -c docker-compose -l skip-hostname-check -d "Don't check the daemon's hostname against the name specified in the client certificate (for example if your docker host is an IP address)"
|
||||
complete -c docker-compose -l no-ansi -d 'Do not print ANSI control characters'
|
||||
complete -c docker-compose -l ansi -a 'never always auto' -d 'Control when to print ANSI control characters'
|
||||
complete -c docker-compose -s h -l help -d 'Print usage'
|
||||
complete -c docker-compose -s v -l version -d 'Print version and exit'
|
||||
|
||||
@@ -342,6 +342,7 @@ _docker-compose() {
|
||||
'--verbose[Show more output]' \
|
||||
'--log-level=[Set log level]:level:(DEBUG INFO WARNING ERROR CRITICAL)' \
|
||||
'--no-ansi[Do not print ANSI control characters]' \
|
||||
'--ansi=[Control when to print ANSI control characters]:when:(never always auto)' \
|
||||
'(-H --host)'{-H,--host}'[Daemon socket to connect to]:host:' \
|
||||
'--tls[Use TLS; implied by --tlsverify]' \
|
||||
'--tlscacert=[Trust certs signed only by this CA]:ca path:' \
|
||||
|
||||
474
docs/Compose file reference (legacy)/compose-versioning.md
Normal file
474
docs/Compose file reference (legacy)/compose-versioning.md
Normal file
@@ -0,0 +1,474 @@
|
||||
There are three legacy versions of the Compose file format:
|
||||
|
||||
- Version 1. This is specified by omitting a `version` key at the root of the YAML.
|
||||
|
||||
- Version 2.x. This is specified with a `version: '2'` or `version: '2.1'`, etc., entry at the root of the YAML.
|
||||
|
||||
- Version 3.x, designed to be cross-compatible between Compose and the Docker Engine's
|
||||
[swarm mode](https://docs.docker.com/engine/swarm/). This is specified with a `version: '3'` or `version: '3.1'`, etc., entry at the root of the YAML.
|
||||
|
||||
The latest and recommended version of the Compose file format is defined by the [Compose Specification](https://docs.docker.com/compose/compose-file/). This format merges the 2.x and 3.x versions and is implemented by **Compose 1.27.0+**.
|
||||
|
||||
|
||||
> **Note**
|
||||
>
|
||||
> If you're using [multiple Compose files](https://docs.docker.com/compose/multiple-compose-files/) or
|
||||
> [extending services](https://docs.docker.com/compose/multiple-compose-files/extends/),
|
||||
> each file must be of the same version - you cannot, for example,
|
||||
> mix version 1 and 2 in a single project.
|
||||
|
||||
Several things differ depending on which version you use:
|
||||
|
||||
- The structure and permitted configuration keys
|
||||
- The minimum Docker Engine version you must be running
|
||||
- Compose's behaviour with regards to networking
|
||||
|
||||
These differences are explained below.
|
||||
|
||||
### Version 2
|
||||
|
||||
Compose files using the version 2 syntax must indicate the version number at
|
||||
the root of the document. All [services](compose-file-v2.md#service-configuration-reference)
|
||||
must be declared under the `services` key.
|
||||
|
||||
Version 2 files are supported by **Compose 1.6.0+** and require a Docker Engine
|
||||
of version **1.10.0+**.
|
||||
|
||||
Named [volumes](compose-file-v2.md#volume-configuration-reference) can be declared under the
|
||||
`volumes` key, and [networks](compose-file-v2.md#network-configuration-reference) can be declared
|
||||
under the `networks` key.
|
||||
|
||||
By default, every container joins an application-wide default network, and is
|
||||
discoverable at a hostname that's the same as the service name. This means
|
||||
[links](compose-file-v2.md#links) are largely unnecessary. For more details, see
|
||||
[Networking in Compose](https://docs.docker.com/compose/networking/).
|
||||
|
||||
> **Note**
|
||||
>
|
||||
> With Compose version 2, when specifying the Compose file version to use, make sure to
|
||||
> specify both the _major_ and _minor_ numbers. If no minor version is given,
|
||||
> `0` is used by default and not the latest minor version. As a result, features added in later versions will not be supported. For example:
|
||||
>
|
||||
> ```yaml
|
||||
> version: "2"
|
||||
> ```
|
||||
>
|
||||
> is equivalent to:
|
||||
>
|
||||
> ```yaml
|
||||
> version: "2.0"
|
||||
> ```
|
||||
|
||||
Simple example:
|
||||
|
||||
version: "{{% param "compose_file_v2" %}}"
|
||||
services:
|
||||
web:
|
||||
build: .
|
||||
ports:
|
||||
- "8000:5000"
|
||||
volumes:
|
||||
- .:/code
|
||||
redis:
|
||||
image: redis
|
||||
|
||||
A more extended example, defining volumes and networks:
|
||||
|
||||
version: "{{% param "compose_file_v2" %}}"
|
||||
services:
|
||||
web:
|
||||
build: .
|
||||
ports:
|
||||
- "8000:5000"
|
||||
volumes:
|
||||
- .:/code
|
||||
networks:
|
||||
- front-tier
|
||||
- back-tier
|
||||
redis:
|
||||
image: redis
|
||||
volumes:
|
||||
- redis-data:/var/lib/redis
|
||||
networks:
|
||||
- back-tier
|
||||
volumes:
|
||||
redis-data:
|
||||
driver: local
|
||||
networks:
|
||||
front-tier:
|
||||
driver: bridge
|
||||
back-tier:
|
||||
driver: bridge
|
||||
|
||||
Several other options were added to support networking, such as:
|
||||
|
||||
* [`aliases`](compose-file-v2.md#aliases)
|
||||
|
||||
* The [`depends_on`](compose-file-v2.md#depends_on) option can be used in place of links to indicate dependencies
|
||||
between services and startup order.
|
||||
|
||||
version: "{{% param "compose_file_v2" %}}"
|
||||
services:
|
||||
web:
|
||||
build: .
|
||||
depends_on:
|
||||
- db
|
||||
- redis
|
||||
redis:
|
||||
image: redis
|
||||
db:
|
||||
image: postgres
|
||||
|
||||
* [`ipv4_address`, `ipv6_address`](compose-file-v2.md#ipv4_address-ipv6_address)
|
||||
|
||||
[Variable substitution](compose-file-v2.md#variable-substitution) also was added in Version 2.
|
||||
|
||||
### Version 2.1
|
||||
|
||||
An upgrade of [version 2](#version-2) that introduces new parameters only
|
||||
available with Docker Engine version **1.12.0+**. Version 2.1 files are
|
||||
supported by **Compose 1.9.0+**.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`link_local_ips`](compose-file-v2.md#link_local_ips)
|
||||
- [`isolation`](compose-file-v2.md#isolation-1) in build configurations and
|
||||
service definitions
|
||||
- `labels` for [volumes](compose-file-v2.md#volume-configuration-reference),
|
||||
[networks](compose-file-v2.md#network-configuration-reference), and
|
||||
[build](compose-file-v3.md#build)
|
||||
- `name` for [volumes](compose-file-v2.md#volume-configuration-reference)
|
||||
- [`userns_mode`](compose-file-v2.md#userns_mode)
|
||||
- [`healthcheck`](compose-file-v2.md#healthcheck)
|
||||
- [`sysctls`](compose-file-v2.md#sysctls)
|
||||
- [`pids_limit`](compose-file-v2.md#pids_limit)
|
||||
- [`oom_kill_disable`](compose-file-v2.md#cpu-and-other-resources)
|
||||
- [`cpu_period`](compose-file-v2.md#cpu-and-other-resources)
|
||||
|
||||
### Version 2.2
|
||||
|
||||
An upgrade of [version 2.1](#version-21) that introduces new parameters only
|
||||
available with Docker Engine version **1.13.0+**. Version 2.2 files are
|
||||
supported by **Compose 1.13.0+**. This version also allows you to specify
|
||||
default scale numbers inside the service's configuration.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`init`](compose-file-v2.md#init)
|
||||
- [`scale`](compose-file-v2.md#scale)
|
||||
- [`cpu_rt_runtime` and `cpu_rt_period`](compose-file-v2.md#cpu_rt_runtime-cpu_rt_period)
|
||||
- [`network`](compose-file-v2.md#network) for [build configurations](compose-file-v2.md#build)
|
||||
|
||||
### Version 2.3
|
||||
|
||||
An upgrade of [version 2.2](#version-22) that introduces new parameters only
|
||||
available with Docker Engine version **17.06.0+**. Version 2.3 files are
|
||||
supported by **Compose 1.16.0+**.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`target`](compose-file-v2.md#target), [`extra_hosts`](compose-file-v2.md#extra_hosts-1) and
|
||||
[`shm_size`](compose-file-v2.md#shm_size) for [build configurations](compose-file-v2.md#build)
|
||||
- `start_period` for [`healthchecks`](compose-file-v2.md#healthcheck)
|
||||
- ["Long syntax" for volumes](compose-file-v2.md#long-syntax)
|
||||
- [`runtime`](compose-file-v2.md#runtime) for service definitions
|
||||
- [`device_cgroup_rules`](compose-file-v2.md#device_cgroup_rules)
|
||||
|
||||
### Version 2.4
|
||||
|
||||
An upgrade of [version 2.3](#version-23) that introduces new parameters only
|
||||
available with Docker Engine version **17.12.0+**. Version 2.4 files are
|
||||
supported by **Compose 1.21.0+**.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`platform`](compose-file-v2.md#platform) for service definitions
|
||||
- Support for extension fields at the root of service, network, and volume
|
||||
definitions
|
||||
|
||||
### Version 3
|
||||
|
||||
Designed to be cross-compatible between Compose and the Docker Engine's
|
||||
[swarm mode](/engine/swarm/), version 3 removes several options and adds
|
||||
several more.
|
||||
|
||||
- Removed: `volume_driver`, `volumes_from`, `cpu_shares`, `cpu_quota`,
|
||||
`cpuset`, `mem_limit`, `memswap_limit`, `extends`, `group_add`. See
|
||||
the [upgrading](#upgrading) guide for how to migrate away from these.
|
||||
|
||||
- Added: [deploy](compose-file-v3.md#deploy)
|
||||
|
||||
If only the major version is given (`version: '3'`),
|
||||
the latest minor version is used by default.
|
||||
|
||||
### Version 3.1
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters only
|
||||
available with Docker Engine version **1.13.1+**, and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`secrets`](compose-file-v3.md#secrets)
|
||||
|
||||
### Version 3.2
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters only
|
||||
available with Docker Engine version **17.04.0+**, and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`cache_from`](compose-file-v3.md#cache_from) in [build configurations](compose-file-v3.md#build)
|
||||
- Long syntax for [ports](compose-file-v3.md#ports) and [volume mounts](compose-file-v3.md#volumes)
|
||||
- [`attachable`](compose-file-v3.md#attachable) network driver option
|
||||
- [deploy `endpoint_mode`](compose-file-v3.md#endpoint_mode)
|
||||
- [deploy placement `preference`](compose-file-v3.md#placement)
|
||||
|
||||
### Version 3.3
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters only
|
||||
available with Docker Engine version **17.06.0+**, and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [build `labels`](compose-file-v3.md#build)
|
||||
- [`credential_spec`](compose-file-v3.md#credential_spec)
|
||||
- [`configs`](compose-file-v3.md#configs)
|
||||
|
||||
### Version 3.4
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters. It is
|
||||
only available with Docker Engine version **17.09.0** and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`target`](compose-file-v3.md#target) and [`network`](compose-file-v3.md#network) in
|
||||
[build configurations](compose-file-v3.md#build)
|
||||
- `start_period` for [`healthchecks`](compose-file-v3.md#healthcheck)
|
||||
- `order` for [update configurations](compose-file-v3.md#update_config)
|
||||
- `name` for [volumes](compose-file-v3.md#volume-configuration-reference)
|
||||
|
||||
### Version 3.5
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters. It is
|
||||
only available with Docker Engine version **17.12.0** and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`isolation`](compose-file-v3.md#isolation) in service definitions
|
||||
- `name` for networks, secrets and configs
|
||||
- `shm_size` in [build configurations](compose-file-v3.md#build)
|
||||
|
||||
### Version 3.6
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters. It is
|
||||
only available with Docker Engine version **18.02.0** and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`tmpfs` size](compose-file-v3.md#long-syntax-3) for `tmpfs`-type mounts
|
||||
|
||||
### Version 3.7
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters. It is
|
||||
only available with Docker Engine version **18.06.0** and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`init`](compose-file-v3.md#init) in service definitions
|
||||
- [`rollback_config`](compose-file-v3.md#rollback_config) in deploy configurations
|
||||
- Support for extension fields at the root of service, network, volume, secret
|
||||
and config definitions
|
||||
|
||||
### Version 3.8
|
||||
|
||||
An upgrade of [version 3](#version-3) that introduces new parameters. It is
|
||||
only available with Docker Engine version **19.03.0** and higher.
|
||||
|
||||
Introduces the following additional parameters:
|
||||
|
||||
- [`max_replicas_per_node`](compose-file-v3.md#max_replicas_per_node) in placement
|
||||
configurations
|
||||
- `template_driver` option for [config](compose-file-v3.md#configs-configuration-reference)
|
||||
and [secret](compose-file-v3.md#secrets-configuration-reference) configurations. This
|
||||
option is only supported when deploying swarm services using
|
||||
`docker stack deploy`.
|
||||
- `driver` and `driver_opts` option for [secret](compose-file-v3.md#secrets-configuration-reference)
|
||||
configurations. This option is only supported when deploying swarm services
|
||||
using `docker stack deploy`.
|
||||
|
||||
### Version 1 (Deprecated)
|
||||
|
||||
Compose versions below 1.6.x are
|
||||
|
||||
Compose files that do not declare a version are considered "version 1". In those
|
||||
files, all the [services](compose-file-v3.md#service-configuration-reference) are
|
||||
declared at the root of the document.
|
||||
|
||||
Version 1 is supported by Compose up to 1.6.x** and has been deprecated.
|
||||
|
||||
Version 1 files cannot declare named
|
||||
[volumes](compose-file-v3.md#volume-configuration-reference), [networks](compose-file-v3.md#network-configuration-reference) or
|
||||
[build arguments](compose-file-v3.md#args).
|
||||
|
||||
Compose does not take advantage of [networking](https://docs.docker.com/compose/networking/) when you
|
||||
use version 1: every container is placed on the default `bridge` network and is
|
||||
reachable from every other container at its IP address. You need to use
|
||||
`links` to enable discovery between containers.
|
||||
|
||||
Example:
|
||||
|
||||
web:
|
||||
build: .
|
||||
ports:
|
||||
- "8000:5000"
|
||||
volumes:
|
||||
- .:/code
|
||||
links:
|
||||
- redis
|
||||
redis:
|
||||
image: redis
|
||||
|
||||
## Upgrading
|
||||
|
||||
### Version 2.x to 3.x
|
||||
|
||||
Between versions 2.x and 3.x, the structure of the Compose file is the same, but
|
||||
several options have been removed:
|
||||
|
||||
- `volume_driver`: Instead of setting the volume driver on the service, define
|
||||
a volume using the
|
||||
[top-level `volumes` option](compose-file-v3.md#volume-configuration-reference)
|
||||
and specify the driver there.
|
||||
|
||||
version: "3.8"
|
||||
services:
|
||||
db:
|
||||
image: postgres
|
||||
volumes:
|
||||
- data:/var/lib/postgresql/data
|
||||
volumes:
|
||||
data:
|
||||
driver: mydriver
|
||||
|
||||
- `volumes_from`: To share a volume between services, define it using the
|
||||
[top-level `volumes` option](compose-file-v3.md#volume-configuration-reference)
|
||||
and reference it from each service that shares it using the
|
||||
[service-level `volumes` option](compose-file-v3.md#driver).
|
||||
|
||||
- `cpu_shares`, `cpu_quota`, `cpuset`, `mem_limit`, `memswap_limit`: These
|
||||
have been replaced by the [resources](compose-file-v3.md#resources) key under
|
||||
`deploy`. `deploy` configuration only takes effect when using
|
||||
`docker stack deploy`, and is ignored by `docker-compose`.
|
||||
|
||||
- `extends`: This option has been removed for `version: "3.x"` Compose files.
|
||||
For more information on `extends`, see
|
||||
[Extending services](https://docs.docker.com/compose/multiple-compose-files/extends/).
|
||||
- `group_add`: This option has been removed for `version: "3.x"` Compose files.
|
||||
- `pids_limit`: This option has not been introduced in `version: "3.x"` Compose files.
|
||||
- `link_local_ips` in `networks`: This option has not been introduced in
|
||||
`version: "3.x"` Compose files.
|
||||
|
||||
#### Compatibility mode
|
||||
|
||||
`docker-compose` 1.20.0 introduces a new `--compatibility` flag designed to
|
||||
help developers transition to version 3 more easily. When enabled,
|
||||
`docker-compose` reads the `deploy` section of each service's definition and
|
||||
attempts to translate it into the equivalent version 2 parameter. Currently,
|
||||
the following deploy keys are translated:
|
||||
|
||||
- [resources](compose-file-v3.md#resources) limits and memory reservations
|
||||
- [replicas](compose-file-v3.md#replicas)
|
||||
- [restart_policy](compose-file-v3.md#restart_policy) `condition` and `max_attempts`
|
||||
|
||||
All other keys are ignored and produce a warning if present. You can review
|
||||
the configuration that will be used to deploy by using the `--compatibility`
|
||||
flag with the `config` command.
|
||||
|
||||
> Do not use this in production
|
||||
>
|
||||
> We recommend against using `--compatibility` mode in production. The
|
||||
> resulting configuration is only an approximate using non-Swarm mode
|
||||
> properties, it may produce unexpected results.
|
||||
|
||||
### Version 1 to 2.x
|
||||
|
||||
In the majority of cases, moving from version 1 to 2 is a very simple process:
|
||||
|
||||
1. Indent the whole file by one level and put a `services:` key at the top.
|
||||
2. Add a `version: '2'` line at the top of the file.
|
||||
|
||||
It's more complicated if you're using particular configuration features:
|
||||
|
||||
- `dockerfile`: This now lives under the `build` key:
|
||||
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile-alternate
|
||||
|
||||
- `log_driver`, `log_opt`: These now live under the `logging` key:
|
||||
|
||||
logging:
|
||||
driver: syslog
|
||||
options:
|
||||
syslog-address: "tcp://192.168.0.42:123"
|
||||
|
||||
- `links` with environment variables: environment variables created by
|
||||
links, such as `CONTAINERNAME_PORT`, ` have been deprecated for some time. In the new Docker network system,
|
||||
they have been removed. You should either connect directly to the
|
||||
appropriate hostname or set the relevant environment variable yourself,
|
||||
using the link hostname:
|
||||
|
||||
web:
|
||||
links:
|
||||
- db
|
||||
environment:
|
||||
- DB_PORT=tcp://db:5432
|
||||
|
||||
- `external_links`: Compose uses Docker networks when running version 2
|
||||
projects, so links behave slightly differently. In particular, two
|
||||
containers must be connected to at least one network in common in order to
|
||||
communicate, even if explicitly linked together.
|
||||
|
||||
Either connect the external container to your app's
|
||||
[default network](https://docs.docker.com/compose/networking/), or connect both the external container and
|
||||
your service's containers to an
|
||||
[external network](https://docs.docker.com/compose/networking/).
|
||||
|
||||
- `net`: This is now replaced by [network_mode](compose-file-v3.md#network_mode):
|
||||
|
||||
net: host -> network_mode: host
|
||||
net: bridge -> network_mode: bridge
|
||||
net: none -> network_mode: none
|
||||
|
||||
If you're using `net: "container:[service name]"`, you must now use
|
||||
`network_mode: "service:[service name]"` instead.
|
||||
|
||||
net: "container:web" -> network_mode: "service:web"
|
||||
|
||||
If you're using `net: "container:[container name/id]"`, the value does not
|
||||
need to change.
|
||||
|
||||
net: "container:cont-name" -> network_mode: "container:cont-name"
|
||||
net: "container:abc12345" -> network_mode: "container:abc12345"
|
||||
|
||||
- `volumes` with named volumes: these must now be explicitly declared in a
|
||||
top-level `volumes` section of your Compose file. If a service mounts a
|
||||
named volume called `data`, you must declare a `data` volume in your
|
||||
top-level `volumes` section. The whole file might look like this:
|
||||
|
||||
version: "{{% param "compose_file_v2" %}}"
|
||||
services:
|
||||
db:
|
||||
image: postgres
|
||||
volumes:
|
||||
- data:/var/lib/postgresql/data
|
||||
volumes:
|
||||
data: {}
|
||||
|
||||
By default, Compose creates a volume whose name is prefixed with your
|
||||
project name. If you want it to just be called `data`, declare it as
|
||||
external:
|
||||
|
||||
volumes:
|
||||
data:
|
||||
external: true
|
||||
2136
docs/Compose file reference (legacy)/version-2.md
Normal file
2136
docs/Compose file reference (legacy)/version-2.md
Normal file
File diff suppressed because it is too large
Load Diff
2913
docs/Compose file reference (legacy)/version-3.md
Normal file
2913
docs/Compose file reference (legacy)/version-3.md
Normal file
File diff suppressed because it is too large
Load Diff
@@ -1,14 +0,0 @@
|
||||
# The docs have been moved!
|
||||
|
||||
The documentation for Compose has been merged into
|
||||
[the general documentation repo](https://github.com/docker/docker.github.io).
|
||||
|
||||
The docs for Compose are now here:
|
||||
https://github.com/docker/docker.github.io/tree/master/compose
|
||||
|
||||
Please submit pull requests for unreleased features/changes on the `master` branch (https://github.com/docker/docker.github.io/tree/master), please prefix the PR title with `[WIP]` to indicate that it relates to an unreleased change.
|
||||
|
||||
If you submit a PR to this codebase that has a docs impact, create a second docs PR on `docker.github.io`. Use the docs PR template provided.
|
||||
|
||||
As always, the docs remain open-source and we appreciate your feedback and
|
||||
pull requests!
|
||||
437
logo.svg
Normal file
437
logo.svg
Normal file
File diff suppressed because one or more lines are too long
|
After Width: | Height: | Size: 132 KiB |
@@ -1,10 +1,9 @@
|
||||
Click==7.1.2
|
||||
coverage==5.2.1
|
||||
ddt==1.4.1
|
||||
coverage==5.5
|
||||
ddt==1.4.2
|
||||
flake8==3.8.3
|
||||
gitpython==3.1.11
|
||||
mock==3.0.5
|
||||
pytest==6.0.1; python_version >= '3.5'
|
||||
pytest==6.2.4; python_version >= '3.5'
|
||||
pytest==4.6.5; python_version < '3.5'
|
||||
pytest-cov==2.10.1
|
||||
PyYAML==5.3.1
|
||||
|
||||
@@ -1,9 +1,9 @@
|
||||
altgraph==0.17
|
||||
appdirs==1.4.4
|
||||
attrs==20.3.0
|
||||
bcrypt==3.1.7
|
||||
bcrypt==3.2.0
|
||||
cffi==1.14.4
|
||||
cryptography==3.2.1
|
||||
cryptography==3.3.2
|
||||
distlib==0.3.1
|
||||
entrypoints==0.3
|
||||
filelock==3.0.12
|
||||
@@ -11,9 +11,9 @@ gitdb2==4.0.2
|
||||
mccabe==0.6.1
|
||||
more-itertools==8.6.0; python_version >= '3.5'
|
||||
more-itertools==5.0.0; python_version < '3.5'
|
||||
packaging==20.4
|
||||
packaging==20.9
|
||||
pluggy==0.13.1
|
||||
py==1.9.0
|
||||
py==1.10.0
|
||||
pycodestyle==2.6.0
|
||||
pycparser==2.20
|
||||
pyflakes==2.2.0
|
||||
@@ -23,6 +23,6 @@ pyrsistent==0.16.0
|
||||
smmap==3.0.4
|
||||
smmap2==3.0.1
|
||||
toml==0.10.1
|
||||
tox==3.20.1
|
||||
virtualenv==20.2.1
|
||||
tox==3.21.2
|
||||
virtualenv==20.4.0
|
||||
wcwidth==0.2.5
|
||||
|
||||
@@ -1,22 +1,22 @@
|
||||
backports.shutil_get_terminal_size==1.0.0
|
||||
cached-property==1.5.1
|
||||
certifi==2020.6.20
|
||||
cached-property==1.5.1; python_version < '3.8'
|
||||
certifi==2021.5.30
|
||||
chardet==3.0.4
|
||||
colorama==0.4.3; sys_platform == 'win32'
|
||||
colorama==0.4.4; sys_platform == 'win32'
|
||||
distro==1.5.0
|
||||
docker==4.4.0
|
||||
docker==5.0.0
|
||||
docker-pycreds==0.4.0
|
||||
dockerpty==0.4.1
|
||||
docopt==0.6.2
|
||||
idna==2.10
|
||||
ipaddress==1.0.23
|
||||
jsonschema==3.2.0
|
||||
paramiko==2.7.1
|
||||
paramiko==2.7.2
|
||||
PySocks==1.7.1
|
||||
python-dotenv==0.14.0
|
||||
pywin32==227; sys_platform == 'win32'
|
||||
PyYAML==5.3.1
|
||||
requests==2.24.0
|
||||
texttable==1.6.2
|
||||
urllib3==1.25.10; python_version == '3.3'
|
||||
websocket-client==0.57.0
|
||||
python-dotenv==0.17.0
|
||||
pywin32==301; sys_platform == 'win32'
|
||||
PyYAML==5.4.1
|
||||
requests==2.25.1
|
||||
texttable==1.6.3
|
||||
urllib3==1.26.5; python_version == '3.3'
|
||||
websocket-client==1.1.0
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
set -ex
|
||||
|
||||
CODE_PATH=/code
|
||||
VENV="${CODE_PATH}"/.tox/py39
|
||||
VENV="${CODE_PATH}"/.tox/py37
|
||||
|
||||
cd "${CODE_PATH}"
|
||||
mkdir -p dist
|
||||
|
||||
@@ -16,7 +16,7 @@
|
||||
#
|
||||
# 4. In Powershell, run the following commands:
|
||||
#
|
||||
# $ pip install 'virtualenv==20.2.1'
|
||||
# $ pip install 'virtualenv==20.2.2'
|
||||
# $ Set-ExecutionPolicy -Scope CurrentUser RemoteSigned
|
||||
#
|
||||
# 5. Clone the repository:
|
||||
|
||||
@@ -20,4 +20,3 @@ This should trigger a new CI build on the new tag. When the CI finishes with the
|
||||
4. In case of a GA version, please update `docker-compose`s release notes and version on [github documentation repository](https://github.com/docker/docker.github.io):
|
||||
- [Release Notes](https://github.com/docker/docker.github.io/blob/master/compose/release-notes.md)
|
||||
- [Config version](https://github.com/docker/docker.github.io/blob/master/_config.yml)
|
||||
- [Config authoring version](https://github.com/docker/docker.github.io/blob/master/_config_authoring.yml)
|
||||
|
||||
@@ -36,7 +36,7 @@ if ! [ -x "$(command -v python3)" ]; then
|
||||
brew install python3
|
||||
fi
|
||||
if ! [ -x "$(command -v virtualenv)" ]; then
|
||||
pip3 install virtualenv==20.2.1
|
||||
pip3 install virtualenv==20.2.2
|
||||
fi
|
||||
|
||||
#
|
||||
|
||||
@@ -11,7 +11,7 @@ docker run --rm \
|
||||
"$TAG" tox -e pre-commit
|
||||
|
||||
get_versions="docker run --rm
|
||||
--entrypoint=/code/.tox/py39/bin/python
|
||||
--entrypoint=/code/.tox/py37/bin/python
|
||||
$TAG
|
||||
/code/script/test/versions.py docker/docker-ce,moby/moby"
|
||||
|
||||
@@ -21,9 +21,8 @@ elif [ "$DOCKER_VERSIONS" == "all" ]; then
|
||||
DOCKER_VERSIONS=$($get_versions -n 2 recent)
|
||||
fi
|
||||
|
||||
|
||||
BUILD_NUMBER=${BUILD_NUMBER-$USER}
|
||||
PY_TEST_VERSIONS=${PY_TEST_VERSIONS:-py39}
|
||||
PY_TEST_VERSIONS=${PY_TEST_VERSIONS:-py37}
|
||||
|
||||
for version in $DOCKER_VERSIONS; do
|
||||
>&2 echo "Running tests against Docker $version"
|
||||
@@ -39,17 +38,23 @@ for version in $DOCKER_VERSIONS; do
|
||||
|
||||
trap "on_exit" EXIT
|
||||
|
||||
repo="dockerswarm/dind"
|
||||
|
||||
docker run \
|
||||
-d \
|
||||
--name "$daemon_container" \
|
||||
--privileged \
|
||||
--volume="/var/lib/docker" \
|
||||
"$repo:$version" \
|
||||
-e "DOCKER_TLS_CERTDIR=" \
|
||||
"docker:$version-dind" \
|
||||
dockerd -H tcp://0.0.0.0:2375 $DOCKER_DAEMON_ARGS \
|
||||
2>&1 | tail -n 10
|
||||
|
||||
docker exec "$daemon_container" sh -c "apk add --no-cache git"
|
||||
|
||||
# copy docker config from host for authentication with Docker Hub
|
||||
docker exec "$daemon_container" sh -c "mkdir /root/.docker"
|
||||
docker cp /root/.docker/config.json $daemon_container:/root/.docker/config.json
|
||||
docker exec "$daemon_container" sh -c "chmod 644 /root/.docker/config.json"
|
||||
|
||||
docker run \
|
||||
--rm \
|
||||
--tty \
|
||||
|
||||
4
setup.py
4
setup.py
@@ -25,14 +25,13 @@ def find_version(*file_paths):
|
||||
|
||||
|
||||
install_requires = [
|
||||
'cached-property >= 1.2.0, < 2',
|
||||
'docopt >= 0.6.1, < 1',
|
||||
'PyYAML >= 3.10, < 6',
|
||||
'requests >= 2.20.0, < 3',
|
||||
'texttable >= 0.9.0, < 2',
|
||||
'websocket-client >= 0.32.0, < 1',
|
||||
'distro >= 1.5.0, < 2',
|
||||
'docker[ssh] >= 4.3.1, < 5',
|
||||
'docker[ssh] >= 5',
|
||||
'dockerpty >= 0.4.1, < 1',
|
||||
'jsonschema >= 2.5.1, < 4',
|
||||
'python-dotenv >= 0.13.0, < 1',
|
||||
@@ -50,6 +49,7 @@ if sys.version_info[:2] < (3, 4):
|
||||
|
||||
extras_require = {
|
||||
':python_version < "3.5"': ['backports.ssl_match_hostname >= 3.5, < 4'],
|
||||
':python_version < "3.8"': ['cached-property >= 1.2.0, < 2'],
|
||||
':sys_platform == "win32"': ['colorama >= 0.4, < 1'],
|
||||
'socks': ['PySocks >= 1.5.6, != 1.5.7, < 2'],
|
||||
'tests': tests_require,
|
||||
|
||||
@@ -58,13 +58,16 @@ COMPOSE_COMPATIBILITY_DICT = {
|
||||
}
|
||||
|
||||
|
||||
def start_process(base_dir, options):
|
||||
def start_process(base_dir, options, executable=None, env=None):
|
||||
executable = executable or DOCKER_COMPOSE_EXECUTABLE
|
||||
proc = subprocess.Popen(
|
||||
[DOCKER_COMPOSE_EXECUTABLE] + options,
|
||||
[executable] + options,
|
||||
stdin=subprocess.PIPE,
|
||||
stdout=subprocess.PIPE,
|
||||
stderr=subprocess.PIPE,
|
||||
cwd=base_dir)
|
||||
cwd=base_dir,
|
||||
env=env,
|
||||
)
|
||||
print("Running process: %s" % proc.pid)
|
||||
return proc
|
||||
|
||||
@@ -78,9 +81,10 @@ def wait_on_process(proc, returncode=0, stdin=None):
|
||||
return ProcessResult(stdout.decode('utf-8'), stderr.decode('utf-8'))
|
||||
|
||||
|
||||
def dispatch(base_dir, options, project_options=None, returncode=0, stdin=None):
|
||||
def dispatch(base_dir, options,
|
||||
project_options=None, returncode=0, stdin=None, executable=None, env=None):
|
||||
project_options = project_options or []
|
||||
proc = start_process(base_dir, project_options + options)
|
||||
proc = start_process(base_dir, project_options + options, executable=executable, env=env)
|
||||
return wait_on_process(proc, returncode=returncode, stdin=stdin)
|
||||
|
||||
|
||||
@@ -233,6 +237,11 @@ class CLITestCase(DockerClientTestCase):
|
||||
result = self.dispatch(['-H=tcp://doesnotexist:8000', 'ps'], returncode=1)
|
||||
assert "Couldn't connect to Docker daemon" in result.stderr
|
||||
|
||||
def test_config_list_profiles(self):
|
||||
self.base_dir = 'tests/fixtures/config-profiles'
|
||||
result = self.dispatch(['config', '--profiles'])
|
||||
assert set(result.stdout.rstrip().split('\n')) == {'debug', 'frontend', 'gui'}
|
||||
|
||||
def test_config_list_services(self):
|
||||
self.base_dir = 'tests/fixtures/v2-full'
|
||||
result = self.dispatch(['config', '--services'])
|
||||
@@ -783,7 +792,11 @@ services:
|
||||
assert BUILD_CACHE_TEXT not in result.stdout
|
||||
assert BUILD_PULL_TEXT in result.stdout
|
||||
|
||||
@mock.patch.dict(os.environ)
|
||||
def test_build_log_level(self):
|
||||
os.environ['COMPOSE_DOCKER_CLI_BUILD'] = '0'
|
||||
os.environ['DOCKER_BUILDKIT'] = '0'
|
||||
self.test_env_file_relative_to_compose_file()
|
||||
self.base_dir = 'tests/fixtures/simple-dockerfile'
|
||||
result = self.dispatch(['--log-level', 'warning', 'build', 'simple'])
|
||||
assert result.stderr == ''
|
||||
@@ -845,13 +858,17 @@ services:
|
||||
for c in self.project.client.containers(all=True):
|
||||
self.addCleanup(self.project.client.remove_container, c, force=True)
|
||||
|
||||
@mock.patch.dict(os.environ)
|
||||
def test_build_shm_size_build_option(self):
|
||||
os.environ['COMPOSE_DOCKER_CLI_BUILD'] = '0'
|
||||
pull_busybox(self.client)
|
||||
self.base_dir = 'tests/fixtures/build-shm-size'
|
||||
result = self.dispatch(['build', '--no-cache'], None)
|
||||
assert 'shm_size: 96' in result.stdout
|
||||
|
||||
@mock.patch.dict(os.environ)
|
||||
def test_build_memory_build_option(self):
|
||||
os.environ['COMPOSE_DOCKER_CLI_BUILD'] = '0'
|
||||
pull_busybox(self.client)
|
||||
self.base_dir = 'tests/fixtures/build-memory'
|
||||
result = self.dispatch(['build', '--no-cache', '--memory', '96m', 'service'], None)
|
||||
|
||||
15
tests/fixtures/config-profiles/docker-compose.yml
vendored
Normal file
15
tests/fixtures/config-profiles/docker-compose.yml
vendored
Normal file
@@ -0,0 +1,15 @@
|
||||
version: '3.8'
|
||||
services:
|
||||
frontend:
|
||||
image: frontend
|
||||
profiles: ["frontend", "gui"]
|
||||
phpmyadmin:
|
||||
image: phpmyadmin
|
||||
depends_on:
|
||||
- db
|
||||
profiles:
|
||||
- debug
|
||||
backend:
|
||||
image: backend
|
||||
db:
|
||||
image: mysql
|
||||
1
tests/fixtures/env-file-override/.env
vendored
Normal file
1
tests/fixtures/env-file-override/.env
vendored
Normal file
@@ -0,0 +1 @@
|
||||
WHEREAMI=default
|
||||
@@ -1,5 +1,6 @@
|
||||
import tempfile
|
||||
|
||||
import pytest
|
||||
from ddt import data
|
||||
from ddt import ddt
|
||||
|
||||
@@ -8,6 +9,7 @@ from ..acceptance.cli_test import dispatch
|
||||
from compose.cli.command import get_project
|
||||
from compose.cli.command import project_from_options
|
||||
from compose.config.environment import Environment
|
||||
from compose.config.errors import EnvFileNotFound
|
||||
from tests.integration.testcases import DockerClientTestCase
|
||||
|
||||
|
||||
@@ -55,13 +57,36 @@ services:
|
||||
class EnvironmentOverrideFileTest(DockerClientTestCase):
|
||||
def test_env_file_override(self):
|
||||
base_dir = 'tests/fixtures/env-file-override'
|
||||
# '--env-file' are relative to the current working dir
|
||||
env = Environment.from_env_file(base_dir, base_dir+'/.env.override')
|
||||
dispatch(base_dir, ['--env-file', '.env.override', 'up'])
|
||||
project = get_project(project_dir=base_dir,
|
||||
config_path=['docker-compose.yml'],
|
||||
environment=Environment.from_env_file(base_dir, '.env.override'),
|
||||
environment=env,
|
||||
override_dir=base_dir)
|
||||
containers = project.containers(stopped=True)
|
||||
assert len(containers) == 1
|
||||
assert "WHEREAMI=override" in containers[0].get('Config.Env')
|
||||
assert "DEFAULT_CONF_LOADED=true" in containers[0].get('Config.Env')
|
||||
dispatch(base_dir, ['--env-file', '.env.override', 'down'], None)
|
||||
|
||||
def test_env_file_not_found_error(self):
|
||||
base_dir = 'tests/fixtures/env-file-override'
|
||||
with pytest.raises(EnvFileNotFound) as excinfo:
|
||||
Environment.from_env_file(base_dir, '.env.override')
|
||||
|
||||
assert "Couldn't find env file" in excinfo.exconly()
|
||||
|
||||
def test_dot_env_file(self):
|
||||
base_dir = 'tests/fixtures/env-file-override'
|
||||
# '.env' is relative to the project_dir (base_dir)
|
||||
env = Environment.from_env_file(base_dir, None)
|
||||
dispatch(base_dir, ['up'])
|
||||
project = get_project(project_dir=base_dir,
|
||||
config_path=['docker-compose.yml'],
|
||||
environment=env,
|
||||
override_dir=base_dir)
|
||||
containers = project.containers(stopped=True)
|
||||
assert len(containers) == 1
|
||||
assert "WHEREAMI=default" in containers[0].get('Config.Env')
|
||||
dispatch(base_dir, ['down'], None)
|
||||
|
||||
125
tests/integration/metrics_test.py
Normal file
125
tests/integration/metrics_test.py
Normal file
@@ -0,0 +1,125 @@
|
||||
import logging
|
||||
import os
|
||||
import socket
|
||||
from http.server import BaseHTTPRequestHandler
|
||||
from http.server import HTTPServer
|
||||
from threading import Thread
|
||||
|
||||
import requests
|
||||
from docker.transport import UnixHTTPAdapter
|
||||
|
||||
from tests.acceptance.cli_test import dispatch
|
||||
from tests.integration.testcases import DockerClientTestCase
|
||||
|
||||
|
||||
TEST_SOCKET_FILE = '/tmp/test-metrics-docker-cli.sock'
|
||||
|
||||
|
||||
class MetricsTest(DockerClientTestCase):
|
||||
test_session = requests.sessions.Session()
|
||||
test_env = None
|
||||
base_dir = 'tests/fixtures/v3-full'
|
||||
|
||||
@classmethod
|
||||
def setUpClass(cls):
|
||||
super().setUpClass()
|
||||
MetricsTest.test_session.mount("http+unix://", UnixHTTPAdapter(TEST_SOCKET_FILE))
|
||||
MetricsTest.test_env = os.environ.copy()
|
||||
MetricsTest.test_env['METRICS_SOCKET_FILE'] = TEST_SOCKET_FILE
|
||||
MetricsServer().start()
|
||||
|
||||
@classmethod
|
||||
def test_metrics_help(cls):
|
||||
# root `docker-compose` command is considered as a `--help`
|
||||
dispatch(cls.base_dir, [], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose --help", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "success"}'
|
||||
dispatch(cls.base_dir, ['help', 'run'], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose help", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "success"}'
|
||||
dispatch(cls.base_dir, ['--help'], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose --help", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "success"}'
|
||||
dispatch(cls.base_dir, ['run', '--help'], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose --help run", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "success"}'
|
||||
dispatch(cls.base_dir, ['up', '--help', 'extra_args'], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose --help up", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "success"}'
|
||||
|
||||
@classmethod
|
||||
def test_metrics_simple_commands(cls):
|
||||
dispatch(cls.base_dir, ['ps'], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose ps", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "success"}'
|
||||
dispatch(cls.base_dir, ['version'], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose version", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "success"}'
|
||||
dispatch(cls.base_dir, ['version', '--yyy'], env=MetricsTest.test_env)
|
||||
assert cls.get_content() == \
|
||||
b'{"command": "compose version", "context": "moby", ' \
|
||||
b'"source": "docker-compose", "status": "failure"}'
|
||||
|
||||
@staticmethod
|
||||
def get_content():
|
||||
resp = MetricsTest.test_session.get("http+unix://localhost")
|
||||
print(resp.content)
|
||||
return resp.content
|
||||
|
||||
|
||||
def start_server(uri=TEST_SOCKET_FILE):
|
||||
try:
|
||||
os.remove(uri)
|
||||
except OSError:
|
||||
pass
|
||||
httpd = HTTPServer(uri, MetricsHTTPRequestHandler, False)
|
||||
sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
|
||||
sock.bind(TEST_SOCKET_FILE)
|
||||
sock.listen(0)
|
||||
httpd.socket = sock
|
||||
print('Serving on ', uri)
|
||||
httpd.serve_forever()
|
||||
sock.shutdown(socket.SHUT_RDWR)
|
||||
sock.close()
|
||||
os.remove(uri)
|
||||
|
||||
|
||||
class MetricsServer:
|
||||
@classmethod
|
||||
def start(cls):
|
||||
t = Thread(target=start_server, daemon=True)
|
||||
t.start()
|
||||
|
||||
|
||||
class MetricsHTTPRequestHandler(BaseHTTPRequestHandler):
|
||||
usages = []
|
||||
|
||||
def do_GET(self):
|
||||
self.client_address = ('',) # avoid exception in BaseHTTPServer.py log_message()
|
||||
self.send_response(200)
|
||||
self.end_headers()
|
||||
for u in MetricsHTTPRequestHandler.usages:
|
||||
self.wfile.write(u)
|
||||
MetricsHTTPRequestHandler.usages = []
|
||||
|
||||
def do_POST(self):
|
||||
self.client_address = ('',) # avoid exception in BaseHTTPServer.py log_message()
|
||||
content_length = int(self.headers['Content-Length'])
|
||||
body = self.rfile.read(content_length)
|
||||
print(body)
|
||||
MetricsHTTPRequestHandler.usages.append(body)
|
||||
self.send_response(200)
|
||||
self.end_headers()
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
logging.getLogger("urllib3").propagate = False
|
||||
logging.getLogger("requests").propagate = False
|
||||
start_server()
|
||||
@@ -25,6 +25,7 @@ from compose.const import COMPOSE_SPEC as VERSION
|
||||
from compose.const import LABEL_PROJECT
|
||||
from compose.const import LABEL_SERVICE
|
||||
from compose.container import Container
|
||||
from compose.errors import CompletedUnsuccessfully
|
||||
from compose.errors import HealthCheckFailed
|
||||
from compose.errors import NoHealthCheckConfigured
|
||||
from compose.project import Project
|
||||
@@ -1899,6 +1900,106 @@ class ProjectTest(DockerClientTestCase):
|
||||
with pytest.raises(NoHealthCheckConfigured):
|
||||
svc1.is_healthy()
|
||||
|
||||
def test_project_up_completed_successfully_dependency(self):
|
||||
config_dict = {
|
||||
'version': '2.1',
|
||||
'services': {
|
||||
'svc1': {
|
||||
'image': BUSYBOX_IMAGE_WITH_TAG,
|
||||
'command': 'true'
|
||||
},
|
||||
'svc2': {
|
||||
'image': BUSYBOX_IMAGE_WITH_TAG,
|
||||
'command': 'top',
|
||||
'depends_on': {
|
||||
'svc1': {'condition': 'service_completed_successfully'},
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
config_data = load_config(config_dict)
|
||||
project = Project.from_config(
|
||||
name='composetest', config_data=config_data, client=self.client
|
||||
)
|
||||
project.up()
|
||||
|
||||
svc1 = project.get_service('svc1')
|
||||
svc2 = project.get_service('svc2')
|
||||
|
||||
assert 'svc1' in svc2.get_dependency_names()
|
||||
assert svc2.containers()[0].is_running
|
||||
assert len(svc1.containers()) == 0
|
||||
assert svc1.is_completed_successfully()
|
||||
|
||||
def test_project_up_completed_unsuccessfully_dependency(self):
|
||||
config_dict = {
|
||||
'version': '2.1',
|
||||
'services': {
|
||||
'svc1': {
|
||||
'image': BUSYBOX_IMAGE_WITH_TAG,
|
||||
'command': 'false'
|
||||
},
|
||||
'svc2': {
|
||||
'image': BUSYBOX_IMAGE_WITH_TAG,
|
||||
'command': 'top',
|
||||
'depends_on': {
|
||||
'svc1': {'condition': 'service_completed_successfully'},
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
config_data = load_config(config_dict)
|
||||
project = Project.from_config(
|
||||
name='composetest', config_data=config_data, client=self.client
|
||||
)
|
||||
with pytest.raises(ProjectError):
|
||||
project.up()
|
||||
|
||||
svc1 = project.get_service('svc1')
|
||||
svc2 = project.get_service('svc2')
|
||||
assert 'svc1' in svc2.get_dependency_names()
|
||||
assert len(svc2.containers()) == 0
|
||||
with pytest.raises(CompletedUnsuccessfully):
|
||||
svc1.is_completed_successfully()
|
||||
|
||||
def test_project_up_completed_differently_dependencies(self):
|
||||
config_dict = {
|
||||
'version': '2.1',
|
||||
'services': {
|
||||
'svc1': {
|
||||
'image': BUSYBOX_IMAGE_WITH_TAG,
|
||||
'command': 'true'
|
||||
},
|
||||
'svc2': {
|
||||
'image': BUSYBOX_IMAGE_WITH_TAG,
|
||||
'command': 'false'
|
||||
},
|
||||
'svc3': {
|
||||
'image': BUSYBOX_IMAGE_WITH_TAG,
|
||||
'command': 'top',
|
||||
'depends_on': {
|
||||
'svc1': {'condition': 'service_completed_successfully'},
|
||||
'svc2': {'condition': 'service_completed_successfully'},
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
config_data = load_config(config_dict)
|
||||
project = Project.from_config(
|
||||
name='composetest', config_data=config_data, client=self.client
|
||||
)
|
||||
with pytest.raises(ProjectError):
|
||||
project.up()
|
||||
|
||||
svc1 = project.get_service('svc1')
|
||||
svc2 = project.get_service('svc2')
|
||||
svc3 = project.get_service('svc3')
|
||||
assert ['svc1', 'svc2'] == svc3.get_dependency_names()
|
||||
assert svc1.is_completed_successfully()
|
||||
assert len(svc3.containers()) == 0
|
||||
with pytest.raises(CompletedUnsuccessfully):
|
||||
svc2.is_completed_successfully()
|
||||
|
||||
def test_project_up_seccomp_profile(self):
|
||||
seccomp_data = {
|
||||
'defaultAction': 'SCMP_ACT_ALLOW',
|
||||
|
||||
@@ -948,7 +948,12 @@ class ServiceTest(DockerClientTestCase):
|
||||
with open(os.path.join(base_dir, 'Dockerfile'), 'w') as f:
|
||||
f.write("FROM busybox\n")
|
||||
|
||||
service = self.create_service('web', build={'context': base_dir})
|
||||
service = self.create_service('web',
|
||||
build={'context': base_dir},
|
||||
environment={
|
||||
'COMPOSE_DOCKER_CLI_BUILD': '0',
|
||||
'DOCKER_BUILDKIT': '0',
|
||||
})
|
||||
service.build()
|
||||
self.addCleanup(self.client.remove_image, service.image_name)
|
||||
|
||||
@@ -964,7 +969,6 @@ class ServiceTest(DockerClientTestCase):
|
||||
service = self.create_service('web',
|
||||
build={'context': base_dir},
|
||||
environment={
|
||||
'COMPOSE_DOCKER_CLI_BUILD': '1',
|
||||
'DOCKER_BUILDKIT': '1',
|
||||
})
|
||||
service.build(cli=True)
|
||||
@@ -1015,7 +1019,6 @@ class ServiceTest(DockerClientTestCase):
|
||||
web = self.create_service('web',
|
||||
build={'context': base_dir},
|
||||
environment={
|
||||
'COMPOSE_DOCKER_CLI_BUILD': '1',
|
||||
'DOCKER_BUILDKIT': '1',
|
||||
})
|
||||
project = Project('composetest', [web], self.client)
|
||||
|
||||
@@ -61,6 +61,7 @@ class DockerClientTestCase(unittest.TestCase):
|
||||
|
||||
@classmethod
|
||||
def tearDownClass(cls):
|
||||
cls.client.close()
|
||||
del cls.client
|
||||
|
||||
def tearDown(self):
|
||||
|
||||
56
tests/unit/cli/colors_test.py
Normal file
56
tests/unit/cli/colors_test.py
Normal file
@@ -0,0 +1,56 @@
|
||||
import os
|
||||
|
||||
import pytest
|
||||
|
||||
from compose.cli.colors import AnsiMode
|
||||
from tests import mock
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def tty_stream():
|
||||
stream = mock.Mock()
|
||||
stream.isatty.return_value = True
|
||||
return stream
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def non_tty_stream():
|
||||
stream = mock.Mock()
|
||||
stream.isatty.return_value = False
|
||||
return stream
|
||||
|
||||
|
||||
class TestAnsiModeTestCase:
|
||||
|
||||
@mock.patch.dict(os.environ)
|
||||
def test_ansi_mode_never(self, tty_stream, non_tty_stream):
|
||||
if "CLICOLOR" in os.environ:
|
||||
del os.environ["CLICOLOR"]
|
||||
assert not AnsiMode.NEVER.use_ansi_codes(tty_stream)
|
||||
assert not AnsiMode.NEVER.use_ansi_codes(non_tty_stream)
|
||||
|
||||
os.environ["CLICOLOR"] = "0"
|
||||
assert not AnsiMode.NEVER.use_ansi_codes(tty_stream)
|
||||
assert not AnsiMode.NEVER.use_ansi_codes(non_tty_stream)
|
||||
|
||||
@mock.patch.dict(os.environ)
|
||||
def test_ansi_mode_always(self, tty_stream, non_tty_stream):
|
||||
if "CLICOLOR" in os.environ:
|
||||
del os.environ["CLICOLOR"]
|
||||
assert AnsiMode.ALWAYS.use_ansi_codes(tty_stream)
|
||||
assert AnsiMode.ALWAYS.use_ansi_codes(non_tty_stream)
|
||||
|
||||
os.environ["CLICOLOR"] = "0"
|
||||
assert AnsiMode.ALWAYS.use_ansi_codes(tty_stream)
|
||||
assert AnsiMode.ALWAYS.use_ansi_codes(non_tty_stream)
|
||||
|
||||
@mock.patch.dict(os.environ)
|
||||
def test_ansi_mode_auto(self, tty_stream, non_tty_stream):
|
||||
if "CLICOLOR" in os.environ:
|
||||
del os.environ["CLICOLOR"]
|
||||
assert AnsiMode.AUTO.use_ansi_codes(tty_stream)
|
||||
assert not AnsiMode.AUTO.use_ansi_codes(non_tty_stream)
|
||||
|
||||
os.environ["CLICOLOR"] = "0"
|
||||
assert not AnsiMode.AUTO.use_ansi_codes(tty_stream)
|
||||
assert not AnsiMode.AUTO.use_ansi_codes(non_tty_stream)
|
||||
@@ -8,7 +8,6 @@ from docker.errors import APIError
|
||||
|
||||
from compose.cli.log_printer import build_log_generator
|
||||
from compose.cli.log_printer import build_log_presenters
|
||||
from compose.cli.log_printer import build_no_log_generator
|
||||
from compose.cli.log_printer import consume_queue
|
||||
from compose.cli.log_printer import QueueItem
|
||||
from compose.cli.log_printer import wait_on_exit
|
||||
@@ -75,14 +74,6 @@ def test_wait_on_exit_raises():
|
||||
assert expected in wait_on_exit(mock_container)
|
||||
|
||||
|
||||
def test_build_no_log_generator(mock_container):
|
||||
mock_container.has_api_logs = False
|
||||
mock_container.log_driver = 'none'
|
||||
output, = build_no_log_generator(mock_container, None)
|
||||
assert "WARNING: no logs are available with the 'none' log driver\n" in output
|
||||
assert "exited with code" not in output
|
||||
|
||||
|
||||
class TestBuildLogGenerator:
|
||||
|
||||
def test_no_log_stream(self, mock_container):
|
||||
|
||||
@@ -137,21 +137,20 @@ class TestCLIMainTestCase:
|
||||
|
||||
class TestSetupConsoleHandlerTestCase:
|
||||
|
||||
def test_with_tty_verbose(self, logging_handler):
|
||||
def test_with_console_formatter_verbose(self, logging_handler):
|
||||
setup_console_handler(logging_handler, True)
|
||||
assert type(logging_handler.formatter) == ConsoleWarningFormatter
|
||||
assert '%(name)s' in logging_handler.formatter._fmt
|
||||
assert '%(funcName)s' in logging_handler.formatter._fmt
|
||||
|
||||
def test_with_tty_not_verbose(self, logging_handler):
|
||||
def test_with_console_formatter_not_verbose(self, logging_handler):
|
||||
setup_console_handler(logging_handler, False)
|
||||
assert type(logging_handler.formatter) == ConsoleWarningFormatter
|
||||
assert '%(name)s' not in logging_handler.formatter._fmt
|
||||
assert '%(funcName)s' not in logging_handler.formatter._fmt
|
||||
|
||||
def test_with_not_a_tty(self, logging_handler):
|
||||
logging_handler.stream.isatty.return_value = False
|
||||
setup_console_handler(logging_handler, False)
|
||||
def test_without_console_formatter(self, logging_handler):
|
||||
setup_console_handler(logging_handler, False, use_console_formatter=False)
|
||||
assert type(logging_handler.formatter) == logging.Formatter
|
||||
|
||||
|
||||
|
||||
@@ -669,7 +669,7 @@ class ConfigTest(unittest.TestCase):
|
||||
|
||||
assert 'Invalid service name \'mong\\o\'' in excinfo.exconly()
|
||||
|
||||
def test_config_duplicate_cache_from_values_validation_error(self):
|
||||
def test_config_duplicate_cache_from_values_no_validation_error(self):
|
||||
with pytest.raises(ConfigurationError) as exc:
|
||||
config.load(
|
||||
build_config_details({
|
||||
@@ -681,7 +681,7 @@ class ConfigTest(unittest.TestCase):
|
||||
})
|
||||
)
|
||||
|
||||
assert 'build.cache_from contains non-unique items' in exc.exconly()
|
||||
assert 'build.cache_from contains non-unique items' not in exc.exconly()
|
||||
|
||||
def test_load_with_multiple_files_v1(self):
|
||||
base_file = config.ConfigFile(
|
||||
@@ -2397,7 +2397,8 @@ web:
|
||||
'image': 'busybox',
|
||||
'depends_on': {
|
||||
'app1': {'condition': 'service_started'},
|
||||
'app2': {'condition': 'service_healthy'}
|
||||
'app2': {'condition': 'service_healthy'},
|
||||
'app3': {'condition': 'service_completed_successfully'}
|
||||
}
|
||||
}
|
||||
override = {}
|
||||
@@ -2409,11 +2410,12 @@ web:
|
||||
'image': 'busybox',
|
||||
'depends_on': {
|
||||
'app1': {'condition': 'service_started'},
|
||||
'app2': {'condition': 'service_healthy'}
|
||||
'app2': {'condition': 'service_healthy'},
|
||||
'app3': {'condition': 'service_completed_successfully'}
|
||||
}
|
||||
}
|
||||
override = {
|
||||
'depends_on': ['app3']
|
||||
'depends_on': ['app4']
|
||||
}
|
||||
|
||||
actual = config.merge_service_dicts(base, override, VERSION)
|
||||
@@ -2422,7 +2424,8 @@ web:
|
||||
'depends_on': {
|
||||
'app1': {'condition': 'service_started'},
|
||||
'app2': {'condition': 'service_healthy'},
|
||||
'app3': {'condition': 'service_started'}
|
||||
'app3': {'condition': 'service_completed_successfully'},
|
||||
'app4': {'condition': 'service_started'},
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3567,9 +3570,11 @@ class InterpolationTest(unittest.TestCase):
|
||||
@mock.patch.dict(os.environ)
|
||||
def test_config_file_with_options_environment_file(self):
|
||||
project_dir = 'tests/fixtures/default-env-file'
|
||||
# env-file is relative to current working dir
|
||||
env = Environment.from_env_file(project_dir, project_dir + '/.env2')
|
||||
service_dicts = config.load(
|
||||
config.find(
|
||||
project_dir, None, Environment.from_env_file(project_dir, '.env2')
|
||||
project_dir, None, env
|
||||
)
|
||||
).services
|
||||
|
||||
@@ -5233,6 +5238,8 @@ class GetDefaultConfigFilesTestCase(unittest.TestCase):
|
||||
files = [
|
||||
'docker-compose.yml',
|
||||
'docker-compose.yaml',
|
||||
'compose.yml',
|
||||
'compose.yaml',
|
||||
]
|
||||
|
||||
def test_get_config_path_default_file_in_basedir(self):
|
||||
@@ -5266,8 +5273,10 @@ def get_config_filename_for_files(filenames, subdir=None):
|
||||
base_dir = tempfile.mkdtemp(dir=project_dir)
|
||||
else:
|
||||
base_dir = project_dir
|
||||
filename, = config.get_default_config_files(base_dir)
|
||||
return os.path.basename(filename)
|
||||
filenames = config.get_default_config_files(base_dir)
|
||||
if not filenames:
|
||||
raise config.ComposeFileNotFound(config.SUPPORTED_FILENAMES)
|
||||
return os.path.basename(filenames[0])
|
||||
finally:
|
||||
shutil.rmtree(project_dir)
|
||||
|
||||
|
||||
@@ -221,34 +221,6 @@ class ContainerTest(unittest.TestCase):
|
||||
container = Container(None, self.container_dict, has_been_inspected=True)
|
||||
assert container.short_id == self.container_id[:12]
|
||||
|
||||
def test_has_api_logs(self):
|
||||
container_dict = {
|
||||
'HostConfig': {
|
||||
'LogConfig': {
|
||||
'Type': 'json-file'
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
container = Container(None, container_dict, has_been_inspected=True)
|
||||
assert container.has_api_logs is True
|
||||
|
||||
container_dict['HostConfig']['LogConfig']['Type'] = 'none'
|
||||
container = Container(None, container_dict, has_been_inspected=True)
|
||||
assert container.has_api_logs is False
|
||||
|
||||
container_dict['HostConfig']['LogConfig']['Type'] = 'syslog'
|
||||
container = Container(None, container_dict, has_been_inspected=True)
|
||||
assert container.has_api_logs is False
|
||||
|
||||
container_dict['HostConfig']['LogConfig']['Type'] = 'journald'
|
||||
container = Container(None, container_dict, has_been_inspected=True)
|
||||
assert container.has_api_logs is True
|
||||
|
||||
container_dict['HostConfig']['LogConfig']['Type'] = 'foobar'
|
||||
container = Container(None, container_dict, has_been_inspected=True)
|
||||
assert container.has_api_logs is False
|
||||
|
||||
|
||||
class GetContainerNameTestCase(unittest.TestCase):
|
||||
|
||||
|
||||
0
tests/unit/metrics/__init__.py
Normal file
0
tests/unit/metrics/__init__.py
Normal file
36
tests/unit/metrics/metrics_test.py
Normal file
36
tests/unit/metrics/metrics_test.py
Normal file
@@ -0,0 +1,36 @@
|
||||
import unittest
|
||||
|
||||
from compose.metrics.client import MetricsCommand
|
||||
from compose.metrics.client import Status
|
||||
|
||||
|
||||
class MetricsTest(unittest.TestCase):
|
||||
@classmethod
|
||||
def test_metrics(cls):
|
||||
assert MetricsCommand('up', 'moby').to_map() == {
|
||||
'command': 'compose up',
|
||||
'context': 'moby',
|
||||
'status': 'success',
|
||||
'source': 'docker-compose',
|
||||
}
|
||||
|
||||
assert MetricsCommand('down', 'local').to_map() == {
|
||||
'command': 'compose down',
|
||||
'context': 'local',
|
||||
'status': 'success',
|
||||
'source': 'docker-compose',
|
||||
}
|
||||
|
||||
assert MetricsCommand('help', 'aci', Status.FAILURE).to_map() == {
|
||||
'command': 'compose help',
|
||||
'context': 'aci',
|
||||
'status': 'failure',
|
||||
'source': 'docker-compose',
|
||||
}
|
||||
|
||||
assert MetricsCommand('run', 'ecs').to_map() == {
|
||||
'command': 'compose run',
|
||||
'context': 'ecs',
|
||||
'status': 'success',
|
||||
'source': 'docker-compose',
|
||||
}
|
||||
@@ -3,6 +3,7 @@ from threading import Lock
|
||||
|
||||
from docker.errors import APIError
|
||||
|
||||
from compose.cli.colors import AnsiMode
|
||||
from compose.parallel import GlobalLimit
|
||||
from compose.parallel import parallel_execute
|
||||
from compose.parallel import parallel_execute_iter
|
||||
@@ -156,7 +157,7 @@ def test_parallel_execute_alignment(capsys):
|
||||
|
||||
def test_parallel_execute_ansi(capsys):
|
||||
ParallelStreamWriter.instance = None
|
||||
ParallelStreamWriter.set_noansi(value=False)
|
||||
ParallelStreamWriter.set_default_ansi_mode(AnsiMode.ALWAYS)
|
||||
results, errors = parallel_execute(
|
||||
objects=["something", "something more"],
|
||||
func=lambda x: x,
|
||||
@@ -172,7 +173,7 @@ def test_parallel_execute_ansi(capsys):
|
||||
|
||||
def test_parallel_execute_noansi(capsys):
|
||||
ParallelStreamWriter.instance = None
|
||||
ParallelStreamWriter.set_noansi()
|
||||
ParallelStreamWriter.set_default_ansi_mode(AnsiMode.NEVER)
|
||||
results, errors = parallel_execute(
|
||||
objects=["something", "something more"],
|
||||
func=lambda x: x,
|
||||
|
||||
@@ -330,7 +330,7 @@ class ServiceTest(unittest.TestCase):
|
||||
assert service.options['environment'] == environment
|
||||
|
||||
assert opts['labels'][LABEL_CONFIG_HASH] == \
|
||||
'689149e6041a85f6fb4945a2146a497ed43c8a5cbd8991753d875b165f1b4de4'
|
||||
'6da0f3ec0d5adf901de304bdc7e0ee44ec5dd7adb08aebc20fe0dd791d4ee5a8'
|
||||
assert opts['environment'] == ['also=real']
|
||||
|
||||
def test_get_container_create_options_sets_affinity_with_binds(self):
|
||||
@@ -700,6 +700,7 @@ class ServiceTest(unittest.TestCase):
|
||||
config_dict = service.config_dict()
|
||||
expected = {
|
||||
'image_id': 'abcd',
|
||||
'ipc_mode': None,
|
||||
'options': {'image': 'example.com/foo'},
|
||||
'links': [('one', 'one')],
|
||||
'net': 'other',
|
||||
@@ -723,6 +724,7 @@ class ServiceTest(unittest.TestCase):
|
||||
config_dict = service.config_dict()
|
||||
expected = {
|
||||
'image_id': 'abcd',
|
||||
'ipc_mode': None,
|
||||
'options': {'image': 'example.com/foo'},
|
||||
'links': [],
|
||||
'networks': {},
|
||||
|
||||
Reference in New Issue
Block a user