1.14.1.2 Common Pipeline Practices
The following are practices that we see a lot of people use in their pipelines. These are by no means "Best" practices, they are simply common and may or may not work for you and your team.
Parallelizing Get Steps in Jobs
All jobs usually have get steps as their first set of steps.
jobs:
- name: awesome-job
plan:
- get: cool-code
- get: funny-binary
- get: the-weather
- task: business-stuffTo reduce the waiting time to the length of the longest running get step, put all get steps under an in_parallel step.
jobs:
- name: awesome-job
plan:
- in_parallel:
- get: cool-code
- get: funny-binary
- get: the-weather
- task: business-stuff
Specify Inputs for Put Steps
By default put step's have all artifacts from a job mounted in their resource container. This can result in long initialization times for put steps. It's likely that a put step only needs a subset of all available artifacts generated throughout the job.
There are two ways to specify which artifacts to send to a put step. You can specify detect as the put step inputs or you can pass in an exact list of all artifacts the put step needs.
Using detect:
jobs:
- name: the-job
plan: # Get some artifacts
- in_parallel:
- get: apples
- get: apple-basket
- get: monkeys
# using detect will result in "apples-location" and "basket" being passed in
# "monkeys" will not be passed in
- put: apples-in-basket
inputs: detect
params:
apples-location: apples/location # matches the first get step
basket: apple-basket # matches the second get stepSpecifying the exact inputs needed for the put step:
jobs:
- name: the-job
plan: # Get some artifacts
- in_parallel:
- get: apples
- get: apple-basket
- get: monkeys
- put: apples-in-basket
inputs: [apples, apple-basket] # specify the exact inputs needed
params:
apples-location: apples/location
basket: apple-basket
Putting Task Configs in Files
A lot of the pipeline examples that you will find on this site and in resource repos will embed a task step config directly in the pipeline. This is a nice way of clearly seeing what inputs/outputs the task uses. Tasks are usually designed to be used in multiple places, maybe with slightly different configuration. To support this scenario, most users store task configs in files instead of embedding the config directly in the pipeline.
Here's what this looks like in practice:
print-date.yml, stored in some git repo under a folder named "tasks"
platform: linux
image_resource: # define a default image for the task to use
type: registry-image
source:
repository: busybox
run:
path: date
args: ["+%Y-%m-%d"]
Using the task in a pipeline:
pipeline.yml
resources:
- name: ci
type: git
source:
uri: https://github.com/concourse/examples.git
jobs:
- name: the-best-job
plan:
- get: ci
- task: today
file: ci/tasks/print-date.yml
Get Images for Tasks Instead of using Anonymous Image Resources
It is easy to let Concourse fetch images for tasks right when they are needed by using the field in a task config. It's the easy out-of-the-box solution. Another way is to pass the image for a task as an input to the job by setting the task step image field. This also allows you to track the version of the image being used by the task and also avoid getting rate-limited by configuring the resource with credentials.
Before - Anonymous Image Fetching for Tasks
jobs:
- name: job
plan:
- task: simple-task
config:
platform: linux
image_resource: # anonymous image resource
type: registry-image
source:
repository: busybox
run:
path: echo
args: ["Hello world!"]After - Passing Task Image as Job Inputs
resources:
- name: busybox
type: registry-image
source:
repository: busybox
username: ((docker.user))
password: ((docker.password))
jobs:
- name: job
plan:
- get: busybox # pass image into job
- task: simple-task
image: busybox # use image for task. Overrides anonymous image
config:
platform: linux
run:
path: echo
args: ["Hello world!"]