2015-04-29 21:20:09 +02:00
|
|
|
<!--GITHUB
|
2015-04-02 17:11:19 +02:00
|
|
|
page_title: Configure a Registry
|
2015-05-05 10:25:42 +02:00
|
|
|
page_description: Explains how to deploy a registry
|
2015-04-02 17:11:19 +02:00
|
|
|
page_keywords: registry, service, images, repository
|
2015-04-29 21:20:09 +02:00
|
|
|
IGNORES-->
|
2015-04-02 17:11:19 +02:00
|
|
|
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
# Registry Configuration Reference
|
2015-04-02 17:11:19 +02:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
You configure a registry server using a YAML file. This page explains the
|
|
|
|
configuration options and the values they can take. You'll also find examples of
|
|
|
|
middleware and development environment configurations.
|
2015-04-02 17:11:19 +02:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
## List of configuration options
|
2015-04-02 17:11:19 +02:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
This section lists all the registry configuration options. Some options in
|
|
|
|
the list are mutually exclusive. So, make sure to read the detailed reference
|
|
|
|
information about each option that appears later in this page.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
```yaml
|
|
|
|
version: 0.1
|
2015-03-24 04:13:35 +01:00
|
|
|
log:
|
|
|
|
level: debug
|
|
|
|
formatter: text
|
2015-03-25 01:12:04 +01:00
|
|
|
fields:
|
|
|
|
service: registry
|
|
|
|
environment: staging
|
2015-03-24 04:13:35 +01:00
|
|
|
loglevel: debug # deprecated: use "log"
|
2015-03-04 01:18:22 +01:00
|
|
|
storage:
|
|
|
|
filesystem:
|
|
|
|
rootdirectory: /tmp/registry
|
|
|
|
azure:
|
|
|
|
accountname: accountname
|
|
|
|
accountkey: base64encodedaccountkey
|
|
|
|
container: containername
|
|
|
|
s3:
|
|
|
|
accesskey: awsaccesskey
|
|
|
|
secretkey: awssecretkey
|
|
|
|
region: us-west-1
|
|
|
|
bucket: bucketname
|
|
|
|
encrypt: true
|
|
|
|
secure: true
|
|
|
|
v4auth: true
|
2015-03-25 00:28:20 +01:00
|
|
|
chunksize: 5242880
|
2015-03-04 01:18:22 +01:00
|
|
|
rootdirectory: /s3/object/name/prefix
|
2015-04-03 01:38:01 +02:00
|
|
|
cache:
|
Refactor Blob Service API
This PR refactors the blob service API to be oriented around blob descriptors.
Identified by digests, blobs become an abstract entity that can be read and
written using a descriptor as a handle. This allows blobs to take many forms,
such as a ReadSeekCloser or a simple byte buffer, allowing blob oriented
operations to better integrate with blob agnostic APIs (such as the `io`
package). The error definitions are now better organized to reflect conditions
that can only be seen when interacting with the blob API.
The main benefit of this is to separate the much smaller metadata from large
file storage. Many benefits also follow from this. Reading and writing has
been separated into discrete services. Backend implementation is also
simplified, by reducing the amount of metadata that needs to be picked up to
simply serve a read. This also improves cacheability.
"Opening" a blob simply consists of an access check (Stat) and a path
calculation. Caching is greatly simplified and we've made the mapping of
provisional to canonical hashes a first-class concept. BlobDescriptorService
and BlobProvider can be combined in different ways to achieve varying effects.
Recommend Review Approach
-------------------------
This is a very large patch. While apologies are in order, we are getting a
considerable amount of refactoring. Most changes follow from the changes to
the root package (distribution), so start there. From there, the main changes
are in storage. Looking at (*repository).Blobs will help to understand the how
the linkedBlobStore is wired. One can explore the internals within and also
branch out into understanding the changes to the caching layer. Following the
descriptions below will also help to guide you.
To reduce the chances for regressions, it was critical that major changes to
unit tests were avoided. Where possible, they are left untouched and where
not, the spirit is hopefully captured. Pay particular attention to where
behavior may have changed.
Storage
-------
The primary changes to the `storage` package, other than the interface
updates, were to merge the layerstore and blobstore. Blob access is now
layered even further. The first layer, blobStore, exposes a global
`BlobStatter` and `BlobProvider`. Operations here provide a fast path for most
read operations that don't take access control into account. The
`linkedBlobStore` layers on top of the `blobStore`, providing repository-
scoped blob link management in the backend. The `linkedBlobStore` implements
the full `BlobStore` suite, providing access-controlled, repository-local blob
writers. The abstraction between the two is slightly broken in that
`linkedBlobStore` is the only channel under which one can write into the global
blob store. The `linkedBlobStore` also provides flexibility in that it can act
over different link sets depending on configuration. This allows us to use the
same code for signature links, manifest links and blob links. Eventually, we
will fully consolidate this storage.
The improved cache flow comes from the `linkedBlobStatter` component
of `linkedBlobStore`. Using a `cachedBlobStatter`, these combine together to
provide a simple cache hierarchy that should streamline access checks on read
and write operations, or at least provide a single path to optimize. The
metrics have been changed in a slightly incompatible way since the former
operations, Fetch and Exists, are no longer relevant.
The fileWriter and fileReader have been slightly modified to support the rest
of the changes. The most interesting is the removal of the `Stat` call from
`newFileReader`. This was the source of unnecessary round trips that were only
present to look up the size of the resulting reader. Now, one must simply pass
in the size, requiring the caller to decide whether or not the `Stat` call is
appropriate. In several cases, it turned out the caller already had the size
already. The `WriterAt` implementation has been removed from `fileWriter`,
since it is no longer required for `BlobWriter`, reducing the number of paths
which writes may take.
Cache
-----
Unfortunately, the `cache` package required a near full rewrite. It was pretty
mechanical in that the cache is oriented around the `BlobDescriptorService`
slightly modified to include the ability to set the values for individual
digests. While the implementation is oriented towards caching, it can act as a
primary store. Provisions are in place to have repository local metadata, in
addition to global metadata. Fallback is implemented as a part of the storage
package to maintain this flexibility.
One unfortunate side-effect is that caching is now repository-scoped, rather
than global. This should have little effect on performance but may increase
memory usage.
Handlers
--------
The `handlers` package has been updated to leverage the new API. For the most
part, the changes are superficial or mechanical based on the API changes. This
did expose a bug in the handling of provisional vs canonical digests that was
fixed in the unit tests.
Configuration
-------------
One user-facing change has been made to the configuration and is updated in
the associated documentation. The `layerinfo` cache parameter has been
deprecated by the `blobdescriptor` cache parameter. Both are equivalent and
configuration files should be backward compatible.
Notifications
-------------
Changes the `notification` package are simply to support the interface
changes.
Context
-------
A small change has been made to the tracing log-level. Traces have been moved
from "info" to "debug" level to reduce output when not needed.
Signed-off-by: Stephen J Day <stephen.day@docker.com>
2015-05-12 09:10:29 +02:00
|
|
|
blobdescriptor: redis
|
2015-04-17 03:34:29 +02:00
|
|
|
maintenance:
|
|
|
|
uploadpurging:
|
|
|
|
enabled: true
|
|
|
|
age: 168h
|
|
|
|
interval: 24h
|
|
|
|
dryrun: false
|
2015-03-04 01:18:22 +01:00
|
|
|
auth:
|
|
|
|
silly:
|
|
|
|
realm: silly-realm
|
|
|
|
service: silly-service
|
|
|
|
token:
|
|
|
|
realm: token-realm
|
|
|
|
service: token-service
|
|
|
|
issuer: registry-token-issuer
|
|
|
|
rootcertbundle: /root/certs/bundle
|
|
|
|
middleware:
|
|
|
|
registry:
|
|
|
|
- name: ARegistryMiddleware
|
|
|
|
options:
|
|
|
|
foo: bar
|
|
|
|
repository:
|
|
|
|
- name: ARepositoryMiddleware
|
|
|
|
options:
|
|
|
|
foo: bar
|
|
|
|
storage:
|
|
|
|
- name: cloudfront
|
|
|
|
options:
|
|
|
|
baseurl: https://my.cloudfronted.domain.com/
|
|
|
|
privatekey: /path/to/pem
|
|
|
|
keypairid: cloudfrontkeypairid
|
|
|
|
duration: 3000
|
|
|
|
reporting:
|
|
|
|
bugsnag:
|
|
|
|
apikey: bugsnagapikey
|
|
|
|
releasestage: bugsnagreleasestage
|
|
|
|
endpoint: bugsnagendpoint
|
|
|
|
newrelic:
|
|
|
|
licensekey: newreliclicensekey
|
|
|
|
name: newrelicname
|
2015-04-13 20:34:07 +02:00
|
|
|
verbose: true
|
2015-03-04 01:18:22 +01:00
|
|
|
http:
|
|
|
|
addr: localhost:5000
|
|
|
|
prefix: /my/nested/registry/
|
|
|
|
secret: asecretforlocaldevelopment
|
|
|
|
tls:
|
|
|
|
certificate: /path/to/x509/public
|
|
|
|
key: /path/to/x509/private
|
2015-03-20 16:19:07 +01:00
|
|
|
clientcas:
|
|
|
|
- /path/to/ca.pem
|
|
|
|
- /path/to/another/ca.pem
|
2015-03-04 01:18:22 +01:00
|
|
|
debug:
|
|
|
|
addr: localhost:5001
|
|
|
|
notifications:
|
2015-05-05 10:25:42 +02:00
|
|
|
endpoints:
|
2015-03-04 01:18:22 +01:00
|
|
|
- name: alistener
|
|
|
|
disabled: false
|
|
|
|
url: https://my.listener.com/event
|
|
|
|
headers: <http.Header>
|
|
|
|
timeout: 500
|
|
|
|
threshold: 5
|
|
|
|
backoff: 1000
|
2015-04-02 01:27:24 +02:00
|
|
|
redis:
|
|
|
|
addr: localhost:6379
|
|
|
|
password: asecret
|
|
|
|
db: 0
|
|
|
|
dialtimeout: 10ms
|
|
|
|
readtimeout: 10ms
|
|
|
|
writetimeout: 10ms
|
|
|
|
pool:
|
|
|
|
maxidle: 16
|
|
|
|
maxactive: 64
|
|
|
|
idletimeout: 300s
|
2015-03-04 01:18:22 +01:00
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
In some instances a configuration option is **optional** but it contains child
|
|
|
|
options marked as **required**. This indicates that you can omit the parent with
|
|
|
|
all its children. However, if the parent is included, you must also include all
|
|
|
|
the children marked **required**.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-28 00:49:19 +02:00
|
|
|
## Override configuration options
|
|
|
|
|
|
|
|
You can use environment variables to override most configuration parameters. The
|
|
|
|
exception is the `version` variable which cannot be overridden. You can set
|
|
|
|
environment variables on the command line using the `-e` flag on `docker run` or
|
|
|
|
from within a Dockerfile using the `ENV` instruction.
|
|
|
|
|
|
|
|
To override a configuration option, create an environment variable named
|
|
|
|
`REGISTRY\variable_` where *`variable`* is the name of the configuration option
|
|
|
|
and the `_` (underscore) represents indention levels. For example, you can
|
|
|
|
configure the `rootdirectory` of the `filesystem` storage backend:
|
|
|
|
|
|
|
|
```
|
|
|
|
storage:
|
|
|
|
filesystem:
|
|
|
|
rootdirectory: /tmp/registry
|
|
|
|
```
|
|
|
|
|
|
|
|
To override this value, set an environment variable like this:
|
|
|
|
|
|
|
|
```
|
|
|
|
REGISTRY_STORAGE_FILESYSTEM_ROOTDIRECTORY=/tmp/registry/test
|
|
|
|
```
|
|
|
|
|
|
|
|
This variable overrides the `/tmp/registry` value to the `/tmp/registry/test`
|
|
|
|
directory.
|
2015-05-07 00:31:12 +02:00
|
|
|
|
|
|
|
>**Note**: If an environment variable changes a map value into a string, such
|
|
|
|
>as replacing the storage driver type with `REGISTRY_STORAGE=filesystem`, then
|
|
|
|
>all sub-fields will be erased. As such, specifying the storage type in the
|
|
|
|
>environment will remove all parameters related to the old storage
|
|
|
|
>configuration.
|
2015-04-28 00:49:19 +02:00
|
|
|
|
|
|
|
|
2015-05-05 10:25:42 +02:00
|
|
|
## version
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
```yaml
|
|
|
|
version: 0.1
|
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `version` option is **required**. It specifies the configuration's version.
|
|
|
|
It is expected to remain a top-level field, to allow for a consistent version
|
2015-05-05 10:25:42 +02:00
|
|
|
check before parsing the remainder of the configuration file.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-03-24 04:13:35 +01:00
|
|
|
## log
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `log` subsection configures the behavior of the logging system. The logging
|
|
|
|
system outputs everything to stdout. You can adjust the granularity and format
|
|
|
|
with this configuration section.
|
2015-03-24 04:13:35 +01:00
|
|
|
|
|
|
|
```yaml
|
|
|
|
log:
|
|
|
|
level: debug
|
|
|
|
formatter: text
|
2015-03-25 01:12:04 +01:00
|
|
|
fields:
|
|
|
|
service: registry
|
|
|
|
environment: staging
|
2015-03-24 04:13:35 +01:00
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>level</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Sets the sensitivity of logging output. Permitted values are
|
|
|
|
<code>error</code>, <code>warn</code>, <code>info</code> and
|
|
|
|
<code>debug</code>. The default is <code>info</code>.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>formatter</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
This selects the format of logging output. The format primarily affects how keyed
|
|
|
|
attributes for a log line are encoded. Options are <code>text</code>, <code>json</code> or
|
|
|
|
<code>logstash</code>. The default is <code>text</code>.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>fields</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
A map of field names to values. These are added to every log line for
|
|
|
|
the context. This is useful for identifying log messages source after
|
|
|
|
being mixed in other systems.
|
|
|
|
</td>
|
|
|
|
</table>
|
|
|
|
|
2015-03-24 04:13:35 +01:00
|
|
|
|
2015-03-04 01:18:22 +01:00
|
|
|
## loglevel
|
|
|
|
|
2015-03-24 04:13:35 +01:00
|
|
|
> **DEPRECATED:** Please use [log](#log) instead.
|
|
|
|
|
2015-03-04 01:18:22 +01:00
|
|
|
```yaml
|
|
|
|
loglevel: debug
|
|
|
|
```
|
|
|
|
|
2015-03-24 04:13:35 +01:00
|
|
|
Permitted values are `error`, `warn`, `info` and `debug`. The default is
|
|
|
|
`info`.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
## storage
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
storage:
|
|
|
|
filesystem:
|
|
|
|
rootdirectory: /tmp/registry
|
|
|
|
azure:
|
|
|
|
accountname: accountname
|
|
|
|
accountkey: base64encodedaccountkey
|
|
|
|
container: containername
|
|
|
|
s3:
|
|
|
|
accesskey: awsaccesskey
|
|
|
|
secretkey: awssecretkey
|
|
|
|
region: us-west-1
|
|
|
|
bucket: bucketname
|
|
|
|
encrypt: true
|
|
|
|
secure: true
|
|
|
|
v4auth: true
|
2015-03-25 00:28:20 +01:00
|
|
|
chunksize: 5242880
|
2015-03-04 01:18:22 +01:00
|
|
|
rootdirectory: /s3/object/name/prefix
|
2015-04-03 01:38:01 +02:00
|
|
|
cache:
|
Refactor Blob Service API
This PR refactors the blob service API to be oriented around blob descriptors.
Identified by digests, blobs become an abstract entity that can be read and
written using a descriptor as a handle. This allows blobs to take many forms,
such as a ReadSeekCloser or a simple byte buffer, allowing blob oriented
operations to better integrate with blob agnostic APIs (such as the `io`
package). The error definitions are now better organized to reflect conditions
that can only be seen when interacting with the blob API.
The main benefit of this is to separate the much smaller metadata from large
file storage. Many benefits also follow from this. Reading and writing has
been separated into discrete services. Backend implementation is also
simplified, by reducing the amount of metadata that needs to be picked up to
simply serve a read. This also improves cacheability.
"Opening" a blob simply consists of an access check (Stat) and a path
calculation. Caching is greatly simplified and we've made the mapping of
provisional to canonical hashes a first-class concept. BlobDescriptorService
and BlobProvider can be combined in different ways to achieve varying effects.
Recommend Review Approach
-------------------------
This is a very large patch. While apologies are in order, we are getting a
considerable amount of refactoring. Most changes follow from the changes to
the root package (distribution), so start there. From there, the main changes
are in storage. Looking at (*repository).Blobs will help to understand the how
the linkedBlobStore is wired. One can explore the internals within and also
branch out into understanding the changes to the caching layer. Following the
descriptions below will also help to guide you.
To reduce the chances for regressions, it was critical that major changes to
unit tests were avoided. Where possible, they are left untouched and where
not, the spirit is hopefully captured. Pay particular attention to where
behavior may have changed.
Storage
-------
The primary changes to the `storage` package, other than the interface
updates, were to merge the layerstore and blobstore. Blob access is now
layered even further. The first layer, blobStore, exposes a global
`BlobStatter` and `BlobProvider`. Operations here provide a fast path for most
read operations that don't take access control into account. The
`linkedBlobStore` layers on top of the `blobStore`, providing repository-
scoped blob link management in the backend. The `linkedBlobStore` implements
the full `BlobStore` suite, providing access-controlled, repository-local blob
writers. The abstraction between the two is slightly broken in that
`linkedBlobStore` is the only channel under which one can write into the global
blob store. The `linkedBlobStore` also provides flexibility in that it can act
over different link sets depending on configuration. This allows us to use the
same code for signature links, manifest links and blob links. Eventually, we
will fully consolidate this storage.
The improved cache flow comes from the `linkedBlobStatter` component
of `linkedBlobStore`. Using a `cachedBlobStatter`, these combine together to
provide a simple cache hierarchy that should streamline access checks on read
and write operations, or at least provide a single path to optimize. The
metrics have been changed in a slightly incompatible way since the former
operations, Fetch and Exists, are no longer relevant.
The fileWriter and fileReader have been slightly modified to support the rest
of the changes. The most interesting is the removal of the `Stat` call from
`newFileReader`. This was the source of unnecessary round trips that were only
present to look up the size of the resulting reader. Now, one must simply pass
in the size, requiring the caller to decide whether or not the `Stat` call is
appropriate. In several cases, it turned out the caller already had the size
already. The `WriterAt` implementation has been removed from `fileWriter`,
since it is no longer required for `BlobWriter`, reducing the number of paths
which writes may take.
Cache
-----
Unfortunately, the `cache` package required a near full rewrite. It was pretty
mechanical in that the cache is oriented around the `BlobDescriptorService`
slightly modified to include the ability to set the values for individual
digests. While the implementation is oriented towards caching, it can act as a
primary store. Provisions are in place to have repository local metadata, in
addition to global metadata. Fallback is implemented as a part of the storage
package to maintain this flexibility.
One unfortunate side-effect is that caching is now repository-scoped, rather
than global. This should have little effect on performance but may increase
memory usage.
Handlers
--------
The `handlers` package has been updated to leverage the new API. For the most
part, the changes are superficial or mechanical based on the API changes. This
did expose a bug in the handling of provisional vs canonical digests that was
fixed in the unit tests.
Configuration
-------------
One user-facing change has been made to the configuration and is updated in
the associated documentation. The `layerinfo` cache parameter has been
deprecated by the `blobdescriptor` cache parameter. Both are equivalent and
configuration files should be backward compatible.
Notifications
-------------
Changes the `notification` package are simply to support the interface
changes.
Context
-------
A small change has been made to the tracing log-level. Traces have been moved
from "info" to "debug" level to reduce output when not needed.
Signed-off-by: Stephen J Day <stephen.day@docker.com>
2015-05-12 09:10:29 +02:00
|
|
|
blobdescriptor: inmemory
|
2015-04-17 03:34:29 +02:00
|
|
|
maintenance:
|
|
|
|
uploadpurging:
|
|
|
|
enabled: true
|
|
|
|
age: 168h
|
|
|
|
interval: 24h
|
|
|
|
dryrun: false
|
2015-03-04 01:18:22 +01:00
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The storage option is **required** and defines which storage backend is in use.
|
|
|
|
You must configure one backend; if you configure more, the registry returns an error.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
### cache
|
2015-04-03 01:38:01 +02:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
Use the `cache` subsection to enable caching of data accessed in the storage
|
|
|
|
backend. Currently, the only available cache provides fast access to layer
|
Refactor Blob Service API
This PR refactors the blob service API to be oriented around blob descriptors.
Identified by digests, blobs become an abstract entity that can be read and
written using a descriptor as a handle. This allows blobs to take many forms,
such as a ReadSeekCloser or a simple byte buffer, allowing blob oriented
operations to better integrate with blob agnostic APIs (such as the `io`
package). The error definitions are now better organized to reflect conditions
that can only be seen when interacting with the blob API.
The main benefit of this is to separate the much smaller metadata from large
file storage. Many benefits also follow from this. Reading and writing has
been separated into discrete services. Backend implementation is also
simplified, by reducing the amount of metadata that needs to be picked up to
simply serve a read. This also improves cacheability.
"Opening" a blob simply consists of an access check (Stat) and a path
calculation. Caching is greatly simplified and we've made the mapping of
provisional to canonical hashes a first-class concept. BlobDescriptorService
and BlobProvider can be combined in different ways to achieve varying effects.
Recommend Review Approach
-------------------------
This is a very large patch. While apologies are in order, we are getting a
considerable amount of refactoring. Most changes follow from the changes to
the root package (distribution), so start there. From there, the main changes
are in storage. Looking at (*repository).Blobs will help to understand the how
the linkedBlobStore is wired. One can explore the internals within and also
branch out into understanding the changes to the caching layer. Following the
descriptions below will also help to guide you.
To reduce the chances for regressions, it was critical that major changes to
unit tests were avoided. Where possible, they are left untouched and where
not, the spirit is hopefully captured. Pay particular attention to where
behavior may have changed.
Storage
-------
The primary changes to the `storage` package, other than the interface
updates, were to merge the layerstore and blobstore. Blob access is now
layered even further. The first layer, blobStore, exposes a global
`BlobStatter` and `BlobProvider`. Operations here provide a fast path for most
read operations that don't take access control into account. The
`linkedBlobStore` layers on top of the `blobStore`, providing repository-
scoped blob link management in the backend. The `linkedBlobStore` implements
the full `BlobStore` suite, providing access-controlled, repository-local blob
writers. The abstraction between the two is slightly broken in that
`linkedBlobStore` is the only channel under which one can write into the global
blob store. The `linkedBlobStore` also provides flexibility in that it can act
over different link sets depending on configuration. This allows us to use the
same code for signature links, manifest links and blob links. Eventually, we
will fully consolidate this storage.
The improved cache flow comes from the `linkedBlobStatter` component
of `linkedBlobStore`. Using a `cachedBlobStatter`, these combine together to
provide a simple cache hierarchy that should streamline access checks on read
and write operations, or at least provide a single path to optimize. The
metrics have been changed in a slightly incompatible way since the former
operations, Fetch and Exists, are no longer relevant.
The fileWriter and fileReader have been slightly modified to support the rest
of the changes. The most interesting is the removal of the `Stat` call from
`newFileReader`. This was the source of unnecessary round trips that were only
present to look up the size of the resulting reader. Now, one must simply pass
in the size, requiring the caller to decide whether or not the `Stat` call is
appropriate. In several cases, it turned out the caller already had the size
already. The `WriterAt` implementation has been removed from `fileWriter`,
since it is no longer required for `BlobWriter`, reducing the number of paths
which writes may take.
Cache
-----
Unfortunately, the `cache` package required a near full rewrite. It was pretty
mechanical in that the cache is oriented around the `BlobDescriptorService`
slightly modified to include the ability to set the values for individual
digests. While the implementation is oriented towards caching, it can act as a
primary store. Provisions are in place to have repository local metadata, in
addition to global metadata. Fallback is implemented as a part of the storage
package to maintain this flexibility.
One unfortunate side-effect is that caching is now repository-scoped, rather
than global. This should have little effect on performance but may increase
memory usage.
Handlers
--------
The `handlers` package has been updated to leverage the new API. For the most
part, the changes are superficial or mechanical based on the API changes. This
did expose a bug in the handling of provisional vs canonical digests that was
fixed in the unit tests.
Configuration
-------------
One user-facing change has been made to the configuration and is updated in
the associated documentation. The `layerinfo` cache parameter has been
deprecated by the `blobdescriptor` cache parameter. Both are equivalent and
configuration files should be backward compatible.
Notifications
-------------
Changes the `notification` package are simply to support the interface
changes.
Context
-------
A small change has been made to the tracing log-level. Traces have been moved
from "info" to "debug" level to reduce output when not needed.
Signed-off-by: Stephen J Day <stephen.day@docker.com>
2015-05-12 09:10:29 +02:00
|
|
|
metadata. This, if configured, uses the `blobdescriptor` field.
|
2015-04-03 01:38:01 +02:00
|
|
|
|
Refactor Blob Service API
This PR refactors the blob service API to be oriented around blob descriptors.
Identified by digests, blobs become an abstract entity that can be read and
written using a descriptor as a handle. This allows blobs to take many forms,
such as a ReadSeekCloser or a simple byte buffer, allowing blob oriented
operations to better integrate with blob agnostic APIs (such as the `io`
package). The error definitions are now better organized to reflect conditions
that can only be seen when interacting with the blob API.
The main benefit of this is to separate the much smaller metadata from large
file storage. Many benefits also follow from this. Reading and writing has
been separated into discrete services. Backend implementation is also
simplified, by reducing the amount of metadata that needs to be picked up to
simply serve a read. This also improves cacheability.
"Opening" a blob simply consists of an access check (Stat) and a path
calculation. Caching is greatly simplified and we've made the mapping of
provisional to canonical hashes a first-class concept. BlobDescriptorService
and BlobProvider can be combined in different ways to achieve varying effects.
Recommend Review Approach
-------------------------
This is a very large patch. While apologies are in order, we are getting a
considerable amount of refactoring. Most changes follow from the changes to
the root package (distribution), so start there. From there, the main changes
are in storage. Looking at (*repository).Blobs will help to understand the how
the linkedBlobStore is wired. One can explore the internals within and also
branch out into understanding the changes to the caching layer. Following the
descriptions below will also help to guide you.
To reduce the chances for regressions, it was critical that major changes to
unit tests were avoided. Where possible, they are left untouched and where
not, the spirit is hopefully captured. Pay particular attention to where
behavior may have changed.
Storage
-------
The primary changes to the `storage` package, other than the interface
updates, were to merge the layerstore and blobstore. Blob access is now
layered even further. The first layer, blobStore, exposes a global
`BlobStatter` and `BlobProvider`. Operations here provide a fast path for most
read operations that don't take access control into account. The
`linkedBlobStore` layers on top of the `blobStore`, providing repository-
scoped blob link management in the backend. The `linkedBlobStore` implements
the full `BlobStore` suite, providing access-controlled, repository-local blob
writers. The abstraction between the two is slightly broken in that
`linkedBlobStore` is the only channel under which one can write into the global
blob store. The `linkedBlobStore` also provides flexibility in that it can act
over different link sets depending on configuration. This allows us to use the
same code for signature links, manifest links and blob links. Eventually, we
will fully consolidate this storage.
The improved cache flow comes from the `linkedBlobStatter` component
of `linkedBlobStore`. Using a `cachedBlobStatter`, these combine together to
provide a simple cache hierarchy that should streamline access checks on read
and write operations, or at least provide a single path to optimize. The
metrics have been changed in a slightly incompatible way since the former
operations, Fetch and Exists, are no longer relevant.
The fileWriter and fileReader have been slightly modified to support the rest
of the changes. The most interesting is the removal of the `Stat` call from
`newFileReader`. This was the source of unnecessary round trips that were only
present to look up the size of the resulting reader. Now, one must simply pass
in the size, requiring the caller to decide whether or not the `Stat` call is
appropriate. In several cases, it turned out the caller already had the size
already. The `WriterAt` implementation has been removed from `fileWriter`,
since it is no longer required for `BlobWriter`, reducing the number of paths
which writes may take.
Cache
-----
Unfortunately, the `cache` package required a near full rewrite. It was pretty
mechanical in that the cache is oriented around the `BlobDescriptorService`
slightly modified to include the ability to set the values for individual
digests. While the implementation is oriented towards caching, it can act as a
primary store. Provisions are in place to have repository local metadata, in
addition to global metadata. Fallback is implemented as a part of the storage
package to maintain this flexibility.
One unfortunate side-effect is that caching is now repository-scoped, rather
than global. This should have little effect on performance but may increase
memory usage.
Handlers
--------
The `handlers` package has been updated to leverage the new API. For the most
part, the changes are superficial or mechanical based on the API changes. This
did expose a bug in the handling of provisional vs canonical digests that was
fixed in the unit tests.
Configuration
-------------
One user-facing change has been made to the configuration and is updated in
the associated documentation. The `layerinfo` cache parameter has been
deprecated by the `blobdescriptor` cache parameter. Both are equivalent and
configuration files should be backward compatible.
Notifications
-------------
Changes the `notification` package are simply to support the interface
changes.
Context
-------
A small change has been made to the tracing log-level. Traces have been moved
from "info" to "debug" level to reduce output when not needed.
Signed-off-by: Stephen J Day <stephen.day@docker.com>
2015-05-12 09:10:29 +02:00
|
|
|
You can set `blobdescriptor` field to `redis` or `inmemory`. The `redis` value uses
|
2015-04-13 20:34:07 +02:00
|
|
|
a Redis pool to cache layer metadata. The `inmemory` value uses an in memory
|
|
|
|
map.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
Refactor Blob Service API
This PR refactors the blob service API to be oriented around blob descriptors.
Identified by digests, blobs become an abstract entity that can be read and
written using a descriptor as a handle. This allows blobs to take many forms,
such as a ReadSeekCloser or a simple byte buffer, allowing blob oriented
operations to better integrate with blob agnostic APIs (such as the `io`
package). The error definitions are now better organized to reflect conditions
that can only be seen when interacting with the blob API.
The main benefit of this is to separate the much smaller metadata from large
file storage. Many benefits also follow from this. Reading and writing has
been separated into discrete services. Backend implementation is also
simplified, by reducing the amount of metadata that needs to be picked up to
simply serve a read. This also improves cacheability.
"Opening" a blob simply consists of an access check (Stat) and a path
calculation. Caching is greatly simplified and we've made the mapping of
provisional to canonical hashes a first-class concept. BlobDescriptorService
and BlobProvider can be combined in different ways to achieve varying effects.
Recommend Review Approach
-------------------------
This is a very large patch. While apologies are in order, we are getting a
considerable amount of refactoring. Most changes follow from the changes to
the root package (distribution), so start there. From there, the main changes
are in storage. Looking at (*repository).Blobs will help to understand the how
the linkedBlobStore is wired. One can explore the internals within and also
branch out into understanding the changes to the caching layer. Following the
descriptions below will also help to guide you.
To reduce the chances for regressions, it was critical that major changes to
unit tests were avoided. Where possible, they are left untouched and where
not, the spirit is hopefully captured. Pay particular attention to where
behavior may have changed.
Storage
-------
The primary changes to the `storage` package, other than the interface
updates, were to merge the layerstore and blobstore. Blob access is now
layered even further. The first layer, blobStore, exposes a global
`BlobStatter` and `BlobProvider`. Operations here provide a fast path for most
read operations that don't take access control into account. The
`linkedBlobStore` layers on top of the `blobStore`, providing repository-
scoped blob link management in the backend. The `linkedBlobStore` implements
the full `BlobStore` suite, providing access-controlled, repository-local blob
writers. The abstraction between the two is slightly broken in that
`linkedBlobStore` is the only channel under which one can write into the global
blob store. The `linkedBlobStore` also provides flexibility in that it can act
over different link sets depending on configuration. This allows us to use the
same code for signature links, manifest links and blob links. Eventually, we
will fully consolidate this storage.
The improved cache flow comes from the `linkedBlobStatter` component
of `linkedBlobStore`. Using a `cachedBlobStatter`, these combine together to
provide a simple cache hierarchy that should streamline access checks on read
and write operations, or at least provide a single path to optimize. The
metrics have been changed in a slightly incompatible way since the former
operations, Fetch and Exists, are no longer relevant.
The fileWriter and fileReader have been slightly modified to support the rest
of the changes. The most interesting is the removal of the `Stat` call from
`newFileReader`. This was the source of unnecessary round trips that were only
present to look up the size of the resulting reader. Now, one must simply pass
in the size, requiring the caller to decide whether or not the `Stat` call is
appropriate. In several cases, it turned out the caller already had the size
already. The `WriterAt` implementation has been removed from `fileWriter`,
since it is no longer required for `BlobWriter`, reducing the number of paths
which writes may take.
Cache
-----
Unfortunately, the `cache` package required a near full rewrite. It was pretty
mechanical in that the cache is oriented around the `BlobDescriptorService`
slightly modified to include the ability to set the values for individual
digests. While the implementation is oriented towards caching, it can act as a
primary store. Provisions are in place to have repository local metadata, in
addition to global metadata. Fallback is implemented as a part of the storage
package to maintain this flexibility.
One unfortunate side-effect is that caching is now repository-scoped, rather
than global. This should have little effect on performance but may increase
memory usage.
Handlers
--------
The `handlers` package has been updated to leverage the new API. For the most
part, the changes are superficial or mechanical based on the API changes. This
did expose a bug in the handling of provisional vs canonical digests that was
fixed in the unit tests.
Configuration
-------------
One user-facing change has been made to the configuration and is updated in
the associated documentation. The `layerinfo` cache parameter has been
deprecated by the `blobdescriptor` cache parameter. Both are equivalent and
configuration files should be backward compatible.
Notifications
-------------
Changes the `notification` package are simply to support the interface
changes.
Context
-------
A small change has been made to the tracing log-level. Traces have been moved
from "info" to "debug" level to reduce output when not needed.
Signed-off-by: Stephen J Day <stephen.day@docker.com>
2015-05-12 09:10:29 +02:00
|
|
|
>**NOTE**: Formerly, `blobdescriptor` was known as `layerinfo`. While these
|
|
|
|
>are equivalent, `layerinfo` has been deprecated, in favor or
|
|
|
|
>`blobdescriptor`.
|
|
|
|
|
2015-03-04 01:18:22 +01:00
|
|
|
### filesystem
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `filesystem` storage backend uses the local disk to store registry files. It
|
|
|
|
is ideal for development and may be appropriate for some small-scale production
|
|
|
|
applications.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
This backend has a single, required `rootdirectory` parameter. The parameter
|
|
|
|
specifies the absolute path to a directory. The registry stores all its data
|
|
|
|
here so make sure there is adequate space available.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
### azure
|
|
|
|
|
2015-05-05 10:25:42 +02:00
|
|
|
This storage backend uses Microsoft's Azure Storage platform.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>accountname</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Azure account name.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>accountkey</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Azure account key.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>container</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Name of the Azure container into which to store data.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
|
|
|
|
|
|
|
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
### S3
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
This storage backend uses Amazon's Simple Storage Service (S3).
|
|
|
|
|
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>accesskey</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Your AWS Access Key.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>secretkey</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Your AWS Secret Key.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>region</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The AWS region in which your bucket exists. For the moment, the Go AWS
|
|
|
|
library in use does not use the newer DNS based bucket routing.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>bucket</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The bucket name in which you want to store the registry's data.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>encrypt</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Specifies whether the registry stores the image in encrypted format or
|
|
|
|
not. A boolean value. The default is false.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>secure</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Indicates whether to use HTTPS instead of HTTP. A boolean value. The
|
|
|
|
default is false.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>v4auth</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Indicates whether the registry uses Version 4 of AWS's authentication.
|
|
|
|
Generally, you should set this to <code>true</code>. By default, this is
|
|
|
|
<code>false</code>.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>chunksize</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The S3 API requires multipart upload chunks to be at least 5MB. This value
|
|
|
|
should be a number that is larger than 5*1024*1024.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>rootdirectory</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
This is a prefix that will be applied to all S3 keys to allow you to segment data in your bucket if necessary.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-17 03:34:29 +02:00
|
|
|
### Maintenance
|
|
|
|
|
|
|
|
Currently the registry can perform one maintenance function: upload purging. This and future
|
|
|
|
maintenance functions which are related to storage can be configured under the maintenance section.
|
|
|
|
|
|
|
|
### Upload Purging
|
|
|
|
|
|
|
|
Upload purging is a background process that periodically removes orphaned files from the upload
|
2015-05-05 10:25:42 +02:00
|
|
|
directories of the registry. Upload purging is enabled by default. To
|
2015-04-17 03:34:29 +02:00
|
|
|
configure upload directory purging, the following parameters
|
|
|
|
must be set.
|
|
|
|
|
|
|
|
|
|
|
|
| Parameter | Required | Description
|
|
|
|
--------- | -------- | -----------
|
|
|
|
`enabled` | yes | Set to true to enable upload purging. Default=true. |
|
|
|
|
`age` | yes | Upload directories which are older than this age will be deleted. Default=168h (1 week)
|
2015-05-05 10:25:42 +02:00
|
|
|
`interval` | yes | The interval between upload directory purging. Default=24h.
|
2015-04-17 03:34:29 +02:00
|
|
|
`dryrun` | yes | dryrun can be set to true to obtain a summary of what directories will be deleted. Default=false.
|
|
|
|
|
2015-05-05 10:25:42 +02:00
|
|
|
Note: `age` and `interval` are strings containing a number with optional fraction and a unit suffix: e.g. 45m, 2h10m, 168h (1 week).
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
## auth
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
auth:
|
|
|
|
silly:
|
|
|
|
realm: silly-realm
|
|
|
|
service: silly-service
|
|
|
|
token:
|
|
|
|
realm: token-realm
|
|
|
|
service: token-service
|
|
|
|
issuer: registry-token-issuer
|
|
|
|
rootcertbundle: /root/certs/bundle
|
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `auth` option is **optional** as there are use cases (i.e. a mirror that
|
|
|
|
only permits pulls) for which authentication may not be desired. There are
|
|
|
|
currently 2 possible auth providers, `silly` and `token`. You can configure only
|
|
|
|
one `auth` provider.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
### silly
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `silly` auth is only for development purposes. It simply checks for the
|
|
|
|
existence of the `Authorization` header in the HTTP request. It has no regard for
|
|
|
|
the header's value. If the header does not exist, the `silly` auth responds with a
|
|
|
|
challenge response, echoing back the realm, service, and scope that access was
|
2015-05-05 10:25:42 +02:00
|
|
|
denied for.
|
2015-04-13 20:34:07 +02:00
|
|
|
|
|
|
|
The following values are used to configure the response:
|
|
|
|
|
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>realm</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The realm in which the registry server authenticates.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>service</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The service being authenticated.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
</table>
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
### token
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
Token based authentication allows the authentication system to be decoupled from
|
|
|
|
the registry. It is a well established authentication paradigm with a high
|
2015-05-05 10:25:42 +02:00
|
|
|
degree of security.
|
2015-04-13 20:34:07 +02:00
|
|
|
|
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>realm</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The realm in which the registry server authenticates.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>service</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The service being authenticated.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>issuer</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The name of the token issuer. The issuer inserts this into
|
|
|
|
the token so it must match the value configured for the issuer.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>rootcertbundle</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
2015-05-05 10:25:42 +02:00
|
|
|
yes
|
2015-04-13 20:34:07 +02:00
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The absolute path to the root certificate bundle. This bundle contains the
|
|
|
|
public part of the certificates that is used to sign authentication tokens.
|
|
|
|
</td>
|
|
|
|
</tr>
|
2015-05-05 10:25:42 +02:00
|
|
|
</table>
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-02 17:11:19 +02:00
|
|
|
For more information about Token based authentication configuration, see the [specification.]
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
## middleware
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `middleware` option is **optional**. Use this option to inject middleware at
|
|
|
|
named hook points. All middlewares must implement the same interface as the
|
|
|
|
object they're wrapping. This means a registry middleware must implement the
|
|
|
|
`distribution.Namespace` interface, repository middleware must implement
|
|
|
|
`distribution.Respository`, and storage middleware must implement
|
|
|
|
`driver.StorageDriver`.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
Currently only one middleware, `cloudfront`, a storage middleware, is supported
|
2015-05-05 10:25:42 +02:00
|
|
|
in the registry implementation.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
```yaml
|
|
|
|
middleware:
|
|
|
|
registry:
|
|
|
|
- name: ARegistryMiddleware
|
|
|
|
options:
|
|
|
|
foo: bar
|
|
|
|
repository:
|
|
|
|
- name: ARepositoryMiddleware
|
|
|
|
options:
|
|
|
|
foo: bar
|
|
|
|
storage:
|
|
|
|
- name: cloudfront
|
|
|
|
options:
|
|
|
|
baseurl: https://my.cloudfronted.domain.com/
|
|
|
|
privatekey: /path/to/pem
|
|
|
|
keypairid: cloudfrontkeypairid
|
|
|
|
duration: 3000
|
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
Each middleware entry has `name` and `options` entries. The `name` must
|
|
|
|
correspond to the name under which the middleware registers itself. The
|
|
|
|
`options` field is a map that details custom configuration required to
|
|
|
|
initialize the middleware. It is treated as a `map[string]interface{}`. As such,
|
|
|
|
it supports any interesting structures desired, leaving it up to the middleware
|
|
|
|
initialization function to best determine how to handle the specific
|
|
|
|
interpretation of the options.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
### cloudfront
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>baseurl</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
<code>SCHEME://HOST[/PATH]</code> at which Cloudfront is served.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>privatekey</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Private Key for Cloudfront provided by AWS.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>keypairid</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Key pair ID provided by AWS.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>duration</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Duration for which a signed URL should be valid.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
</table>
|
|
|
|
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
## reporting
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
reporting:
|
|
|
|
bugsnag:
|
|
|
|
apikey: bugsnagapikey
|
|
|
|
releasestage: bugsnagreleasestage
|
|
|
|
endpoint: bugsnagendpoint
|
|
|
|
newrelic:
|
|
|
|
licensekey: newreliclicensekey
|
|
|
|
name: newrelicname
|
2015-04-13 20:34:07 +02:00
|
|
|
verbose: true
|
2015-03-04 01:18:22 +01:00
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `reporting` option is **optional** and configures error and metrics
|
|
|
|
reporting tools. At the moment only two services are supported, [New
|
|
|
|
Relic](http://newrelic.com/) and [Bugsnag](http://bugsnag.com), a valid
|
|
|
|
configuration may contain both.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
### bugsnag
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>apikey</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
API Key provided by Bugsnag
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>releasestage</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Tracks where the registry is deployed, for example,
|
|
|
|
<codde>production</code>,<codde>staging</code>, or
|
|
|
|
<codde>development</code>.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>endpoint</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
2015-05-05 10:25:42 +02:00
|
|
|
Specify the enterprise Bugsnag endpoint.
|
2015-04-13 20:34:07 +02:00
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
|
|
|
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
### newrelic
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>licensekey</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
License key provided by New Relic.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>name</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
New Relic application name.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>verbose</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Enable New Relic debugging output on stdout.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
## http
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
http:
|
|
|
|
addr: localhost:5000
|
2015-05-05 10:25:42 +02:00
|
|
|
net: tcp
|
2015-03-04 01:18:22 +01:00
|
|
|
prefix: /my/nested/registry/
|
|
|
|
secret: asecretforlocaldevelopment
|
|
|
|
tls:
|
|
|
|
certificate: /path/to/x509/public
|
|
|
|
key: /path/to/x509/private
|
2015-03-20 16:19:07 +01:00
|
|
|
clientcas:
|
|
|
|
- /path/to/ca.pem
|
|
|
|
- /path/to/another/ca.pem
|
2015-03-04 01:18:22 +01:00
|
|
|
debug:
|
|
|
|
addr: localhost:5001
|
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `http` option details the configuration for the HTTP server that hosts the registry.
|
|
|
|
|
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>addr</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
2015-05-05 10:25:42 +02:00
|
|
|
The address for which the server should accept connections. The form depends on a network type (see <code>net</code> option):
|
|
|
|
<code>HOST:PORT</code> for tcp and <code>FILE</code> for a unix socket.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>net</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The network which is used to create a listening socket. Known networks are <code>unix</code> and <code>tcp</code>.
|
|
|
|
The default empty value means tcp.
|
2015-04-13 20:34:07 +02:00
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>prefix</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
If the server does not run at the root path use this value to specify the
|
|
|
|
prefix. The root path is the section before <code>v2</code>. It
|
|
|
|
should have both preceding and trailing slashes, for example <code>/path/</code>.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>secret</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
A random piece of data. This is used to sign state that may be stored with the
|
|
|
|
client to protect against tampering. For production environments you should generate a
|
|
|
|
random piece of data using a cryptographically secure random generator.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
</table>
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
|
|
|
|
### tls
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `tls` struct within `http` is **optional**. Use this to configure TLS
|
|
|
|
for the server. If you already have a server such as Nginx or Apache running on
|
|
|
|
the same host as the registry, you may prefer to configure TLS termination there
|
|
|
|
and proxy connections to the registry server.
|
|
|
|
|
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>certificate</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Absolute path to x509 cert file
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>key</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Absolute path to x509 private key file.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>clientcas</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
An array of absolute paths to a x509 CA file
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
|
|
|
|
### debug
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `debug` option is **optional** . Use it to configure a debug server that can
|
|
|
|
be helpful in diagnosing problems. Contributors to the distribution repository
|
|
|
|
should find the debug server useful. Docker recommends disabling it in
|
|
|
|
production environments.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The `debug` section takes a single, required `addr` parameter. This parameter
|
|
|
|
specifies the `HOST:PORT` on which the debug server should accept connections.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
|
|
|
|
## notifications
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
notifications:
|
2015-05-05 10:25:42 +02:00
|
|
|
endpoints:
|
2015-03-04 01:18:22 +01:00
|
|
|
- name: alistener
|
|
|
|
disabled: false
|
|
|
|
url: https://my.listener.com/event
|
|
|
|
headers: <http.Header>
|
|
|
|
timeout: 500
|
|
|
|
threshold: 5
|
|
|
|
backoff: 1000
|
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
The notifications option is **optional** and currently may contain a single
|
|
|
|
option, `endpoints`.
|
2015-03-04 01:18:22 +01:00
|
|
|
|
|
|
|
### endpoints
|
|
|
|
|
|
|
|
Endpoints is a list of named services (URLs) that can accept event notifications.
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>name</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
2015-05-05 10:25:42 +02:00
|
|
|
A human readable name for the service.
|
2015-04-13 20:34:07 +02:00
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>disabled</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
A boolean to enable/disable notifications for a service.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>url</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
The URL to which events should be published.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>headers</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Static headers to add to each request.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>timeout</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
An HTTP timeout value. This field takes a positive integer and an optional
|
|
|
|
suffix indicating the unit of time. Possible units are:
|
|
|
|
<ul>
|
|
|
|
<li><code>ns</code> (nanoseconds)</li>
|
|
|
|
<li><code>us</code> (microseconds)</li>
|
|
|
|
<li><code>ms</code> (milliseconds)</li>
|
|
|
|
<li><code>s</code> (seconds)</li>
|
|
|
|
<li><code>m</code> (minutes)</li>
|
|
|
|
<li><code>h</code> (hours)</li>
|
|
|
|
</ul>
|
|
|
|
If you omit the suffix, the system interprets the value as nanoseconds.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>threshold</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
An integer specifying how long to wait before backing off a failure.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>backoff</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
How long the system backs off before retrying. This field takes a positive
|
|
|
|
integer and an optional suffix indicating the unit of time. Possible units
|
|
|
|
are:
|
|
|
|
<ul>
|
|
|
|
<li><code>ns</code> (nanoseconds)</li>
|
|
|
|
<li><code>us</code> (microseconds)</li>
|
|
|
|
<li><code>ms</code> (milliseconds)</li>
|
|
|
|
<li><code>s</code> (seconds)</li>
|
|
|
|
<li><code>m</code> (minutes)</li>
|
|
|
|
<li><code>h</code> (hours)</li>
|
|
|
|
</ul>
|
|
|
|
If you omit the suffix, the system interprets the value as nanoseconds.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
|
|
|
|
2015-03-04 01:18:22 +01:00
|
|
|
|
2015-04-02 01:27:24 +02:00
|
|
|
## redis
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
redis:
|
|
|
|
addr: localhost:6379
|
|
|
|
password: asecret
|
|
|
|
db: 0
|
|
|
|
dialtimeout: 10ms
|
|
|
|
readtimeout: 10ms
|
|
|
|
writetimeout: 10ms
|
|
|
|
pool:
|
|
|
|
maxidle: 16
|
|
|
|
maxactive: 64
|
|
|
|
idletimeout: 300s
|
|
|
|
```
|
|
|
|
|
|
|
|
Declare parameters for constructing the redis connections. Registry instances
|
2015-04-13 20:34:07 +02:00
|
|
|
may use the Redis instance for several applications. The current purpose is
|
2015-04-02 01:27:24 +02:00
|
|
|
caching information about immutable blobs. Most of the options below control
|
2015-04-13 20:34:07 +02:00
|
|
|
how the registry connects to redis. You can control the pool's behavior
|
2015-04-02 01:27:24 +02:00
|
|
|
with the [pool](#pool) subsection.
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>addr</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
yes
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Address (host and port) of redis instance.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>password</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
A password used to authenticate to the redis instance.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>db</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Selects the db for each connection.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>dialtimeout</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Timeout for connecting to a redis instance.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>readtimeout</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Timeout for reading from redis connections.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>writetimeout</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Timeout for writing to redis connections.
|
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
|
|
|
|
2015-04-02 01:27:24 +02:00
|
|
|
|
|
|
|
### pool
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
pool:
|
|
|
|
maxidle: 16
|
|
|
|
maxactive: 64
|
|
|
|
idletimeout: 300s
|
|
|
|
```
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
Configure the behavior of the Redis connection pool.
|
|
|
|
|
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Required</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>maxidle</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
Sets the maximum number of idle connections.
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>maxactive</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
sets the maximum number of connections that should
|
2015-04-02 01:27:24 +02:00
|
|
|
be opened before blocking a connection request.
|
2015-04-13 20:34:07 +02:00
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td>
|
|
|
|
<code>idletimeout</code>
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
no
|
|
|
|
</td>
|
|
|
|
<td>
|
|
|
|
sets the amount time to wait before closing
|
2015-04-02 01:27:24 +02:00
|
|
|
inactive connections.
|
2015-04-13 20:34:07 +02:00
|
|
|
</td>
|
2015-05-05 10:25:42 +02:00
|
|
|
</tr>
|
2015-04-13 20:34:07 +02:00
|
|
|
</table>
|
2015-04-17 03:34:29 +02:00
|
|
|
|
|
|
|
|
2015-04-13 20:34:07 +02:00
|
|
|
## Example: Development configuration
|
|
|
|
|
|
|
|
The following is a simple example you can use for local development:
|
|
|
|
|
|
|
|
```yaml
|
|
|
|
version: 0.1
|
2015-05-05 10:25:42 +02:00
|
|
|
log:
|
2015-04-13 20:34:07 +02:00
|
|
|
level: debug
|
|
|
|
storage:
|
|
|
|
filesystem:
|
|
|
|
rootdirectory: /tmp/registry-dev
|
|
|
|
http:
|
|
|
|
addr: localhost:5000
|
|
|
|
secret: asecretforlocaldevelopment
|
|
|
|
debug:
|
|
|
|
addr: localhost:5001
|
|
|
|
```
|
|
|
|
|
|
|
|
The above configures the registry instance to run on port `5000`, binding to
|
2015-05-05 10:25:42 +02:00
|
|
|
`localhost`, with the `debug` server enabled. Registry data storage is in the
|
2015-04-13 20:34:07 +02:00
|
|
|
`/tmp/registry-dev` directory. Logging is in `debug` mode, which is the most
|
|
|
|
verbose.
|
|
|
|
|
|
|
|
A similar simple configuration is available at
|
2015-04-17 23:05:07 +02:00
|
|
|
[config.yml](https://github.com/docker/distribution/blob/master/cmd/registry/config.yml).
|
|
|
|
Both are generally useful for local development.
|
2015-04-13 20:34:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
## Example: Middleware configuration
|
|
|
|
|
|
|
|
This example illustrates how to configure storage middleware in a registry.
|
|
|
|
Middleware allows the registry to serve layers via a content delivery network
|
2015-05-05 10:25:42 +02:00
|
|
|
(CDN). This is useful for reducing requests to the storage layer.
|
2015-04-13 20:34:07 +02:00
|
|
|
|
|
|
|
Currently, the registry supports [Amazon
|
|
|
|
Cloudfront](http://aws.amazon.com/cloudfront/). You can only use Cloudfront in
|
|
|
|
conjunction with the S3 storage driver.
|
|
|
|
|
|
|
|
<table>
|
|
|
|
<tr>
|
|
|
|
<th>Parameter</th>
|
|
|
|
<th>Description</th>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td><code>name</code></td>
|
|
|
|
<td>The storage middleware name. Currently <code>cloudfront</code> is an accepted value.</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td><code>disabled<code></td>
|
|
|
|
<td>Set to <code>false</code> to easily disable the middleware.</td>
|
|
|
|
</tr>
|
|
|
|
<tr>
|
|
|
|
<td><code>options:</code></td>
|
2015-05-05 10:25:42 +02:00
|
|
|
<td>
|
2015-04-13 20:34:07 +02:00
|
|
|
A set of key/value options to configure the middleware.
|
|
|
|
<ul>
|
|
|
|
<li><code>baseurl:</code> The Cloudfront base URL.</li>
|
|
|
|
<li><code>privatekey:</code> The location of your AWS private key on the filesystem. </li>
|
|
|
|
<li><code>keypairid:</code> The ID of your Cloudfront keypair. </li>
|
|
|
|
<li><code>duration:</code> The duration in minutes for which the URL is valid. Default is 20. </li>
|
|
|
|
</ul>
|
|
|
|
</td>
|
|
|
|
</tr>
|
|
|
|
</table>
|
|
|
|
|
|
|
|
The following example illustrates these values:
|
|
|
|
|
|
|
|
```
|
|
|
|
middleware:
|
|
|
|
storage:
|
|
|
|
- name: cloudfront
|
|
|
|
disabled: false
|
|
|
|
options:
|
|
|
|
baseurl: http://d111111abcdef8.cloudfront.net
|
|
|
|
privatekey: /path/to/asecret.pem
|
|
|
|
keypairid: asecret
|
|
|
|
duration: 60
|
|
|
|
```
|
|
|
|
|
|
|
|
|
|
|
|
>**Note**: Cloudfront keys exist separately to other AWS keys. See
|
2015-04-17 23:05:07 +02:00
|
|
|
>[the documentation on AWS credentials](http://docs.aws.amazon.com/AWSSecurityCredentials/1.0/AboutAWSCredentials.html#KeyPairs)
|
|
|
|
>for more information.
|