Verdaccio is a lightweight private Node.js proxy registry.
$ helm repo add verdaccio https://charts.verdaccio.org
$ helm repo update
$ helm install verdaccio verdaccio/verdaccio
⚠️ If you are using stable/verdaccio chart, be aware is deprecated, forward all new PR and or issues to this repository.
If you need support for Helm v2, please use
<=v0.19.0
, be aware we do not support Helm v2 anymore.
This chart bootstraps a Verdaccio deployment on a Kubernetes cluster using the Helm package manager.
- Kubernetes 1.7+ with Beta APIs enabled
- PV provisioner support in the underlying infrastructure
helm repo add verdaccio https://charts.verdaccio.org
In this example we use npm
as release name:
# Helm v3+
helm install npm verdaccio/verdaccio
Note: Avoid release name
verdaccio
, otherwise, Kubernetes-generated environment variables may get into conflict with Verdaccio's own environment variables in the Pod itself. In case you insist naming the releaseverdaccio
, to mitigate the problem, you can use eithernameOverride
orfullnameOverride
to have a different name for the service.
# Helm v3+
helm install npm --set image.tag=6.0.0 verdaccio/verdaccio
helm upgrade npm verdaccio/verdaccio
The command deploys Verdaccio on the Kubernetes cluster in the default configuration. The configuration section lists the parameters that can be configured during installation.
Tip: List all releases using
helm list
To uninstall/delete the npm
deployment:
# Helm v3+
helm uninstall npm
The command removes all the Kubernetes components associated with the chart and deletes the release.
The following table lists the configurable parameters of the Verdaccio chart and their default values.
Parameter | Description | Default |
---|---|---|
type |
The type of resource to create. Either deployment or statefulset . Note: Statefulset is primarly useful when Verdaccio is being used as an edge cache |
deployment |
annotations |
Annotations to set on the deployment | {} |
affinity |
Affinity for pod assignment | {} |
existingConfigMap |
Name of custom ConfigMap to use | false |
image.pullPolicy |
Image pull policy | IfNotPresent |
image.pullSecrets |
Image pull secrets | [] |
image.repository |
Verdaccio container image repository | verdaccio/verdaccio |
image.tag |
Verdaccio container image tag | 5.21.1 |
nodeSelector |
Node labels for pod assignment | {} |
tolerations |
List of node taints to tolerate | [] |
persistence.accessMode |
PVC Access Mode for Verdaccio volume | ReadWriteOnce |
persistence.annotations |
Annotations to add to the PVC | {} |
persistence.enabled |
Enable persistence using PVC | true |
persistence.existingClaim |
Use existing PVC. Ignored when type is statefuleset |
nil |
persistence.mounts |
Additional mounts | nil |
persistence.resourcePolicy |
Set "keep" to avoid removing PVC during a helm delete operation | "" |
persistence.selector |
Selector to match an existing Persistent Volume | {} (evaluated as a template) |
persistence.size |
PVC Storage Request for Verdaccio volume | 8Gi |
persistence.storageClass |
PVC Storage Class for Verdaccio volume | nil |
persistence.volumes |
Additional volumes | nil |
topologySpreadConstraints |
Topology Spread Constraints for pod assignment | [] |
podLabels |
Additional pod labels | {} (evaluated as a template) |
podAnnotations |
Annotations to add to each pod | {} |
priorityClass.enabled |
Enable specifying pod priorityClassName | false |
priorityClass.name |
PriorityClassName to be specified in pod spec | "" |
replicaCount |
Desired number of pods | 1 |
replicaCountEnabled |
Enable the replicaCount field | true |
livenessProbe |
Configuration of liveness probe | {} |
readinessProbe |
Configuration of readiness probe | {} |
resources |
CPU/Memory resource requests/limits | {} |
service.annotations |
Annotations to add to service | none |
service.clusterIP |
IP address to assign to service | "" |
service.externalIPs |
Service external IP addresses | [] |
service.loadBalancerIP |
IP address to assign to load balancer (if supported) | "" |
service.loadBalancerSourceRanges |
List of IP CIDRs allowed access to load balancer (if supported) | [] |
service.port |
Service port to expose | 4873 |
service.nodePort |
Service port to expose | none |
service.type |
Type of service to create | ClusterIP |
serviceAccount.create |
Create service account | false |
serviceAccount.name |
Service account Name | none |
extraEnvVars |
Define environment variables to be passed to the container | [] |
secretEnvVars |
Define sensitive environment variables to be passed to the container | {} |
existingSecret |
Existing secret containing environment variables to be passed to the container | "" |
extraInitContainers |
Define additional initContainers to be added to the deployment | [] |
securityContext |
Define Container Security Context | {runAsUser=10001} |
podSecurityContext |
Define Pod Security Context | {fsGroup=101} |
nameOverride |
Set resource name override | "" |
fullnameOverride |
Set resource fullname override | "" |
useSecretHtpasswd |
Use htpasswd from .Values.secrets.htpasswd . This require helm v3.2.0 or above. |
false |
secrets.htpasswd |
user and password list to generate htpasswd. | [] |
ingress.enabled |
Enable/Disable Ingress | false |
ingress.className |
Ingress Class Name (k8s >=1.18 required) |
"" |
ingress.annotations |
Ingress Annotations | {} |
ingress.hosts |
List of Ingress Hosts | [] |
ingress.paths |
List of Ingress Paths | ["/"] |
ingress.extraPaths |
List of extra Ingress Paths | [] |
extraPorts |
List of extra ports to expose from the pods | [] |
extraManifests |
List of extra manifests to deploy within the chart | [] |
Specify each parameter using the --set key=value[,key=value]
argument to helm install
. For example,
# Helm v3+
$ helm install my-release \
--set service.type=LoadBalancer \
verdaccio/verdaccio
The above command sets the service type LoadBalancer.
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
# Helm v3+
$ helm install npm -f values.yaml verdaccio/verdaccio
Tip: You can use the default values.yaml
This requires helm v3.2.0 or above. You can list all username and password in
.Values.secrets.htpasswd
. Helm will generate secret with htpaswd format. This
file is mounted on pod in this path /verdaccio/auth/htpasswd
. The Default
config uses this. The conditional statement {{- if .Values.secrets.htpasswd }}
is evaluated as false if the list is an empty collection.
(Source helm flow control)
Tip: These values are in plaintext. So don't forget to put additional encryption.
secrets:
# list of users and password for htpasswd plugin
# This this is mounted as /verdaccio/auth/htpasswd on pods
htpasswd:
- username: "verdaccio"
password: "verdaccio"
This config will create a htpasswd file with user "verdaccio", If in config 'htpasswd' auth is used. You can login using this credentials.
When creating a new chart with this chart as a dependency, CustomConfigMap can
be used to override the default config.yaml provided. It also allows for
providing additional configuration files that will be copied into
/verdaccio/conf
. In the parent chart's values.yaml, set the value to true and
provide the file templates/config.yaml
for your use case.
The Verdaccio image stores persistence under /verdaccio/storage
path of the
container. A dynamically managed Persistent Volume Claim is used to keep the
data across deployments, by default. This is known to work in GCE, AWS, and
minikube.
Alternatively, a previously configured Persistent Volume Claim can be used.
It is possible to mount several volumes using Persistence.volumes
and
Persistence.mounts
parameters.
- Create the PersistentVolume
- Create the PersistentVolumeClaim
- Install the chart
# Helm v3+
$ helm install npm \
--set persistence.existingClaim=PVC_NAME \
verdaccio/verdaccio
Due to some breaking changes in Selector Labels and Security Contexts in Chart 3.0.0
you will need to migrate when upgrading.
First off, the securityContext.enabled
field has been removed.
In addition to this, fsGroup
is not a valid Container Security Context field and has been migrated to the podSecurityContext
instead.
# values.yaml
podSecurityContext:
+ fsGroup: 101
securityContext:
- enabled: true
- fsGroup: 101
runAsUser: 10001
Secondly, the apps.v1.Deployment.spec.selector
field is immutable and changes were made to Selector Labels which tries to update this.
To get around this, you will need to kubectl delete deployment $deploymentName
before doing a helm upgrade
So long as your PVC is not destroyed, the new deployment will be rolled out with the same PVC as before and your data will remain intact.
Due the major release Verdaccio 5 has some breaking changes to be aware of, please read the migration guide here.