Skip to content

magda-io/minio-charts

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MinIO

MinIO is a High Performance Object Storage released under Apache License v2.0. It is API compatible with Amazon S3 cloud storage service. Use MinIO to build high performance infrastructure for machine learning, analytics and application data workloads.

MinIO supports distributed mode. In distributed mode, you can pool multiple drives (even on different machines) into a single object storage server.

For more detailed documentation please visit here

Introduction

This chart bootstraps MinIO deployment on a Kubernetes cluster using the Helm package manager.

Prerequisites

  • Kubernetes 1.4+ with Beta APIs enabled for default standalone mode.
  • Kubernetes 1.5+ with Beta APIs enabled to run MinIO in distributed mode.
  • PV provisioner support in the underlying infrastructure.

Configure MinIO Helm repo

$ helm repo add minio https://helm.min.io/

Installing the Chart

Install this chart using:

$ helm install --namespace minio --generate-name minio/minio

The command deploys MinIO on the Kubernetes cluster in the default configuration. The configuration section lists the parameters that can be configured during installation.

Release name

An instance of a chart running in a Kubernetes cluster is called a release. Each release is identified by a unique name within the cluster. Helm automatically assigns a unique release name after installing the chart. You can also set your preferred name by:

$ helm install my-release minio/minio

Access and Secret keys

By default a pre-generated access and secret key will be used. To override the default keys, pass the access and secret keys as arguments to helm install.

$ helm install --set accessKey=myaccesskey,secretKey=mysecretkey --generate-name minio/minio

Updating MinIO configuration via Helm

ConfigMap allows injecting containers with configuration data even while a Helm release is deployed.

To update your MinIO server configuration while it is deployed in a release, you need to

  1. Check all the configurable values in the MinIO chart using helm inspect values minio/minio.
  2. Override the minio_server_config settings in a YAML formatted file, and then pass that file like this helm upgrade -f config.yaml minio/minio.
  3. Restart the MinIO server(s) for the changes to take effect.

You can also check the history of upgrades to a release using helm history my-release. Replace my-release with the actual release name.

Uninstalling the Chart

Assuming your release is named as my-release, delete it using the command:

$ helm delete my-release

or

$ helm uninstall my-release

The command removes all the Kubernetes components associated with the chart and deletes the release.

Upgrading the Chart

You can use Helm to update MinIO version in a live release. Assuming your release is named as my-release, get the values using the command:

$ helm get values my-release > old_values.yaml

Then change the field image.tag in old_values.yaml file with MinIO image tag you want to use. Now update the chart using

$ helm upgrade -f old_values.yaml my-release minio/minio

Default upgrade strategies are specified in the values.yaml file. Update these fields if you'd like to use a different strategy.

Configuration

The following table lists the configurable parameters of the MinIO chart and their default values.

Parameter Description Default
nameOverride Provide a name in place of minio ""
fullnameOverride Provide a name to substitute for the full names of resources ""
image.repository Image repository minio/minio
image.tag MinIO image tag. Possible values listed here. RELEASE.2020-09-26T03-44-56Z
image.pullPolicy Image pull policy IfNotPresent
imagePullSecrets List of container registry secrets []
mcImage.repository Client image repository minio/mc
mcImage.tag mc image tag. Possible values listed here. RELEASE.2020-09-23T20-02-13Z
mcImage.pullPolicy mc Image pull policy IfNotPresent
ingress.enabled Enables Ingress false
ingress.labels Ingress labels {}
ingress.annotations Ingress annotations {}
ingress.hosts Ingress accepted hostnames []
ingress.tls Ingress TLS configuration []
mode MinIO server mode (standalone or distributed) standalone
extraArgs Additional command line arguments to pass to the MinIO server []
replicas Number of nodes (applicable only for MinIO distributed mode). 4
zones Number of zones (applicable only for MinIO distributed mode). 1
drivesPerNode Number of drives per node (applicable only for MinIO distributed mode). 1
existingSecret Name of existing secret with access and secret key. ""
accessKey Default access key (5 to 20 characters) YOURACCESSKEY
secretKey Default secret key (8 to 40 characters) YOURSECRETKEY
certsPath Default certs path location /etc/minio/certs
configPathmc Default config file location for MinIO client - mc /etc/minio/mc
mountPath Default mount location for persistent drive /export
bucketRoot Directory from where minio should serve buckets. Value of .mountPath
clusterDomain domain name of kubernetes cluster where pod is running. cluster.local
service.type Kubernetes service type ClusterIP
service.port Kubernetes port where service is exposed 9000
service.externalIPs service external IP addresses nil
service.annotations Service annotations {}
serviceAccount.create Toggle creation of new service account true
serviceAccount.name Name of service account to create and/or use ""
persistence.enabled Use persistent volume to store data true
persistence.size Size of persistent volume claim 500Gi
persistence.existingClaim Use an existing PVC to persist data nil
persistence.storageClass Storage class name of PVC nil
persistence.accessMode ReadWriteOnce or ReadOnly ReadWriteOnce
persistence.subPath Mount a sub directory of the persistent volume if set ""
resources Memory resource requests Memory: 4Gi
priorityClassName Pod priority settings ""
securityContext.enabled Enable to run containers as non-root. NOTE: if persistence.enabled=false then securityContext will be automatically disabled true
securityContext.runAsUser User id of the user for the container 1000
securityContext.runAsGroup Group id of the user for the container 1000
securityContext.fsGroup Group id of the persistent volume mount for the container 1000
nodeSelector Node labels for pod assignment {}
affinity Affinity settings for pod assignment {}
tolerations Toleration labels for pod assignment []
podAnnotations Pod annotations {}
podLabels Pod Labels {}
tls.enabled Enable TLS for MinIO server false
tls.certSecret Kubernetes Secret with public.crt and private.key files. ""
livenessProbe.initialDelaySeconds Delay before liveness probe is initiated 120
livenessProbe.periodSeconds How often to perform the probe 15
livenessProbe.timeoutSeconds When the probe times out 10
livenessProbe.successThreshold Minimum consecutive successes for the probe to be considered successful after having failed. 1
livenessProbe.failureThreshold Minimum consecutive failures for the probe to be considered failed after having succeeded. 3
defaultBucket.enabled If set to true, a bucket will be created after MinIO install false
defaultBucket.name Bucket name bucket
defaultBucket.policy Bucket policy none
defaultBucket.purge Purge the bucket if already exists false
defaultBucket.versioning Enable / Suspend versioning for bucket nil
buckets List of buckets to create after MinIO install []
makeBucketJob.annotations Additional annotations for the Kubernetes Batch (make-bucket-job) ""
makeBucketJob.podAnnotations Additional annotations for the pods of the Kubernetes Batch (make-bucket-job) ""
makeBucketJob.securityContext.enabled Enable to run Kubernetes Batch (make-bucket-job) containers as non-root. false
makeBucketJob.securityContext.runAsUser User id of the user for the container 1000
makeBucketJob.securityContext.runAsGroup Group id of the user for the container 1000
makeBucketJob.securityContext.fsGroup Group id of the persistent volume mount for the container 1000
makeBucketJob.resources.requests.memory Memory resource requests for 'make bucket' job 128Mi
updatePrometheusJob.podAnnotations Additional annotations for the pods of the Kubernetes Batch (update-prometheus-secret) ""
updatePrometheusJob.securityContext.enabled Enable to run Kubernetes Batch (update-prometheus-secret) containers as non-root. false
updatePrometheusJob.securityContext.runAsUser User id of the user for the container 1000
updatePrometheusJob.securityContext.runAsGroup Group id of the user for the container 1000
updatePrometheusJob.securityContext.fsGroup Group id of the persistent volume mount for the container 1000
s3gateway.enabled Use MinIO as a s3 gateway false
s3gateway.replicas Number of s3 gateway instances to run in parallel 4
s3gateway.serviceEndpoint Endpoint to the S3 compatible service ""
s3gateway.accessKey Access key of S3 compatible service ""
s3gateway.secretKey Secret key of S3 compatible service ""
azuregateway.enabled Use MinIO as an azure gateway false
azuregateway.replicas Number of azure gateway instances to run in parallel 4
gcsgateway.enabled Use MinIO as a Google Cloud Storage gateway false
gcsgateway.gcsKeyJson credential json file of service account key ""
gcsgateway.projectId Google cloud project id ""
nasgateway.enabled Use MinIO as a NAS gateway false
nasgateway.replicas Number of NAS gateway instances to be run in parallel on a PV 4
environment Set MinIO server relevant environment variables in values.yaml file. MinIO containers will be passed these variables when they start. MINIO_API_READY_DEADLINE: "5s"
metrics.serviceMonitor.enabled Set this to true to create ServiceMonitor for Prometheus operator false
metrics.serviceMonitor.additionalLabels Additional labels that can be used so ServiceMonitor will be discovered by Prometheus {}
metrics.serviceMonitor.namespace Optional namespace in which to create ServiceMonitor nil
metrics.serviceMonitor.interval Scrape interval. If not set, the Prometheus default scrape interval is used nil
metrics.serviceMonitor.scrapeTimeout Scrape timeout. If not set, the Prometheus default scrape timeout is used nil
etcd.endpoints Enpoints of etcd []
etcd.pathPrefix Prefix for all etcd keys ""
etcd.corednsPathPrefix Prefix for all CoreDNS etcd keys ""
etcd.clientCert Certificate used for SSL/TLS connections to etcd (etcd Security) ""
etcd.clientCertKey Key for the certificate (etcd Security) ""

Some of the parameters above map to the env variables defined in the MinIO DockerHub image.

You can specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

$ helm install --name my-release --set persistence.size=1Ti minio/minio

The above command deploys MinIO server with a 100Gi backing persistent volume.

Alternately, you can provide a YAML file that specifies parameter values while installing the chart. For example,

$ helm install --name my-release -f values.yaml minio/minio

Tip: You can use the default values.yaml

Distributed MinIO

This chart provisions a MinIO server in standalone mode, by default. To provision MinIO server in distributed mode, set the mode field to distributed,

$ helm install --set mode=distributed minio/minio

This provisions MinIO server in distributed mode with 4 nodes. To change the number of nodes in your distributed MinIO server, set the replicas field,

$ helm install --set mode=distributed,replicas=8 minio/minio

This provisions MinIO server in distributed mode with 8 nodes. Note that the replicas value should be a minimum value of 4, there is no limit on number of servers you can run.

You can also expand an existing deployment by adding new zones, following command will create a total of 16 nodes with each zone running 8 nodes.

$ helm install --set mode=distributed,replicas=8,zones=2 minio/minio

StatefulSet limitations applicable to distributed MinIO

  1. StatefulSets need persistent storage, so the persistence.enabled flag is ignored when mode is set to distributed.
  2. When uninstalling a distributed MinIO release, you'll need to manually delete volumes associated with the StatefulSet.

NAS Gateway

Prerequisites

MinIO in NAS gateway mode can be used to create multiple MinIO instances backed by single PV in ReadWriteMany mode. Currently few Kubernetes volume plugins support ReadWriteMany mode. To deploy MinIO NAS gateway with Helm chart you'll need to have a Persistent Volume running with one of the supported volume plugins. This document outlines steps to create a NFS PV in Kubernetes cluster.

Provision NAS Gateway MinIO instances

To provision MinIO servers in NAS gateway mode, set the nasgateway.enabled field to true,

$ helm install --set nasgateway.enabled=true minio/minio

This provisions 4 MinIO NAS gateway instances backed by single storage. To change the number of instances in your MinIO deployment, set the replicas field,

$ helm install --set nasgateway.enabled=true,nasgateway.replicas=8 minio/minio

This provisions MinIO NAS gateway with 8 instances.

Persistence

This chart provisions a PersistentVolumeClaim and mounts corresponding persistent volume to default location /export. You'll need physical storage available in the Kubernetes cluster for this to work. If you'd rather use emptyDir, disable PersistentVolumeClaim by:

$ helm install --set persistence.enabled=false minio/minio

"An emptyDir volume is first created when a Pod is assigned to a Node, and exists as long as that Pod is running on that node. When a Pod is removed from a node for any reason, the data in the emptyDir is deleted forever."

Existing PersistentVolumeClaim

If a Persistent Volume Claim already exists, specify it during installation.

  1. Create the PersistentVolume
  2. Create the PersistentVolumeClaim
  3. Install the chart
$ helm install --set persistence.existingClaim=PVC_NAME minio/minio

NetworkPolicy

To enable network policy for MinIO, install a networking plugin that implements the Kubernetes NetworkPolicy spec, and set networkPolicy.enabled to true.

For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the DefaultDeny namespace annotation. Note: this will enforce policy for all pods in the namespace:

kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"

With NetworkPolicy enabled, traffic will be limited to just port 9000.

For more precise policy, set networkPolicy.allowExternal=true. This will only allow pods with the generated client label to connect to MinIO. This label will be displayed in the output of a successful install.

Existing secret

Instead of having this chart create the secret for you, you can supply a preexisting secret, much like an existing PersistentVolumeClaim.

First, create the secret:

$ kubectl create secret generic my-minio-secret --from-literal=accesskey=foobarbaz --from-literal=secretkey=foobarbazqux

Then install the chart, specifying that you want to use an existing secret:

$ helm install --set existingSecret=my-minio-secret minio/minio

The following fields are expected in the secret

  1. accesskey - the access key ID
  2. secretkey - the secret key
  3. gcs_key.json - The GCS key if you are using the GCS gateway feature. This is optional.

Configure TLS

To enable TLS for MinIO containers, acquire TLS certificates from a CA or create self-signed certificates. While creating / acquiring certificates ensure the corresponding domain names are set as per the standard DNS naming conventions in a Kubernetes StatefulSet (for a distributed MinIO setup). Then create a secret using

$ kubectl create secret generic tls-ssl-minio --from-file=path/to/private.key --from-file=path/to/public.crt

Then install the chart, specifying that you want to use the TLS secret:

$ helm install --set tls.enabled=true,tls.certSecret=tls-ssl-minio minio/minio

Pass environment variables to MinIO containers

To pass environment variables to MinIO containers when deploying via Helm chart, use the below command line format

$ helm install --set environment.MINIO_BROWSER=on,environment.MINIO_DOMAIN=domain-name minio/minio

You can add as many environment variables as required, using the above format. Just add environment.<VARIABLE_NAME>=<value> under set flag.

Create buckets after install

Install the chart, specifying the buckets you want to create after install:

$ helm install --set buckets[0].name=bucket1,buckets[0].policy=none,buckets[0].purge=false minio/minio

Description of the configuration parameters used above -

  • buckets[].name - name of the bucket to create, must be a string with length > 0
  • buckets[].policy - can be one of none|download|upload|public
  • buckets[].purge - purge if bucket exists already

About

No description, website, or topics provided.

Resources

License

AGPL-3.0, Apache-2.0 licenses found

Licenses found

AGPL-3.0
LICENSE
Apache-2.0
LICENSE.helm

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Mustache 60.9%
  • Shell 39.1%