Skip to content

Commit

Permalink
Merge pull request kubernetes#12441 from vlajos/typofixes-vlajos-2015…
Browse files Browse the repository at this point in the history
  • Loading branch information
Marek Grabowski committed Aug 10, 2015
2 parents 08f67e5 + 9f77e49 commit 5f9cefc
Show file tree
Hide file tree
Showing 108 changed files with 150 additions and 150 deletions.
4 changes: 2 additions & 2 deletions api/swagger-spec/v1.json
Original file line number Diff line number Diff line change
Expand Up @@ -11429,7 +11429,7 @@
},
"lastTimestamp": {
"type": "string",
"description": "the time at which the most recent occurance of this event was recorded"
"description": "the time at which the most recent occurrence of this event was recorded"
},
"count": {
"type": "integer",
Expand Down Expand Up @@ -11985,7 +11985,7 @@
},
"hostPath": {
"$ref": "v1.HostPathVolumeSource",
"description": "a HostPath provisioned by a developer or tester; for develment use only; see http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath"
"description": "a HostPath provisioned by a developer or tester; for development use only; see http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath"
},
"glusterfs": {
"$ref": "v1.GlusterfsVolumeSource",
Expand Down
2 changes: 1 addition & 1 deletion cluster/aws/config-default.sh
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ SERVICE_CLUSTER_IP_RANGE="10.0.0.0/16" # formerly PORTAL_NET
CLUSTER_IP_RANGE="${CLUSTER_IP_RANGE:-10.244.0.0/16}"
MASTER_IP_RANGE="${MASTER_IP_RANGE:-10.246.0.0/24}"
# If set to Elastic IP, master instance will be associated with this IP.
# If set to auto, a new Elastic IP will be aquired
# If set to auto, a new Elastic IP will be acquired
# Otherwise amazon-given public ip will be used (it'll change with reboot).
MASTER_RESERVED_IP="${MASTER_RESERVED_IP:-}"

Expand Down
2 changes: 1 addition & 1 deletion cluster/aws/config-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ SERVICE_CLUSTER_IP_RANGE="10.0.0.0/16" # formerly PORTAL_NET
CLUSTER_IP_RANGE="${CLUSTER_IP_RANGE:-10.245.0.0/16}"
MASTER_IP_RANGE="${MASTER_IP_RANGE:-10.246.0.0/24}"
# If set to Elastic IP, master instance will be associated with this IP.
# If set to auto, a new Elastic IP will be aquired
# If set to auto, a new Elastic IP will be acquired
# Otherwise amazon-given public ip will be used (it'll change with reboot).
MASTER_RESERVED_IP="${MASTER_RESERVED_IP:-}"

Expand Down
2 changes: 1 addition & 1 deletion cluster/aws/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -876,7 +876,7 @@ function kube-up {

# This is a race between instance start and volume attachment. There appears to be no way to start an AWS instance with a volume attached.
# To work around this, we wait for volume to be ready in setup-master-pd.sh
echo "Attaching peristent data volume (${MASTER_DISK_ID}) to master"
echo "Attaching persistent data volume (${MASTER_DISK_ID}) to master"
$AWS_CMD attach-volume --volume-id ${MASTER_DISK_ID} --device /dev/sdb --instance-id ${master_id}

sleep 10
Expand Down
2 changes: 1 addition & 1 deletion cluster/common.sh
Original file line number Diff line number Diff line change
Expand Up @@ -172,7 +172,7 @@ function get-kubeconfig-bearertoken() {
}

# Sets KUBE_VERSION variable to the version passed in as an argument, or if argument is
# latest_stable, latest_release, or latest_ci fetches and sets the correponding version number
# latest_stable, latest_release, or latest_ci fetches and sets the corresponding version number
#
# Args:
# $1 version string from command line
Expand Down
2 changes: 1 addition & 1 deletion cluster/gce/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -425,7 +425,7 @@ function create-node-template {
}

# Robustly try to add metadata on an instance.
# $1: The name of the instace.
# $1: The name of the instance.
# $2...$n: The metadata key=value pairs to add.
function add-instance-metadata {
local -r instance=$1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ def test_update_rc_files(self, pmock):
pmock.return_value.write_lines.assert_called_with(['line1', 'line2',
'test1', 'test2'])

def test_update_rc_files_with_nonexistant_path(self):
def test_update_rc_files_with_nonexistent_path(self):
"""
Test an unhappy path if the bashrc/users do not exist.
"""
Expand Down
2 changes: 1 addition & 1 deletion cluster/juju/charms/trusty/kubernetes/hooks/start
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
set -ex

# Start is guaranteed to be called once when after the unit is installed
# *AND* once everytime a machine is rebooted.
# *AND* once every time a machine is rebooted.

if [ ! -f $CHARM_DIR/.unit-state ]
then
Expand Down
2 changes: 1 addition & 1 deletion cluster/juju/return-node-ips.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@

import json
import sys
# This script helps parse out the private IP addreses from the
# This script helps parse out the private IP addresses from the
# `juju run` command's JSON object, see cluster/juju/util.sh

if len(sys.argv) > 1:
Expand Down
2 changes: 1 addition & 1 deletion cluster/juju/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -135,7 +135,7 @@ function sleep-status() {

# sleep because we cannot get the status back of where the minions are in the deploy phase
# thanks to a generic "started" state and our service not actually coming online until the
# minions have recieved the binary from the master distribution hub during relations
# minions have received the binary from the master distribution hub during relations
echo "Sleeping an additional minute to allow the cluster to settle"
sleep 60
}
2 changes: 1 addition & 1 deletion cluster/libvirt-coreos/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -296,7 +296,7 @@ function kube-push-internal {
local)
kube-push-local;;
*)
echo "The only known push methods are \"release\" to use the relase tarball or \"local\" to use the binaries built by make. KUBE_PUSH is set \"$KUBE_PUSH\"" >&2
echo "The only known push methods are \"release\" to use the release tarball or \"local\" to use the binaries built by make. KUBE_PUSH is set \"$KUBE_PUSH\"" >&2
return 1;;
esac
}
Expand Down
2 changes: 1 addition & 1 deletion cluster/mesos/docker/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -218,7 +218,7 @@ function validate-cluster {
echo "Validating ${KUBERNETES_PROVIDER} cluster" 1>&2

# Do not validate cluster size. There will be zero k8s minions until a pod is created.
# TODO(karlkfi): use componentstatuses or equivelent when it supports non-localhost core components
# TODO(karlkfi): use componentstatuses or equivalent when it supports non-localhost core components

# Validate immediate cluster reachability and responsiveness
echo "KubeDNS: $(cluster::mesos::docker::addon_status 'kube-dns')"
Expand Down
2 changes: 1 addition & 1 deletion cluster/saltbase/salt/kube-addons/kube-addon-update.sh
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
# LIMITATIONS
# 1. controllers are not updated unless their name is changed
# 3. Services will not be updated unless their name is changed,
# but for services we acually want updates without name change.
# but for services we actually want updates without name change.
# 4. Json files are not handled at all. Currently addons must be
# in yaml files
# 5. exit code is probably not always correct (I haven't checked
Expand Down
2 changes: 1 addition & 1 deletion cluster/saltbase/salt/kube-master-addons/init.sls
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
- mode: 755

# Used to restart kube-master-addons service each time salt is run
# Actually, it doens't work (the service is not restarted),
# Actually, it does not work (the service is not restarted),
# but master-addon service always terminates after it does it job,
# so it is (usually) not running and it will be started when
# salt is run.
Expand Down
2 changes: 1 addition & 1 deletion cluster/ubuntu/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ function test-build-release {
"${KUBE_ROOT}/build/release.sh"
}

# From user input set the necessary k8s and etcd configuration infomation
# From user input set the necessary k8s and etcd configuration information
function setClusterInfo() {
# Initialize MINION_IPS in setClusterInfo function
# MINION_IPS is defined as a global variable, and is concatenated with other nodeIP
Expand Down
4 changes: 2 additions & 2 deletions cmd/kube-apiserver/app/server.go
Original file line number Diff line number Diff line change
Expand Up @@ -479,7 +479,7 @@ func (s *APIServer) Run(_ []string) error {
}
// err == systemd.SdNotifyNoSocket when not running on a systemd system
if err := systemd.SdNotify("READY=1\n"); err != nil && err != systemd.SdNotifyNoSocket {
glog.Errorf("Unable to send systemd daemon sucessful start message: %v\n", err)
glog.Errorf("Unable to send systemd daemon successful start message: %v\n", err)
}
if err := secureServer.ListenAndServeTLS(s.TLSCertFile, s.TLSPrivateKeyFile); err != nil {
glog.Errorf("Unable to listen for secure (%v); will try again.", err)
Expand All @@ -497,7 +497,7 @@ func (s *APIServer) Run(_ []string) error {
if secureLocation == "" {
// err == systemd.SdNotifyNoSocket when not running on a systemd system
if err := systemd.SdNotify("READY=1\n"); err != nil && err != systemd.SdNotifyNoSocket {
glog.Errorf("Unable to send systemd daemon sucessful start message: %v\n", err)
glog.Errorf("Unable to send systemd daemon successful start message: %v\n", err)
}
}
glog.Infof("Serving insecurely on %s", insecureLocation)
Expand Down
2 changes: 1 addition & 1 deletion cmd/kube-controller-manager/app/controllermanager.go
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,7 @@ func (s *CMServer) AddFlags(fs *pflag.FlagSet) {
fs.IntVar(&s.DeletingPodsBurst, "deleting-pods-burst", 10, "Number of nodes on which pods are bursty deleted in case of node failure. For more details look into RateLimiter.")
fs.IntVar(&s.RegisterRetryCount, "register-retry-count", s.RegisterRetryCount, ""+
"The number of retries for initial node registration. Retry interval equals node-sync-period.")
fs.MarkDeprecated("register-retry-count", "This flag is currenty no-op and will be deleted.")
fs.MarkDeprecated("register-retry-count", "This flag is currently no-op and will be deleted.")
fs.DurationVar(&s.NodeMonitorGracePeriod, "node-monitor-grace-period", 40*time.Second,
"Amount of time which we allow running Node to be unresponsive before marking it unhealty. "+
"Must be N times more than kubelet's nodeStatusUpdateFrequency, "+
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
# LIMITATIONS
# 1. controllers are not updated unless their name is changed
# 3. Services will not be updated unless their name is changed,
# but for services we acually want updates without name change.
# but for services we actually want updates without name change.
# 4. Json files are not handled at all. Currently addons must be
# in yaml files
# 5. exit code is probably not always correct (I haven't checked
Expand Down
2 changes: 1 addition & 1 deletion contrib/ansible/roles/kubernetes/defaults/main.yml
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ kube_token_dir: "{{ kube_config_dir }}/tokens"
kube_manifest_dir: "{{ kube_config_dir }}/manifests"

# This is the group that the cert creation scripts chgrp the
# cert files to. Not really changable...
# cert files to. Not really changeable...
kube_cert_group: kube-cert

# Internal DNS domain name.
Expand Down
2 changes: 1 addition & 1 deletion contrib/diurnal/time.go
Original file line number Diff line number Diff line change
Expand Up @@ -182,7 +182,7 @@ func parseTimeISO8601(s string) (time.Time, error) {
theTime.offMinute = v
s = s[2:]
default:
return time.Time{}, errors.New("an unknown error occured")
return time.Time{}, errors.New("an unknown error occurred")
}
state++
}
Expand Down
2 changes: 1 addition & 1 deletion contrib/for-demos/proxy-to-service/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# Alpine linux would be great for this, but it's DNS does not use seach paths.
# Alpine linux would be great for this, but it's DNS does not use search paths.
FROM progrium/busybox
MAINTAINER Tim Hockin "thockin@google.com"

Expand Down
4 changes: 2 additions & 2 deletions contrib/mesos/pkg/executor/executor_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -263,7 +263,7 @@ func TestExecutorRegister(t *testing.T) {
}

// TestExecutorDisconnect ensures that the executor thinks that it is not
// connected after a call to Disconnected has occured.
// connected after a call to Disconnected has occurred.
func TestExecutorDisconnect(t *testing.T) {
mockDriver := &MockExecutorDriver{}
executor := NewTestKubernetesExecutor()
Expand Down Expand Up @@ -346,7 +346,7 @@ func TestExecutorLaunchAndKillTask(t *testing.T) {
select {
case <-updates:
case <-time.After(time.Second):
t.Fatalf("Executor should send an intial update on Registration")
t.Fatalf("Executor should send an initial update on Registration")
}

pod := NewTestPod(1)
Expand Down
2 changes: 1 addition & 1 deletion contrib/mesos/pkg/queue/interface.go
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ type FIFO interface {
// Pop waits until an item is ready and returns it. If multiple items are
// ready, they are returned in the order in which they were added/updated.
// The item is removed from the queue (and the store) before it is returned,
// so if you don't succesfully process it, you need to add it back with Add().
// so if you don't successfully process it, you need to add it back with Add().
Pop() interface{}

// Await attempts to Pop within the given interval; upon success the non-nil
Expand Down
2 changes: 1 addition & 1 deletion contrib/mesos/pkg/scheduler/plugin.go
Original file line number Diff line number Diff line change
Expand Up @@ -350,7 +350,7 @@ func (k *kubeScheduler) doSchedule(task *podtask.T, err error) (string, error) {
return "", fmt.Errorf("task.offer assignment must be idempotent, task %+v: offer %+v", task, offer)
}

// write resource limits into the pod spec which is transfered to the executor. From here
// write resource limits into the pod spec which is transferred to the executor. From here
// on we can expect that the pod spec of a task has proper limits for CPU and memory.
// TODO(sttts): For a later separation of the kubelet and the executor also patch the pod on the apiserver
if unlimitedCPU := mresource.LimitPodCPU(&task.Pod, k.defaultContainerCPULimit); unlimitedCPU {
Expand Down
2 changes: 1 addition & 1 deletion contrib/service-loadbalancer/service_loadbalancer_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -141,7 +141,7 @@ func TestGetServices(t *testing.T) {
{Port: 20, TargetPort: util.NewIntOrStringFromInt(ports[1])},
}

// 2 services targetting the same endpoints, one of which is declared as a tcp service.
// 2 services targeting the same endpoints, one of which is declared as a tcp service.
svc1 := getService(servicePorts)
svc2 := getService(servicePorts)
endpoints := []*api.Endpoints{
Expand Down
2 changes: 1 addition & 1 deletion docs/admin/cluster-large.md
Original file line number Diff line number Diff line change
Expand Up @@ -88,7 +88,7 @@ To avoid running into cluster addon resource issues, when creating a cluster wit
* [Kibana](http://releases.k8s.io/HEAD/cluster/addons/fluentd-elasticsearch/kibana-controller.yaml)
* Scale number of replicas for the following addons, if used, along with the size of cluster (there are multiple replicas of each so increasing replicas should help handle increased load, but, since load per replica also increases slightly, also consider increasing CPU/memory limits):
* [elasticsearch](http://releases.k8s.io/HEAD/cluster/addons/fluentd-elasticsearch/es-controller.yaml)
* Increase memory and CPU limits sligthly for each of the following addons, if used, along with the size of cluster (there is one replica per node but CPU/memory usage increases slightly along with cluster load/size as well):
* Increase memory and CPU limits slightly for each of the following addons, if used, along with the size of cluster (there is one replica per node but CPU/memory usage increases slightly along with cluster load/size as well):
* [FluentD with ElasticSearch Plugin](http://releases.k8s.io/HEAD/cluster/saltbase/salt/fluentd-es/fluentd-es.yaml)
* [FluentD with GCP Plugin](http://releases.k8s.io/HEAD/cluster/saltbase/salt/fluentd-gcp/fluentd-gcp.yaml)
Expand Down
4 changes: 2 additions & 2 deletions docs/api-reference/definitions.html
Original file line number Diff line number Diff line change
Expand Up @@ -4224,7 +4224,7 @@ <h3 id="_v1_persistentvolumespec">v1.PersistentVolumeSpec</h3>
</tr>
<tr>
<td class="tableblock halign-left valign-top"><p class="tableblock">hostPath</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">a HostPath provisioned by a developer or tester; for develment use only; see <a href="http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath">http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath</a></p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">a HostPath provisioned by a developer or tester; for development use only; see <a href="http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath">http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath</a></p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">false</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock"><a href="#_v1_hostpathvolumesource">v1.HostPathVolumeSource</a></p></td>
<td class="tableblock halign-left valign-top"></td>
Expand Down Expand Up @@ -5099,7 +5099,7 @@ <h3 id="_v1_event">v1.Event</h3>
</tr>
<tr>
<td class="tableblock halign-left valign-top"><p class="tableblock">lastTimestamp</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">the time at which the most recent occurance of this event was recorded</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">the time at which the most recent occurrence of this event was recorded</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">false</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">string</p></td>
<td class="tableblock halign-left valign-top"></td>
Expand Down
2 changes: 1 addition & 1 deletion docs/devel/making-release-notes.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ This documents the process for making release notes for a release.
Find the most-recent PR that was merged with the previous .0 release. Remember this as $LASTPR.
_TODO_: Figure out a way to record this somewhere to save the next release engineer time.

Find the most-recent PR that was merged with the current .0 release. Remeber this as $CURRENTPR.
Find the most-recent PR that was merged with the current .0 release. Remember this as $CURRENTPR.

### 2) Run the release-notes tool

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
## This file is used as input to deployment script, which ammends it as needed.
## This file is used as input to deployment script, which amends it as needed.
## More specifically, we need to add peer hosts for each but the elected peer.

coreos:
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
## This file is used as input to deployment script, which ammends it as needed.
## This file is used as input to deployment script, which amends it as needed.
## More specifically, we need to add environment files for as many nodes as we
## are going to deploy.

Expand Down
6 changes: 3 additions & 3 deletions docs/man/man1/kubectl-namespace.1
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@

.SH NAME
.PP
kubectl namespace \- SUPERCEDED: Set and view the current Kubernetes namespace
kubectl namespace \- SUPERSEDED: Set and view the current Kubernetes namespace


.SH SYNOPSIS
Expand All @@ -13,10 +13,10 @@ kubectl namespace \- SUPERCEDED: Set and view the current Kubernetes namespace

.SH DESCRIPTION
.PP
SUPERCEDED: Set and view the current Kubernetes namespace scope for command line requests.
SUPERSEDED: Set and view the current Kubernetes namespace scope for command line requests.

.PP
namespace has been superceded by the context.namespace field of .kubeconfig files. See 'kubectl config set\-context \-\-help' for more details.
namespace has been superseded by the context.namespace field of .kubeconfig files. See 'kubectl config set\-context \-\-help' for more details.


.SH OPTIONS
Expand Down
2 changes: 1 addition & 1 deletion docs/user-guide/kubectl/kubectl.md
Original file line number Diff line number Diff line change
Expand Up @@ -91,7 +91,7 @@ kubectl
* [kubectl get](kubectl_get.md) - Display one or many resources
* [kubectl label](kubectl_label.md) - Update the labels on a resource
* [kubectl logs](kubectl_logs.md) - Print the logs for a container in a pod.
* [kubectl namespace](kubectl_namespace.md) - SUPERCEDED: Set and view the current Kubernetes namespace
* [kubectl namespace](kubectl_namespace.md) - SUPERSEDED: Set and view the current Kubernetes namespace
* [kubectl patch](kubectl_patch.md) - Update field(s) of a resource by stdin.
* [kubectl port-forward](kubectl_port-forward.md) - Forward one or more local ports to a pod.
* [kubectl proxy](kubectl_proxy.md) - Run a proxy to the Kubernetes API server
Expand Down
6 changes: 3 additions & 3 deletions docs/user-guide/kubectl/kubectl_namespace.md
Original file line number Diff line number Diff line change
Expand Up @@ -33,14 +33,14 @@ Documentation for other releases can be found at

## kubectl namespace

SUPERCEDED: Set and view the current Kubernetes namespace
SUPERSEDED: Set and view the current Kubernetes namespace

### Synopsis


SUPERCEDED: Set and view the current Kubernetes namespace scope for command line requests.
SUPERSEDED: Set and view the current Kubernetes namespace scope for command line requests.

namespace has been superceded by the context.namespace field of .kubeconfig files. See 'kubectl config set-context --help' for more details.
namespace has been superseded by the context.namespace field of .kubeconfig files. See 'kubectl config set-context --help' for more details.


```
Expand Down
2 changes: 1 addition & 1 deletion docs/user-guide/secrets.md
Original file line number Diff line number Diff line change
Expand Up @@ -160,7 +160,7 @@ See another example of creating a secret and a pod that consumes that secret in

### Manually specifying an imagePullSecret

Use of imagePullSecrets is desribed in the [images documentation](images.md#specifying-imagepullsecrets-on-a-pod)
Use of imagePullSecrets is described in the [images documentation](images.md#specifying-imagepullsecrets-on-a-pod)

### Automatic use of Manually Created Secrets

Expand Down
2 changes: 1 addition & 1 deletion examples/cassandra/image/cassandra.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -506,7 +506,7 @@ tombstone_failure_threshold: 100000
# Increase if your rows are large, or if you have a very large
# number of rows per partition. The competing goals are these:
# 1) a smaller granularity means more index entries are generated
# and looking up rows withing the partition by collation column
# and looking up rows within the partition by collation column
# is faster
# 2) but, Cassandra will keep the collation index in memory for hot
# rows (as part of the key cache), so a larger granularity means
Expand Down
2 changes: 1 addition & 1 deletion examples/high-availability/provision.sh
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ function write_api_server_config {
mkdir -p -m 777 /srv/kubernetes/

### We will move files back and forth between the /srv/kube.. directory.
### That is how we modulate leader. Each node will continously either
### That is how we modulate leader. Each node will continuously either
### ensure that the manifests are in this dir, or else, are in the kubelet manifest dir.
cp /vagrant/kube-scheduler.manifest /vagrant/kube-controller-manager.manifest /srv/kubernetes

Expand Down
Loading

0 comments on commit 5f9cefc

Please sign in to comment.