summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--contributors/design-proposals/instrumentation/events-redesign.md24
-rw-r--r--contributors/design-proposals/instrumentation/performance-related-monitoring.md4
-rw-r--r--contributors/design-proposals/instrumentation/resource-metrics-api.md2
-rw-r--r--contributors/design-proposals/multicluster/federation-lite.md6
-rw-r--r--contributors/design-proposals/network/external-lb-source-ip-preservation.md2
-rw-r--r--contributors/design-proposals/node/dynamic-kubelet-configuration.md2
-rw-r--r--contributors/design-proposals/node/kubelet-cri-logging.md2
-rw-r--r--contributors/design-proposals/node/kubelet-rkt-runtime.md2
8 files changed, 22 insertions, 22 deletions
diff --git a/contributors/design-proposals/instrumentation/events-redesign.md b/contributors/design-proposals/instrumentation/events-redesign.md
index cf9fbdd6..6540d61f 100644
--- a/contributors/design-proposals/instrumentation/events-redesign.md
+++ b/contributors/design-proposals/instrumentation/events-redesign.md
@@ -100,13 +100,13 @@ We'd like to propose following structure in Events object in the new events API
```golang
type Event struct {
-// <type and object metadata>
+ // <type and object metadata>
// Time when this Event was first observed.
EventTime metav1.MicroTime
// Data about the Event series this event represents or nil if it's
-// a singleton Event.
+ // a singleton Event.
// +optional
Series *EventSeries
@@ -123,7 +123,7 @@ type Event struct {
Reason string
// The object this Event is “about”. In most cases it's the object that the
-// given controller implements.
+ // given controller implements.
// +optional
Regarding ObjectReference
@@ -132,24 +132,24 @@ type Event struct {
Related *ObjectReference
// Human readable description of the Event. Possibly discarded when and
-// Event series is being deduplicated.
+ // Event series is being deduplicated.
// +optional
Note string
// Type of this event (Normal, Warning), new types could be added in the
-// future.
-// +optional
-Type string
+ // future.
+ // +optional
+ Type string
}
type EventSeries struct {
-Count int32
-LastObservedTime MicroTime
-State EventSeriesState
+ Count int32
+ LastObservedTime MicroTime
+ State EventSeriesState
}
const (
-EventSeriesStateOngoing = "Ongoing"
+ EventSeriesStateOngoing = "Ongoing"
EventSeriesStateFinished = "Finished"
EventSeriesStateUnknown = "Unknown"
)
@@ -161,7 +161,7 @@ EventSeriesStateOngoing = "Ongoing"
| ----------| -------| -------| --------------------|---------|
| Node X | BecameUnreachable | HeartbeatTooOld | kubernetes.io/node-ctrl | <nil> |
| Node Y | FailedToAttachVolume | Unknown | kubernetes.io/pv-attach-ctrl | PVC X |
-| ReplicaSet X | FailedToInstantiantePod | QuotaExceeded | kubernetes.io/replica-set-ctrl | <nil> |
+| ReplicaSet X | FailedToInstantiatePod | QuotaExceeded | kubernetes.io/replica-set-ctrl | <nil> |
| ReplicaSet X | InstantiatedPod | | kubernetes.io/replica-set-ctrl | Pod Y |
| Ingress X | CreatedLoadBalancer | | kubernetes.io/ingress-ctrl | <nil> |
| Pod X | ScheduledOn | | kubernetes.io/scheduler | Node Y |
diff --git a/contributors/design-proposals/instrumentation/performance-related-monitoring.md b/contributors/design-proposals/instrumentation/performance-related-monitoring.md
index 1279cd29..91b799f1 100644
--- a/contributors/design-proposals/instrumentation/performance-related-monitoring.md
+++ b/contributors/design-proposals/instrumentation/performance-related-monitoring.md
@@ -32,7 +32,7 @@ how careful we need to be.
### Huge number of handshakes slows down API server
It was a long standing issue for performance and is/was an important bottleneck for scalability (https://github.com/kubernetes/kubernetes/issues/13671). The bug directly
-causing this problem was incorrect (from the golangs standpoint) handling of TCP connections. Secondary issue was that elliptic curve encryption (only one available in go 1.4)
+causing this problem was incorrect (from the golang's standpoint) handling of TCP connections. Secondary issue was that elliptic curve encryption (only one available in go 1.4)
is unbelievably slow.
## Proposed metrics/statistics to gather/compute to avoid problems
@@ -42,7 +42,7 @@ is unbelievably slow.
Basic ideas:
- number of Pods/ReplicationControllers/Services in the cluster
- number of running replicas of master components (if they are replicated)
-- current elected master of ectd cluster (if running distributed version)
+- current elected master of etcd cluster (if running distributed version)
- number of master component restarts
- number of lost Nodes
diff --git a/contributors/design-proposals/instrumentation/resource-metrics-api.md b/contributors/design-proposals/instrumentation/resource-metrics-api.md
index b5840b75..075c6180 100644
--- a/contributors/design-proposals/instrumentation/resource-metrics-api.md
+++ b/contributors/design-proposals/instrumentation/resource-metrics-api.md
@@ -20,7 +20,7 @@ Use cases which are not listed below are out of the scope of MVP version of Reso
HPA uses the latest value of cpu usage as an average aggregated across 1 minute
(the window may change in the future). The data for a given set of pods
-(defined either by pod list or label selector) should be accesible in one request
+(defined either by pod list or label selector) should be accessible in one request
due to performance issues.
#### Scheduler
diff --git a/contributors/design-proposals/multicluster/federation-lite.md b/contributors/design-proposals/multicluster/federation-lite.md
index afdd6c02..3dd1f4d8 100644
--- a/contributors/design-proposals/multicluster/federation-lite.md
+++ b/contributors/design-proposals/multicluster/federation-lite.md
@@ -86,7 +86,7 @@ service of type LoadBalancer. The native cloud load-balancers on both AWS &
GCE are region-level, and support load-balancing across instances in multiple
zones (in the same region). For both clouds, the behaviour of the native cloud
load-balancer is reasonable in the face of failures (indeed, this is why clouds
-provide load-balancing as a primitve).
+provide load-balancing as a primitive).
For multi-AZ clusters we will therefore simply rely on the native cloud provider
load balancer behaviour, and we do not anticipate substantial code changes.
@@ -170,8 +170,8 @@ GCE. If you had two volumes both named `myvolume` in two different GCE zones,
this would not be ambiguous when Kubernetes is operating only in a single zone.
But, when operating a cluster across multiple zones, `myvolume` is no longer
sufficient to specify a volume uniquely. Worse, the fact that a volume happens
-to be unambigious at a particular time is no guarantee that it will continue to
-be unambigious in future, because a volume with the same name could
+to be unambiguous at a particular time is no guarantee that it will continue to
+be unambiguous in future, because a volume with the same name could
subsequently be created in a second zone. While perhaps unlikely in practice,
we cannot automatically enable multi-AZ clusters for GCE users if this then causes
volume mounts to stop working.
diff --git a/contributors/design-proposals/network/external-lb-source-ip-preservation.md b/contributors/design-proposals/network/external-lb-source-ip-preservation.md
index ac48ed5d..50140a0e 100644
--- a/contributors/design-proposals/network/external-lb-source-ip-preservation.md
+++ b/contributors/design-proposals/network/external-lb-source-ip-preservation.md
@@ -162,7 +162,7 @@ For the 1.4 release, this feature will be implemented for the GCE cloud provider
- Node: On the node, we expect to see the real source IP of the client. Destination IP will be the Service Virtual External IP.
-- Pod: For processes running inside the Pod network namepsace, the source IP will be the real client source IP. The destination address will the be Pod IP.
+- Pod: For processes running inside the Pod network namespace, the source IP will be the real client source IP. The destination address will the be Pod IP.
#### GCE Expected Packet Destination IP (HealthCheck path)
diff --git a/contributors/design-proposals/node/dynamic-kubelet-configuration.md b/contributors/design-proposals/node/dynamic-kubelet-configuration.md
index 188fb84f..3dc8870f 100644
--- a/contributors/design-proposals/node/dynamic-kubelet-configuration.md
+++ b/contributors/design-proposals/node/dynamic-kubelet-configuration.md
@@ -75,7 +75,7 @@ There are two types of configuration that are stored on disk:
- cached configurations from a remote source, e.g. `ConfigMaps` from etcd.
- the local "init" configuration, e.g. the set of config files the node is provisioned with.
-The Kubelet should accept a `--dynamic-config-dir` flag, which specifies a directory for storing all of the information necessary for dynamic configuraiton from remote sources; e.g. the cached configurations, which configuration is currently in use, which configurations are known to be bad, etc.
+The Kubelet should accept a `--dynamic-config-dir` flag, which specifies a directory for storing all of the information necessary for dynamic configuration from remote sources; e.g. the cached configurations, which configuration is currently in use, which configurations are known to be bad, etc.
- When the Kubelet downloads a `ConfigMap`, it will checkpoint a serialization of the `ConfigMap` object to a file at `{dynamic-config-dir}/checkpoints/{UID}`.
- We checkpoint the entire object, rather than unpacking the contents to disk, because the former is less complex and reduces chance for errors during the checkpoint process.
diff --git a/contributors/design-proposals/node/kubelet-cri-logging.md b/contributors/design-proposals/node/kubelet-cri-logging.md
index 37f3e3da..a19ff3f5 100644
--- a/contributors/design-proposals/node/kubelet-cri-logging.md
+++ b/contributors/design-proposals/node/kubelet-cri-logging.md
@@ -199,7 +199,7 @@ clients attaching as well.
There are ad-hoc solutions/discussions that addresses one or two of the
requirements, but no comprehensive solution for CRI specifically has been
-proposed so far (with the excpetion of @tmrtfs's proposal
+proposed so far (with the exception of @tmrtfs's proposal
[#33111](https://github.com/kubernetes/kubernetes/pull/33111), which has a much
wider scope). It has come up in discussions that kubelet can delegate all the
logging management to the runtime to allow maximum flexibility. However, it is
diff --git a/contributors/design-proposals/node/kubelet-rkt-runtime.md b/contributors/design-proposals/node/kubelet-rkt-runtime.md
index 98a06187..1bc6435b 100644
--- a/contributors/design-proposals/node/kubelet-rkt-runtime.md
+++ b/contributors/design-proposals/node/kubelet-rkt-runtime.md
@@ -74,7 +74,7 @@ In addition, the rkt cli has historically been the primary interface to the rkt
The initial integration will execute the rkt binary directly for app creation/start/stop/removal, as well as image pulling/removal.
-The creation of pod sanbox is also done via rkt command line, but it will run under `systemd-run` so it's monitored by the init process.
+The creation of pod sandbox is also done via rkt command line, but it will run under `systemd-run` so it's monitored by the init process.
In the future, some of these decisions are expected to be changed such that rkt is vendored as a library dependency for all operations, and other init systems will be supported as well.