diff options
| author | Kubernetes Submit Queue <k8s-merge-robot@users.noreply.github.com> | 2017-11-10 07:42:31 -0800 |
|---|---|---|
| committer | GitHub <noreply@github.com> | 2017-11-10 07:42:31 -0800 |
| commit | f6b7936faef34ac71e9c6f00304df138b315697c (patch) | |
| tree | 0bfb5438c2efad11fa2ee7c8b8d42d87666a0e24 | |
| parent | 3907591a7d98d041c1bee242cf5abbc29a75aeda (diff) | |
| parent | e4db35f630c56dda6fad1820956bf17c65c91b82 (diff) | |
Merge pull request #1388 from zhxcai/typos
Automatic merge from submit-queue.
Update a lot of typos
8 files changed, 22 insertions, 22 deletions
diff --git a/contributors/design-proposals/instrumentation/events-redesign.md b/contributors/design-proposals/instrumentation/events-redesign.md index cf9fbdd6..6540d61f 100644 --- a/contributors/design-proposals/instrumentation/events-redesign.md +++ b/contributors/design-proposals/instrumentation/events-redesign.md @@ -100,13 +100,13 @@ We'd like to propose following structure in Events object in the new events API ```golang type Event struct { -// <type and object metadata> + // <type and object metadata> // Time when this Event was first observed. EventTime metav1.MicroTime // Data about the Event series this event represents or nil if it's -// a singleton Event. + // a singleton Event. // +optional Series *EventSeries @@ -123,7 +123,7 @@ type Event struct { Reason string // The object this Event is “about”. In most cases it's the object that the -// given controller implements. + // given controller implements. // +optional Regarding ObjectReference @@ -132,24 +132,24 @@ type Event struct { Related *ObjectReference // Human readable description of the Event. Possibly discarded when and -// Event series is being deduplicated. + // Event series is being deduplicated. // +optional Note string // Type of this event (Normal, Warning), new types could be added in the -// future. -// +optional -Type string + // future. + // +optional + Type string } type EventSeries struct { -Count int32 -LastObservedTime MicroTime -State EventSeriesState + Count int32 + LastObservedTime MicroTime + State EventSeriesState } const ( -EventSeriesStateOngoing = "Ongoing" + EventSeriesStateOngoing = "Ongoing" EventSeriesStateFinished = "Finished" EventSeriesStateUnknown = "Unknown" ) @@ -161,7 +161,7 @@ EventSeriesStateOngoing = "Ongoing" | ----------| -------| -------| --------------------|---------| | Node X | BecameUnreachable | HeartbeatTooOld | kubernetes.io/node-ctrl | <nil> | | Node Y | FailedToAttachVolume | Unknown | kubernetes.io/pv-attach-ctrl | PVC X | -| ReplicaSet X | FailedToInstantiantePod | QuotaExceeded | kubernetes.io/replica-set-ctrl | <nil> | +| ReplicaSet X | FailedToInstantiatePod | QuotaExceeded | kubernetes.io/replica-set-ctrl | <nil> | | ReplicaSet X | InstantiatedPod | | kubernetes.io/replica-set-ctrl | Pod Y | | Ingress X | CreatedLoadBalancer | | kubernetes.io/ingress-ctrl | <nil> | | Pod X | ScheduledOn | | kubernetes.io/scheduler | Node Y | diff --git a/contributors/design-proposals/instrumentation/performance-related-monitoring.md b/contributors/design-proposals/instrumentation/performance-related-monitoring.md index 1279cd29..91b799f1 100644 --- a/contributors/design-proposals/instrumentation/performance-related-monitoring.md +++ b/contributors/design-proposals/instrumentation/performance-related-monitoring.md @@ -32,7 +32,7 @@ how careful we need to be. ### Huge number of handshakes slows down API server It was a long standing issue for performance and is/was an important bottleneck for scalability (https://github.com/kubernetes/kubernetes/issues/13671). The bug directly -causing this problem was incorrect (from the golangs standpoint) handling of TCP connections. Secondary issue was that elliptic curve encryption (only one available in go 1.4) +causing this problem was incorrect (from the golang's standpoint) handling of TCP connections. Secondary issue was that elliptic curve encryption (only one available in go 1.4) is unbelievably slow. ## Proposed metrics/statistics to gather/compute to avoid problems @@ -42,7 +42,7 @@ is unbelievably slow. Basic ideas: - number of Pods/ReplicationControllers/Services in the cluster - number of running replicas of master components (if they are replicated) -- current elected master of ectd cluster (if running distributed version) +- current elected master of etcd cluster (if running distributed version) - number of master component restarts - number of lost Nodes diff --git a/contributors/design-proposals/instrumentation/resource-metrics-api.md b/contributors/design-proposals/instrumentation/resource-metrics-api.md index b5840b75..075c6180 100644 --- a/contributors/design-proposals/instrumentation/resource-metrics-api.md +++ b/contributors/design-proposals/instrumentation/resource-metrics-api.md @@ -20,7 +20,7 @@ Use cases which are not listed below are out of the scope of MVP version of Reso HPA uses the latest value of cpu usage as an average aggregated across 1 minute (the window may change in the future). The data for a given set of pods -(defined either by pod list or label selector) should be accesible in one request +(defined either by pod list or label selector) should be accessible in one request due to performance issues. #### Scheduler diff --git a/contributors/design-proposals/multicluster/federation-lite.md b/contributors/design-proposals/multicluster/federation-lite.md index afdd6c02..3dd1f4d8 100644 --- a/contributors/design-proposals/multicluster/federation-lite.md +++ b/contributors/design-proposals/multicluster/federation-lite.md @@ -86,7 +86,7 @@ service of type LoadBalancer. The native cloud load-balancers on both AWS & GCE are region-level, and support load-balancing across instances in multiple zones (in the same region). For both clouds, the behaviour of the native cloud load-balancer is reasonable in the face of failures (indeed, this is why clouds -provide load-balancing as a primitve). +provide load-balancing as a primitive). For multi-AZ clusters we will therefore simply rely on the native cloud provider load balancer behaviour, and we do not anticipate substantial code changes. @@ -170,8 +170,8 @@ GCE. If you had two volumes both named `myvolume` in two different GCE zones, this would not be ambiguous when Kubernetes is operating only in a single zone. But, when operating a cluster across multiple zones, `myvolume` is no longer sufficient to specify a volume uniquely. Worse, the fact that a volume happens -to be unambigious at a particular time is no guarantee that it will continue to -be unambigious in future, because a volume with the same name could +to be unambiguous at a particular time is no guarantee that it will continue to +be unambiguous in future, because a volume with the same name could subsequently be created in a second zone. While perhaps unlikely in practice, we cannot automatically enable multi-AZ clusters for GCE users if this then causes volume mounts to stop working. diff --git a/contributors/design-proposals/network/external-lb-source-ip-preservation.md b/contributors/design-proposals/network/external-lb-source-ip-preservation.md index ac48ed5d..50140a0e 100644 --- a/contributors/design-proposals/network/external-lb-source-ip-preservation.md +++ b/contributors/design-proposals/network/external-lb-source-ip-preservation.md @@ -162,7 +162,7 @@ For the 1.4 release, this feature will be implemented for the GCE cloud provider - Node: On the node, we expect to see the real source IP of the client. Destination IP will be the Service Virtual External IP. -- Pod: For processes running inside the Pod network namepsace, the source IP will be the real client source IP. The destination address will the be Pod IP. +- Pod: For processes running inside the Pod network namespace, the source IP will be the real client source IP. The destination address will the be Pod IP. #### GCE Expected Packet Destination IP (HealthCheck path) diff --git a/contributors/design-proposals/node/dynamic-kubelet-configuration.md b/contributors/design-proposals/node/dynamic-kubelet-configuration.md index 188fb84f..3dc8870f 100644 --- a/contributors/design-proposals/node/dynamic-kubelet-configuration.md +++ b/contributors/design-proposals/node/dynamic-kubelet-configuration.md @@ -75,7 +75,7 @@ There are two types of configuration that are stored on disk: - cached configurations from a remote source, e.g. `ConfigMaps` from etcd. - the local "init" configuration, e.g. the set of config files the node is provisioned with. -The Kubelet should accept a `--dynamic-config-dir` flag, which specifies a directory for storing all of the information necessary for dynamic configuraiton from remote sources; e.g. the cached configurations, which configuration is currently in use, which configurations are known to be bad, etc. +The Kubelet should accept a `--dynamic-config-dir` flag, which specifies a directory for storing all of the information necessary for dynamic configuration from remote sources; e.g. the cached configurations, which configuration is currently in use, which configurations are known to be bad, etc. - When the Kubelet downloads a `ConfigMap`, it will checkpoint a serialization of the `ConfigMap` object to a file at `{dynamic-config-dir}/checkpoints/{UID}`. - We checkpoint the entire object, rather than unpacking the contents to disk, because the former is less complex and reduces chance for errors during the checkpoint process. diff --git a/contributors/design-proposals/node/kubelet-cri-logging.md b/contributors/design-proposals/node/kubelet-cri-logging.md index 37f3e3da..a19ff3f5 100644 --- a/contributors/design-proposals/node/kubelet-cri-logging.md +++ b/contributors/design-proposals/node/kubelet-cri-logging.md @@ -199,7 +199,7 @@ clients attaching as well. There are ad-hoc solutions/discussions that addresses one or two of the requirements, but no comprehensive solution for CRI specifically has been -proposed so far (with the excpetion of @tmrtfs's proposal +proposed so far (with the exception of @tmrtfs's proposal [#33111](https://github.com/kubernetes/kubernetes/pull/33111), which has a much wider scope). It has come up in discussions that kubelet can delegate all the logging management to the runtime to allow maximum flexibility. However, it is diff --git a/contributors/design-proposals/node/kubelet-rkt-runtime.md b/contributors/design-proposals/node/kubelet-rkt-runtime.md index 98a06187..1bc6435b 100644 --- a/contributors/design-proposals/node/kubelet-rkt-runtime.md +++ b/contributors/design-proposals/node/kubelet-rkt-runtime.md @@ -74,7 +74,7 @@ In addition, the rkt cli has historically been the primary interface to the rkt The initial integration will execute the rkt binary directly for app creation/start/stop/removal, as well as image pulling/removal. -The creation of pod sanbox is also done via rkt command line, but it will run under `systemd-run` so it's monitored by the init process. +The creation of pod sandbox is also done via rkt command line, but it will run under `systemd-run` so it's monitored by the init process. In the future, some of these decisions are expected to be changed such that rkt is vendored as a library dependency for all operations, and other init systems will be supported as well. |
