Correct spelling typos in several code comments. - alignement -> alignment - wont -> won't - orignal -> original - substract -> subtract - Segement -> Segment - splitted -> split Signed-off-by: Shi Hao --- changes v2: - Update subject message - Add additional spelling typo fixes - Combined all typo fixes into single patch as requested --- net/ipv4/ip_fragment.c | 2 +- net/ipv4/netfilter/arp_tables.c | 2 +- net/ipv4/syncookies.c | 2 +- net/ipv4/tcp.c | 2 +- net/ipv4/tcp_input.c | 4 ++-- net/ipv4/tcp_output.c | 2 +- 6 files changed, 7 insertions(+), 7 deletions(-) diff --git a/net/ipv4/ip_fragment.c b/net/ipv4/ip_fragment.c index f7012479713b..116ebdd1ec86 100644 --- a/net/ipv4/ip_fragment.c +++ b/net/ipv4/ip_fragment.c @@ -335,7 +335,7 @@ static int ip_frag_queue(struct ipq *qp, struct sk_buff *skb, int *refs) /* Note : skb->rbnode and skb->dev share the same location. */ dev = skb->dev; - /* Makes sure compiler wont do silly aliasing games */ + /* Makes sure compiler won't do silly aliasing games */ barrier(); prev_tail = qp->q.fragments_tail; diff --git a/net/ipv4/netfilter/arp_tables.c b/net/ipv4/netfilter/arp_tables.c index 1cdd9c28ab2d..f4a339921a8d 100644 --- a/net/ipv4/netfilter/arp_tables.c +++ b/net/ipv4/netfilter/arp_tables.c @@ -59,7 +59,7 @@ static inline int arp_devaddr_compare(const struct arpt_devaddr_info *ap, /* * Unfortunately, _b and _mask are not aligned to an int (or long int) * Some arches dont care, unrolling the loop is a win on them. - * For other arches, we only have a 16bit alignement. + * For other arches, we only have a 16bit alignment. */ static unsigned long ifname_compare(const char *_a, const char *_b, const char *_mask) { diff --git a/net/ipv4/syncookies.c b/net/ipv4/syncookies.c index 569befcf021b..7d0d34329259 100644 --- a/net/ipv4/syncookies.c +++ b/net/ipv4/syncookies.c @@ -131,7 +131,7 @@ static __u32 check_tcp_syn_cookie(__u32 cookie, __be32 saddr, __be32 daddr, /* * MSS Values are chosen based on the 2011 paper - * 'An Analysis of TCP Maximum Segement Sizes' by S. Alcock and R. Nelson. + * 'An Analysis of TCP Maximum Segment Sizes' by S. Alcock and R. Nelson. * Values .. * .. lower than 536 are rare (< 0.2%) * .. between 537 and 1299 account for less than < 1.5% of observed values diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c index 8a18aeca7ab0..0d1c8e805d24 100644 --- a/net/ipv4/tcp.c +++ b/net/ipv4/tcp.c @@ -1600,7 +1600,7 @@ struct sk_buff *tcp_recv_skb(struct sock *sk, u32 seq, u32 *off) return skb; } /* This looks weird, but this can happen if TCP collapsing - * splitted a fat GRO packet, while we released socket lock + * split a fat GRO packet, while we released socket lock * in skb_splice_bits() */ tcp_eat_recv_skb(sk, skb); diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c index e4a979b75cc6..6c09018b3900 100644 --- a/net/ipv4/tcp_input.c +++ b/net/ipv4/tcp_input.c @@ -1129,7 +1129,7 @@ static void tcp_update_pacing_rate(struct sock *sk) do_div(rate, tp->srtt_us); /* WRITE_ONCE() is needed because sch_fq fetches sk_pacing_rate - * without any lock. We want to make sure compiler wont store + * without any lock. We want to make sure compiler won't store * intermediate values in this location. */ WRITE_ONCE(sk->sk_pacing_rate, @@ -4868,7 +4868,7 @@ void tcp_sack_compress_send_ack(struct sock *sk) __sock_put(sk); /* Since we have to send one ack finally, - * substract one from tp->compressed_ack to keep + * subtract one from tp->compressed_ack to keep * LINUX_MIB_TCPACKCOMPRESSED accurate. */ NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, diff --git a/net/ipv4/tcp_output.c b/net/ipv4/tcp_output.c index b94efb3050d2..483d6b578503 100644 --- a/net/ipv4/tcp_output.c +++ b/net/ipv4/tcp_output.c @@ -2629,7 +2629,7 @@ static int tcp_mtu_probe(struct sock *sk) interval = icsk->icsk_mtup.search_high - icsk->icsk_mtup.search_low; /* When misfortune happens, we are reprobing actively, * and then reprobe timer has expired. We stick with current - * probing process by not resetting search range to its orignal. + * probing process by not resetting search range to its original. */ if (probe_size > tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_high) || interval < READ_ONCE(net->ipv4.sysctl_tcp_probe_threshold)) { -- 2.51.0