On Mon, 2015-12-28 at 16:38 +0100, Michal Kubecek wrote:
> On Mon, Dec 28, 2015 at 03:29:42PM +0100, Sabrina Dubroca wrote:
> > 2015-12-28, 15:01:57 +0100, Michal Kubecek wrote:
> > > Recent fix "net: add length argument to
> > > skb_copy_and_csum_datagram_iovec" added to some pre-3.19 stable
> > > branches, namely
> > > 
> > >   stable-3.2.y: commit 127500d724f8
> > >   stable-3.12.y: commit 3e1ac3aafbd0
> > > 
> > > doesn't handle truncated reads correctly. If read length is shorter than
> > > incoming datagram (but non-zero) and first segment of target iovec is
> > > sufficient for read length, skb_copy_and_csum_datagram() is used to copy
> > > checksum the data while copying it. For truncated reads this means only
> > > the copied part is checksummed (rather than the whole datagram) so that
> > > the check almost always fails.
> > 
> > I just ran into this issue too, sorry I didn't notice it earlier :(
> > 
> > > Add checksum of the remaining part so that the proper checksum of the
> > > whole datagram is computed and checked. Special care must be taken if
> > > the copied length is odd.
> > > 
> > > For zero read length, we don't have to copy anything but we still should
> > > check the checksum so that a peek doesn't return with a datagram which
> > > is invalid and wouldn't be returned by an actual read.
> > > 
> > > Signed-off-by: Michal Kubecek <mkube...@suse.cz>
> > > ---
> > >  net/core/datagram.c | 26 +++++++++++++++++++++-----
> > >  1 file changed, 21 insertions(+), 5 deletions(-)
> > > 
> > > diff --git a/net/core/datagram.c b/net/core/datagram.c
> > > index f22f120771ef..af4bf368257c 100644
> > > --- a/net/core/datagram.c
> > > +++ b/net/core/datagram.c
> > > @@ -809,13 +809,14 @@ int skb_copy_and_csum_datagram_iovec(struct sk_buff 
> > > *skb,
> > >                                int hlen, struct iovec *iov, int len)
> > >  {
> > >   __wsum csum;
> > > - int chunk = skb->len - hlen;
> > > + int full_chunk = skb->len - hlen;
> > > + int chunk = min_t(int, full_chunk, len);
> > >  
> > > - if (chunk > len)
> > > -         chunk = len;
> > > -
> > > - if (!chunk)
> > > + if (!chunk) {
> > > +         if (__skb_checksum_complete(skb))
> > > +                 goto csum_error;
> > >           return 0;
> > > + }
> > >  
> > >   /* Skip filled elements.
> > >    * Pretty silly, look at memcpy_toiovec, though 8)
> > > @@ -833,6 +834,21 @@ int skb_copy_and_csum_datagram_iovec(struct sk_buff 
> > > *skb,
> > >           if (skb_copy_and_csum_datagram(skb, hlen, iov->iov_base,
> > >                                          chunk, &csum))
> > >                   goto fault;
> > > +         if (full_chunk > chunk) {
> > > +                 if (chunk % 2) {
> > > +                         __be16 odd = 0;
> > > +
> > > +                         if (skb_copy_bits(skb, hlen + chunk,
> > > +                                           (char *)&odd + 1, 1))
> > > +                                 goto fault;
> > > +                         csum = add32_with_carry(odd, csum);
> > > +                         csum = skb_checksum(skb, hlen + chunk + 1,
> > > +                                             full_chunk - chunk - 1,
> > > +                                             csum);
> > > +                 } else
> > > +                         csum = skb_checksum(skb, hlen + chunk,
> > > +                                             full_chunk - chunk, csum);
> > > +         }
> > >           if (csum_fold(csum))
> > >                   goto csum_error;
> > >           if (unlikely(skb->ip_summed == CHECKSUM_COMPLETE))
> > > -- 
> > > 2.6.4
> > 
> > 
> > This adds quite a bit of complexity.
> 
> I'm not really happy about it either. :-( Most of the complexity comes
> from the corner case when one 16-bit word is divided between the copied
> part and the rest - but I can't see a nicer way to handle it.
> 
> There is another option: in the case of truncated read, we could simply
> take the first branch where copying is separated from checksumming. This
> would be less efficient, obviously, but I must admit I have no idea how
> much.
> 
> > I am considering a revert of my
> > buggy patch, and use what Eric Dumazet suggested instead:
> > 
> > https://patchwork.ozlabs.org/patch/543562/
> > 
> > What do you think?
> 
> I believe that would work. I have a little bad feeling about such
> solution as it would keep the function broken and just rely on not
> hitting it in the case when it matters. But it worked that way for quite
> some time so it's probably safe.

For the record, this is what we are using here at Google on our prod
kernel.

Sabrina, I certainly can send the patch for net-next kernel, as this
does not fix a bug for current kernels, but backporting it would be
indeed a way to fix the issue for old kernels.






--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to