net/tstun,wgengine/netstack: make inbound synthetic packet injection GSO-aware (#13266)
Updates tailscale/corp#22511 Signed-off-by: Jordan Whited <jordan@tailscale.com>
This commit is contained in:
parent
6d4973e1e0
commit
d097096ddc
|
@ -994,6 +994,13 @@ func stackGSOToTunGSO(pkt []byte, gso stack.GSO) (tun.GSOOptions, error) {
|
|||
return options, nil
|
||||
}
|
||||
|
||||
// invertGSOChecksum inverts the transport layer checksum in pkt if gVisor
|
||||
// handed us a segment with a partial checksum. A partial checksum is not a
|
||||
// ones' complement of the sum, and incremental checksum updating is not yet
|
||||
// partial checksum aware. This may be called twice for a single packet,
|
||||
// both before and after partial checksum updates where later checksum
|
||||
// offloading still expects a partial checksum.
|
||||
// TODO(jwhited): plumb partial checksum awareness into net/packet/checksum.
|
||||
func invertGSOChecksum(pkt []byte, gso stack.GSO) {
|
||||
if gso.NeedsCsum != true {
|
||||
return
|
||||
|
@ -1030,13 +1037,6 @@ func (t *Wrapper) injectedRead(res tunInjectedRead, outBuffs [][]byte, sizes []i
|
|||
defer parsedPacketPool.Put(p)
|
||||
p.Decode(pkt)
|
||||
|
||||
// We invert the transport layer checksum before and after snat() if gVisor
|
||||
// handed us a segment with a partial checksum. A partial checksum is not a
|
||||
// ones' complement of the sum, and incremental checksum updating that could
|
||||
// occur as a result of snat() is not aware of this. Alternatively we could
|
||||
// plumb partial transport layer checksum awareness down through snat(),
|
||||
// but the surface area of such a change is much larger, and not yet
|
||||
// justified by this singular case.
|
||||
invertGSOChecksum(pkt, gso)
|
||||
pc.snat(p)
|
||||
invertGSOChecksum(pkt, gso)
|
||||
|
@ -1241,36 +1241,73 @@ func (t *Wrapper) SetJailedFilter(filt *filter.Filter) {
|
|||
}
|
||||
|
||||
// InjectInboundPacketBuffer makes the Wrapper device behave as if a packet
|
||||
// with the given contents was received from the network.
|
||||
// It takes ownership of one reference count on the packet. The injected
|
||||
// (pkt) with the given contents was received from the network.
|
||||
// It takes ownership of one reference count on pkt. The injected
|
||||
// packet will not pass through inbound filters.
|
||||
//
|
||||
// pkt will be copied into buffs before writing to the underlying tun.Device.
|
||||
// Therefore, callers must allocate and pass a buffs slice that is sized
|
||||
// appropriately for holding pkt.Size() + PacketStartOffset as either a single
|
||||
// element (buffs[0]), or split across multiple elements if the originating
|
||||
// stack supports GSO. sizes must be sized with similar consideration,
|
||||
// len(buffs) should be equal to len(sizes). If any len(buffs[<index>]) was
|
||||
// mutated by InjectInboundPacketBuffer it will be reset to cap(buffs[<index>])
|
||||
// before returning.
|
||||
//
|
||||
// This path is typically used to deliver synthesized packets to the
|
||||
// host networking stack.
|
||||
func (t *Wrapper) InjectInboundPacketBuffer(pkt *stack.PacketBuffer) error {
|
||||
buf := make([]byte, PacketStartOffset+pkt.Size())
|
||||
func (t *Wrapper) InjectInboundPacketBuffer(pkt *stack.PacketBuffer, buffs [][]byte, sizes []int) error {
|
||||
buf := buffs[0][PacketStartOffset:]
|
||||
|
||||
n := copy(buf[PacketStartOffset:], pkt.NetworkHeader().Slice())
|
||||
n += copy(buf[PacketStartOffset+n:], pkt.TransportHeader().Slice())
|
||||
n += copy(buf[PacketStartOffset+n:], pkt.Data().AsRange().ToSlice())
|
||||
if n != pkt.Size() {
|
||||
bufN := copy(buf, pkt.NetworkHeader().Slice())
|
||||
bufN += copy(buf[bufN:], pkt.TransportHeader().Slice())
|
||||
bufN += copy(buf[bufN:], pkt.Data().AsRange().ToSlice())
|
||||
if bufN != pkt.Size() {
|
||||
panic("unexpected packet size after copy")
|
||||
}
|
||||
pkt.DecRef()
|
||||
buf = buf[:bufN]
|
||||
defer pkt.DecRef()
|
||||
|
||||
pc := t.peerConfig.Load()
|
||||
|
||||
p := parsedPacketPool.Get().(*packet.Parsed)
|
||||
defer parsedPacketPool.Put(p)
|
||||
p.Decode(buf[PacketStartOffset:])
|
||||
p.Decode(buf)
|
||||
captHook := t.captureHook.Load()
|
||||
if captHook != nil {
|
||||
captHook(capture.SynthesizedToLocal, t.now(), p.Buffer(), p.CaptureMeta)
|
||||
}
|
||||
|
||||
invertGSOChecksum(buf, pkt.GSOOptions)
|
||||
pc.dnat(p)
|
||||
invertGSOChecksum(buf, pkt.GSOOptions)
|
||||
|
||||
return t.InjectInboundDirect(buf, PacketStartOffset)
|
||||
gso, err := stackGSOToTunGSO(buf, pkt.GSOOptions)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// TODO(jwhited): support GSO passthrough to t.tdev. If t.tdev supports
|
||||
// GSO we don't need to split here and coalesce inside wireguard-go,
|
||||
// we can pass a coalesced segment all the way through.
|
||||
n, err := tun.GSOSplit(buf, gso, buffs, sizes, PacketStartOffset)
|
||||
if err != nil {
|
||||
if errors.Is(err, tun.ErrTooManySegments) {
|
||||
t.limitedLogf("InjectInboundPacketBuffer: GSO split overflows buffs")
|
||||
} else {
|
||||
return err
|
||||
}
|
||||
}
|
||||
for i := 0; i < n; i++ {
|
||||
buffs[i] = buffs[i][:PacketStartOffset+sizes[i]]
|
||||
}
|
||||
defer func() {
|
||||
for i := 0; i < n; i++ {
|
||||
buffs[i] = buffs[i][:cap(buffs[i])]
|
||||
}
|
||||
}()
|
||||
_, err = t.tdevWrite(buffs[:n], PacketStartOffset)
|
||||
return err
|
||||
}
|
||||
|
||||
// InjectInboundDirect makes the Wrapper device behave as if a packet
|
||||
|
|
|
@ -882,7 +882,10 @@ func TestCaptureHook(t *testing.T) {
|
|||
packetBuf := stack.NewPacketBuffer(stack.PacketBufferOptions{
|
||||
Payload: buffer.MakeWithData([]byte("InjectInboundPacketBuffer")),
|
||||
})
|
||||
w.InjectInboundPacketBuffer(packetBuf)
|
||||
buffs := make([][]byte, 1)
|
||||
buffs[0] = make([]byte, PacketStartOffset+packetBuf.Size())
|
||||
sizes := make([]int, 1)
|
||||
w.InjectInboundPacketBuffer(packetBuf, buffs, sizes)
|
||||
|
||||
packetBuf = stack.NewPacketBuffer(stack.PacketBufferOptions{
|
||||
Payload: buffer.MakeWithData([]byte("InjectOutboundPacketBuffer")),
|
||||
|
|
|
@ -19,6 +19,7 @@ import (
|
|||
"sync/atomic"
|
||||
"time"
|
||||
|
||||
"github.com/tailscale/wireguard-go/conn"
|
||||
"gvisor.dev/gvisor/pkg/refs"
|
||||
"gvisor.dev/gvisor/pkg/tcpip"
|
||||
"gvisor.dev/gvisor/pkg/tcpip/adapters/gonet"
|
||||
|
@ -819,9 +820,27 @@ func (ns *Impl) DialContextUDP(ctx context.Context, ipp netip.AddrPort) (*gonet.
|
|||
return gonet.DialUDP(ns.ipstack, nil, remoteAddress, ipType)
|
||||
}
|
||||
|
||||
// getInjectInboundBuffsSizes returns packet memory and a sizes slice for usage
|
||||
// when calling tstun.Wrapper.InjectInboundPacketBuffer(). These are sized with
|
||||
// consideration for MTU and GSO support on ns.linkEP. They should be recycled
|
||||
// across subsequent inbound packet injection calls.
|
||||
func (ns *Impl) getInjectInboundBuffsSizes() (buffs [][]byte, sizes []int) {
|
||||
batchSize := 1
|
||||
if ns.linkEP.SupportedGSO() == stack.HostGSOSupported {
|
||||
batchSize = conn.IdealBatchSize
|
||||
}
|
||||
buffs = make([][]byte, batchSize)
|
||||
sizes = make([]int, batchSize)
|
||||
for i := 0; i < batchSize; i++ {
|
||||
buffs[i] = make([]byte, tstun.PacketStartOffset+tstun.DefaultTUNMTU())
|
||||
}
|
||||
return buffs, sizes
|
||||
}
|
||||
|
||||
// The inject goroutine reads in packets that netstack generated, and delivers
|
||||
// them to the correct path.
|
||||
func (ns *Impl) inject() {
|
||||
inboundBuffs, inboundBuffsSizes := ns.getInjectInboundBuffsSizes()
|
||||
for {
|
||||
pkt := ns.linkEP.ReadContext(ns.ctx)
|
||||
if pkt == nil {
|
||||
|
@ -847,7 +866,7 @@ func (ns *Impl) inject() {
|
|||
// pkt has a non-zero refcount, so injection methods takes
|
||||
// ownership of one count and will decrement on completion.
|
||||
if sendToHost {
|
||||
if err := ns.tundev.InjectInboundPacketBuffer(pkt); err != nil {
|
||||
if err := ns.tundev.InjectInboundPacketBuffer(pkt, inboundBuffs, inboundBuffsSizes); err != nil {
|
||||
ns.logf("netstack inject inbound: %v", err)
|
||||
return
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue