netdev.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* 3.6 routing cache regression, multicast loopback broken
@ 2012-11-05 16:37 Maxime Bizon
  2012-11-21  0:35 ` Julian Anastasov
  0 siblings, 1 reply; 4+ messages in thread
From: Maxime Bizon @ 2012-11-05 16:37 UTC (permalink / raw)
  To: David Miller; +Cc: netdev


Hi David & all,


kernel 3.6 to 3.6.5 (3.5 is working fine)

I have a "sender" sample app that does:

  - socket(dgram)
  - setsockopt mcast ttl 8
  - setsockopt mcast loopback
  - sendto() to 239.0.0.x

and a "receiver" sample app:

  - socket(dgram)
  - bind(239.0.0.x)
  - add membership (239.0.0.x)
  - loop on recv()


My setup: no default route, "ip route add 239.0.0.0/8 dev eth0", sender
& receiver running on same host.

If I first start one sender app (on 239.0.0.1), and after a receiver app
on 239.0.0.1 => no problem

If I first start two or more sender apps (239.0.0.1/239.0.0.2/...), then
assuming I don't start as many matching receivers, receivers sometimes
get all data or nothing at all.


After digging in __mkroute_output(), I found that unless I'm using a
default route (fi != NULL), rtable is cached and shared by all senders
(even those with different mcast addresses)

Since ip_check_mc_rcu() returns different results depending on whether
fl->daddr is present in device mc_list or not, I don't think we can
cache this.

The random working/not working effect I get is because add_membership
flushes the rt_cache, so depending on which sender does sendto() first
after flush, the cached entry will either use ip_mc_output() or
ip_output().

-- 
Maxime

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2012-11-22 19:28 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2012-11-05 16:37 3.6 routing cache regression, multicast loopback broken Maxime Bizon
2012-11-21  0:35 ` Julian Anastasov
2012-11-21 17:25   ` David Miller
2012-11-22 14:04   ` Maxime Bizon

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).