Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit c85bb41e authored by Flavio Leitner's avatar Flavio Leitner Committed by David S. Miller
Browse files

igmp: fix ip_mc_sf_allow race [v5]



Almost all igmp functions accessing inet->mc_list are protected by
rtnl_lock(), but there is one exception which is ip_mc_sf_allow(),
so there is a chance of either ip_mc_drop_socket or ip_mc_leave_group
remove an entry while ip_mc_sf_allow is running causing a crash.

Signed-off-by: default avatarFlavio Leitner <fleitner@redhat.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 8b64056d
Loading
Loading
Loading
Loading
+2 −0
Original line number Original line Diff line number Diff line
@@ -153,6 +153,7 @@ extern int sysctl_igmp_max_msf;
struct ip_sf_socklist {
struct ip_sf_socklist {
	unsigned int		sl_max;
	unsigned int		sl_max;
	unsigned int		sl_count;
	unsigned int		sl_count;
	struct rcu_head		rcu;
	__be32			sl_addr[0];
	__be32			sl_addr[0];
};
};


@@ -170,6 +171,7 @@ struct ip_mc_socklist {
	struct ip_mreqn		multi;
	struct ip_mreqn		multi;
	unsigned int		sfmode;		/* MCAST_{INCLUDE,EXCLUDE} */
	unsigned int		sfmode;		/* MCAST_{INCLUDE,EXCLUDE} */
	struct ip_sf_socklist	*sflist;
	struct ip_sf_socklist	*sflist;
	struct rcu_head		rcu;
};
};


struct ip_sf_list {
struct ip_sf_list {
+62 −21
Original line number Original line Diff line number Diff line
@@ -1799,7 +1799,7 @@ int ip_mc_join_group(struct sock *sk , struct ip_mreqn *imr)
	iml->next = inet->mc_list;
	iml->next = inet->mc_list;
	iml->sflist = NULL;
	iml->sflist = NULL;
	iml->sfmode = MCAST_EXCLUDE;
	iml->sfmode = MCAST_EXCLUDE;
	inet->mc_list = iml;
	rcu_assign_pointer(inet->mc_list, iml);
	ip_mc_inc_group(in_dev, addr);
	ip_mc_inc_group(in_dev, addr);
	err = 0;
	err = 0;
done:
done:
@@ -1807,24 +1807,46 @@ int ip_mc_join_group(struct sock *sk , struct ip_mreqn *imr)
	return err;
	return err;
}
}


static void ip_sf_socklist_reclaim(struct rcu_head *rp)
{
	struct ip_sf_socklist *psf;

	psf = container_of(rp, struct ip_sf_socklist, rcu);
	/* sk_omem_alloc should have been decreased by the caller*/
	kfree(psf);
}

static int ip_mc_leave_src(struct sock *sk, struct ip_mc_socklist *iml,
static int ip_mc_leave_src(struct sock *sk, struct ip_mc_socklist *iml,
			   struct in_device *in_dev)
			   struct in_device *in_dev)
{
{
	struct ip_sf_socklist *psf = iml->sflist;
	int err;
	int err;


	if (iml->sflist == NULL) {
	if (psf == NULL) {
		/* any-source empty exclude case */
		/* any-source empty exclude case */
		return ip_mc_del_src(in_dev, &iml->multi.imr_multiaddr.s_addr,
		return ip_mc_del_src(in_dev, &iml->multi.imr_multiaddr.s_addr,
			iml->sfmode, 0, NULL, 0);
			iml->sfmode, 0, NULL, 0);
	}
	}
	err = ip_mc_del_src(in_dev, &iml->multi.imr_multiaddr.s_addr,
	err = ip_mc_del_src(in_dev, &iml->multi.imr_multiaddr.s_addr,
			iml->sfmode, iml->sflist->sl_count,
			iml->sfmode, psf->sl_count, psf->sl_addr, 0);
			iml->sflist->sl_addr, 0);
	rcu_assign_pointer(iml->sflist, NULL);
	sock_kfree_s(sk, iml->sflist, IP_SFLSIZE(iml->sflist->sl_max));
	/* decrease mem now to avoid the memleak warning */
	iml->sflist = NULL;
	atomic_sub(IP_SFLSIZE(psf->sl_max), &sk->sk_omem_alloc);
	call_rcu(&psf->rcu, ip_sf_socklist_reclaim);
	return err;
	return err;
}
}



static void ip_mc_socklist_reclaim(struct rcu_head *rp)
{
	struct ip_mc_socklist *iml;

	iml = container_of(rp, struct ip_mc_socklist, rcu);
	/* sk_omem_alloc should have been decreased by the caller*/
	kfree(iml);
}


/*
/*
 *	Ask a socket to leave a group.
 *	Ask a socket to leave a group.
 */
 */
@@ -1854,12 +1876,14 @@ int ip_mc_leave_group(struct sock *sk, struct ip_mreqn *imr)


		(void) ip_mc_leave_src(sk, iml, in_dev);
		(void) ip_mc_leave_src(sk, iml, in_dev);


		*imlp = iml->next;
		rcu_assign_pointer(*imlp, iml->next);


		if (in_dev)
		if (in_dev)
			ip_mc_dec_group(in_dev, group);
			ip_mc_dec_group(in_dev, group);
		rtnl_unlock();
		rtnl_unlock();
		sock_kfree_s(sk, iml, sizeof(*iml));
		/* decrease mem now to avoid the memleak warning */
		atomic_sub(sizeof(*iml), &sk->sk_omem_alloc);
		call_rcu(&iml->rcu, ip_mc_socklist_reclaim);
		return 0;
		return 0;
	}
	}
	if (!in_dev)
	if (!in_dev)
@@ -1974,9 +1998,12 @@ int ip_mc_source(int add, int omode, struct sock *sk, struct
		if (psl) {
		if (psl) {
			for (i=0; i<psl->sl_count; i++)
			for (i=0; i<psl->sl_count; i++)
				newpsl->sl_addr[i] = psl->sl_addr[i];
				newpsl->sl_addr[i] = psl->sl_addr[i];
			sock_kfree_s(sk, psl, IP_SFLSIZE(psl->sl_max));
			/* decrease mem now to avoid the memleak warning */
			atomic_sub(IP_SFLSIZE(psl->sl_max), &sk->sk_omem_alloc);
			call_rcu(&psl->rcu, ip_sf_socklist_reclaim);
		}
		}
		pmc->sflist = psl = newpsl;
		rcu_assign_pointer(pmc->sflist, newpsl);
		psl = newpsl;
	}
	}
	rv = 1;	/* > 0 for insert logic below if sl_count is 0 */
	rv = 1;	/* > 0 for insert logic below if sl_count is 0 */
	for (i=0; i<psl->sl_count; i++) {
	for (i=0; i<psl->sl_count; i++) {
@@ -2072,11 +2099,13 @@ int ip_mc_msfilter(struct sock *sk, struct ip_msfilter *msf, int ifindex)
	if (psl) {
	if (psl) {
		(void) ip_mc_del_src(in_dev, &msf->imsf_multiaddr, pmc->sfmode,
		(void) ip_mc_del_src(in_dev, &msf->imsf_multiaddr, pmc->sfmode,
			psl->sl_count, psl->sl_addr, 0);
			psl->sl_count, psl->sl_addr, 0);
		sock_kfree_s(sk, psl, IP_SFLSIZE(psl->sl_max));
		/* decrease mem now to avoid the memleak warning */
		atomic_sub(IP_SFLSIZE(psl->sl_max), &sk->sk_omem_alloc);
		call_rcu(&psl->rcu, ip_sf_socklist_reclaim);
	} else
	} else
		(void) ip_mc_del_src(in_dev, &msf->imsf_multiaddr, pmc->sfmode,
		(void) ip_mc_del_src(in_dev, &msf->imsf_multiaddr, pmc->sfmode,
			0, NULL, 0);
			0, NULL, 0);
	pmc->sflist = newpsl;
	rcu_assign_pointer(pmc->sflist, newpsl);
	pmc->sfmode = msf->imsf_fmode;
	pmc->sfmode = msf->imsf_fmode;
	err = 0;
	err = 0;
done:
done:
@@ -2209,30 +2238,40 @@ int ip_mc_sf_allow(struct sock *sk, __be32 loc_addr, __be32 rmt_addr, int dif)
	struct ip_mc_socklist *pmc;
	struct ip_mc_socklist *pmc;
	struct ip_sf_socklist *psl;
	struct ip_sf_socklist *psl;
	int i;
	int i;
	int ret;


	ret = 1;
	if (!ipv4_is_multicast(loc_addr))
	if (!ipv4_is_multicast(loc_addr))
		return 1;
		goto out;


	for (pmc=inet->mc_list; pmc; pmc=pmc->next) {
	rcu_read_lock();
	for (pmc=rcu_dereference(inet->mc_list); pmc; pmc=rcu_dereference(pmc->next)) {
		if (pmc->multi.imr_multiaddr.s_addr == loc_addr &&
		if (pmc->multi.imr_multiaddr.s_addr == loc_addr &&
		    pmc->multi.imr_ifindex == dif)
		    pmc->multi.imr_ifindex == dif)
			break;
			break;
	}
	}
	ret = inet->mc_all;
	if (!pmc)
	if (!pmc)
		return inet->mc_all;
		goto unlock;
	psl = pmc->sflist;
	psl = pmc->sflist;
	ret = (pmc->sfmode == MCAST_EXCLUDE);
	if (!psl)
	if (!psl)
		return pmc->sfmode == MCAST_EXCLUDE;
		goto unlock;


	for (i=0; i<psl->sl_count; i++) {
	for (i=0; i<psl->sl_count; i++) {
		if (psl->sl_addr[i] == rmt_addr)
		if (psl->sl_addr[i] == rmt_addr)
			break;
			break;
	}
	}
	ret = 0;
	if (pmc->sfmode == MCAST_INCLUDE && i >= psl->sl_count)
	if (pmc->sfmode == MCAST_INCLUDE && i >= psl->sl_count)
		return 0;
		goto unlock;
	if (pmc->sfmode == MCAST_EXCLUDE && i < psl->sl_count)
	if (pmc->sfmode == MCAST_EXCLUDE && i < psl->sl_count)
		return 0;
		goto unlock;
	return 1;
	ret = 1;
unlock:
	rcu_read_unlock();
out:
	return ret;
}
}


/*
/*
@@ -2251,7 +2290,7 @@ void ip_mc_drop_socket(struct sock *sk)
	rtnl_lock();
	rtnl_lock();
	while ((iml = inet->mc_list) != NULL) {
	while ((iml = inet->mc_list) != NULL) {
		struct in_device *in_dev;
		struct in_device *in_dev;
		inet->mc_list = iml->next;
		rcu_assign_pointer(inet->mc_list, iml->next);


		in_dev = inetdev_by_index(net, iml->multi.imr_ifindex);
		in_dev = inetdev_by_index(net, iml->multi.imr_ifindex);
		(void) ip_mc_leave_src(sk, iml, in_dev);
		(void) ip_mc_leave_src(sk, iml, in_dev);
@@ -2259,7 +2298,9 @@ void ip_mc_drop_socket(struct sock *sk)
			ip_mc_dec_group(in_dev, iml->multi.imr_multiaddr.s_addr);
			ip_mc_dec_group(in_dev, iml->multi.imr_multiaddr.s_addr);
			in_dev_put(in_dev);
			in_dev_put(in_dev);
		}
		}
		sock_kfree_s(sk, iml, sizeof(*iml));
		/* decrease mem now to avoid the memleak warning */
		atomic_sub(sizeof(*iml), &sk->sk_omem_alloc);
		call_rcu(&iml->rcu, ip_mc_socklist_reclaim);
	}
	}
	rtnl_unlock();
	rtnl_unlock();
}
}