Skip site navigation (1)Skip section navigation (2)
Date:      Sun, 15 May 2011 11:16:03 -0400
From:      John Baldwin <jhb@FreeBSD.org>
To:        Andriy Gapon <avg@FreeBSD.org>
Cc:        Max Laier <max@love2party.net>, FreeBSD current <freebsd-current@FreeBSD.org>, neel@FreeBSD.org, Peter Grehan <grehan@FreeBSD.org>
Subject:   Re: proposed smp_rendezvous change
Message-ID:  <4DCFEE33.5090808@FreeBSD.org>
In-Reply-To: <4DCFE8FA.6080005@FreeBSD.org>
References:  <4DCD357D.6000109@FreeBSD.org> <4DCE9EF0.3050803@FreeBSD.org> <4DCF7CF0.1080508@FreeBSD.org> <4DCFE8FA.6080005@FreeBSD.org>

next in thread | previous in thread | raw e-mail | index | archive | help
On 5/15/11 10:53 AM, Andriy Gapon wrote:
> on 15/05/2011 10:12 Andriy Gapon said the following:
>> on 14/05/2011 18:25 John Baldwin said the following:
>>> Hmmm, so this is not actually sufficient.  NetApp ran into a very similar race
>>> with virtual CPUs in BHyVe.  In their case because virtual CPUs are threads that
>>> can be preempted, they have a chance at a longer race.
>>>
>>> The problem that they see is that even though the values have been updated, the
>>> next CPU to start a rendezvous can clear smp_rv_waiters[2] to zero before one of
>>> the other CPUs notices that it has finished.
>>
>> As a follow up to my previous question.  Have you noticed that in my patch no
>> slave CPU actually waits/spins on smp_rv_waiters[2]?  It's always only master
>> CPU (and under smp_ipi_mtx).
>>
>
> Here's a cleaner version of my approach to the fix.
> This one does not remove the initial wait on smp_rv_waiters[0] in
> smp_rendezvous_action() and thus does not renumber all smp_rv_waiters[] members
> and thus hopefully should be clearer.
>
> Index: sys/kern/subr_smp.c
> ===================================================================
> --- sys/kern/subr_smp.c	(revision 221943)
> +++ sys/kern/subr_smp.c	(working copy)
> @@ -110,7 +110,7 @@ static void (*volatile smp_rv_setup_func)(void *ar
>   static void (*volatile smp_rv_action_func)(void *arg);
>   static void (*volatile smp_rv_teardown_func)(void *arg);
>   static void *volatile smp_rv_func_arg;
> -static volatile int smp_rv_waiters[3];
> +static volatile int smp_rv_waiters[4];
>
>   /*
>    * Shared mutex to restrict busywaits between smp_rendezvous() and
> @@ -338,11 +338,15 @@ smp_rendezvous_action(void)
>
>   	/* spin on exit rendezvous */
>   	atomic_add_int(&smp_rv_waiters[2], 1);
> -	if (local_teardown_func == smp_no_rendevous_barrier)
> +	if (local_teardown_func == smp_no_rendevous_barrier) {
> +		atomic_add_int(&smp_rv_waiters[3], 1);
>                   return;
> +	}
>   	while (smp_rv_waiters[2]<  smp_rv_ncpus)
>   		cpu_spinwait();
>
> +	atomic_add_int(&smp_rv_waiters[3], 1);
> +
>   	/* teardown function */
>   	if (local_teardown_func != NULL)
>   		local_teardown_func(local_func_arg);
> @@ -377,6 +381,9 @@ smp_rendezvous_cpus(cpumask_t map,
>   	/* obtain rendezvous lock */
>   	mtx_lock_spin(&smp_ipi_mtx);
>
> +	while (smp_rv_waiters[3]<  smp_rv_ncpus)
> +		cpu_spinwait();
> +
>   	/* set static function pointers */
>   	smp_rv_ncpus = ncpus;
>   	smp_rv_setup_func = setup_func;
> @@ -385,6 +392,7 @@ smp_rendezvous_cpus(cpumask_t map,
>   	smp_rv_func_arg = arg;
>   	smp_rv_waiters[1] = 0;
>   	smp_rv_waiters[2] = 0;
> +	smp_rv_waiters[3] = 0;
>   	atomic_store_rel_int(&smp_rv_waiters[0], 0);
>
>   	/* signal other processors, which will enter the IPI with interrupts off */

Ahh, so the bump is after the change.  I do think this will still be ok
and I probably just didn't explain it well to Neel.  I wonder though
if the bump shouldn't happen until after the call of the local teardown
function?

-- 
John Baldwin



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?4DCFEE33.5090808>