Message ID | f6116069b730c0c3a74ff627fa818b98dc4f1491.1395379422.git.viresh.kumar@linaro.org |
---|---|
State | New |
Headers | show |
On 03/21/2014 11:04 AM, Viresh Kumar wrote: > From: "Srivatsa S. Bhat" <srivatsa.bhat@linux.vnet.ibm.com> > > Whenever we change the frequency of a CPU, we call the PRECHANGE and POSTCHANGE > notifiers. They must be serialized, i.e. PRECHANGE and POSTCHANGE notifiers > should strictly alternate, thereby preventing two different sets of PRECHANGE or > POSTCHANGE notifiers from interleaving arbitrarily. > > The following examples illustrate why this is important: > > Scenario 1: > ----------- > A thread reading the value of cpuinfo_cur_freq, will call > __cpufreq_cpu_get()->cpufreq_out_of_sync()->cpufreq_notify_transition() > > The ondemand governor can decide to change the frequency of the CPU at the same > time and hence it can end up sending the notifications via ->target(). > > If the notifiers are not serialized, the following sequence can occur: > - PRECHANGE Notification for freq A (from cpuinfo_cur_freq) > - PRECHANGE Notification for freq B (from target()) > - Freq changed by target() to B > - POSTCHANGE Notification for freq B > - POSTCHANGE Notification for freq A > > We can see from the above that the last POSTCHANGE Notification happens for freq > A but the hardware is set to run at freq B. > > Where would we break then?: adjust_jiffies() in cpufreq.c & cpufreq_callback() > in arch/arm/kernel/smp.c (which also adjusts the jiffies). All the > loops_per_jiffy calculations will get messed up. > > Scenario 2: > ----------- > The governor calls __cpufreq_driver_target() to change the frequency. At the > same time, if we change scaling_{min|max}_freq from sysfs, it will end up > calling the governor's CPUFREQ_GOV_LIMITS notification, which will also call > __cpufreq_driver_target(). And hence we end up issuing concurrent calls to > ->target(). > > Typically, platforms have the following logic in their ->target() routines: > (Eg: cpufreq-cpu0, omap, exynos, etc) > > A. If new freq is more than old: Increase voltage > B. Change freq > C. If new freq is less than old: decrease voltage > > Now, if the two concurrent calls to ->target() are X and Y, where X is trying to > increase the freq and Y is trying to decrease it, we get the following race > condition: > > X.A: voltage gets increased for larger freq > Y.A: nothing happens > Y.B: freq gets decreased > Y.C: voltage gets decreased > X.B: freq gets increased > X.C: nothing happens > > Thus we can end up setting a freq which is not supported by the voltage we have > set. That will probably make the clock to the CPU unstable and the system might > not work properly anymore. > > This patch introduces a set of synchronization primitives to serialize frequency > transitions, which are to be used as shown below: > > cpufreq_freq_transition_begin(); > > //Perform the frequency change > > cpufreq_freq_transition_end(); > > The _begin() call sends the PRECHANGE notification whereas the _end() call sends > the POSTCHANGE notification. Also, all the necessary synchronization is handled > within these calls. In particular, even drivers which set the ASYNC_NOTIFICATION > flag can also use these APIs for performing frequency transitions (ie., you can > call _begin() from one task, and call the corresponding _end() from a different > task). > > The actual synchronization underneath is not that complicated: > > The key challenge is to allow drivers to begin the transition from one thread > and end it in a completely different thread (this is to enable drivers that do > asynchronous POSTCHANGE notification from bottom-halves, to also use the same > interface). > > To achieve this, a 'transition_ongoing' flag, a 'transition_lock' spinlock and a > wait-queue are added per-policy. The flag and the wait-queue are used in > conjunction to create an "uninterrupted flow" from _begin() to _end(). The > spinlock is used to ensure that only one such "flow" is in flight at any given > time. Put together, this provides us all the necessary synchronization. > > Based-on-patch-by: Viresh Kumar <viresh.kumar@linaro.org> > Signed-off-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com> > Signed-off-by: Viresh Kumar <viresh.kumar@linaro.org> > --- > > I have kept your Authorship for this patch as is and did few modifications: > - removed 'state' parameter from begin/end routines. > - added 'trasition_failed' parameter to end routine. > - changed mutex with spinlock as discussed earlier. > - Added WARN_ON() as discussed. > - Exported these new routines. > - Removed locks from end. > Wonderful! I was going to do this today, but thanks a lot for taking care of this for me! The patch looks good, but I have one comment below. > drivers/cpufreq/cpufreq.c | 37 +++++++++++++++++++++++++++++++++++++ > include/linux/cpufreq.h | 10 ++++++++++ > 2 files changed, 47 insertions(+) > > diff --git a/drivers/cpufreq/cpufreq.c b/drivers/cpufreq/cpufreq.c > index b349406..4279cc9 100644 > --- a/drivers/cpufreq/cpufreq.c > +++ b/drivers/cpufreq/cpufreq.c > @@ -353,6 +353,41 @@ void cpufreq_notify_post_transition(struct cpufreq_policy *policy, > } > EXPORT_SYMBOL_GPL(cpufreq_notify_post_transition); > > +void cpufreq_freq_transition_begin(struct cpufreq_policy *policy, > + struct cpufreq_freqs *freqs) > +{ > +wait: > + wait_event(policy->transition_wait, !policy->transition_ongoing); > + > + spin_lock(&policy->transition_lock); > + > + if (unlikely(policy->transition_ongoing)) { > + spin_unlock(&policy->transition_lock); > + goto wait; > + } > + > + policy->transition_ongoing = true; > + > + spin_unlock(&policy->transition_lock); > + > + cpufreq_notify_transition(policy, freqs, CPUFREQ_PRECHANGE); > +} > +EXPORT_SYMBOL_GPL(cpufreq_freq_transition_begin); > + > +void cpufreq_freq_transition_end(struct cpufreq_policy *policy, > + struct cpufreq_freqs *freqs, int transition_failed) > +{ > + if (unlikely(WARN_ON(!policy->transition_ongoing))) > + return; > + > + cpufreq_notify_post_transition(policy, freqs, transition_failed); > + > + policy->transition_ongoing = false; We need this assignment to happen exactly at this point, that is, *after* the call to post_transition() completes and before calling wake_up(). If the compiler or the CPU reorders the instructions and moves this assignment to some other place, then we will be in trouble! We might need memory barriers to ensure this doesn't get reordered. Alternatively, we can simply hold the spin-lock around this assignment, since locks automatically imply memory barriers. As an added advantage, the code will then look more intuitive and easier to understand as well. Thoughts? Regards, Srivatsa S. Bhat > + > + wake_up(&policy->transition_wait); > +} > +EXPORT_SYMBOL_GPL(cpufreq_freq_transition_end); > + > > /********************************************************************* > * SYSFS INTERFACE * > @@ -982,6 +1017,8 @@ static struct cpufreq_policy *cpufreq_policy_alloc(void) > > INIT_LIST_HEAD(&policy->policy_list); > init_rwsem(&policy->rwsem); > + spin_lock_init(&policy->transition_lock); > + init_waitqueue_head(&policy->transition_wait); > > return policy; > > diff --git a/include/linux/cpufreq.h b/include/linux/cpufreq.h > index 70929bc..263173d 100644 > --- a/include/linux/cpufreq.h > +++ b/include/linux/cpufreq.h > @@ -16,6 +16,7 @@ > #include <linux/completion.h> > #include <linux/kobject.h> > #include <linux/notifier.h> > +#include <linux/spinlock.h> > #include <linux/sysfs.h> > > /********************************************************************* > @@ -104,6 +105,11 @@ struct cpufreq_policy { > * __cpufreq_governor(data, CPUFREQ_GOV_POLICY_EXIT); > */ > struct rw_semaphore rwsem; > + > + /* Synchronization for frequency transitions */ > + bool transition_ongoing; /* Tracks transition status */ > + spinlock_t transition_lock; > + wait_queue_head_t transition_wait; > }; > > /* Only for ACPI */ > @@ -336,6 +342,10 @@ void cpufreq_notify_transition(struct cpufreq_policy *policy, > struct cpufreq_freqs *freqs, unsigned int state); > void cpufreq_notify_post_transition(struct cpufreq_policy *policy, > struct cpufreq_freqs *freqs, int transition_failed); > +void cpufreq_freq_transition_begin(struct cpufreq_policy *policy, > + struct cpufreq_freqs *freqs); > +void cpufreq_freq_transition_end(struct cpufreq_policy *policy, > + struct cpufreq_freqs *freqs, int transition_failed); > > #else /* CONFIG_CPU_FREQ */ > static inline int cpufreq_register_notifier(struct notifier_block *nb, > -- To unsubscribe from this list: send the line "unsubscribe linux-pm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 21 March 2014 13:16, Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com> wrote: > Wonderful! I was going to do this today, but thanks a lot for taking > care of this for me! I just wanted to finish this long lasting thread as soon as possible. > We need this assignment to happen exactly at this point, that is, *after* > the call to post_transition() completes and before calling wake_up(). > > If the compiler or the CPU reorders the instructions and moves this > assignment to some other place, then we will be in trouble! > > We might need memory barriers to ensure this doesn't get reordered. > Alternatively, we can simply hold the spin-lock around this assignment, > since locks automatically imply memory barriers. As an added advantage, > the code will then look more intuitive and easier to understand as well. > > Thoughts? I may be wrong but this is how I understand locks. Yes, spinlocks have memory barriers implemented but it wouldn't guarantee what you are asking for in the above explanation. It will guarantee that transition_ongoing will be updated after the lock is taken but the notification() can happen after the lock is taken and also after the variable is modified. You can find some information on this in Documentation/memory-barriers.txt I don't think compiler or CPU will reorder calls to a function and updates of a variable. And so this code might simply work. And I hope there would be plenty of such code in kernel. -- To unsubscribe from this list: send the line "unsubscribe linux-pm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 03/21/2014 01:28 PM, Viresh Kumar wrote: > On 21 March 2014 13:16, Srivatsa S. Bhat > <srivatsa.bhat@linux.vnet.ibm.com> wrote: >> We need this assignment to happen exactly at this point, that is, *after* >> the call to post_transition() completes and before calling wake_up(). >> >> If the compiler or the CPU reorders the instructions and moves this >> assignment to some other place, then we will be in trouble! >> >> We might need memory barriers to ensure this doesn't get reordered. >> Alternatively, we can simply hold the spin-lock around this assignment, >> since locks automatically imply memory barriers. As an added advantage, >> the code will then look more intuitive and easier to understand as well. >> >> Thoughts? > > I may be wrong but this is how I understand locks. Yes, spinlocks have > memory barriers implemented but it wouldn't guarantee what you are > asking for in the above explanation. > > It will guarantee that transition_ongoing will be updated after the lock > is taken but the notification() can happen after the lock is taken and > also after the variable is modified. > Actually, yes, that's true. The lock and unlock act as one-way barriers, hence they ensure that the critical section doesn't seep outside of the locks, but don't necessarily ensure that pieces of code outside the critical section don't seep -into- the critical section. IOW, my reasoning was not quite correct, but see below. > You can find some information on this in > Documentation/memory-barriers.txt > Yep, I know, I have read it several times, but I'm no expert ;-) I found this interesting section on "SLEEP AND WAKE-UP FUNCTIONS". It says that doing: policy->transition_ongoing = false; wake_up(&policy->transition_wait); is safe (as long as some tasks are woken up). So we don't have to worry about that part. So only the first part remains to be solved: ensuring that the assignment occurs _after_ completing the invocation of the POSTCHANGE notifiers. For that, we can do: cpufreq_notify_post_transition(); smp_mb(); policy->transition_ongoing = false; That should take care of everything. > I don't think compiler or CPU will reorder calls to a function and > updates of a variable. I'm not sure about that. I think it is free to do so if it finds that there is no dependency that prevents it from reordering. In this case the update to the flag has no "visible" dependency on the call to post_transition(). > And so this code might simply work. And > I hope there would be plenty of such code in kernel. > Sure, there are plenty of examples in the kernel where we call functions and update variables. But in this particular case, our synchronization _depends_ on those operations happening in a particular order. Hence we need to ensure the ordering is right. Otherwise the synchronization might get broken. Here are some examples where memory barriers are inserted to avoid reordering of variable updates and function calls: kernel/rcu/torture.c: rcu_torture_barrier_cbs() kernel/smp.c: kick_all_cpus_sync() Regards, Srivatsa S. Bhat -- To unsubscribe from this list: send the line "unsubscribe cpufreq" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 21 March 2014 14:12, Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com> wrote: >> You can find some information on this in >> Documentation/memory-barriers.txt > > Yep, I know, I have read it several times, but I'm no expert ;-) Not me either :) .. That file has so complex stuff in there that its difficult to understand what's all it says.. I read it several times the last time I went for a interview (Almost 2 years back) and don't remember anything now :) > I found this interesting section on "SLEEP AND WAKE-UP FUNCTIONS". It > says that doing: > > policy->transition_ongoing = false; > wake_up(&policy->transition_wait); > > is safe (as long as some tasks are woken up). So we don't have to worry > about that part. Okay.. > So only the first part remains to be solved: ensuring > that the assignment occurs _after_ completing the invocation of the > POSTCHANGE notifiers. > > For that, we can do: > > cpufreq_notify_post_transition(); > > smp_mb(); > > policy->transition_ongoing = false; > > That should take care of everything. > >> I don't think compiler or CPU will reorder calls to a function and >> updates of a variable. > > I'm not sure about that. I think it is free to do so if it finds > that there is no dependency that prevents it from reordering. In this > case the update to the flag has no "visible" dependency on the call > to post_transition(). > >> And so this code might simply work. And >> I hope there would be plenty of such code in kernel. >> > > Sure, there are plenty of examples in the kernel where we call functions > and update variables. But in this particular case, our synchronization > _depends_ on those operations happening in a particular order. Hence > we need to ensure the ordering is right. Otherwise the synchronization > might get broken. I still don't buy that.. Lets call an expert :) > Here are some examples where memory barriers are inserted to avoid > reordering of variable updates and function calls: > > kernel/rcu/torture.c: rcu_torture_barrier_cbs() rcutorture.c instead. > kernel/smp.c: kick_all_cpus_sync() These examples are a bit different than what we have here.. @Catalin: We have a problem here and need your expert advice. After changing CPU frequency we need to call this code: cpufreq_notify_post_transition(); policy->transition_ongoing = false; And the sequence must be like this only. Is this guaranteed without any memory barriers? cpufreq_notify_post_transition() isn't touching transition_ongoing at all.. -- thanks.. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
On 21 March 2014 14:51, Viresh Kumar <viresh.kumar@linaro.org> wrote: > @Catalin: We have a problem here and need your expert advice. After changing > CPU frequency we need to call this code: > > cpufreq_notify_post_transition(); > policy->transition_ongoing = false; > > And the sequence must be like this only. Is this guaranteed without any > memory barriers? cpufreq_notify_post_transition() isn't touching > transition_ongoing at all.. For others this is what we discussed on IRC (rmk: Russell King) <rmk> I'm no barrier expert, but the compiler can't reorder that assignment across a function call which it knows nothing about (which it can't know anything about because it calls other functions through function pointers) <rmk> however, the CPU could re-order the effects with respect to other agents (cpus/devices) when they look at the memory <rmk> for the local CPU, the question is really: what does the C language virtual machine say about this - that's what really matters. If the CPU does speculative stuff, it still has to make the machine behaviour fit that model. -- To unsubscribe from this list: send the line "unsubscribe linux-pm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Fri, Mar 21, 2014 at 09:21:02AM +0000, Viresh Kumar wrote: > @Catalin: We have a problem here and need your expert advice. After changing > CPU frequency we need to call this code: > > cpufreq_notify_post_transition(); > policy->transition_ongoing = false; > > And the sequence must be like this only. Is this guaranteed without any > memory barriers? cpufreq_notify_post_transition() isn't touching > transition_ongoing at all.. The above sequence doesn't say much. As rmk said, the compiler wouldn't reorder the transition_ongoing write before the function call. I think most architectures (not sure about Alpha) don't do speculative stores, so hardware wouldn't reorder them either. However, other stores inside the cpufreq_notify_post_transition() could be reordered after transition_ongoing store. The same for memory accesses after the transition_ongoing update, they could be reordered before. So what we actually need to know is what are the other relevant memory accesses that require strict ordering with transition_ongoing. What I find strange in your patch is that cpufreq_freq_transition_begin() uses spinlocks around transition_ongoing update but cpufreq_freq_transition_end() doesn't.
On 03/21/2014 04:35 PM, Catalin Marinas wrote: > On Fri, Mar 21, 2014 at 09:21:02AM +0000, Viresh Kumar wrote: >> @Catalin: We have a problem here and need your expert advice. After changing >> CPU frequency we need to call this code: >> >> cpufreq_notify_post_transition(); >> policy->transition_ongoing = false; >> >> And the sequence must be like this only. Is this guaranteed without any >> memory barriers? cpufreq_notify_post_transition() isn't touching >> transition_ongoing at all.. > > The above sequence doesn't say much. As rmk said, the compiler wouldn't > reorder the transition_ongoing write before the function call. I think > most architectures (not sure about Alpha) don't do speculative stores, > so hardware wouldn't reorder them either. However, other stores inside > the cpufreq_notify_post_transition() could be reordered after > transition_ongoing store. The same for memory accesses after the > transition_ongoing update, they could be reordered before. > > So what we actually need to know is what are the other relevant memory > accesses that require strict ordering with transition_ongoing. > Hmm.. The thing is, _everything_ inside the post_transition() function should complete before writing to transition_ongoing. Because, setting the flag to 'false' indicates the end of the critical section, and the next contending task can enter the critical section. So, I think we should use smp_mb() before setting transition_ongoing = false. That way we'll be safe. > What I find strange in your patch is that > cpufreq_freq_transition_begin() uses spinlocks around transition_ongoing > update but cpufreq_freq_transition_end() doesn't. > The reason is that, by the time we drop the spinlock, we would have set the transition_ongoing flag to true, which prevents any other task from entering the critical section. Hence, when we call the _end() function, we are 100% sure that only one task is executing it. Hence locks are not necessary around that second update. In fact, that very update marks the end of the critical section (which acts much like a spin_unlock(&lock) in a "regular" critical section). I know the "critical section" and the synchronization used in this patch is a bit unconventional, but that's because the scenario itself is unconventional : we need to able to start the critical section in one task, and end it in another task! That's where all the complication arises :-) It sounds weird, but in this cpufreq case, its actually valid and surprisingly, makes sense too! :-) Regards, Srivatsa S. Bhat -- To unsubscribe from this list: send the line "unsubscribe linux-pm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Fri, Mar 21, 2014 at 11:24:16AM +0000, Srivatsa S. Bhat wrote: > On 03/21/2014 04:35 PM, Catalin Marinas wrote: > > On Fri, Mar 21, 2014 at 09:21:02AM +0000, Viresh Kumar wrote: > >> @Catalin: We have a problem here and need your expert advice. After changing > >> CPU frequency we need to call this code: > >> > >> cpufreq_notify_post_transition(); > >> policy->transition_ongoing = false; > >> > >> And the sequence must be like this only. Is this guaranteed without any > >> memory barriers? cpufreq_notify_post_transition() isn't touching > >> transition_ongoing at all.. > > > > The above sequence doesn't say much. As rmk said, the compiler wouldn't > > reorder the transition_ongoing write before the function call. I think > > most architectures (not sure about Alpha) don't do speculative stores, > > so hardware wouldn't reorder them either. However, other stores inside > > the cpufreq_notify_post_transition() could be reordered after > > transition_ongoing store. The same for memory accesses after the > > transition_ongoing update, they could be reordered before. > > > > So what we actually need to know is what are the other relevant memory > > accesses that require strict ordering with transition_ongoing. > > Hmm.. The thing is, _everything_ inside the post_transition() function > should complete before writing to transition_ongoing. Because, setting the > flag to 'false' indicates the end of the critical section, and the next > contending task can enter the critical section. smp_mb() is all about relative ordering. So if you want memory accesses in post_transition() to be visible to other observers before transition_ongoing = false, you also need to make sure that the readers of transition_ongoing have a barrier before subsequent memory accesses. > > What I find strange in your patch is that > > cpufreq_freq_transition_begin() uses spinlocks around transition_ongoing > > update but cpufreq_freq_transition_end() doesn't. > > The reason is that, by the time we drop the spinlock, we would have set > the transition_ongoing flag to true, which prevents any other task from > entering the critical section. Hence, when we call the _end() function, > we are 100% sure that only one task is executing it. Hence locks are not > necessary around that second update. In fact, that very update marks the > end of the critical section (which acts much like a spin_unlock(&lock) > in a "regular" critical section). OK, I start to get it. Is there a risk of missing a wake_up event? E.g. one thread waking up earlier, noticing that transition is in progress and waiting indefinitely?
On 21 March 2014 23:37, Catalin Marinas <catalin.marinas@arm.com> wrote: > smp_mb() is all about relative ordering. So if you want memory accesses > in post_transition() to be visible to other observers before > transition_ongoing = false, you also need to make sure that the readers > of transition_ongoing have a barrier before subsequent memory accesses. I don't think this is a requirement in our case. We are just trying to serialize frequency transitions here and just want to make sure that second one start after first one is over. And so this query. > OK, I start to get it. Is there a risk of missing a wake_up event? E.g. > one thread waking up earlier, noticing that transition is in progress > and waiting indefinitely? I don't think so. The only requirement is that second thread wakes up after this variable is set to false. -- To unsubscribe from this list: send the line "unsubscribe cpufreq" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 21 March 2014 16:35, Catalin Marinas <catalin.marinas@arm.com> wrote: > The above sequence doesn't say much. As rmk said, the compiler wouldn't > reorder the transition_ongoing write before the function call. I think > most architectures (not sure about Alpha) don't do speculative stores, > so hardware wouldn't reorder them either. However, other stores inside > the cpufreq_notify_post_transition() could be reordered after > transition_ongoing store. The same for memory accesses after the > transition_ongoing update, they could be reordered before. I got confused again. If we see what cpufreq_notify_post_transition() does: Just calling a list of routines from a notifiers chain. And going by the above statements from you, we aren't going to reorder this with function calls or a branch instructions. And even if for some reason, there is a bit of reorder, it doesn't look harmless at all to me. We are more concerned about serialization of frequency translations here. And it still looks to me like we don't really need a barrier at all.. Probably we can keep it as is for now and maybe later add a barrier if required. -- To unsubscribe from this list: send the line "unsubscribe cpufreq" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 03/21/2014 11:37 PM, Catalin Marinas wrote: > On Fri, Mar 21, 2014 at 11:24:16AM +0000, Srivatsa S. Bhat wrote: >> On 03/21/2014 04:35 PM, Catalin Marinas wrote: >>> On Fri, Mar 21, 2014 at 09:21:02AM +0000, Viresh Kumar wrote: >>>> @Catalin: We have a problem here and need your expert advice. After changing >>>> CPU frequency we need to call this code: >>>> >>>> cpufreq_notify_post_transition(); >>>> policy->transition_ongoing = false; >>>> >>>> And the sequence must be like this only. Is this guaranteed without any >>>> memory barriers? cpufreq_notify_post_transition() isn't touching >>>> transition_ongoing at all.. >>> >>> The above sequence doesn't say much. As rmk said, the compiler wouldn't >>> reorder the transition_ongoing write before the function call. I think >>> most architectures (not sure about Alpha) don't do speculative stores, >>> so hardware wouldn't reorder them either. However, other stores inside >>> the cpufreq_notify_post_transition() could be reordered after >>> transition_ongoing store. The same for memory accesses after the >>> transition_ongoing update, they could be reordered before. >>> >>> So what we actually need to know is what are the other relevant memory >>> accesses that require strict ordering with transition_ongoing. >> >> Hmm.. The thing is, _everything_ inside the post_transition() function >> should complete before writing to transition_ongoing. Because, setting the >> flag to 'false' indicates the end of the critical section, and the next >> contending task can enter the critical section. > > smp_mb() is all about relative ordering. So if you want memory accesses > in post_transition() to be visible to other observers before > transition_ongoing = false, you also need to make sure that the readers > of transition_ongoing have a barrier before subsequent memory accesses. > The reader takes a spin-lock before reading the flag.. won't that suffice? +wait: + wait_event(policy->transition_wait, !policy->transition_ongoing); + + spin_lock(&policy->transition_lock); + + if (unlikely(policy->transition_ongoing)) { + spin_unlock(&policy->transition_lock); + goto wait; + } >>> What I find strange in your patch is that >>> cpufreq_freq_transition_begin() uses spinlocks around transition_ongoing >>> update but cpufreq_freq_transition_end() doesn't. >> >> The reason is that, by the time we drop the spinlock, we would have set >> the transition_ongoing flag to true, which prevents any other task from >> entering the critical section. Hence, when we call the _end() function, >> we are 100% sure that only one task is executing it. Hence locks are not >> necessary around that second update. In fact, that very update marks the >> end of the critical section (which acts much like a spin_unlock(&lock) >> in a "regular" critical section). > > OK, I start to get it. Is there a risk of missing a wake_up event? E.g. > one thread waking up earlier, noticing that transition is in progress > and waiting indefinitely? > No, the only downside to having the CPU reorder the assignment to the flag is that a new transition can begin while the old one is still finishing up the frequency transition by calling the _post_transition() notifiers. Regards, Srivatsa S. Bhat -- To unsubscribe from this list: send the line "unsubscribe cpufreq" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/drivers/cpufreq/cpufreq.c b/drivers/cpufreq/cpufreq.c index b349406..4279cc9 100644 --- a/drivers/cpufreq/cpufreq.c +++ b/drivers/cpufreq/cpufreq.c @@ -353,6 +353,41 @@ void cpufreq_notify_post_transition(struct cpufreq_policy *policy, } EXPORT_SYMBOL_GPL(cpufreq_notify_post_transition); +void cpufreq_freq_transition_begin(struct cpufreq_policy *policy, + struct cpufreq_freqs *freqs) +{ +wait: + wait_event(policy->transition_wait, !policy->transition_ongoing); + + spin_lock(&policy->transition_lock); + + if (unlikely(policy->transition_ongoing)) { + spin_unlock(&policy->transition_lock); + goto wait; + } + + policy->transition_ongoing = true; + + spin_unlock(&policy->transition_lock); + + cpufreq_notify_transition(policy, freqs, CPUFREQ_PRECHANGE); +} +EXPORT_SYMBOL_GPL(cpufreq_freq_transition_begin); + +void cpufreq_freq_transition_end(struct cpufreq_policy *policy, + struct cpufreq_freqs *freqs, int transition_failed) +{ + if (unlikely(WARN_ON(!policy->transition_ongoing))) + return; + + cpufreq_notify_post_transition(policy, freqs, transition_failed); + + policy->transition_ongoing = false; + + wake_up(&policy->transition_wait); +} +EXPORT_SYMBOL_GPL(cpufreq_freq_transition_end); + /********************************************************************* * SYSFS INTERFACE * @@ -982,6 +1017,8 @@ static struct cpufreq_policy *cpufreq_policy_alloc(void) INIT_LIST_HEAD(&policy->policy_list); init_rwsem(&policy->rwsem); + spin_lock_init(&policy->transition_lock); + init_waitqueue_head(&policy->transition_wait); return policy; diff --git a/include/linux/cpufreq.h b/include/linux/cpufreq.h index 70929bc..263173d 100644 --- a/include/linux/cpufreq.h +++ b/include/linux/cpufreq.h @@ -16,6 +16,7 @@ #include <linux/completion.h> #include <linux/kobject.h> #include <linux/notifier.h> +#include <linux/spinlock.h> #include <linux/sysfs.h> /********************************************************************* @@ -104,6 +105,11 @@ struct cpufreq_policy { * __cpufreq_governor(data, CPUFREQ_GOV_POLICY_EXIT); */ struct rw_semaphore rwsem; + + /* Synchronization for frequency transitions */ + bool transition_ongoing; /* Tracks transition status */ + spinlock_t transition_lock; + wait_queue_head_t transition_wait; }; /* Only for ACPI */ @@ -336,6 +342,10 @@ void cpufreq_notify_transition(struct cpufreq_policy *policy, struct cpufreq_freqs *freqs, unsigned int state); void cpufreq_notify_post_transition(struct cpufreq_policy *policy, struct cpufreq_freqs *freqs, int transition_failed); +void cpufreq_freq_transition_begin(struct cpufreq_policy *policy, + struct cpufreq_freqs *freqs); +void cpufreq_freq_transition_end(struct cpufreq_policy *policy, + struct cpufreq_freqs *freqs, int transition_failed); #else /* CONFIG_CPU_FREQ */ static inline int cpufreq_register_notifier(struct notifier_block *nb,