[PATCH 3.19-rc2 v15 4/8] sched_clock: Avoid deadlock during read from NMI
Thomas Gleixner
tglx at linutronix.de
Sat Jan 24 14:40:46 PST 2015
On Fri, 23 Jan 2015, Daniel Thompson wrote:
> This patch fixes that problem by providing banked clock data in a
> similar manner to Thomas Gleixner's 4396e058c52e("timekeeping: Provide
> fast and NMI safe access to CLOCK_MONOTONIC").
By some definition of similar.
> -struct clock_data {
> - ktime_t wrap_kt;
> +struct clock_data_banked {
> u64 epoch_ns;
> u64 epoch_cyc;
> - seqcount_t seq;
> - unsigned long rate;
> + u64 (*read_sched_clock)(void);
> + u64 sched_clock_mask;
> u32 mult;
> u32 shift;
> bool suspended;
> };
>
> +struct clock_data {
> + ktime_t wrap_kt;
> + seqcount_t seq;
> + unsigned long rate;
> + struct clock_data_banked bank[2];
> +};
....
> -static u64 __read_mostly (*read_sched_clock)(void) = jiffy_sched_clock_read;
> +static struct clock_data cd = {
> + .bank = {
> + [0] = {
> + .mult = NSEC_PER_SEC / HZ,
> + .read_sched_clock = jiffy_sched_clock_read,
> + },
> + },
> +};
If you had carefully studied the changes which made it possible to do
the nmi safe clock monotonic accessor then you'd had noticed that I
went a great way to optimize the cache foot print first and then add
this new fangled thing.
So in the first place 'cd' lacks ____cacheline_aligned. It should have
been there before, but that's a different issue. You should have
noticed.
Secondly, I don't see any hint that you actually thought about the
cache foot print of the result struct clock_data.
struct clock_data {
ktime_t wrap_kt;
seqcount_t seq;
unsigned long rate;
struct clock_data_banked bank[2];
};
wrap_kt and rate are completely irrelevant for the hotpath. The whole
thing up to the last member of bank[0] still fits into 64 byte on both
32 and 64bit, but that's not by design and not documented so anyone
who is aware of cache foot print issues will go WTF when the first
member of a hot path data structure is completely irrelevant.
> static inline u64 notrace cyc_to_ns(u64 cyc, u32 mult, u32 shift)
> {
> @@ -58,50 +65,82 @@ static inline u64 notrace cyc_to_ns(u64 cyc, u32 mult, u32 shift)
>
> unsigned long long notrace sched_clock(void)
> {
> - u64 epoch_ns;
> - u64 epoch_cyc;
> u64 cyc;
> unsigned long seq;
> -
> - if (cd.suspended)
> - return cd.epoch_ns;
> + struct clock_data_banked *b;
> + u64 res;
So we now have
u64 cyc;
unsigned long seq;
struct clock_data_banked *b;
u64 res;
Let me try a different version of that:
struct clock_data_banked *b;
unsigned long seq;
u64 res, cyc;
Can you spot the difference in the reading experience?
> do {
> - seq = raw_read_seqcount_begin(&cd.seq);
> - epoch_cyc = cd.epoch_cyc;
> - epoch_ns = cd.epoch_ns;
> + seq = raw_read_seqcount(&cd.seq);
> + b = cd.bank + (seq & 1);
> + if (b->suspended) {
> + res = b->epoch_ns;
So now we have read_sched_clock as a pointer in the bank. Why do you
still need b->suspended?
What's wrong with setting b->read_sched_clock to NULL at suspend and
restore the proper pointer on resume and use that as a conditional?
It would allow the compiler to generate better code, but that's
obviously not the goal here. Darn, this is hot path code and not some
random driver.
> + } else {
> + cyc = b->read_sched_clock();
> + cyc = (cyc - b->epoch_cyc) & b->sched_clock_mask;
> + res = b->epoch_ns + cyc_to_ns(cyc, b->mult, b->shift);
It would allow the following optimization as well:
res = b->epoch_ns;
if (b->read_sched_clock) {
...
}
If you think that compilers are smart enough to figure all that out
for you, you might get surprised. The more clear your code is the
better is the chance that the compiler gets it right. We have seen the
opposite of that as well, but that's clearly a compiler bug.
> +/*
> + * Start updating the banked clock data.
> + *
> + * sched_clock will never observe mis-matched data even if called from
> + * an NMI. We do this by maintaining an odd/even copy of the data and
> + * steering sched_clock to one or the other using a sequence counter.
> + * In order to preserve the data cache profile of sched_clock as much
> + * as possible the system reverts back to the even copy when the update
> + * completes; the odd copy is used *only* during an update.
> + *
> + * The caller is responsible for avoiding simultaneous updates.
> + */
> +static struct clock_data_banked *update_bank_begin(void)
> +{
> + /* update the backup (odd) bank and steer readers towards it */
> + memcpy(cd.bank + 1, cd.bank, sizeof(struct clock_data_banked));
> + raw_write_seqcount_latch(&cd.seq);
> +
> + return cd.bank;
> +}
> +
> +/*
> + * Finalize update of banked clock data.
> + *
> + * This is just a trivial switch back to the primary (even) copy.
> + */
> +static void update_bank_end(void)
> +{
> + raw_write_seqcount_latch(&cd.seq);
> }
What's wrong with having a master struct
struct master_data {
struct clock_data_banked master_data;
ktime_t wrap_kt;
unsigned long rate;
u64 (*real_read_sched_clock)(void);
};
Then you only have to care about the serialization of the master_data
update and then the hotpath data update would be the same simple
function as update_fast_timekeeper(). And it would have the same
ordering scheme and aside of that the resulting code would be simpler,
more intuitive to read and I'm pretty sure faster.
Thanks,
tglx
More information about the linux-arm-kernel
mailing list