Lines Matching refs:stores

177 Furthermore, the stores committed by a CPU to the memory system may not be
178 perceived by the loads made by another CPU in the same order as the stores were
247 (*) Overlapping loads and stores within a particular CPU will appear to be
264 (Loads and stores overlap if they are targeted at overlapping pieces of
275 (*) It _must_not_ be assumed that independent loads and stores will be issued
387 A write barrier is a partial ordering on stores only; it is not required
391 memory system as time progresses. All stores _before_ a write barrier
392 will occur _before_ all the stores after the write barrier.
408 loads only; it is not required to have any effect on stores, independent
412 committing sequences of stores to the memory system that the CPU being
415 if that load touches one of a sequence of stores from another CPU, then
416 by the time the barrier completes, the effects of all the stores prior to
446 have any effect on stores.
462 A general memory barrier is a partial ordering over both loads and stores.
703 However, stores are not speculated. This means that ordering -is- provided
715 the compiler might combine the store to 'b' with other stores to 'b'.
728 It is tempting to try to enforce ordering on identical stores on both
773 ordering is guaranteed only when the stores differ, for example:
825 Please note once again that the stores to 'b' differ. If they were
877 In short, control dependencies apply only to the stores in the then-clause
889 (*) Control dependencies can order prior loads against later stores.
891 Not prior loads against later loads, nor prior stores against
893 use smp_rmb(), smp_wmb(), or, in the case of prior stores and
896 (*) If both legs of the "if" statement begin with identical stores to
897 the same variable, then those stores must be ordered, either by
899 to carry out the stores. Please note that it is -not- sufficient
982 [!] Note that the stores before the write barrier would normally be expected to
1024 | | +------+ } requires all stores prior to the
1026 | | : +------+ } further stores may take place
1031 | Sequence in which stores are committed to the
1369 CPUs agree on the order in which all stores become visible. However,
1385 Suppose that CPU 2's load from X returns 1, which it then stores to Y,
1481 at least aside from stores. Therefore, the following outcome is possible:
1493 and smp_store_release() are not required to order prior stores against
1553 (*) The compiler is within its rights to reorder loads and stores
1741 (*) The compiler is within its rights to invent stores to a variable,
1809 loads followed by a pair of 32-bit stores. This would result in
1953 This is for use with persistent memory to ensure that stores for which
1958 to ensure that stores have reached a platform durability domain. This ensures
1959 that stores have updated persistent storage before any data access or
2258 order multiple stores before the wake-up with respect to loads of those stored
2739 their own loads and stores as if they had happened in program order.
2808 execution progress, whereas stores can often be deferred without a
2820 (*) loads and stores may be combined to improve performance when talking to