X-Git-Url: https://git.sur5r.net/?a=blobdiff_plain;f=include%2Fasm-i386%2Fbitops.h;h=b768e20fbb66484cbccc42235500e4c1651db278;hb=1a9eeb78b825bfade31d7606a2fe3b9eca9e35be;hp=7d29335fdee8824c2050d4fbe2ed1321f651e56a;hpb=7a8e9bed17d7924a9c5c4699b1f6a3a0359524ed;p=u-boot diff --git a/include/asm-i386/bitops.h b/include/asm-i386/bitops.h index 7d29335fde..b768e20fbb 100644 --- a/include/asm-i386/bitops.h +++ b/include/asm-i386/bitops.h @@ -116,7 +116,7 @@ static __inline__ void change_bit(int nr, volatile void * addr) * @nr: Bit to set * @addr: Address to count from * - * This operation is atomic and cannot be reordered. + * This operation is atomic and cannot be reordered. * It also implies a memory barrier. */ static __inline__ int test_and_set_bit(int nr, volatile void * addr) @@ -135,7 +135,7 @@ static __inline__ int test_and_set_bit(int nr, volatile void * addr) * @nr: Bit to set * @addr: Address to count from * - * This operation is non-atomic and can be reordered. + * This operation is non-atomic and can be reordered. * If two examples of this operation race, one can appear to succeed * but actually fail. You must protect multiple accesses with a lock. */ @@ -155,7 +155,7 @@ static __inline__ int __test_and_set_bit(int nr, volatile void * addr) * @nr: Bit to set * @addr: Address to count from * - * This operation is atomic and cannot be reordered. + * This operation is atomic and cannot be reordered. * It also implies a memory barrier. */ static __inline__ int test_and_clear_bit(int nr, volatile void * addr) @@ -174,7 +174,7 @@ static __inline__ int test_and_clear_bit(int nr, volatile void * addr) * @nr: Bit to set * @addr: Address to count from * - * This operation is non-atomic and can be reordered. + * This operation is non-atomic and can be reordered. * If two examples of this operation race, one can appear to succeed * but actually fail. You must protect multiple accesses with a lock. */ @@ -206,7 +206,7 @@ static __inline__ int __test_and_change_bit(int nr, volatile void * addr) * @nr: Bit to set * @addr: Address to count from * - * This operation is atomic and cannot be reordered. + * This operation is atomic and cannot be reordered. * It also implies a memory barrier. */ static __inline__ int test_and_change_bit(int nr, volatile void * addr) @@ -292,7 +292,7 @@ static __inline__ int find_next_zero_bit (void * addr, int size, int offset) { unsigned long * p = ((unsigned long *) addr) + (offset >> 5); int set = 0, bit = offset & 31, res; - + if (bit) { /* * Look for zero in first byte