WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-changelog

[Xen-changelog] [xen-unstable] x86/atomic.h: Clean up for Xen code style

To: xen-changelog@xxxxxxxxxxxxxxxxxxx
Subject: [Xen-changelog] [xen-unstable] x86/atomic.h: Clean up for Xen code style; remove unused smp_mb__*
From: Xen patchbot-unstable <patchbot-unstable@xxxxxxxxxxxxxxxxxxx>
Date: Thu, 23 Dec 2010 05:35:09 -0800
Delivery-date: Thu, 23 Dec 2010 05:47:15 -0800
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
List-help: <mailto:xen-changelog-request@lists.xensource.com?subject=help>
List-id: BK change log <xen-changelog.lists.xensource.com>
List-post: <mailto:xen-changelog@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-changelog>, <mailto:xen-changelog-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-changelog>, <mailto:xen-changelog-request@lists.xensource.com?subject=unsubscribe>
Reply-to: xen-devel@xxxxxxxxxxxxxxxxxxx
Sender: xen-changelog-bounces@xxxxxxxxxxxxxxxxxxx
# HG changeset patch
# User Keir Fraser <keir@xxxxxxx>
# Date 1292526095 0
# Node ID 6754771bf10181cc83de84778be40bb62ec2093d
# Parent  04f15c2de8fa8e22104ce0dfc591b57ce02dadac
x86/atomic.h: Clean up for Xen code style; remove unused smp_mb__*

Signed-off-by: Keir Fraser <keir@xxxxxxx>
---
 xen/include/asm-x86/atomic.h |  123 ++++++++++++++++++++-----------------------
 1 files changed, 58 insertions(+), 65 deletions(-)

diff -r 04f15c2de8fa -r 6754771bf101 xen/include/asm-x86/atomic.h
--- a/xen/include/asm-x86/atomic.h      Thu Dec 16 18:46:55 2010 +0000
+++ b/xen/include/asm-x86/atomic.h      Thu Dec 16 19:01:35 2010 +0000
@@ -11,7 +11,7 @@
  */
 typedef struct { int counter; } atomic_t;
 
-#define ATOMIC_INIT(i) { (i) }
+#define ATOMIC_INIT(i) { (i) }
 
 /**
  * atomic_read - read atomic variable
@@ -19,8 +19,8 @@ typedef struct { int counter; } atomic_t
  * 
  * Atomically reads the value of @v.
  */
-#define _atomic_read(v)                ((v).counter)
-#define atomic_read(v)         (*(volatile int *)&((v)->counter))
+#define _atomic_read(v)  ((v).counter)
+#define atomic_read(v)   (*(volatile int *)&((v)->counter))
 
 /**
  * atomic_set - set atomic variable
@@ -29,8 +29,8 @@ typedef struct { int counter; } atomic_t
  * 
  * Atomically sets the value of @v to @i.
  */ 
-#define _atomic_set(v,i)       (((v).counter) = (i))
-#define atomic_set(v,i)                (*(volatile int *)&((v)->counter) = (i))
+#define _atomic_set(v,i) (((v).counter) = (i))
+#define atomic_set(v,i)  (*(volatile int *)&((v)->counter) = (i))
 
 /**
  * atomic_add - add integer to atomic variable
@@ -39,12 +39,12 @@ typedef struct { int counter; } atomic_t
  * 
  * Atomically adds @i to @v.
  */
-static __inline__ void atomic_add(int i, atomic_t *v)
+static inline void atomic_add(int i, atomic_t *v)
 {
-       asm volatile(
-               "lock; addl %1,%0"
-               :"=m" (*(volatile int *)&v->counter)
-               :"ir" (i), "m" (*(volatile int *)&v->counter));
+    asm volatile (
+        "lock; addl %1,%0"
+        : "=m" (*(volatile int *)&v->counter)
+        : "ir" (i), "m" (*(volatile int *)&v->counter) );
 }
 
 /**
@@ -54,12 +54,12 @@ static __inline__ void atomic_add(int i,
  * 
  * Atomically subtracts @i from @v.
  */
-static __inline__ void atomic_sub(int i, atomic_t *v)
+static inline void atomic_sub(int i, atomic_t *v)
 {
-       asm volatile(
-               "lock; subl %1,%0"
-               :"=m" (*(volatile int *)&v->counter)
-               :"ir" (i), "m" (*(volatile int *)&v->counter));
+    asm volatile (
+        "lock; subl %1,%0"
+        : "=m" (*(volatile int *)&v->counter)
+        : "ir" (i), "m" (*(volatile int *)&v->counter) );
 }
 
 /**
@@ -71,15 +71,15 @@ static __inline__ void atomic_sub(int i,
  * true if the result is zero, or false for all
  * other cases.
  */
-static __inline__ int atomic_sub_and_test(int i, atomic_t *v)
+static inline int atomic_sub_and_test(int i, atomic_t *v)
 {
-       unsigned char c;
+    unsigned char c;
 
-       asm volatile(
-               "lock; subl %2,%0; sete %1"
-               :"=m" (*(volatile int *)&v->counter), "=qm" (c)
-               :"ir" (i), "m" (*(volatile int *)&v->counter) : "memory");
-       return c;
+    asm volatile (
+        "lock; subl %2,%0; sete %1"
+        : "=m" (*(volatile int *)&v->counter), "=qm" (c)
+        : "ir" (i), "m" (*(volatile int *)&v->counter) : "memory" );
+    return c;
 }
 
 /**
@@ -88,12 +88,12 @@ static __inline__ int atomic_sub_and_tes
  * 
  * Atomically increments @v by 1.
  */ 
-static __inline__ void atomic_inc(atomic_t *v)
+static inline void atomic_inc(atomic_t *v)
 {
-       asm volatile(
-               "lock; incl %0"
-               :"=m" (*(volatile int *)&v->counter)
-               :"m" (*(volatile int *)&v->counter));
+    asm volatile (
+        "lock; incl %0"
+        : "=m" (*(volatile int *)&v->counter)
+        : "m" (*(volatile int *)&v->counter) );
 }
 
 /**
@@ -102,12 +102,12 @@ static __inline__ void atomic_inc(atomic
  * 
  * Atomically decrements @v by 1.
  */ 
-static __inline__ void atomic_dec(atomic_t *v)
+static inline void atomic_dec(atomic_t *v)
 {
-       asm volatile(
-               "lock; decl %0"
-               :"=m" (*(volatile int *)&v->counter)
-               :"m" (*(volatile int *)&v->counter));
+    asm volatile (
+        "lock; decl %0"
+        : "=m" (*(volatile int *)&v->counter)
+        : "m" (*(volatile int *)&v->counter) );
 }
 
 /**
@@ -118,15 +118,15 @@ static __inline__ void atomic_dec(atomic
  * returns true if the result is 0, or false for all other
  * cases.
  */ 
-static __inline__ int atomic_dec_and_test(atomic_t *v)
+static inline int atomic_dec_and_test(atomic_t *v)
 {
-       unsigned char c;
+    unsigned char c;
 
-       asm volatile(
-               "lock; decl %0; sete %1"
-               :"=m" (*(volatile int *)&v->counter), "=qm" (c)
-               :"m" (*(volatile int *)&v->counter) : "memory");
-       return c != 0;
+    asm volatile (
+        "lock; decl %0; sete %1"
+        : "=m" (*(volatile int *)&v->counter), "=qm" (c)
+        : "m" (*(volatile int *)&v->counter) : "memory" );
+    return c != 0;
 }
 
 /**
@@ -137,15 +137,15 @@ static __inline__ int atomic_dec_and_tes
  * and returns true if the result is zero, or false for all
  * other cases.
  */ 
-static __inline__ int atomic_inc_and_test(atomic_t *v)
+static inline int atomic_inc_and_test(atomic_t *v)
 {
-       unsigned char c;
+    unsigned char c;
 
-       asm volatile(
-               "lock; incl %0; sete %1"
-               :"=m" (*(volatile int *)&v->counter), "=qm" (c)
-               :"m" (*(volatile int *)&v->counter) : "memory");
-       return c != 0;
+    asm volatile (
+        "lock; incl %0; sete %1"
+        : "=m" (*(volatile int *)&v->counter), "=qm" (c)
+        : "m" (*(volatile int *)&v->counter) : "memory" );
+    return c != 0;
 }
 
 /**
@@ -157,30 +157,23 @@ static __inline__ int atomic_inc_and_tes
  * if the result is negative, or false when
  * result is greater than or equal to zero.
  */ 
-static __inline__ int atomic_add_negative(int i, atomic_t *v)
+static inline int atomic_add_negative(int i, atomic_t *v)
 {
-       unsigned char c;
+    unsigned char c;
 
-       asm volatile(
-               "lock; addl %2,%0; sets %1"
-               :"=m" (*(volatile int *)&v->counter), "=qm" (c)
-               :"ir" (i), "m" (*(volatile int *)&v->counter) : "memory");
-       return c;
+    asm volatile (
+        "lock; addl %2,%0; sets %1"
+        : "=m" (*(volatile int *)&v->counter), "=qm" (c)
+        : "ir" (i), "m" (*(volatile int *)&v->counter) : "memory" );
+    return c;
 }
 
-static __inline__ atomic_t atomic_compareandswap(
-       atomic_t old, atomic_t new, atomic_t *v)
+static inline atomic_t atomic_compareandswap(
+    atomic_t old, atomic_t new, atomic_t *v)
 {
-       atomic_t rc;
-       rc.counter = 
-               __cmpxchg(&v->counter, old.counter, new.counter, sizeof(int));
-       return rc;
+    atomic_t rc;
+    rc.counter = __cmpxchg(&v->counter, old.counter, new.counter, sizeof(int));
+    return rc;
 }
 
-/* Atomic operations are already serializing on x86 */
-#define smp_mb__before_atomic_dec()    barrier()
-#define smp_mb__after_atomic_dec()     barrier()
-#define smp_mb__before_atomic_inc()    barrier()
-#define smp_mb__after_atomic_inc()     barrier()
-
 #endif /* __ARCH_X86_ATOMIC__ */

_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-changelog

<Prev in Thread] Current Thread [Next in Thread>
  • [Xen-changelog] [xen-unstable] x86/atomic.h: Clean up for Xen code style; remove unused smp_mb__*, Xen patchbot-unstable <=