/* xchg */
static __attribute__((always_inline))
-unsigned long _atomic_exchange(volatile void *addr, unsigned long val, int len)
+unsigned long _atomic_exchange(void *addr, unsigned long val, int len)
{
switch (len) {
case 4:
/* cmpxchg */
static __attribute__((always_inline))
-unsigned long _atomic_cmpxchg(volatile void *addr, unsigned long old,
+unsigned long _atomic_cmpxchg(void *addr, unsigned long old,
unsigned long _new, int len)
{
switch (len) {
/* atomic_add_return */
static __attribute__((always_inline))
-unsigned long _atomic_add_return(volatile void *addr, unsigned long val,
+unsigned long _atomic_add_return(void *addr, unsigned long val,
int len)
{
switch (len) {
/* cmpxchg */
static inline __attribute__((always_inline))
-unsigned long _atomic_cmpxchg(volatile void *addr, unsigned long old,
+unsigned long _atomic_cmpxchg(void *addr, unsigned long old,
unsigned long _new, int len)
{
switch (len) {
/* xchg */
static inline __attribute__((always_inline))
-unsigned long _atomic_exchange(volatile void *addr, unsigned long val, int len)
+unsigned long _atomic_exchange(void *addr, unsigned long val, int len)
{
/* Note: the "xchg" instruction does not need a "lock" prefix. */
switch (len) {
/* atomic_add_return, atomic_sub_return */
static inline __attribute__((always_inline))
-unsigned long _atomic_add_return(volatile void *addr, unsigned long val,
+unsigned long _atomic_add_return(void *addr, unsigned long val,
int len)
{
switch (len) {
/* atomic_add, atomic_sub */
static inline __attribute__((always_inline))
-void _atomic_add(volatile void *addr, unsigned long val, int len)
+void _atomic_add(void *addr, unsigned long val, int len)
{
switch (len) {
case 1:
/* atomic_inc */
static inline __attribute__((always_inline))
-void _atomic_inc(volatile void *addr, int len)
+void _atomic_inc(void *addr, int len)
{
switch (len) {
case 1:
/* atomic_dec */
static inline __attribute__((always_inline))
-void _atomic_dec(volatile void *addr, int len)
+void _atomic_dec(void *addr, int len)
{
switch (len) {
case 1: