Commit 144b2a91 authored by Harvey Harrison's avatar Harvey Harrison Committed by Linus Torvalds

asm-generic: remove fastcall

Signed-off-by: default avatarHarvey Harrison <harvey.harrison@gmail.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent ec701584
...@@ -25,17 +25,17 @@ ...@@ -25,17 +25,17 @@
* in the low address range. Architectures for which this is not * in the low address range. Architectures for which this is not
* true can't use this generic implementation. * true can't use this generic implementation.
*/ */
extern unsigned int fastcall ioread8(void __iomem *); extern unsigned int ioread8(void __iomem *);
extern unsigned int fastcall ioread16(void __iomem *); extern unsigned int ioread16(void __iomem *);
extern unsigned int fastcall ioread16be(void __iomem *); extern unsigned int ioread16be(void __iomem *);
extern unsigned int fastcall ioread32(void __iomem *); extern unsigned int ioread32(void __iomem *);
extern unsigned int fastcall ioread32be(void __iomem *); extern unsigned int ioread32be(void __iomem *);
extern void fastcall iowrite8(u8, void __iomem *); extern void iowrite8(u8, void __iomem *);
extern void fastcall iowrite16(u16, void __iomem *); extern void iowrite16(u16, void __iomem *);
extern void fastcall iowrite16be(u16, void __iomem *); extern void iowrite16be(u16, void __iomem *);
extern void fastcall iowrite32(u32, void __iomem *); extern void iowrite32(u32, void __iomem *);
extern void fastcall iowrite32be(u32, void __iomem *); extern void iowrite32be(u32, void __iomem *);
/* /*
* "string" versions of the above. Note that they * "string" versions of the above. Note that they
...@@ -48,13 +48,13 @@ extern void fastcall iowrite32be(u32, void __iomem *); ...@@ -48,13 +48,13 @@ extern void fastcall iowrite32be(u32, void __iomem *);
* memory across multiple ports, use "memcpy_toio()" * memory across multiple ports, use "memcpy_toio()"
* and friends. * and friends.
*/ */
extern void fastcall ioread8_rep(void __iomem *port, void *buf, unsigned long count); extern void ioread8_rep(void __iomem *port, void *buf, unsigned long count);
extern void fastcall ioread16_rep(void __iomem *port, void *buf, unsigned long count); extern void ioread16_rep(void __iomem *port, void *buf, unsigned long count);
extern void fastcall ioread32_rep(void __iomem *port, void *buf, unsigned long count); extern void ioread32_rep(void __iomem *port, void *buf, unsigned long count);
extern void fastcall iowrite8_rep(void __iomem *port, const void *buf, unsigned long count); extern void iowrite8_rep(void __iomem *port, const void *buf, unsigned long count);
extern void fastcall iowrite16_rep(void __iomem *port, const void *buf, unsigned long count); extern void iowrite16_rep(void __iomem *port, const void *buf, unsigned long count);
extern void fastcall iowrite32_rep(void __iomem *port, const void *buf, unsigned long count); extern void iowrite32_rep(void __iomem *port, const void *buf, unsigned long count);
/* Create a virtual mapping cookie for an IO port range */ /* Create a virtual mapping cookie for an IO port range */
extern void __iomem *ioport_map(unsigned long port, unsigned int nr); extern void __iomem *ioport_map(unsigned long port, unsigned int nr);
......
...@@ -18,7 +18,7 @@ ...@@ -18,7 +18,7 @@
* 1 even when the "1" assertion wasn't true. * 1 even when the "1" assertion wasn't true.
*/ */
static inline void static inline void
__mutex_fastpath_lock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *)) __mutex_fastpath_lock(atomic_t *count, void (*fail_fn)(atomic_t *))
{ {
if (unlikely(atomic_dec_return(count) < 0)) if (unlikely(atomic_dec_return(count) < 0))
fail_fn(count); fail_fn(count);
...@@ -37,7 +37,7 @@ __mutex_fastpath_lock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *)) ...@@ -37,7 +37,7 @@ __mutex_fastpath_lock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *))
* or anything the slow path function returns. * or anything the slow path function returns.
*/ */
static inline int static inline int
__mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t *)) __mutex_fastpath_lock_retval(atomic_t *count, int (*fail_fn)(atomic_t *))
{ {
if (unlikely(atomic_dec_return(count) < 0)) if (unlikely(atomic_dec_return(count) < 0))
return fail_fn(count); return fail_fn(count);
...@@ -61,7 +61,7 @@ __mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t * ...@@ -61,7 +61,7 @@ __mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t *
* to return 0 otherwise. * to return 0 otherwise.
*/ */
static inline void static inline void
__mutex_fastpath_unlock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *)) __mutex_fastpath_unlock(atomic_t *count, void (*fail_fn)(atomic_t *))
{ {
smp_mb(); smp_mb();
if (unlikely(atomic_inc_return(count) <= 0)) if (unlikely(atomic_inc_return(count) <= 0))
......
...@@ -23,7 +23,7 @@ ...@@ -23,7 +23,7 @@
* even when the "1" assertion wasn't true. * even when the "1" assertion wasn't true.
*/ */
static inline void static inline void
__mutex_fastpath_lock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *)) __mutex_fastpath_lock(atomic_t *count, void (*fail_fn)(atomic_t *))
{ {
if (unlikely(atomic_xchg(count, 0) != 1)) if (unlikely(atomic_xchg(count, 0) != 1))
fail_fn(count); fail_fn(count);
...@@ -42,7 +42,7 @@ __mutex_fastpath_lock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *)) ...@@ -42,7 +42,7 @@ __mutex_fastpath_lock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *))
* or anything the slow path function returns * or anything the slow path function returns
*/ */
static inline int static inline int
__mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t *)) __mutex_fastpath_lock_retval(atomic_t *count, int (*fail_fn)(atomic_t *))
{ {
if (unlikely(atomic_xchg(count, 0) != 1)) if (unlikely(atomic_xchg(count, 0) != 1))
return fail_fn(count); return fail_fn(count);
...@@ -65,7 +65,7 @@ __mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t * ...@@ -65,7 +65,7 @@ __mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t *
* to return 0 otherwise. * to return 0 otherwise.
*/ */
static inline void static inline void
__mutex_fastpath_unlock(atomic_t *count, fastcall void (*fail_fn)(atomic_t *)) __mutex_fastpath_unlock(atomic_t *count, void (*fail_fn)(atomic_t *))
{ {
smp_mb(); smp_mb();
if (unlikely(atomic_xchg(count, 1) != 0)) if (unlikely(atomic_xchg(count, 1) != 0))
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment