[PATCH] fuse: uninline some functions
Inline keyword is unnecessary in most cases. Clean them up. Signed-off-by: Miklos Szeredi <miklos@szeredi.hu> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
committed by
Linus Torvalds
parent
b3bebd94bb
commit
8bfc016d2e
@ -21,7 +21,7 @@ MODULE_ALIAS_MISCDEV(FUSE_MINOR);
|
||||
|
||||
static kmem_cache_t *fuse_req_cachep;
|
||||
|
||||
static inline struct fuse_conn *fuse_get_conn(struct file *file)
|
||||
static struct fuse_conn *fuse_get_conn(struct file *file)
|
||||
{
|
||||
struct fuse_conn *fc;
|
||||
spin_lock(&fuse_lock);
|
||||
@ -32,7 +32,7 @@ static inline struct fuse_conn *fuse_get_conn(struct file *file)
|
||||
return fc;
|
||||
}
|
||||
|
||||
static inline void fuse_request_init(struct fuse_req *req)
|
||||
static void fuse_request_init(struct fuse_req *req)
|
||||
{
|
||||
memset(req, 0, sizeof(*req));
|
||||
INIT_LIST_HEAD(&req->list);
|
||||
@ -53,7 +53,7 @@ void fuse_request_free(struct fuse_req *req)
|
||||
kmem_cache_free(fuse_req_cachep, req);
|
||||
}
|
||||
|
||||
static inline void block_sigs(sigset_t *oldset)
|
||||
static void block_sigs(sigset_t *oldset)
|
||||
{
|
||||
sigset_t mask;
|
||||
|
||||
@ -61,7 +61,7 @@ static inline void block_sigs(sigset_t *oldset)
|
||||
sigprocmask(SIG_BLOCK, &mask, oldset);
|
||||
}
|
||||
|
||||
static inline void restore_sigs(sigset_t *oldset)
|
||||
static void restore_sigs(sigset_t *oldset)
|
||||
{
|
||||
sigprocmask(SIG_SETMASK, oldset, NULL);
|
||||
}
|
||||
@ -385,7 +385,7 @@ void fuse_send_init(struct fuse_conn *fc)
|
||||
* anything that could cause a page-fault. If the request was already
|
||||
* interrupted bail out.
|
||||
*/
|
||||
static inline int lock_request(struct fuse_req *req)
|
||||
static int lock_request(struct fuse_req *req)
|
||||
{
|
||||
int err = 0;
|
||||
if (req) {
|
||||
@ -404,7 +404,7 @@ static inline int lock_request(struct fuse_req *req)
|
||||
* requester thread is currently waiting for it to be unlocked, so
|
||||
* wake it up.
|
||||
*/
|
||||
static inline void unlock_request(struct fuse_req *req)
|
||||
static void unlock_request(struct fuse_req *req)
|
||||
{
|
||||
if (req) {
|
||||
spin_lock(&fuse_lock);
|
||||
@ -440,7 +440,7 @@ static void fuse_copy_init(struct fuse_copy_state *cs, int write,
|
||||
}
|
||||
|
||||
/* Unmap and put previous page of userspace buffer */
|
||||
static inline void fuse_copy_finish(struct fuse_copy_state *cs)
|
||||
static void fuse_copy_finish(struct fuse_copy_state *cs)
|
||||
{
|
||||
if (cs->mapaddr) {
|
||||
kunmap_atomic(cs->mapaddr, KM_USER0);
|
||||
@ -489,8 +489,7 @@ static int fuse_copy_fill(struct fuse_copy_state *cs)
|
||||
}
|
||||
|
||||
/* Do as much copy to/from userspace buffer as we can */
|
||||
static inline int fuse_copy_do(struct fuse_copy_state *cs, void **val,
|
||||
unsigned *size)
|
||||
static int fuse_copy_do(struct fuse_copy_state *cs, void **val, unsigned *size)
|
||||
{
|
||||
unsigned ncpy = min(*size, cs->len);
|
||||
if (val) {
|
||||
@ -510,8 +509,8 @@ static inline int fuse_copy_do(struct fuse_copy_state *cs, void **val,
|
||||
* Copy a page in the request to/from the userspace buffer. Must be
|
||||
* done atomically
|
||||
*/
|
||||
static inline int fuse_copy_page(struct fuse_copy_state *cs, struct page *page,
|
||||
unsigned offset, unsigned count, int zeroing)
|
||||
static int fuse_copy_page(struct fuse_copy_state *cs, struct page *page,
|
||||
unsigned offset, unsigned count, int zeroing)
|
||||
{
|
||||
if (page && zeroing && count < PAGE_SIZE) {
|
||||
void *mapaddr = kmap_atomic(page, KM_USER1);
|
||||
|
Reference in New Issue
Block a user