diff options
author | Peter Gavin <pete@gentoo.org> | 2001-08-07 01:25:08 +0000 |
---|---|---|
committer | Peter Gavin <pete@gentoo.org> | 2001-08-07 01:25:08 +0000 |
commit | 51cd75bb13417a541002b096183f1d8580bd54a6 (patch) | |
tree | 38972d914c1f406f0f6c1b1b857e55970e70cbcf /sys-kernel | |
parent | petrovich - a file integrity checker (diff) | |
download | gentoo-2-51cd75bb13417a541002b096183f1d8580bd54a6.tar.gz gentoo-2-51cd75bb13417a541002b096183f1d8580bd54a6.tar.bz2 gentoo-2-51cd75bb13417a541002b096183f1d8580bd54a6.zip |
yay, should work now :)
Diffstat (limited to 'sys-kernel')
21 files changed, 7178 insertions, 40 deletions
diff --git a/sys-kernel/linux-extras/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch b/sys-kernel/linux-extras/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch new file mode 100644 index 000000000000..5f3830c8ffe1 --- /dev/null +++ b/sys-kernel/linux-extras/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch @@ -0,0 +1,647 @@ +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/inode.c linux-2.4.6-mmap/fs/nfs/inode.c +--- linux-2.4.6-file/fs/nfs/inode.c Wed Jul 4 17:30:28 2001 ++++ linux-2.4.6-mmap/fs/nfs/inode.c Wed Jul 4 20:48:01 2001 +@@ -48,6 +48,7 @@ + static void nfs_write_inode(struct inode *,int); + static void nfs_delete_inode(struct inode *); + static void nfs_put_super(struct super_block *); ++static void nfs_clear_inode(struct inode *); + static void nfs_umount_begin(struct super_block *); + static int nfs_statfs(struct super_block *, struct statfs *); + +@@ -57,6 +58,7 @@ + delete_inode: nfs_delete_inode, + put_super: nfs_put_super, + statfs: nfs_statfs, ++ clear_inode: nfs_clear_inode, + umount_begin: nfs_umount_begin, + }; + +@@ -141,6 +143,21 @@ + clear_inode(inode); + } + ++/* ++ * For the moment, the only task for the NFS clear_inode method is to ++ * release the mmap credential ++ */ ++static void ++nfs_clear_inode(struct inode *inode) ++{ ++ struct rpc_cred *cred = NFS_I(inode)->mm_cred; ++ ++ if (cred) { ++ put_rpccred(cred); ++ NFS_I(inode)->mm_cred = 0; ++ } ++} ++ + void + nfs_put_super(struct super_block *sb) + { +@@ -600,7 +617,6 @@ + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + NFS_CACHE_CTIME(inode) = fattr->ctime; + NFS_CACHE_MTIME(inode) = fattr->mtime; +- NFS_CACHE_ATIME(inode) = fattr->atime; + NFS_CACHE_ISIZE(inode) = fattr->size; + NFS_ATTRTIMEO(inode) = NFS_MINATTRTIMEO(inode); + NFS_ATTRTIMEO_UPDATE(inode) = jiffies; +@@ -794,6 +810,20 @@ + } + + /* ++ * Ensure that mmap has a recent RPC credential for use when writing out ++ * shared pages ++ */ ++static inline void ++nfs_set_mmcred(struct inode *inode, struct rpc_cred *cred) ++{ ++ struct rpc_cred *oldcred = NFS_I(inode)->mm_cred; ++ ++ NFS_I(inode)->mm_cred = get_rpccred(cred); ++ if (oldcred) ++ put_rpccred(oldcred); ++} ++ ++/* + * These are probably going to contain hooks for + * allocating and releasing RPC credentials for + * the file. I'll have to think about Tronds patch +@@ -808,20 +838,20 @@ + auth = NFS_CLIENT(inode)->cl_auth; + cred = rpcauth_lookupcred(auth, 0); + filp->private_data = cred; ++ if (filp->f_mode & FMODE_WRITE) ++ nfs_set_mmcred(inode, cred); + unlock_kernel(); + return 0; + } + + int nfs_release(struct inode *inode, struct file *filp) + { +- struct rpc_auth *auth; + struct rpc_cred *cred; + + lock_kernel(); +- auth = NFS_CLIENT(inode)->cl_auth; + cred = nfs_file_cred(filp); + if (cred) +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + unlock_kernel(); + return 0; + } +@@ -976,7 +1006,6 @@ + NFS_CACHE_CTIME(inode) = fattr->ctime; + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + +- NFS_CACHE_ATIME(inode) = fattr->atime; + inode->i_atime = nfs_time_to_secs(fattr->atime); + + NFS_CACHE_MTIME(inode) = new_mtime; +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/unlink.c linux-2.4.6-mmap/fs/nfs/unlink.c +--- linux-2.4.6-file/fs/nfs/unlink.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/fs/nfs/unlink.c Wed Jul 4 20:48:47 2001 +@@ -128,7 +128,7 @@ + dir_i = dir->d_inode; + nfs_zap_caches(dir_i); + NFS_PROTO(dir_i)->unlink_done(dir, &task->tk_msg); +- rpcauth_releasecred(task->tk_auth, data->cred); ++ put_rpccred(data->cred); + data->cred = NULL; + dput(dir); + } +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/write.c linux-2.4.6-mmap/fs/nfs/write.c +--- linux-2.4.6-file/fs/nfs/write.c Mon May 21 21:29:14 2001 ++++ linux-2.4.6-mmap/fs/nfs/write.c Wed Jul 4 20:48:24 2001 +@@ -181,7 +181,9 @@ + + + if (file) +- cred = nfs_file_cred(file); ++ cred = get_rpccred(nfs_file_cred(file)); ++ if (!cred) ++ cred = get_rpccred(NFS_I(inode)->mm_cred); + + dprintk("NFS: nfs_writepage_sync(%x/%Ld %d@%Ld)\n", + inode->i_dev, (long long)NFS_FILEID(inode), +@@ -226,6 +228,8 @@ + + io_error: + kunmap(page); ++ if (cred) ++ put_rpccred(cred); + + return written? written : result; + } +@@ -241,6 +245,9 @@ + status = (IS_ERR(req)) ? PTR_ERR(req) : 0; + if (status < 0) + goto out; ++ if (!req->wb_cred) ++ req->wb_cred = get_rpccred(NFS_I(inode)->mm_cred); ++ nfs_unlock_request(req); + nfs_release_request(req); + nfs_strategy(inode); + out: +@@ -557,13 +564,11 @@ + req->wb_bytes = count; + req->wb_file = file; + +- /* If we have a struct file, use its cached credentials +- * else cache the current process' credentials. */ ++ /* If we have a struct file, use its cached credentials */ + if (file) { + get_file(file); + req->wb_cred = nfs_file_cred(file); +- } else +- req->wb_cred = rpcauth_lookupcred(NFS_CLIENT(inode)->cl_auth, 0); ++ } + req->wb_inode = inode; + req->wb_count = 1; + +@@ -608,8 +613,8 @@ + /* Release struct file or cached credential */ + if (req->wb_file) + fput(req->wb_file); +- else +- rpcauth_releasecred(NFS_CLIENT(inode)->cl_auth, req->wb_cred); ++ else if (req->wb_cred) ++ put_rpccred(req->wb_cred); + page_cache_release(page); + nfs_page_free(req); + /* wake up anyone waiting to allocate a request */ +@@ -927,8 +932,6 @@ + if (end > rqend) + req->wb_bytes = end - req->wb_offset; + +- nfs_unlock_request(req); +- + return req; + } + +@@ -1049,6 +1052,7 @@ + goto done; + + status = 0; ++ nfs_unlock_request(req); + /* If we wrote past the end of the page. + * Call the strategy routine so it can send out a bunch + * of requests. +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs.h linux-2.4.6-mmap/include/linux/nfs_fs.h +--- linux-2.4.6-file/include/linux/nfs_fs.h Sat May 26 03:02:11 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs.h Wed Jul 4 17:32:16 2001 +@@ -63,6 +63,11 @@ + */ + #define NFS_SUPER_MAGIC 0x6969 + ++static inline struct nfs_inode_info *NFS_I(struct inode *inode) ++{ ++ return &inode->u.nfs_i; ++} ++ + #define NFS_FH(inode) (&(inode)->u.nfs_i.fh) + #define NFS_SERVER(inode) (&(inode)->i_sb->u.nfs_sb.s_server) + #define NFS_CLIENT(inode) (NFS_SERVER(inode)->client) +@@ -74,7 +79,6 @@ + #define NFS_READTIME(inode) ((inode)->u.nfs_i.read_cache_jiffies) + #define NFS_CACHE_CTIME(inode) ((inode)->u.nfs_i.read_cache_ctime) + #define NFS_CACHE_MTIME(inode) ((inode)->u.nfs_i.read_cache_mtime) +-#define NFS_CACHE_ATIME(inode) ((inode)->u.nfs_i.read_cache_atime) + #define NFS_CACHE_ISIZE(inode) ((inode)->u.nfs_i.read_cache_isize) + #define NFS_NEXTSCAN(inode) ((inode)->u.nfs_i.nextscan) + #define NFS_CACHEINV(inode) \ +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs_i.h linux-2.4.6-mmap/include/linux/nfs_fs_i.h +--- linux-2.4.6-file/include/linux/nfs_fs_i.h Thu May 31 15:27:10 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs_i.h Wed Jul 4 17:32:16 2001 +@@ -45,7 +45,6 @@ + unsigned long read_cache_jiffies; + __u64 read_cache_ctime; + __u64 read_cache_mtime; +- __u64 read_cache_atime; + __u64 read_cache_isize; + unsigned long attrtimeo; + unsigned long attrtimeo_timestamp; +@@ -73,6 +72,9 @@ + struct inode *hash_next, + *hash_prev; + unsigned long nextscan; ++ ++ /* Credentials for shared mmap */ ++ struct rpc_cred *mm_cred; + }; + + /* +diff -u --recursive --new-file linux-2.4.6-file/include/linux/sunrpc/auth.h linux-2.4.6-mmap/include/linux/sunrpc/auth.h +--- linux-2.4.6-file/include/linux/sunrpc/auth.h Fri Jun 22 19:30:22 2001 ++++ linux-2.4.6-mmap/include/linux/sunrpc/auth.h Wed Jul 4 20:51:35 2001 +@@ -14,6 +14,8 @@ + #include <linux/config.h> + #include <linux/sunrpc/sched.h> + ++#include <asm/atomic.h> ++ + /* size of the nodename buffer */ + #define UNX_MAXNODENAME 32 + +@@ -22,8 +24,10 @@ + */ + struct rpc_cred { + struct rpc_cred * cr_next; /* linked list */ ++ struct rpc_auth * cr_auth; ++ struct rpc_credops * cr_ops; + unsigned long cr_expire; /* when to gc */ +- unsigned short cr_count; /* ref count */ ++ atomic_t cr_count; /* ref count */ + unsigned short cr_flags; /* various flags */ + #ifdef RPC_DEBUG + unsigned long cr_magic; /* 0x0f4aa4f0 */ +@@ -71,6 +75,9 @@ + void (*destroy)(struct rpc_auth *); + + struct rpc_cred * (*crcreate)(int); ++}; ++ ++struct rpc_credops { + void (*crdestroy)(struct rpc_cred *); + + int (*crmatch)(struct rpc_cred *, int); +@@ -92,8 +99,7 @@ + struct rpc_cred * rpcauth_lookupcred(struct rpc_auth *, int); + struct rpc_cred * rpcauth_bindcred(struct rpc_task *); + void rpcauth_holdcred(struct rpc_task *); +-void rpcauth_releasecred(struct rpc_auth *, +- struct rpc_cred *); ++void put_rpccred(struct rpc_cred *); + void rpcauth_unbindcred(struct rpc_task *); + int rpcauth_matchcred(struct rpc_auth *, + struct rpc_cred *, int); +@@ -106,6 +112,13 @@ + void rpcauth_free_credcache(struct rpc_auth *); + void rpcauth_insert_credcache(struct rpc_auth *, + struct rpc_cred *); ++ ++static inline ++struct rpc_cred * get_rpccred(struct rpc_cred *cred) ++{ ++ atomic_inc(&cred->cr_count); ++ return cred; ++} + + #endif /* __KERNEL__ */ + #endif /* _LINUX_SUNRPC_AUTH_H */ +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth.c linux-2.4.6-mmap/net/sunrpc/auth.c +--- linux-2.4.6-file/net/sunrpc/auth.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth.c Wed Jul 4 20:51:04 2001 +@@ -81,42 +81,61 @@ + auth->au_nextgc = jiffies + (auth->au_expire >> 1); + } + ++/* ++ * Destroy an unreferenced credential ++ */ + static inline void +-rpcauth_crdestroy(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_crdestroy(struct rpc_cred *cred) + { + #ifdef RPC_DEBUG + if (cred->cr_magic != RPCAUTH_CRED_MAGIC) + BUG(); + cred->cr_magic = 0; ++ if (atomic_read(&cred->cr_count) || cred->cr_auth) ++ BUG(); + #endif +- if (auth->au_ops->crdestroy) +- auth->au_ops->crdestroy(cred); +- else +- rpc_free(cred); ++ cred->cr_ops->crdestroy(cred); + } + + /* +- * Clear the RPC credential cache ++ * Destroy a list of credentials ++ */ ++static inline ++void rpcauth_destroy_credlist(struct rpc_cred *head) ++{ ++ struct rpc_cred *cred; ++ ++ while ((cred = head) != NULL) { ++ head = cred->cr_next; ++ rpcauth_crdestroy(cred); ++ } ++} ++ ++/* ++ * Clear the RPC credential cache, and delete those credentials ++ * that are not referenced. + */ + void + rpcauth_free_credcache(struct rpc_auth *auth) + { +- struct rpc_cred **q, *cred; +- void (*destroy)(struct rpc_cred *); ++ struct rpc_cred **q, *cred, *free = NULL; + int i; + +- if (!(destroy = auth->au_ops->crdestroy)) +- destroy = (void (*)(struct rpc_cred *)) rpc_free; +- + spin_lock(&rpc_credcache_lock); + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { + *q = cred->cr_next; +- destroy(cred); ++ cred->cr_auth = NULL; ++ if (atomic_read(&cred->cr_count) == 0) { ++ cred->cr_next = free; ++ free = cred; ++ } else ++ cred->cr_next = NULL; + } + } + spin_unlock(&rpc_credcache_lock); ++ rpcauth_destroy_credlist(free); + } + + /* +@@ -133,9 +152,10 @@ + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { +- if (!cred->cr_count && ++ if (!atomic_read(&cred->cr_count) && + time_before(cred->cr_expire, jiffies)) { + *q = cred->cr_next; ++ cred->cr_auth = NULL; + cred->cr_next = free; + free = cred; + continue; +@@ -144,10 +164,7 @@ + } + } + spin_unlock(&rpc_credcache_lock); +- while ((cred = free) != NULL) { +- free = cred->cr_next; +- rpcauth_crdestroy(auth, cred); +- } ++ rpcauth_destroy_credlist(free); + auth->au_nextgc = jiffies + auth->au_expire; + } + +@@ -163,8 +180,8 @@ + spin_lock(&rpc_credcache_lock); + cred->cr_next = auth->au_credcache[nr]; + auth->au_credcache[nr] = cred; +- cred->cr_count++; +- cred->cr_expire = jiffies + auth->au_expire; ++ cred->cr_auth = auth; ++ get_rpccred(cred); + spin_unlock(&rpc_credcache_lock); + } + +@@ -187,7 +204,7 @@ + q = &auth->au_credcache[nr]; + while ((cred = *q) != NULL) { + if (!(cred->cr_flags & RPCAUTH_CRED_DEAD) && +- auth->au_ops->crmatch(cred, taskflags)) { ++ cred->cr_ops->crmatch(cred, taskflags)) { + *q = cred->cr_next; + break; + } +@@ -213,23 +230,23 @@ + * Remove cred handle from cache + */ + static void +-rpcauth_remove_credcache(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_remove_credcache(struct rpc_cred *cred) + { ++ struct rpc_auth *auth = cred->cr_auth; + struct rpc_cred **q, *cr; + int nr; + + nr = (cred->cr_uid & RPC_CREDCACHE_MASK); +- spin_lock(&rpc_credcache_lock); + q = &auth->au_credcache[nr]; + while ((cr = *q) != NULL) { + if (cred == cr) { + *q = cred->cr_next; + cred->cr_next = NULL; ++ cred->cr_auth = NULL; + break; + } + q = &cred->cr_next; + } +- spin_unlock(&rpc_credcache_lock); + } + + struct rpc_cred * +@@ -258,7 +275,7 @@ + { + dprintk("RPC: matching %s cred %d\n", + auth->au_ops->au_name, taskflags); +- return auth->au_ops->crmatch(cred, taskflags); ++ return cred->cr_ops->crmatch(cred, taskflags); + } + + void +@@ -266,26 +283,25 @@ + { + dprintk("RPC: %4d holding %s cred %p\n", + task->tk_pid, task->tk_auth->au_ops->au_name, task->tk_msg.rpc_cred); +- if (task->tk_msg.rpc_cred) { +- spin_lock(&rpc_credcache_lock); +- task->tk_msg.rpc_cred->cr_count++; +- task->tk_msg.rpc_cred->cr_expire = jiffies + task->tk_auth->au_expire; +- spin_unlock(&rpc_credcache_lock); +- } ++ if (task->tk_msg.rpc_cred) ++ get_rpccred(task->tk_msg.rpc_cred); + } + + void +-rpcauth_releasecred(struct rpc_auth *auth, struct rpc_cred *cred) ++put_rpccred(struct rpc_cred *cred) + { +- spin_lock(&rpc_credcache_lock); +- if (cred != NULL && cred->cr_count > 0) { +- if (!--cred->cr_count && (cred->cr_flags & RPCAUTH_CRED_DEAD)) { +- spin_unlock(&rpc_credcache_lock); +- rpcauth_remove_credcache(auth, cred); +- rpcauth_crdestroy(auth, cred); +- return; +- } ++ if (!atomic_dec_and_lock(&cred->cr_count, &rpc_credcache_lock)) ++ return; ++ ++ if (cred->cr_auth && cred->cr_flags & RPCAUTH_CRED_DEAD) ++ rpcauth_remove_credcache(cred); ++ ++ if (!cred->cr_auth) { ++ spin_unlock(&rpc_credcache_lock); ++ rpcauth_crdestroy(cred); ++ return; + } ++ cred->cr_expire = jiffies + cred->cr_auth->au_expire; + spin_unlock(&rpc_credcache_lock); + } + +@@ -298,7 +314,7 @@ + dprintk("RPC: %4d releasing %s cred %p\n", + task->tk_pid, auth->au_ops->au_name, cred); + +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + task->tk_msg.rpc_cred = NULL; + } + +@@ -306,10 +322,11 @@ + rpcauth_marshcred(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d marshaling %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crmarshal(task, p, ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crmarshal(task, p, + task->tk_flags & RPC_CALL_REALUID); + } + +@@ -317,20 +334,22 @@ + rpcauth_checkverf(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d validating %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crvalidate(task, p); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crvalidate(task, p); + } + + int + rpcauth_refreshcred(struct rpc_task *task) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d refreshing %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- task->tk_status = auth->au_ops->crrefresh(task); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ task->tk_status = cred->cr_ops->crrefresh(task); + return task->tk_status; + } + +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_null.c linux-2.4.6-mmap/net/sunrpc/auth_null.c +--- linux-2.4.6-file/net/sunrpc/auth_null.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_null.c Wed Jul 4 19:27:38 2001 +@@ -17,6 +17,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops null_credops; ++ + static struct rpc_auth * + nul_create(struct rpc_clnt *clnt) + { +@@ -52,9 +54,10 @@ + + if (!(cred = (struct rpc_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; +- cred->cr_count = 0; ++ atomic_set(&cred->cr_count, 0); + cred->cr_flags = RPCAUTH_CRED_UPTODATE; + cred->cr_uid = current->uid; ++ cred->cr_ops = &null_credops; + + return cred; + } +@@ -124,7 +127,11 @@ + #endif + nul_create, + nul_destroy, +- nul_create_cred, ++ nul_create_cred ++}; ++ ++static ++struct rpc_credops null_credops = { + nul_destroy_cred, + nul_match, + nul_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_unix.c linux-2.4.6-mmap/net/sunrpc/auth_unix.c +--- linux-2.4.6-file/net/sunrpc/auth_unix.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_unix.c Wed Jul 4 19:29:07 2001 +@@ -33,6 +33,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops unix_credops; ++ + static struct rpc_auth * + unx_create(struct rpc_clnt *clnt) + { +@@ -71,7 +73,7 @@ + if (!(cred = (struct unx_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 0; ++ atomic_set(&cred->uc_count, 0); + cred->uc_flags = RPCAUTH_CRED_UPTODATE; + if (flags & RPC_TASK_ROOTCREDS) { + cred->uc_uid = cred->uc_fsuid = 0; +@@ -91,6 +93,7 @@ + if (i < NFS_NGROUPS) + cred->uc_gids[i] = NOGROUP; + } ++ cred->uc_base.cr_ops = &unix_credops; + + return (struct rpc_cred *) cred; + } +@@ -106,7 +109,7 @@ + if (!(cred = (struct unx_cred *) rpc_malloc(task, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 1; ++ atomic_set(&cred->uc_count, 1); + cred->uc_flags = RPCAUTH_CRED_DEAD|RPCAUTH_CRED_UPTODATE; + cred->uc_uid = uid; + cred->uc_gid = gid; +@@ -236,7 +239,11 @@ + #endif + unx_create, + unx_destroy, +- unx_create_cred, ++ unx_create_cred ++}; ++ ++static ++struct rpc_credops unix_credops = { + unx_destroy_cred, + unx_match, + unx_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/sunrpc_syms.c linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c +--- linux-2.4.6-file/net/sunrpc/sunrpc_syms.c Fri Jan 12 00:53:02 2001 ++++ linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c Wed Jul 4 20:55:19 2001 +@@ -65,7 +65,7 @@ + EXPORT_SYMBOL(rpcauth_lookupcred); + EXPORT_SYMBOL(rpcauth_bindcred); + EXPORT_SYMBOL(rpcauth_matchcred); +-EXPORT_SYMBOL(rpcauth_releasecred); ++EXPORT_SYMBOL(put_rpccred); + + /* RPC server stuff */ + EXPORT_SYMBOL(svc_create); + +_______________________________________________ +NFS maillist - NFS@lists.sourceforge.net +http://lists.sourceforge.net/lists/listinfo/nfs + + diff --git a/sys-kernel/linux-extras/files/2.4.6-r2/lvm.c.diff b/sys-kernel/linux-extras/files/2.4.6-r2/lvm.c.diff new file mode 100644 index 000000000000..6392671b8f7e --- /dev/null +++ b/sys-kernel/linux-extras/files/2.4.6-r2/lvm.c.diff @@ -0,0 +1,20 @@ +--- lvm.c.orig Sat Apr 28 10:48:37 2001 ++++ lvm.c Sat Apr 28 11:07:07 2001 +@@ -1791,7 +1791,7 @@ + int max_hardblocksize = 0, hardblocksize; + + for (le = 0; le < lv->lv_allocated_le; le++) { +- hardblocksize = get_hardblocksize(lv->lv_current_pe[le].dev); ++ hardblocksize = get_hardsect_size(lv->lv_current_pe[le].dev); + if (hardblocksize == 0) + hardblocksize = 512; + if (hardblocksize > max_hardblocksize) +@@ -1801,7 +1801,7 @@ + if (lv->lv_access & LV_SNAPSHOT) { + for (e = 0; e < lv->lv_remap_end; e++) { + hardblocksize = +- get_hardblocksize( ++ get_hardsect_size( + lv->lv_block_exception[e].rdev_new); + if (hardblocksize == 0) + hardblocksize = 512; diff --git a/sys-kernel/linux-extras/files/2.4.6-r2/mosix.init b/sys-kernel/linux-extras/files/2.4.6-r2/mosix.init new file mode 100644 index 000000000000..f339dabb666e --- /dev/null +++ b/sys-kernel/linux-extras/files/2.4.6-r2/mosix.init @@ -0,0 +1,118 @@ +#!/bin/sh - +#RCUPDATE:2 3 4:95:This line is required for script management + +# description: MOSIX is am extension of the operating system, +# supporting scalable and transparent cluster computing. +# +# mosix Script to stop/start MOSIX +# +# Author: Amnon Shiloh + +. /etc/rc.d/config/functions + +SERVICE="mosix" +opts="start stop status restart" + +mosix_stop() { + echo 0 > /proc/mosix/admin/mospe + rm -f /var/lock/subsys/mosix +} + +mosix_start() { + if [ -s /etc/overheads -a -f /proc/mosix/admin/overheads ] + then + /bin/grep -v '^#' /etc/overheads > /proc/mosix/admin/overheads + fi + if [ -s /etc/mfscosts -a -f /proc/mosix/admin/mfscosts ] + then + /bin/grep -v '^#' /etc/mfscosts > /proc/mosix/admin/mfscosts + fi + a1= + [ -s /etc/mospe ] && a1="-p `cat /etc/mospe`" + a2= + [ -s /etc/mosgates ] && a2="-g `cat /etc/mosgates`" + /sbin/setpe -W $a1 $a2 -f /etc/mosix.map + touch /var/lock/subsys/mosix +} + +alarm() { + t=$2 + while : + do + case "$t" in 0) break ;; esac + sleep 1 + t=$(($t-1)) + done + kill -1 $1 + exit 0 +} + +# See how we were called. +start() { + eecho "Initializing MOSIX" + if [ ! -f /etc/mosix.map ] + then + echo You have no MOSIX configuration - please edit yours now: + echo -n "Editor to use [q to quit] - [vi] :- " + trap 'echo Timed-Out ; exit 1' 1 + alarm $$ 120 & + killer=$! + read editor + kill $killer + trap '' + case "$editor" in [qQ]) exit ;; + "") editor=/usr/bin/vi + [ -f $editor ] || editor=/bin/vi + ;; + esac + echo "# MOSIX CONFIGURATION" > /etc/mosix.map + echo "# ===================" >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# Each line should contain 3 fields, mapping IP addresses to MOSIX node-numbers:" >> /etc/mosix.map + echo "# 1) first MOSIX node-number in range." >> /etc/mosix.map + echo "# 2) IP address of the above node (or node-name from /etc/hosts)." >> /etc/mosix.map + echo "# 3) number of nodes in this range." >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# MOSIX-# IP number-of-nodes" >> /etc/mosix.map + echo "# ============================" >> /etc/mosix.map + $editor /etc/mosix.map + echo + while : + do + echo "If this node's standard IP address is not part of the table that" + echo "you just edited, because MOSIX uses a different or separate network," + echo "you need to type this node's MOSIX-number now." + echo + /bin/echo -n "Otherwise please press only <Enter> :- " + read me + case "$me" in "") /bin/rm -f /etc/mospe ; break ;; + [1-9] | [1-9][0-9] | [1-9][0-9][0-9] | \ + [1-9][0-9][0-9][0-9] | \ + [1-6][0-9][0-9][0-9][0-9]) + echo $me > /etc/mospe ; break ;; + *) echo Improper MOSIX number - please try again: ;; + esac + done + fi + start_mosix + eend +} + +stop() { + eecho "Stopping MOSIX" + stop_mosix + eend +} + +status() { + /sbin/setpe -r +} + +restart() { + eecho "Restarting MOSIX" + stop_mosix + start_mosix + eend +} + +doservice ${@} diff --git a/sys-kernel/linux-extras/files/2.4.6-r2/sched.h b/sys-kernel/linux-extras/files/2.4.6-r2/sched.h new file mode 100644 index 000000000000..438203f84cdd --- /dev/null +++ b/sys-kernel/linux-extras/files/2.4.6-r2/sched.h @@ -0,0 +1,997 @@ +#ifndef _LINUX_SCHED_H +#define _LINUX_SCHED_H + +#include <asm/param.h> /* for HZ */ + +extern unsigned long event; + +#include <linux/config.h> +#include <linux/binfmts.h> +#include <linux/personality.h> +#include <linux/threads.h> +#include <linux/kernel.h> +#include <linux/types.h> +#include <linux/times.h> +#include <linux/timex.h> + +#include <asm/system.h> +#include <asm/semaphore.h> +#include <asm/page.h> +#include <asm/ptrace.h> +#include <asm/mmu.h> + +#include <linux/smp.h> +#include <linux/tty.h> +#include <linux/sem.h> +#include <linux/signal.h> +#include <linux/securebits.h> + +#ifdef CONFIG_MOSIX +#include <mos/mosixtask.h> +#else +#define MOSIX_INIT_TASK(tsk) +#endif CONFIG_MOSIX + +#include <linux/fs_struct.h> + +/* + * cloning flags: + */ +#define CSIGNAL 0x000000ff /* signal mask to be sent at exit */ +#define CLONE_VM 0x00000100 /* set if VM shared between processes */ +#define CLONE_FS 0x00000200 /* set if fs info shared between processes */ +#define CLONE_FILES 0x00000400 /* set if open files shared between processes */ +#define CLONE_SIGHAND 0x00000800 /* set if signal handlers and blocked signals shared */ +#define CLONE_PID 0x00001000 /* set if pid shared */ +#define CLONE_PTRACE 0x00002000 /* set if we want to let tracing continue on the child too */ +#define CLONE_VFORK 0x00004000 /* set if the parent wants the child to wake it up on mm_release */ +#define CLONE_PARENT 0x00008000 /* set if we want to have the same parent as the cloner */ +#define CLONE_THREAD 0x00010000 /* Same thread group? */ + +#define CLONE_SIGNAL (CLONE_SIGHAND | CLONE_THREAD) + +/* + * These are the constant used to fake the fixed-point load-average + * counting. Some notes: + * - 11 bit fractions expand to 22 bits by the multiplies: this gives + * a load-average precision of 10 bits integer + 11 bits fractional + * - if you want to count load-averages more often, you need more + * precision, or rounding will get you. With 2-second counting freq, + * the EXP_n values would be 1981, 2034 and 2043 if still using only + * 11 bit fractions. + */ +extern unsigned long avenrun[]; /* Load averages */ + +#define FSHIFT 11 /* nr of bits of precision */ +#define FIXED_1 (1<<FSHIFT) /* 1.0 as fixed-point */ +#define LOAD_FREQ (5*HZ) /* 5 sec intervals */ +#define EXP_1 1884 /* 1/exp(5sec/1min) as fixed-point */ +#define EXP_5 2014 /* 1/exp(5sec/5min) */ +#define EXP_15 2037 /* 1/exp(5sec/15min) */ + +#define CALC_LOAD(load,exp,n) \ + load *= exp; \ + load += n*(FIXED_1-exp); \ + load >>= FSHIFT; + +#define CT_TO_SECS(x) ((x) / HZ) +#define CT_TO_USECS(x) (((x) % HZ) * 1000000/HZ) + +extern int nr_running, nr_threads; +extern int last_pid; + +#include <linux/fs.h> +#include <linux/time.h> +#include <linux/param.h> +#include <linux/resource.h> +#include <linux/timer.h> + +#include <asm/processor.h> + +#define TASK_RUNNING 0 +#define TASK_INTERRUPTIBLE 1 +#define TASK_UNINTERRUPTIBLE 2 +#define TASK_ZOMBIE 4 +#define TASK_STOPPED 8 +#ifdef CONFIG_MOSIX +#define TASK_SAME 16 +#endif CONFIG_MOSIX + +#define __set_task_state(tsk, state_value) \ + do { (tsk)->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_task_state(tsk, state_value) \ + set_mb((tsk)->state, (state_value)) +#else +#define set_task_state(tsk, state_value) \ + __set_task_state((tsk), (state_value)) +#endif + +#define __set_current_state(state_value) \ + do { current->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_current_state(state_value) \ + set_mb(current->state, (state_value)) +#else +#define set_current_state(state_value) \ + __set_current_state(state_value) +#endif + +/* + * Scheduling policies + */ +#define SCHED_OTHER 0 +#define SCHED_FIFO 1 +#define SCHED_RR 2 + +/* + * This is an additional bit set when we want to + * yield the CPU for one re-schedule.. + */ +#define SCHED_YIELD 0x10 + +struct sched_param { + int sched_priority; +}; + +#ifdef __KERNEL__ + +#include <linux/spinlock.h> + +/* + * This serializes "schedule()" and also protects + * the run-queue from deletions/modifications (but + * _adding_ to the beginning of the run-queue has + * a separate lock). + */ +extern rwlock_t tasklist_lock; +extern spinlock_t runqueue_lock; +extern spinlock_t mmlist_lock; + +extern void sched_init(void); +extern void init_idle(void); +extern void show_state(void); +extern void cpu_init (void); +extern void trap_init(void); +extern void update_process_times(int user); +extern void update_one_process(struct task_struct *p, unsigned long user, + unsigned long system, int cpu); + +#define MAX_SCHEDULE_TIMEOUT LONG_MAX +extern signed long FASTCALL(schedule_timeout(signed long timeout)); +asmlinkage void schedule(void); + +extern int schedule_task(struct tq_struct *task); +extern void flush_scheduled_tasks(void); +extern int start_context_thread(void); +extern int current_is_keventd(void); + +/* + * The default fd array needs to be at least BITS_PER_LONG, + * as this is the granularity returned by copy_fdset(). + */ +#define NR_OPEN_DEFAULT BITS_PER_LONG + +/* + * Open file table structure + */ +struct files_struct { + atomic_t count; + rwlock_t file_lock; /* Protects all the below members. Nests inside tsk->alloc_lock */ + int max_fds; + int max_fdset; + int next_fd; + struct file ** fd; /* current fd array */ + fd_set *close_on_exec; + fd_set *open_fds; + fd_set close_on_exec_init; + fd_set open_fds_init; + struct file * fd_array[NR_OPEN_DEFAULT]; +#ifdef CONFIG_MOSIX_DFSA + atomic_t users; /* the real count, excluding /proc grabs */ + int count_dropped_to_one; + int maxclosed; + fd_set *closed; + fd_set closed_fds_init; + int maxopened; + fd_set *opened; + fd_set opened_fds_init; + int maxmod; + fd_set *modified; + fd_set modified_fds_init; +#define INIT_FILES_DFSA , users : ATOMIC_INIT(1), \ + closed: &init_files.closed_fds_init, \ + opened: &init_files.opened_fds_init, \ + modified: &init_files.modified_fds_init, +#else +#define INIT_FILES_DFSA +#endif CONFIG_MOSIX_DFSA +}; + +#define INIT_FILES \ +{ \ + count: ATOMIC_INIT(1), \ + file_lock: RW_LOCK_UNLOCKED, \ + max_fds: NR_OPEN_DEFAULT, \ + max_fdset: __FD_SETSIZE, \ + next_fd: 0, \ + fd: &init_files.fd_array[0], \ + close_on_exec: &init_files.close_on_exec_init, \ + open_fds: &init_files.open_fds_init, \ + close_on_exec_init: { { 0, } }, \ + open_fds_init: { { 0, } }, \ + fd_array: { NULL, } \ + INIT_FILES_DFSA \ +} + +/* Maximum number of active map areas.. This is a random (large) number */ +#define MAX_MAP_COUNT (65536) + +/* Number of map areas at which the AVL tree is activated. This is arbitrary. */ +#define AVL_MIN_MAP_COUNT 32 + +struct mm_struct { + struct vm_area_struct * mmap; /* list of VMAs */ + struct vm_area_struct * mmap_avl; /* tree of VMAs */ + struct vm_area_struct * mmap_cache; /* last find_vma result */ + pgd_t * pgd; + atomic_t mm_users; /* How many users with user space? */ + atomic_t mm_count; /* How many references to "struct mm_struct" (users count as 1) */ + int map_count; /* number of VMAs */ + struct rw_semaphore mmap_sem; + spinlock_t page_table_lock; /* Protects task page tables and mm->rss */ + + struct list_head mmlist; /* List of all active mm's. These are globally strung + * together off init_mm.mmlist, and are protected + * by mmlist_lock + */ + + unsigned long start_code, end_code, start_data, end_data; + unsigned long start_brk, brk, start_stack; + unsigned long arg_start, arg_end, env_start, env_end; + unsigned long rss, total_vm, locked_vm; + unsigned long def_flags; + unsigned long cpu_vm_mask; + unsigned long swap_address; + + /* Architecture-specific MM context */ + mm_context_t context; +#ifdef CONFIG_MOSIX + atomic_t mm_kiocount; + atomic_t mm_realusers; + unsigned long last_memsort; + int used, unused, swapped; /* break-down of memory pages */ + int private_unused; /* unused but not active/inactive */ + int mark; +#define INIT_MM_MOSIX mm_kiocount: ATOMIC_INIT(0), mm_realusers: ATOMIC_INIT(1), +#else +#define INIT_MM_MOSIX +#endif CONFIG_MOSIX +}; + +extern int mmlist_nr; + +#define INIT_MM(name) \ +{ \ + mmap: &init_mmap, \ + mmap_avl: NULL, \ + mmap_cache: NULL, \ + pgd: swapper_pg_dir, \ + mm_users: ATOMIC_INIT(2), \ + mm_count: ATOMIC_INIT(1), \ + map_count: 1, \ + mmap_sem: __RWSEM_INITIALIZER(name.mmap_sem), \ + page_table_lock: SPIN_LOCK_UNLOCKED, \ + mmlist: LIST_HEAD_INIT(name.mmlist), \ + INIT_MM_MOSIX \ +} + +struct signal_struct { + atomic_t count; + struct k_sigaction action[_NSIG]; + spinlock_t siglock; +}; + + +#define INIT_SIGNALS { \ + count: ATOMIC_INIT(1), \ + action: { {{0,}}, }, \ + siglock: SPIN_LOCK_UNLOCKED \ +} + +/* + * Some day this will be a full-fledged user tracking system.. + */ +struct user_struct { + atomic_t __count; /* reference count */ + atomic_t processes; /* How many processes does this user have? */ + atomic_t files; /* How many open files does this user have? */ + + /* Hash table maintenance information */ + struct user_struct *next, **pprev; + uid_t uid; +}; + +#define get_current_user() ({ \ + struct user_struct *__user = current->user; \ + atomic_inc(&__user->__count); \ + __user; }) + +extern struct user_struct root_user; +#define INIT_USER (&root_user) + +struct task_struct { + /* + * offsets of these are hardcoded elsewhere - touch with care + */ + volatile long state; /* -1 unrunnable, 0 runnable, >0 stopped */ + unsigned long flags; /* per process flags, defined below */ + int sigpending; + mm_segment_t addr_limit; /* thread address space: + 0-0xBFFFFFFF for user-thead + 0-0xFFFFFFFF for kernel-thread + */ + struct exec_domain *exec_domain; + volatile long need_resched; + unsigned long ptrace; + + int lock_depth; /* Lock depth */ + +/* + * offset 32 begins here on 32-bit platforms. We keep + * all fields in a single cacheline that are needed for + * the goodness() loop in schedule(). + */ + long counter; + long nice; + unsigned long policy; + struct mm_struct *mm; + int has_cpu, processor; + unsigned long cpus_allowed; + /* + * (only the 'next' pointer fits into the cacheline, but + * that's just fine.) + */ + struct list_head run_list; + unsigned long sleep_time; + + struct task_struct *next_task, *prev_task; + struct mm_struct *active_mm; + +/* task state */ + struct linux_binfmt *binfmt; + int exit_code, exit_signal; + int pdeath_signal; /* The signal sent when the parent dies */ + /* ??? */ + unsigned long personality; + int dumpable:1; + int did_exec:1; + pid_t pid; + pid_t pgrp; + pid_t tty_old_pgrp; + pid_t session; + pid_t tgid; + /* boolean value for session group leader */ + int leader; + /* + * pointers to (original) parent process, youngest child, younger sibling, + * older sibling, respectively. (p->father can be replaced with + * p->p_pptr->pid) + */ + struct task_struct *p_opptr, *p_pptr, *p_cptr, *p_ysptr, *p_osptr; + struct list_head thread_group; + + /* PID hash table linkage. */ + struct task_struct *pidhash_next; + struct task_struct **pidhash_pprev; + + wait_queue_head_t wait_chldexit; /* for wait4() */ + struct semaphore *vfork_sem; /* for vfork() */ + unsigned long rt_priority; + unsigned long it_real_value, it_prof_value, it_virt_value; + unsigned long it_real_incr, it_prof_incr, it_virt_incr; + struct timer_list real_timer; + struct tms times; + unsigned long start_time; + long per_cpu_utime[NR_CPUS], per_cpu_stime[NR_CPUS]; +/* mm fault and swap info: this can arguably be seen as either mm-specific or thread-specific */ + unsigned long min_flt, maj_flt, nswap, cmin_flt, cmaj_flt, cnswap; + int swappable:1; +/* process credentials */ + uid_t uid,euid,suid,fsuid; + gid_t gid,egid,sgid,fsgid; + int ngroups; + gid_t groups[NGROUPS]; + kernel_cap_t cap_effective, cap_inheritable, cap_permitted; + int keep_capabilities:1; + struct user_struct *user; +/* limits */ + struct rlimit rlim[RLIM_NLIMITS]; + unsigned short used_math; + char comm[16]; +/* file system info */ + int link_count; + struct tty_struct *tty; /* NULL if no tty */ + unsigned int locks; /* How many file locks are being held */ +/* ipc stuff */ + struct sem_undo *semundo; + struct sem_queue *semsleeping; +/* CPU-specific state of this task */ + struct thread_struct thread; +/* filesystem information */ + struct fs_struct *fs; +/* open file information */ + struct files_struct *files; +/* signal handlers */ + spinlock_t sigmask_lock; /* Protects signal and blocked */ + struct signal_struct *sig; + + sigset_t blocked; + struct sigpending pending; + + unsigned long sas_ss_sp; + size_t sas_ss_size; + int (*notifier)(void *priv); + void *notifier_data; + sigset_t *notifier_mask; + +/* Thread group tracking */ + u32 parent_exec_id; + u32 self_exec_id; +/* Protection of (de-)allocation: mm, files, fs, tty */ + spinlock_t alloc_lock; +#ifdef CONFIG_MOSIX + struct mosix_task mosix; +#endif CONFIG_MOSIX + void *journal_info; +}; + +/* + * Per process flags + */ +#define PF_ALIGNWARN 0x00000001 /* Print alignment warning msgs */ + /* Not implemented yet, only for 486*/ +#define PF_STARTING 0x00000002 /* being created */ +#define PF_EXITING 0x00000004 /* getting shut down */ +#define PF_FORKNOEXEC 0x00000040 /* forked but didn't exec */ +#define PF_SUPERPRIV 0x00000100 /* used super-user privileges */ +#define PF_DUMPCORE 0x00000200 /* dumped core */ +#define PF_SIGNALED 0x00000400 /* killed by a signal */ +#define PF_MEMALLOC 0x00000800 /* Allocating memory */ +#define PF_VFORK 0x00001000 /* Wake up parent in mm_release */ + +#define PF_USEDFPU 0x00100000 /* task used FPU this quantum (SMP) */ + +/* + * Ptrace flags + */ + +#define PT_PTRACED 0x00000001 +#define PT_TRACESYS 0x00000002 +#define PT_DTRACE 0x00000004 /* delayed trace (used on m68k, i386) */ +#define PT_TRACESYSGOOD 0x00000008 + +/* + * Limit the stack by to some sane default: root can always + * increase this limit if needed.. 8MB seems reasonable. + */ +#define _STK_LIM (8*1024*1024) + +#define DEF_COUNTER (10*HZ/100) /* 100 ms time slice */ +#define MAX_COUNTER (20*HZ/100) +#define DEF_NICE (0) + +/* + * INIT_TASK is used to set up the first task table, touch at + * your own risk!. Base=0, limit=0x1fffff (=2MB) + */ +#define INIT_TASK(tsk) \ +{ \ + state: 0, \ + flags: 0, \ + sigpending: 0, \ + addr_limit: KERNEL_DS, \ + exec_domain: &default_exec_domain, \ + lock_depth: -1, \ + counter: DEF_COUNTER, \ + nice: DEF_NICE, \ + policy: SCHED_OTHER, \ + mm: NULL, \ + active_mm: &init_mm, \ + cpus_allowed: -1, \ + run_list: LIST_HEAD_INIT(tsk.run_list), \ + next_task: &tsk, \ + prev_task: &tsk, \ + p_opptr: &tsk, \ + p_pptr: &tsk, \ + thread_group: LIST_HEAD_INIT(tsk.thread_group), \ + wait_chldexit: __WAIT_QUEUE_HEAD_INITIALIZER(tsk.wait_chldexit),\ + real_timer: { \ + function: it_real_fn \ + }, \ + cap_effective: CAP_INIT_EFF_SET, \ + cap_inheritable: CAP_INIT_INH_SET, \ + cap_permitted: CAP_FULL_SET, \ + keep_capabilities: 0, \ + rlim: INIT_RLIMITS, \ + user: INIT_USER, \ + comm: "swapper", \ + thread: INIT_THREAD, \ + fs: &init_fs, \ + files: &init_files, \ + sigmask_lock: SPIN_LOCK_UNLOCKED, \ + sig: &init_signals, \ + pending: { NULL, &tsk.pending.head, {{0}}}, \ + blocked: {{0}}, \ + alloc_lock: SPIN_LOCK_UNLOCKED, \ + MOSIX_INIT_TASK(tsk), \ + journal_info: NULL \ +} + + +#ifndef INIT_TASK_SIZE +# define INIT_TASK_SIZE 2048*sizeof(long) +#endif + +union task_union { + struct task_struct task; + unsigned long stack[INIT_TASK_SIZE/sizeof(long)]; +}; + +extern union task_union init_task_union; + +extern struct mm_struct init_mm; +extern struct task_struct *init_tasks[NR_CPUS]; + +/* PID hashing. (shouldnt this be dynamic?) */ +#define PIDHASH_SZ (4096 >> 2) +extern struct task_struct *pidhash[PIDHASH_SZ]; + +#define pid_hashfn(x) ((((x) >> 8) ^ (x)) & (PIDHASH_SZ - 1)) + +static inline void hash_pid(struct task_struct *p) +{ + struct task_struct **htable = &pidhash[pid_hashfn(p->pid)]; + + if((p->pidhash_next = *htable) != NULL) + (*htable)->pidhash_pprev = &p->pidhash_next; + *htable = p; + p->pidhash_pprev = htable; +} + +static inline void unhash_pid(struct task_struct *p) +{ + if(p->pidhash_next) + p->pidhash_next->pidhash_pprev = p->pidhash_pprev; + *p->pidhash_pprev = p->pidhash_next; +} + +static inline struct task_struct *find_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; +#ifdef CONFIG_MOSIX + if(p && (p->mosix.dflags & DREMOTE)) + p = NULL; +#endif CONFIG_MOSIX + return p; +} + +#ifdef CONFIG_MOSIX +static inline struct task_struct *find_any_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; + return p; +} +#endif CONFIG_MOSIX + +/* per-UID process charging. */ +extern struct user_struct * alloc_uid(uid_t); +extern void free_uid(struct user_struct *); + +#include <asm/current.h> + +extern unsigned long volatile jiffies; +extern unsigned long itimer_ticks; +extern unsigned long itimer_next; +extern struct timeval xtime; +extern void do_timer(struct pt_regs *); + +extern unsigned int * prof_buffer; +extern unsigned long prof_len; +extern unsigned long prof_shift; + +#define CURRENT_TIME (xtime.tv_sec) + +extern void FASTCALL(__wake_up(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(__wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern void FASTCALL(interruptible_sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(interruptible_sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern int FASTCALL(wake_up_process(struct task_struct * tsk)); + +#define wake_up(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_nr(x, nr) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_all(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 0) +#define wake_up_sync(x) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_sync_nr(x, nr) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible(x) __wake_up((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_nr(x, nr) __wake_up((x),TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible_all(x) __wake_up((x),TASK_INTERRUPTIBLE, 0) +#define wake_up_interruptible_sync(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_sync_nr(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, nr) +asmlinkage long sys_wait4(pid_t pid,unsigned int * stat_addr, int options, struct rusage * ru); + +extern int in_group_p(gid_t); +extern int in_egroup_p(gid_t); + +extern void proc_caches_init(void); +extern void flush_signals(struct task_struct *); +extern void flush_signal_handlers(struct task_struct *); +extern int dequeue_signal(sigset_t *, siginfo_t *); +extern void block_all_signals(int (*notifier)(void *priv), void *priv, + sigset_t *mask); +extern void unblock_all_signals(void); +extern int send_sig_info(int, struct siginfo *, struct task_struct *); +extern int force_sig_info(int, struct siginfo *, struct task_struct *); +extern int kill_pg_info(int, struct siginfo *, pid_t); +extern int kill_sl_info(int, struct siginfo *, pid_t); +extern int kill_proc_info(int, struct siginfo *, pid_t); +extern void notify_parent(struct task_struct *, int); +extern void do_notify_parent(struct task_struct *, int); +extern void force_sig(int, struct task_struct *); +extern int send_sig(int, struct task_struct *, int); +extern int kill_pg(pid_t, int, int); +extern int kill_sl(pid_t, int, int); +extern int kill_proc(pid_t, int, int); +extern int do_sigaction(int, const struct k_sigaction *, struct k_sigaction *); +extern int do_sigaltstack(const stack_t *, stack_t *, unsigned long); + +static inline int signal_pending(struct task_struct *p) +{ + return (p->sigpending != 0); +} + +/* + * Re-calculate pending state from the set of locally pending + * signals, globally pending signals, and blocked signals. + */ +static inline int has_pending_signals(sigset_t *signal, sigset_t *blocked) +{ + unsigned long ready; + long i; + + switch (_NSIG_WORDS) { + default: + for (i = _NSIG_WORDS, ready = 0; --i >= 0 ;) + ready |= signal->sig[i] &~ blocked->sig[i]; + break; + + case 4: ready = signal->sig[3] &~ blocked->sig[3]; + ready |= signal->sig[2] &~ blocked->sig[2]; + ready |= signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 2: ready = signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 1: ready = signal->sig[0] &~ blocked->sig[0]; + } + return ready != 0; +} + +/* Reevaluate whether the task has signals pending delivery. + This is required every time the blocked sigset_t changes. + All callers should have t->sigmask_lock. */ + +static inline void recalc_sigpending(struct task_struct *t) +{ +#ifdef CONFIG_MOSIX + t->mosix.ignoreoldsigs = 0; + if(t->mosix.dflags & DHEAVYSLEEP) + t->sigpending = 0; + else if(t->mosix.dflags & DFAKESIGNAL) + t->sigpending = 1; + else +#endif CONFIG_MOSIX + t->sigpending = has_pending_signals(&t->pending.signal, &t->blocked); +} + +/* True if we are on the alternate signal stack. */ + +static inline int on_sig_stack(unsigned long sp) +{ + return (sp - current->sas_ss_sp < current->sas_ss_size); +} + +static inline int sas_ss_flags(unsigned long sp) +{ + return (current->sas_ss_size == 0 ? SS_DISABLE + : on_sig_stack(sp) ? SS_ONSTACK : 0); +} + +extern int request_irq(unsigned int, + void (*handler)(int, void *, struct pt_regs *), + unsigned long, const char *, void *); +extern void free_irq(unsigned int, void *); + +/* + * This has now become a routine instead of a macro, it sets a flag if + * it returns true (to do BSD-style accounting where the process is flagged + * if it uses root privs). The implication of this is that you should do + * normal permissions checks first, and check suser() last. + * + * [Dec 1997 -- Chris Evans] + * For correctness, the above considerations need to be extended to + * fsuser(). This is done, along with moving fsuser() checks to be + * last. + * + * These will be removed, but in the mean time, when the SECURE_NOROOT + * flag is set, uids don't grant privilege. + */ +static inline int suser(void) +{ + if (!issecure(SECURE_NOROOT) && current->euid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +static inline int fsuser(void) +{ + if (!issecure(SECURE_NOROOT) && current->fsuid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * capable() checks for a particular capability. + * New privilege checks should use this interface, rather than suser() or + * fsuser(). See include/linux/capability.h for defined capabilities. + */ + +static inline int capable(int cap) +{ +#if 1 /* ok now */ + if (cap_raised(current->cap_effective, cap)) +#else + if (cap_is_fs_cap(cap) ? current->fsuid == 0 : current->euid == 0) +#endif + { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * Routines for handling mm_structs + */ +extern struct mm_struct * mm_alloc(void); + +extern struct mm_struct * start_lazy_tlb(void); +extern void end_lazy_tlb(struct mm_struct *mm); + +/* mmdrop drops the mm and the page tables */ +extern inline void FASTCALL(__mmdrop(struct mm_struct *)); +static inline void mmdrop(struct mm_struct * mm) +{ + if (atomic_dec_and_test(&mm->mm_count)) + __mmdrop(mm); +} + +/* mmput gets rid of the mappings and all user-space */ +extern void mmput(struct mm_struct *); +/* Remove the current tasks stale references to the old mm_struct */ +extern void mm_release(void); + +/* + * Routines for handling the fd arrays + */ +extern struct file ** alloc_fd_array(int); +extern int expand_fd_array(struct files_struct *, int nr); +extern void free_fd_array(struct file **, int); + +extern fd_set *alloc_fdset(int); +extern int expand_fdset(struct files_struct *, int nr); +extern void free_fdset(fd_set *, int); + +extern int copy_thread(int, unsigned long, unsigned long, unsigned long, struct task_struct *, struct pt_regs *); +extern void flush_thread(void); +extern void exit_thread(void); + +extern void exit_mm(struct task_struct *); +extern void exit_files(struct task_struct *); +extern void exit_sighand(struct task_struct *); + +extern void daemonize(void); + +extern int do_execve(char *, char **, char **, struct pt_regs *); +extern int do_fork(unsigned long, unsigned long, struct pt_regs *, unsigned long); + +extern void FASTCALL(add_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(add_wait_queue_exclusive(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(remove_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); + +#define __wait_event(wq, condition) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_UNINTERRUPTIBLE); \ + if (condition) \ + break; \ + schedule(); \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event(wq, condition) \ +do { \ + if (condition) \ + break; \ + __wait_event(wq, condition); \ +} while (0) + +#define __wait_event_interruptible(wq, condition, ret) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_INTERRUPTIBLE); \ + if (condition) \ + break; \ + if (!signal_pending(current)) { \ + schedule(); \ + continue; \ + } \ + ret = -ERESTARTSYS; \ + break; \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event_interruptible(wq, condition) \ +({ \ + int __ret = 0; \ + if (!(condition)) \ + __wait_event_interruptible(wq, condition, __ret); \ + __ret; \ +}) + +#define REMOVE_LINKS(p) do { \ + (p)->next_task->prev_task = (p)->prev_task; \ + (p)->prev_task->next_task = (p)->next_task; \ + if ((p)->p_osptr) \ + (p)->p_osptr->p_ysptr = (p)->p_ysptr; \ + if ((p)->p_ysptr) \ + (p)->p_ysptr->p_osptr = (p)->p_osptr; \ + else \ + (p)->p_pptr->p_cptr = (p)->p_osptr; \ + } while (0) + +#define SET_LINKS(p) do { \ + (p)->next_task = &init_task; \ + (p)->prev_task = init_task.prev_task; \ + init_task.prev_task->next_task = (p); \ + init_task.prev_task = (p); \ + (p)->p_ysptr = NULL; \ + if (((p)->p_osptr = (p)->p_pptr->p_cptr) != NULL) \ + (p)->p_osptr->p_ysptr = p; \ + (p)->p_pptr->p_cptr = p; \ + } while (0) + +#define for_each_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) + +#define next_thread(p) \ + list_entry((p)->thread_group.next, struct task_struct, thread_group) + +#ifdef CONFIG_MOSIX +#define for_each_local_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) \ + if(!(p->mosix.dflags & DREMOTE)) + +static inline long +LOGICAL_STATE(struct task_struct *p) +{ + int result; + unsigned long flags; + + read_lock_irqsave(&p->mosix.state_lock, flags); + result = (p->mosix.bstate == TASK_SAME) ? p->state : p->mosix.bstate; + read_unlock_irqrestore(&p->mosix.state_lock, flags); + return(result); +} + +extern void run_on(struct task_struct *); +extern void run_off(struct task_struct *); + +#define need_interim_while_asleep() mosix_need_while_asleep() +#define run_interim_while_asleep() mosix_run_while_asleep() +#endif CONFIG_MOSIX + +static inline void del_from_runqueue(struct task_struct * p) +{ + nr_running--; + p->sleep_time = jiffies; +#ifdef CONFIG_MOSIX + run_off(p); +#endif CONFIG_MOSIX + list_del(&p->run_list); + p->run_list.next = NULL; +} + +static inline int task_on_runqueue(struct task_struct *p) +{ + return (p->run_list.next != NULL); +} + +static inline void unhash_process(struct task_struct *p) +{ + if (task_on_runqueue(p)) BUG(); + write_lock_irq(&tasklist_lock); + nr_threads--; + unhash_pid(p); + REMOVE_LINKS(p); + list_del(&p->thread_group); + write_unlock_irq(&tasklist_lock); +} + +/* Protects ->fs, ->files, ->mm, and synchronises with wait4(). Nests inside tasklist_lock */ +static inline void task_lock(struct task_struct *p) +{ + spin_lock(&p->alloc_lock); +} + +static inline void task_unlock(struct task_struct *p) +{ + spin_unlock(&p->alloc_lock); +} + +/* write full pathname into buffer and return start of pathname */ +static inline char * d_path(struct dentry *dentry, struct vfsmount *vfsmnt, + char *buf, int buflen) +{ + char *res; + struct vfsmount *rootmnt; + struct dentry *root; +#if defined(CONFIG_MOSIX_DFSA) || defined(CONFIG_MOSIX_FS) + if(dentry->d_inode && dentry->d_inode->i_op && + dentry->d_inode->i_op->check_path) + dentry->d_inode->i_op->check_path(dentry); +#endif CONFIG_MOSIX_DFSA || CONFIG_MOSIX_FS + read_lock(¤t->fs->lock); + rootmnt = mntget(current->fs->rootmnt); + root = dget(current->fs->root); + read_unlock(¤t->fs->lock); + spin_lock(&dcache_lock); + res = __d_path(dentry, vfsmnt, root, rootmnt, buf, buflen); + spin_unlock(&dcache_lock); + dput(root); + mntput(rootmnt); + return res; +} + +#endif /* __KERNEL__ */ + +#endif diff --git a/sys-kernel/linux-extras/linux-extras-2.4.6-r2.ebuild b/sys-kernel/linux-extras/linux-extras-2.4.6-r2.ebuild index 6a686433b00d..55e4ed5de076 100644 --- a/sys-kernel/linux-extras/linux-extras-2.4.6-r2.ebuild +++ b/sys-kernel/linux-extras/linux-extras-2.4.6-r2.ebuild @@ -128,13 +128,13 @@ src_unpack() { cd MOSIX-${MOSV} tar xzf MOSIX-${MOSV}.tar.gz patches.${OKV} kernel.new.${OKV}.tar cd ${S} - try cat ${S2}/MOSIX-${MOSV}/patches.2.4.6 | patch -p0 - tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.2.4.6.tar + try cat ${S2}/MOSIX-${MOSV}/patches.${KV} | patch -p0 + tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.${KV}.tar fi cd ${S} echo "Applying reiserfs-NFS fix..." - try cat ${FILESDIR}/2.4.6/linux-2.4.6-reiserfs-NFS.patch | patch -N -p1 + try cat ${FILESDIR}/${PVR}/linux-${KV}-reiserfs-NFS.patch | patch -N -p1 if [ "`use lvm`" ] then @@ -146,14 +146,14 @@ src_unpack() { # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${S}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" + CFLAGS="${CFLAGS} -I${S}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" cd PATCHES try make KERNEL_VERSION=${KV} KERNEL_DIR=${S} cd ${S} # the -l option allows this patch to apply cleanly (ignore whitespace changes) try patch -l -p1 < ${S2}/LVM/${LVMV}/PATCHES/lvm-${LVMV}-${KV}.patch cd ${S}/drivers/md - try patch -p0 < ${FILESDIR}/${KV}/lvm.c.diff + try patch -p0 < ${FILESDIR}/${PVR}/lvm.c.diff fi # if [ "`use lm_sensors`" ] @@ -193,7 +193,7 @@ src_unpack() { # echo "Unpacking pcmcia-cs tools..." # cd ${S2} # unpack pcmcia-cs-${PCV}.tar.gz - # patch -p0 < ${FILESDIR}/${KV}/pcmcia-cs-${PCV}-gentoo.diff + # patch -p0 < ${FILESDIR}/${PVR}/pcmcia-cs-${PCV}-gentoo.diff # fi #JFS patch works; commented out because it's not ready for production use @@ -223,7 +223,7 @@ src_unpack() { echo echo "Fixing reject in include/linux/sched.h..." echo - cp ${FILESDIR}/${KV}/sched.h include/linux + cp ${FILESDIR}/${PVR}/sched.h include/linux fi fi @@ -271,8 +271,8 @@ src_compile() { fi # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${KS}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" - try make + CFLAGS="${CFLAGS} -I${KS}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" + try make fi # if [ "`use lm_sensors`" ] @@ -294,6 +294,7 @@ src_compile() { if [ "$PN" == "linux" ] then + cd ${KS} try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" dep try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" bzImage try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" modules @@ -326,7 +327,7 @@ src_install() { then cd ${KS2}/LVM/${LVMV}/tools - try CFLAGS="${CFLAGS} -I${KS}/include" make install -e prefix=${D} mandir=${D}/usr/share/man \ + CFLAGS="${CFLAGS} -I${KS}/include" try make install -e prefix=${D} mandir=${D}/usr/share/man \ sbindir=${D}/sbin libdir=${D}/lib #no need for a static library in /lib mv ${D}/lib/*.a ${D}/usr/lib diff --git a/sys-kernel/linux-headers/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch b/sys-kernel/linux-headers/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch new file mode 100644 index 000000000000..5f3830c8ffe1 --- /dev/null +++ b/sys-kernel/linux-headers/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch @@ -0,0 +1,647 @@ +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/inode.c linux-2.4.6-mmap/fs/nfs/inode.c +--- linux-2.4.6-file/fs/nfs/inode.c Wed Jul 4 17:30:28 2001 ++++ linux-2.4.6-mmap/fs/nfs/inode.c Wed Jul 4 20:48:01 2001 +@@ -48,6 +48,7 @@ + static void nfs_write_inode(struct inode *,int); + static void nfs_delete_inode(struct inode *); + static void nfs_put_super(struct super_block *); ++static void nfs_clear_inode(struct inode *); + static void nfs_umount_begin(struct super_block *); + static int nfs_statfs(struct super_block *, struct statfs *); + +@@ -57,6 +58,7 @@ + delete_inode: nfs_delete_inode, + put_super: nfs_put_super, + statfs: nfs_statfs, ++ clear_inode: nfs_clear_inode, + umount_begin: nfs_umount_begin, + }; + +@@ -141,6 +143,21 @@ + clear_inode(inode); + } + ++/* ++ * For the moment, the only task for the NFS clear_inode method is to ++ * release the mmap credential ++ */ ++static void ++nfs_clear_inode(struct inode *inode) ++{ ++ struct rpc_cred *cred = NFS_I(inode)->mm_cred; ++ ++ if (cred) { ++ put_rpccred(cred); ++ NFS_I(inode)->mm_cred = 0; ++ } ++} ++ + void + nfs_put_super(struct super_block *sb) + { +@@ -600,7 +617,6 @@ + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + NFS_CACHE_CTIME(inode) = fattr->ctime; + NFS_CACHE_MTIME(inode) = fattr->mtime; +- NFS_CACHE_ATIME(inode) = fattr->atime; + NFS_CACHE_ISIZE(inode) = fattr->size; + NFS_ATTRTIMEO(inode) = NFS_MINATTRTIMEO(inode); + NFS_ATTRTIMEO_UPDATE(inode) = jiffies; +@@ -794,6 +810,20 @@ + } + + /* ++ * Ensure that mmap has a recent RPC credential for use when writing out ++ * shared pages ++ */ ++static inline void ++nfs_set_mmcred(struct inode *inode, struct rpc_cred *cred) ++{ ++ struct rpc_cred *oldcred = NFS_I(inode)->mm_cred; ++ ++ NFS_I(inode)->mm_cred = get_rpccred(cred); ++ if (oldcred) ++ put_rpccred(oldcred); ++} ++ ++/* + * These are probably going to contain hooks for + * allocating and releasing RPC credentials for + * the file. I'll have to think about Tronds patch +@@ -808,20 +838,20 @@ + auth = NFS_CLIENT(inode)->cl_auth; + cred = rpcauth_lookupcred(auth, 0); + filp->private_data = cred; ++ if (filp->f_mode & FMODE_WRITE) ++ nfs_set_mmcred(inode, cred); + unlock_kernel(); + return 0; + } + + int nfs_release(struct inode *inode, struct file *filp) + { +- struct rpc_auth *auth; + struct rpc_cred *cred; + + lock_kernel(); +- auth = NFS_CLIENT(inode)->cl_auth; + cred = nfs_file_cred(filp); + if (cred) +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + unlock_kernel(); + return 0; + } +@@ -976,7 +1006,6 @@ + NFS_CACHE_CTIME(inode) = fattr->ctime; + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + +- NFS_CACHE_ATIME(inode) = fattr->atime; + inode->i_atime = nfs_time_to_secs(fattr->atime); + + NFS_CACHE_MTIME(inode) = new_mtime; +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/unlink.c linux-2.4.6-mmap/fs/nfs/unlink.c +--- linux-2.4.6-file/fs/nfs/unlink.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/fs/nfs/unlink.c Wed Jul 4 20:48:47 2001 +@@ -128,7 +128,7 @@ + dir_i = dir->d_inode; + nfs_zap_caches(dir_i); + NFS_PROTO(dir_i)->unlink_done(dir, &task->tk_msg); +- rpcauth_releasecred(task->tk_auth, data->cred); ++ put_rpccred(data->cred); + data->cred = NULL; + dput(dir); + } +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/write.c linux-2.4.6-mmap/fs/nfs/write.c +--- linux-2.4.6-file/fs/nfs/write.c Mon May 21 21:29:14 2001 ++++ linux-2.4.6-mmap/fs/nfs/write.c Wed Jul 4 20:48:24 2001 +@@ -181,7 +181,9 @@ + + + if (file) +- cred = nfs_file_cred(file); ++ cred = get_rpccred(nfs_file_cred(file)); ++ if (!cred) ++ cred = get_rpccred(NFS_I(inode)->mm_cred); + + dprintk("NFS: nfs_writepage_sync(%x/%Ld %d@%Ld)\n", + inode->i_dev, (long long)NFS_FILEID(inode), +@@ -226,6 +228,8 @@ + + io_error: + kunmap(page); ++ if (cred) ++ put_rpccred(cred); + + return written? written : result; + } +@@ -241,6 +245,9 @@ + status = (IS_ERR(req)) ? PTR_ERR(req) : 0; + if (status < 0) + goto out; ++ if (!req->wb_cred) ++ req->wb_cred = get_rpccred(NFS_I(inode)->mm_cred); ++ nfs_unlock_request(req); + nfs_release_request(req); + nfs_strategy(inode); + out: +@@ -557,13 +564,11 @@ + req->wb_bytes = count; + req->wb_file = file; + +- /* If we have a struct file, use its cached credentials +- * else cache the current process' credentials. */ ++ /* If we have a struct file, use its cached credentials */ + if (file) { + get_file(file); + req->wb_cred = nfs_file_cred(file); +- } else +- req->wb_cred = rpcauth_lookupcred(NFS_CLIENT(inode)->cl_auth, 0); ++ } + req->wb_inode = inode; + req->wb_count = 1; + +@@ -608,8 +613,8 @@ + /* Release struct file or cached credential */ + if (req->wb_file) + fput(req->wb_file); +- else +- rpcauth_releasecred(NFS_CLIENT(inode)->cl_auth, req->wb_cred); ++ else if (req->wb_cred) ++ put_rpccred(req->wb_cred); + page_cache_release(page); + nfs_page_free(req); + /* wake up anyone waiting to allocate a request */ +@@ -927,8 +932,6 @@ + if (end > rqend) + req->wb_bytes = end - req->wb_offset; + +- nfs_unlock_request(req); +- + return req; + } + +@@ -1049,6 +1052,7 @@ + goto done; + + status = 0; ++ nfs_unlock_request(req); + /* If we wrote past the end of the page. + * Call the strategy routine so it can send out a bunch + * of requests. +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs.h linux-2.4.6-mmap/include/linux/nfs_fs.h +--- linux-2.4.6-file/include/linux/nfs_fs.h Sat May 26 03:02:11 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs.h Wed Jul 4 17:32:16 2001 +@@ -63,6 +63,11 @@ + */ + #define NFS_SUPER_MAGIC 0x6969 + ++static inline struct nfs_inode_info *NFS_I(struct inode *inode) ++{ ++ return &inode->u.nfs_i; ++} ++ + #define NFS_FH(inode) (&(inode)->u.nfs_i.fh) + #define NFS_SERVER(inode) (&(inode)->i_sb->u.nfs_sb.s_server) + #define NFS_CLIENT(inode) (NFS_SERVER(inode)->client) +@@ -74,7 +79,6 @@ + #define NFS_READTIME(inode) ((inode)->u.nfs_i.read_cache_jiffies) + #define NFS_CACHE_CTIME(inode) ((inode)->u.nfs_i.read_cache_ctime) + #define NFS_CACHE_MTIME(inode) ((inode)->u.nfs_i.read_cache_mtime) +-#define NFS_CACHE_ATIME(inode) ((inode)->u.nfs_i.read_cache_atime) + #define NFS_CACHE_ISIZE(inode) ((inode)->u.nfs_i.read_cache_isize) + #define NFS_NEXTSCAN(inode) ((inode)->u.nfs_i.nextscan) + #define NFS_CACHEINV(inode) \ +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs_i.h linux-2.4.6-mmap/include/linux/nfs_fs_i.h +--- linux-2.4.6-file/include/linux/nfs_fs_i.h Thu May 31 15:27:10 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs_i.h Wed Jul 4 17:32:16 2001 +@@ -45,7 +45,6 @@ + unsigned long read_cache_jiffies; + __u64 read_cache_ctime; + __u64 read_cache_mtime; +- __u64 read_cache_atime; + __u64 read_cache_isize; + unsigned long attrtimeo; + unsigned long attrtimeo_timestamp; +@@ -73,6 +72,9 @@ + struct inode *hash_next, + *hash_prev; + unsigned long nextscan; ++ ++ /* Credentials for shared mmap */ ++ struct rpc_cred *mm_cred; + }; + + /* +diff -u --recursive --new-file linux-2.4.6-file/include/linux/sunrpc/auth.h linux-2.4.6-mmap/include/linux/sunrpc/auth.h +--- linux-2.4.6-file/include/linux/sunrpc/auth.h Fri Jun 22 19:30:22 2001 ++++ linux-2.4.6-mmap/include/linux/sunrpc/auth.h Wed Jul 4 20:51:35 2001 +@@ -14,6 +14,8 @@ + #include <linux/config.h> + #include <linux/sunrpc/sched.h> + ++#include <asm/atomic.h> ++ + /* size of the nodename buffer */ + #define UNX_MAXNODENAME 32 + +@@ -22,8 +24,10 @@ + */ + struct rpc_cred { + struct rpc_cred * cr_next; /* linked list */ ++ struct rpc_auth * cr_auth; ++ struct rpc_credops * cr_ops; + unsigned long cr_expire; /* when to gc */ +- unsigned short cr_count; /* ref count */ ++ atomic_t cr_count; /* ref count */ + unsigned short cr_flags; /* various flags */ + #ifdef RPC_DEBUG + unsigned long cr_magic; /* 0x0f4aa4f0 */ +@@ -71,6 +75,9 @@ + void (*destroy)(struct rpc_auth *); + + struct rpc_cred * (*crcreate)(int); ++}; ++ ++struct rpc_credops { + void (*crdestroy)(struct rpc_cred *); + + int (*crmatch)(struct rpc_cred *, int); +@@ -92,8 +99,7 @@ + struct rpc_cred * rpcauth_lookupcred(struct rpc_auth *, int); + struct rpc_cred * rpcauth_bindcred(struct rpc_task *); + void rpcauth_holdcred(struct rpc_task *); +-void rpcauth_releasecred(struct rpc_auth *, +- struct rpc_cred *); ++void put_rpccred(struct rpc_cred *); + void rpcauth_unbindcred(struct rpc_task *); + int rpcauth_matchcred(struct rpc_auth *, + struct rpc_cred *, int); +@@ -106,6 +112,13 @@ + void rpcauth_free_credcache(struct rpc_auth *); + void rpcauth_insert_credcache(struct rpc_auth *, + struct rpc_cred *); ++ ++static inline ++struct rpc_cred * get_rpccred(struct rpc_cred *cred) ++{ ++ atomic_inc(&cred->cr_count); ++ return cred; ++} + + #endif /* __KERNEL__ */ + #endif /* _LINUX_SUNRPC_AUTH_H */ +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth.c linux-2.4.6-mmap/net/sunrpc/auth.c +--- linux-2.4.6-file/net/sunrpc/auth.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth.c Wed Jul 4 20:51:04 2001 +@@ -81,42 +81,61 @@ + auth->au_nextgc = jiffies + (auth->au_expire >> 1); + } + ++/* ++ * Destroy an unreferenced credential ++ */ + static inline void +-rpcauth_crdestroy(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_crdestroy(struct rpc_cred *cred) + { + #ifdef RPC_DEBUG + if (cred->cr_magic != RPCAUTH_CRED_MAGIC) + BUG(); + cred->cr_magic = 0; ++ if (atomic_read(&cred->cr_count) || cred->cr_auth) ++ BUG(); + #endif +- if (auth->au_ops->crdestroy) +- auth->au_ops->crdestroy(cred); +- else +- rpc_free(cred); ++ cred->cr_ops->crdestroy(cred); + } + + /* +- * Clear the RPC credential cache ++ * Destroy a list of credentials ++ */ ++static inline ++void rpcauth_destroy_credlist(struct rpc_cred *head) ++{ ++ struct rpc_cred *cred; ++ ++ while ((cred = head) != NULL) { ++ head = cred->cr_next; ++ rpcauth_crdestroy(cred); ++ } ++} ++ ++/* ++ * Clear the RPC credential cache, and delete those credentials ++ * that are not referenced. + */ + void + rpcauth_free_credcache(struct rpc_auth *auth) + { +- struct rpc_cred **q, *cred; +- void (*destroy)(struct rpc_cred *); ++ struct rpc_cred **q, *cred, *free = NULL; + int i; + +- if (!(destroy = auth->au_ops->crdestroy)) +- destroy = (void (*)(struct rpc_cred *)) rpc_free; +- + spin_lock(&rpc_credcache_lock); + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { + *q = cred->cr_next; +- destroy(cred); ++ cred->cr_auth = NULL; ++ if (atomic_read(&cred->cr_count) == 0) { ++ cred->cr_next = free; ++ free = cred; ++ } else ++ cred->cr_next = NULL; + } + } + spin_unlock(&rpc_credcache_lock); ++ rpcauth_destroy_credlist(free); + } + + /* +@@ -133,9 +152,10 @@ + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { +- if (!cred->cr_count && ++ if (!atomic_read(&cred->cr_count) && + time_before(cred->cr_expire, jiffies)) { + *q = cred->cr_next; ++ cred->cr_auth = NULL; + cred->cr_next = free; + free = cred; + continue; +@@ -144,10 +164,7 @@ + } + } + spin_unlock(&rpc_credcache_lock); +- while ((cred = free) != NULL) { +- free = cred->cr_next; +- rpcauth_crdestroy(auth, cred); +- } ++ rpcauth_destroy_credlist(free); + auth->au_nextgc = jiffies + auth->au_expire; + } + +@@ -163,8 +180,8 @@ + spin_lock(&rpc_credcache_lock); + cred->cr_next = auth->au_credcache[nr]; + auth->au_credcache[nr] = cred; +- cred->cr_count++; +- cred->cr_expire = jiffies + auth->au_expire; ++ cred->cr_auth = auth; ++ get_rpccred(cred); + spin_unlock(&rpc_credcache_lock); + } + +@@ -187,7 +204,7 @@ + q = &auth->au_credcache[nr]; + while ((cred = *q) != NULL) { + if (!(cred->cr_flags & RPCAUTH_CRED_DEAD) && +- auth->au_ops->crmatch(cred, taskflags)) { ++ cred->cr_ops->crmatch(cred, taskflags)) { + *q = cred->cr_next; + break; + } +@@ -213,23 +230,23 @@ + * Remove cred handle from cache + */ + static void +-rpcauth_remove_credcache(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_remove_credcache(struct rpc_cred *cred) + { ++ struct rpc_auth *auth = cred->cr_auth; + struct rpc_cred **q, *cr; + int nr; + + nr = (cred->cr_uid & RPC_CREDCACHE_MASK); +- spin_lock(&rpc_credcache_lock); + q = &auth->au_credcache[nr]; + while ((cr = *q) != NULL) { + if (cred == cr) { + *q = cred->cr_next; + cred->cr_next = NULL; ++ cred->cr_auth = NULL; + break; + } + q = &cred->cr_next; + } +- spin_unlock(&rpc_credcache_lock); + } + + struct rpc_cred * +@@ -258,7 +275,7 @@ + { + dprintk("RPC: matching %s cred %d\n", + auth->au_ops->au_name, taskflags); +- return auth->au_ops->crmatch(cred, taskflags); ++ return cred->cr_ops->crmatch(cred, taskflags); + } + + void +@@ -266,26 +283,25 @@ + { + dprintk("RPC: %4d holding %s cred %p\n", + task->tk_pid, task->tk_auth->au_ops->au_name, task->tk_msg.rpc_cred); +- if (task->tk_msg.rpc_cred) { +- spin_lock(&rpc_credcache_lock); +- task->tk_msg.rpc_cred->cr_count++; +- task->tk_msg.rpc_cred->cr_expire = jiffies + task->tk_auth->au_expire; +- spin_unlock(&rpc_credcache_lock); +- } ++ if (task->tk_msg.rpc_cred) ++ get_rpccred(task->tk_msg.rpc_cred); + } + + void +-rpcauth_releasecred(struct rpc_auth *auth, struct rpc_cred *cred) ++put_rpccred(struct rpc_cred *cred) + { +- spin_lock(&rpc_credcache_lock); +- if (cred != NULL && cred->cr_count > 0) { +- if (!--cred->cr_count && (cred->cr_flags & RPCAUTH_CRED_DEAD)) { +- spin_unlock(&rpc_credcache_lock); +- rpcauth_remove_credcache(auth, cred); +- rpcauth_crdestroy(auth, cred); +- return; +- } ++ if (!atomic_dec_and_lock(&cred->cr_count, &rpc_credcache_lock)) ++ return; ++ ++ if (cred->cr_auth && cred->cr_flags & RPCAUTH_CRED_DEAD) ++ rpcauth_remove_credcache(cred); ++ ++ if (!cred->cr_auth) { ++ spin_unlock(&rpc_credcache_lock); ++ rpcauth_crdestroy(cred); ++ return; + } ++ cred->cr_expire = jiffies + cred->cr_auth->au_expire; + spin_unlock(&rpc_credcache_lock); + } + +@@ -298,7 +314,7 @@ + dprintk("RPC: %4d releasing %s cred %p\n", + task->tk_pid, auth->au_ops->au_name, cred); + +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + task->tk_msg.rpc_cred = NULL; + } + +@@ -306,10 +322,11 @@ + rpcauth_marshcred(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d marshaling %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crmarshal(task, p, ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crmarshal(task, p, + task->tk_flags & RPC_CALL_REALUID); + } + +@@ -317,20 +334,22 @@ + rpcauth_checkverf(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d validating %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crvalidate(task, p); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crvalidate(task, p); + } + + int + rpcauth_refreshcred(struct rpc_task *task) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d refreshing %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- task->tk_status = auth->au_ops->crrefresh(task); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ task->tk_status = cred->cr_ops->crrefresh(task); + return task->tk_status; + } + +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_null.c linux-2.4.6-mmap/net/sunrpc/auth_null.c +--- linux-2.4.6-file/net/sunrpc/auth_null.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_null.c Wed Jul 4 19:27:38 2001 +@@ -17,6 +17,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops null_credops; ++ + static struct rpc_auth * + nul_create(struct rpc_clnt *clnt) + { +@@ -52,9 +54,10 @@ + + if (!(cred = (struct rpc_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; +- cred->cr_count = 0; ++ atomic_set(&cred->cr_count, 0); + cred->cr_flags = RPCAUTH_CRED_UPTODATE; + cred->cr_uid = current->uid; ++ cred->cr_ops = &null_credops; + + return cred; + } +@@ -124,7 +127,11 @@ + #endif + nul_create, + nul_destroy, +- nul_create_cred, ++ nul_create_cred ++}; ++ ++static ++struct rpc_credops null_credops = { + nul_destroy_cred, + nul_match, + nul_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_unix.c linux-2.4.6-mmap/net/sunrpc/auth_unix.c +--- linux-2.4.6-file/net/sunrpc/auth_unix.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_unix.c Wed Jul 4 19:29:07 2001 +@@ -33,6 +33,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops unix_credops; ++ + static struct rpc_auth * + unx_create(struct rpc_clnt *clnt) + { +@@ -71,7 +73,7 @@ + if (!(cred = (struct unx_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 0; ++ atomic_set(&cred->uc_count, 0); + cred->uc_flags = RPCAUTH_CRED_UPTODATE; + if (flags & RPC_TASK_ROOTCREDS) { + cred->uc_uid = cred->uc_fsuid = 0; +@@ -91,6 +93,7 @@ + if (i < NFS_NGROUPS) + cred->uc_gids[i] = NOGROUP; + } ++ cred->uc_base.cr_ops = &unix_credops; + + return (struct rpc_cred *) cred; + } +@@ -106,7 +109,7 @@ + if (!(cred = (struct unx_cred *) rpc_malloc(task, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 1; ++ atomic_set(&cred->uc_count, 1); + cred->uc_flags = RPCAUTH_CRED_DEAD|RPCAUTH_CRED_UPTODATE; + cred->uc_uid = uid; + cred->uc_gid = gid; +@@ -236,7 +239,11 @@ + #endif + unx_create, + unx_destroy, +- unx_create_cred, ++ unx_create_cred ++}; ++ ++static ++struct rpc_credops unix_credops = { + unx_destroy_cred, + unx_match, + unx_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/sunrpc_syms.c linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c +--- linux-2.4.6-file/net/sunrpc/sunrpc_syms.c Fri Jan 12 00:53:02 2001 ++++ linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c Wed Jul 4 20:55:19 2001 +@@ -65,7 +65,7 @@ + EXPORT_SYMBOL(rpcauth_lookupcred); + EXPORT_SYMBOL(rpcauth_bindcred); + EXPORT_SYMBOL(rpcauth_matchcred); +-EXPORT_SYMBOL(rpcauth_releasecred); ++EXPORT_SYMBOL(put_rpccred); + + /* RPC server stuff */ + EXPORT_SYMBOL(svc_create); + +_______________________________________________ +NFS maillist - NFS@lists.sourceforge.net +http://lists.sourceforge.net/lists/listinfo/nfs + + diff --git a/sys-kernel/linux-headers/files/2.4.6-r2/lvm.c.diff b/sys-kernel/linux-headers/files/2.4.6-r2/lvm.c.diff new file mode 100644 index 000000000000..6392671b8f7e --- /dev/null +++ b/sys-kernel/linux-headers/files/2.4.6-r2/lvm.c.diff @@ -0,0 +1,20 @@ +--- lvm.c.orig Sat Apr 28 10:48:37 2001 ++++ lvm.c Sat Apr 28 11:07:07 2001 +@@ -1791,7 +1791,7 @@ + int max_hardblocksize = 0, hardblocksize; + + for (le = 0; le < lv->lv_allocated_le; le++) { +- hardblocksize = get_hardblocksize(lv->lv_current_pe[le].dev); ++ hardblocksize = get_hardsect_size(lv->lv_current_pe[le].dev); + if (hardblocksize == 0) + hardblocksize = 512; + if (hardblocksize > max_hardblocksize) +@@ -1801,7 +1801,7 @@ + if (lv->lv_access & LV_SNAPSHOT) { + for (e = 0; e < lv->lv_remap_end; e++) { + hardblocksize = +- get_hardblocksize( ++ get_hardsect_size( + lv->lv_block_exception[e].rdev_new); + if (hardblocksize == 0) + hardblocksize = 512; diff --git a/sys-kernel/linux-headers/files/2.4.6-r2/mosix.init b/sys-kernel/linux-headers/files/2.4.6-r2/mosix.init new file mode 100644 index 000000000000..f339dabb666e --- /dev/null +++ b/sys-kernel/linux-headers/files/2.4.6-r2/mosix.init @@ -0,0 +1,118 @@ +#!/bin/sh - +#RCUPDATE:2 3 4:95:This line is required for script management + +# description: MOSIX is am extension of the operating system, +# supporting scalable and transparent cluster computing. +# +# mosix Script to stop/start MOSIX +# +# Author: Amnon Shiloh + +. /etc/rc.d/config/functions + +SERVICE="mosix" +opts="start stop status restart" + +mosix_stop() { + echo 0 > /proc/mosix/admin/mospe + rm -f /var/lock/subsys/mosix +} + +mosix_start() { + if [ -s /etc/overheads -a -f /proc/mosix/admin/overheads ] + then + /bin/grep -v '^#' /etc/overheads > /proc/mosix/admin/overheads + fi + if [ -s /etc/mfscosts -a -f /proc/mosix/admin/mfscosts ] + then + /bin/grep -v '^#' /etc/mfscosts > /proc/mosix/admin/mfscosts + fi + a1= + [ -s /etc/mospe ] && a1="-p `cat /etc/mospe`" + a2= + [ -s /etc/mosgates ] && a2="-g `cat /etc/mosgates`" + /sbin/setpe -W $a1 $a2 -f /etc/mosix.map + touch /var/lock/subsys/mosix +} + +alarm() { + t=$2 + while : + do + case "$t" in 0) break ;; esac + sleep 1 + t=$(($t-1)) + done + kill -1 $1 + exit 0 +} + +# See how we were called. +start() { + eecho "Initializing MOSIX" + if [ ! -f /etc/mosix.map ] + then + echo You have no MOSIX configuration - please edit yours now: + echo -n "Editor to use [q to quit] - [vi] :- " + trap 'echo Timed-Out ; exit 1' 1 + alarm $$ 120 & + killer=$! + read editor + kill $killer + trap '' + case "$editor" in [qQ]) exit ;; + "") editor=/usr/bin/vi + [ -f $editor ] || editor=/bin/vi + ;; + esac + echo "# MOSIX CONFIGURATION" > /etc/mosix.map + echo "# ===================" >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# Each line should contain 3 fields, mapping IP addresses to MOSIX node-numbers:" >> /etc/mosix.map + echo "# 1) first MOSIX node-number in range." >> /etc/mosix.map + echo "# 2) IP address of the above node (or node-name from /etc/hosts)." >> /etc/mosix.map + echo "# 3) number of nodes in this range." >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# MOSIX-# IP number-of-nodes" >> /etc/mosix.map + echo "# ============================" >> /etc/mosix.map + $editor /etc/mosix.map + echo + while : + do + echo "If this node's standard IP address is not part of the table that" + echo "you just edited, because MOSIX uses a different or separate network," + echo "you need to type this node's MOSIX-number now." + echo + /bin/echo -n "Otherwise please press only <Enter> :- " + read me + case "$me" in "") /bin/rm -f /etc/mospe ; break ;; + [1-9] | [1-9][0-9] | [1-9][0-9][0-9] | \ + [1-9][0-9][0-9][0-9] | \ + [1-6][0-9][0-9][0-9][0-9]) + echo $me > /etc/mospe ; break ;; + *) echo Improper MOSIX number - please try again: ;; + esac + done + fi + start_mosix + eend +} + +stop() { + eecho "Stopping MOSIX" + stop_mosix + eend +} + +status() { + /sbin/setpe -r +} + +restart() { + eecho "Restarting MOSIX" + stop_mosix + start_mosix + eend +} + +doservice ${@} diff --git a/sys-kernel/linux-headers/files/2.4.6-r2/sched.h b/sys-kernel/linux-headers/files/2.4.6-r2/sched.h new file mode 100644 index 000000000000..438203f84cdd --- /dev/null +++ b/sys-kernel/linux-headers/files/2.4.6-r2/sched.h @@ -0,0 +1,997 @@ +#ifndef _LINUX_SCHED_H +#define _LINUX_SCHED_H + +#include <asm/param.h> /* for HZ */ + +extern unsigned long event; + +#include <linux/config.h> +#include <linux/binfmts.h> +#include <linux/personality.h> +#include <linux/threads.h> +#include <linux/kernel.h> +#include <linux/types.h> +#include <linux/times.h> +#include <linux/timex.h> + +#include <asm/system.h> +#include <asm/semaphore.h> +#include <asm/page.h> +#include <asm/ptrace.h> +#include <asm/mmu.h> + +#include <linux/smp.h> +#include <linux/tty.h> +#include <linux/sem.h> +#include <linux/signal.h> +#include <linux/securebits.h> + +#ifdef CONFIG_MOSIX +#include <mos/mosixtask.h> +#else +#define MOSIX_INIT_TASK(tsk) +#endif CONFIG_MOSIX + +#include <linux/fs_struct.h> + +/* + * cloning flags: + */ +#define CSIGNAL 0x000000ff /* signal mask to be sent at exit */ +#define CLONE_VM 0x00000100 /* set if VM shared between processes */ +#define CLONE_FS 0x00000200 /* set if fs info shared between processes */ +#define CLONE_FILES 0x00000400 /* set if open files shared between processes */ +#define CLONE_SIGHAND 0x00000800 /* set if signal handlers and blocked signals shared */ +#define CLONE_PID 0x00001000 /* set if pid shared */ +#define CLONE_PTRACE 0x00002000 /* set if we want to let tracing continue on the child too */ +#define CLONE_VFORK 0x00004000 /* set if the parent wants the child to wake it up on mm_release */ +#define CLONE_PARENT 0x00008000 /* set if we want to have the same parent as the cloner */ +#define CLONE_THREAD 0x00010000 /* Same thread group? */ + +#define CLONE_SIGNAL (CLONE_SIGHAND | CLONE_THREAD) + +/* + * These are the constant used to fake the fixed-point load-average + * counting. Some notes: + * - 11 bit fractions expand to 22 bits by the multiplies: this gives + * a load-average precision of 10 bits integer + 11 bits fractional + * - if you want to count load-averages more often, you need more + * precision, or rounding will get you. With 2-second counting freq, + * the EXP_n values would be 1981, 2034 and 2043 if still using only + * 11 bit fractions. + */ +extern unsigned long avenrun[]; /* Load averages */ + +#define FSHIFT 11 /* nr of bits of precision */ +#define FIXED_1 (1<<FSHIFT) /* 1.0 as fixed-point */ +#define LOAD_FREQ (5*HZ) /* 5 sec intervals */ +#define EXP_1 1884 /* 1/exp(5sec/1min) as fixed-point */ +#define EXP_5 2014 /* 1/exp(5sec/5min) */ +#define EXP_15 2037 /* 1/exp(5sec/15min) */ + +#define CALC_LOAD(load,exp,n) \ + load *= exp; \ + load += n*(FIXED_1-exp); \ + load >>= FSHIFT; + +#define CT_TO_SECS(x) ((x) / HZ) +#define CT_TO_USECS(x) (((x) % HZ) * 1000000/HZ) + +extern int nr_running, nr_threads; +extern int last_pid; + +#include <linux/fs.h> +#include <linux/time.h> +#include <linux/param.h> +#include <linux/resource.h> +#include <linux/timer.h> + +#include <asm/processor.h> + +#define TASK_RUNNING 0 +#define TASK_INTERRUPTIBLE 1 +#define TASK_UNINTERRUPTIBLE 2 +#define TASK_ZOMBIE 4 +#define TASK_STOPPED 8 +#ifdef CONFIG_MOSIX +#define TASK_SAME 16 +#endif CONFIG_MOSIX + +#define __set_task_state(tsk, state_value) \ + do { (tsk)->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_task_state(tsk, state_value) \ + set_mb((tsk)->state, (state_value)) +#else +#define set_task_state(tsk, state_value) \ + __set_task_state((tsk), (state_value)) +#endif + +#define __set_current_state(state_value) \ + do { current->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_current_state(state_value) \ + set_mb(current->state, (state_value)) +#else +#define set_current_state(state_value) \ + __set_current_state(state_value) +#endif + +/* + * Scheduling policies + */ +#define SCHED_OTHER 0 +#define SCHED_FIFO 1 +#define SCHED_RR 2 + +/* + * This is an additional bit set when we want to + * yield the CPU for one re-schedule.. + */ +#define SCHED_YIELD 0x10 + +struct sched_param { + int sched_priority; +}; + +#ifdef __KERNEL__ + +#include <linux/spinlock.h> + +/* + * This serializes "schedule()" and also protects + * the run-queue from deletions/modifications (but + * _adding_ to the beginning of the run-queue has + * a separate lock). + */ +extern rwlock_t tasklist_lock; +extern spinlock_t runqueue_lock; +extern spinlock_t mmlist_lock; + +extern void sched_init(void); +extern void init_idle(void); +extern void show_state(void); +extern void cpu_init (void); +extern void trap_init(void); +extern void update_process_times(int user); +extern void update_one_process(struct task_struct *p, unsigned long user, + unsigned long system, int cpu); + +#define MAX_SCHEDULE_TIMEOUT LONG_MAX +extern signed long FASTCALL(schedule_timeout(signed long timeout)); +asmlinkage void schedule(void); + +extern int schedule_task(struct tq_struct *task); +extern void flush_scheduled_tasks(void); +extern int start_context_thread(void); +extern int current_is_keventd(void); + +/* + * The default fd array needs to be at least BITS_PER_LONG, + * as this is the granularity returned by copy_fdset(). + */ +#define NR_OPEN_DEFAULT BITS_PER_LONG + +/* + * Open file table structure + */ +struct files_struct { + atomic_t count; + rwlock_t file_lock; /* Protects all the below members. Nests inside tsk->alloc_lock */ + int max_fds; + int max_fdset; + int next_fd; + struct file ** fd; /* current fd array */ + fd_set *close_on_exec; + fd_set *open_fds; + fd_set close_on_exec_init; + fd_set open_fds_init; + struct file * fd_array[NR_OPEN_DEFAULT]; +#ifdef CONFIG_MOSIX_DFSA + atomic_t users; /* the real count, excluding /proc grabs */ + int count_dropped_to_one; + int maxclosed; + fd_set *closed; + fd_set closed_fds_init; + int maxopened; + fd_set *opened; + fd_set opened_fds_init; + int maxmod; + fd_set *modified; + fd_set modified_fds_init; +#define INIT_FILES_DFSA , users : ATOMIC_INIT(1), \ + closed: &init_files.closed_fds_init, \ + opened: &init_files.opened_fds_init, \ + modified: &init_files.modified_fds_init, +#else +#define INIT_FILES_DFSA +#endif CONFIG_MOSIX_DFSA +}; + +#define INIT_FILES \ +{ \ + count: ATOMIC_INIT(1), \ + file_lock: RW_LOCK_UNLOCKED, \ + max_fds: NR_OPEN_DEFAULT, \ + max_fdset: __FD_SETSIZE, \ + next_fd: 0, \ + fd: &init_files.fd_array[0], \ + close_on_exec: &init_files.close_on_exec_init, \ + open_fds: &init_files.open_fds_init, \ + close_on_exec_init: { { 0, } }, \ + open_fds_init: { { 0, } }, \ + fd_array: { NULL, } \ + INIT_FILES_DFSA \ +} + +/* Maximum number of active map areas.. This is a random (large) number */ +#define MAX_MAP_COUNT (65536) + +/* Number of map areas at which the AVL tree is activated. This is arbitrary. */ +#define AVL_MIN_MAP_COUNT 32 + +struct mm_struct { + struct vm_area_struct * mmap; /* list of VMAs */ + struct vm_area_struct * mmap_avl; /* tree of VMAs */ + struct vm_area_struct * mmap_cache; /* last find_vma result */ + pgd_t * pgd; + atomic_t mm_users; /* How many users with user space? */ + atomic_t mm_count; /* How many references to "struct mm_struct" (users count as 1) */ + int map_count; /* number of VMAs */ + struct rw_semaphore mmap_sem; + spinlock_t page_table_lock; /* Protects task page tables and mm->rss */ + + struct list_head mmlist; /* List of all active mm's. These are globally strung + * together off init_mm.mmlist, and are protected + * by mmlist_lock + */ + + unsigned long start_code, end_code, start_data, end_data; + unsigned long start_brk, brk, start_stack; + unsigned long arg_start, arg_end, env_start, env_end; + unsigned long rss, total_vm, locked_vm; + unsigned long def_flags; + unsigned long cpu_vm_mask; + unsigned long swap_address; + + /* Architecture-specific MM context */ + mm_context_t context; +#ifdef CONFIG_MOSIX + atomic_t mm_kiocount; + atomic_t mm_realusers; + unsigned long last_memsort; + int used, unused, swapped; /* break-down of memory pages */ + int private_unused; /* unused but not active/inactive */ + int mark; +#define INIT_MM_MOSIX mm_kiocount: ATOMIC_INIT(0), mm_realusers: ATOMIC_INIT(1), +#else +#define INIT_MM_MOSIX +#endif CONFIG_MOSIX +}; + +extern int mmlist_nr; + +#define INIT_MM(name) \ +{ \ + mmap: &init_mmap, \ + mmap_avl: NULL, \ + mmap_cache: NULL, \ + pgd: swapper_pg_dir, \ + mm_users: ATOMIC_INIT(2), \ + mm_count: ATOMIC_INIT(1), \ + map_count: 1, \ + mmap_sem: __RWSEM_INITIALIZER(name.mmap_sem), \ + page_table_lock: SPIN_LOCK_UNLOCKED, \ + mmlist: LIST_HEAD_INIT(name.mmlist), \ + INIT_MM_MOSIX \ +} + +struct signal_struct { + atomic_t count; + struct k_sigaction action[_NSIG]; + spinlock_t siglock; +}; + + +#define INIT_SIGNALS { \ + count: ATOMIC_INIT(1), \ + action: { {{0,}}, }, \ + siglock: SPIN_LOCK_UNLOCKED \ +} + +/* + * Some day this will be a full-fledged user tracking system.. + */ +struct user_struct { + atomic_t __count; /* reference count */ + atomic_t processes; /* How many processes does this user have? */ + atomic_t files; /* How many open files does this user have? */ + + /* Hash table maintenance information */ + struct user_struct *next, **pprev; + uid_t uid; +}; + +#define get_current_user() ({ \ + struct user_struct *__user = current->user; \ + atomic_inc(&__user->__count); \ + __user; }) + +extern struct user_struct root_user; +#define INIT_USER (&root_user) + +struct task_struct { + /* + * offsets of these are hardcoded elsewhere - touch with care + */ + volatile long state; /* -1 unrunnable, 0 runnable, >0 stopped */ + unsigned long flags; /* per process flags, defined below */ + int sigpending; + mm_segment_t addr_limit; /* thread address space: + 0-0xBFFFFFFF for user-thead + 0-0xFFFFFFFF for kernel-thread + */ + struct exec_domain *exec_domain; + volatile long need_resched; + unsigned long ptrace; + + int lock_depth; /* Lock depth */ + +/* + * offset 32 begins here on 32-bit platforms. We keep + * all fields in a single cacheline that are needed for + * the goodness() loop in schedule(). + */ + long counter; + long nice; + unsigned long policy; + struct mm_struct *mm; + int has_cpu, processor; + unsigned long cpus_allowed; + /* + * (only the 'next' pointer fits into the cacheline, but + * that's just fine.) + */ + struct list_head run_list; + unsigned long sleep_time; + + struct task_struct *next_task, *prev_task; + struct mm_struct *active_mm; + +/* task state */ + struct linux_binfmt *binfmt; + int exit_code, exit_signal; + int pdeath_signal; /* The signal sent when the parent dies */ + /* ??? */ + unsigned long personality; + int dumpable:1; + int did_exec:1; + pid_t pid; + pid_t pgrp; + pid_t tty_old_pgrp; + pid_t session; + pid_t tgid; + /* boolean value for session group leader */ + int leader; + /* + * pointers to (original) parent process, youngest child, younger sibling, + * older sibling, respectively. (p->father can be replaced with + * p->p_pptr->pid) + */ + struct task_struct *p_opptr, *p_pptr, *p_cptr, *p_ysptr, *p_osptr; + struct list_head thread_group; + + /* PID hash table linkage. */ + struct task_struct *pidhash_next; + struct task_struct **pidhash_pprev; + + wait_queue_head_t wait_chldexit; /* for wait4() */ + struct semaphore *vfork_sem; /* for vfork() */ + unsigned long rt_priority; + unsigned long it_real_value, it_prof_value, it_virt_value; + unsigned long it_real_incr, it_prof_incr, it_virt_incr; + struct timer_list real_timer; + struct tms times; + unsigned long start_time; + long per_cpu_utime[NR_CPUS], per_cpu_stime[NR_CPUS]; +/* mm fault and swap info: this can arguably be seen as either mm-specific or thread-specific */ + unsigned long min_flt, maj_flt, nswap, cmin_flt, cmaj_flt, cnswap; + int swappable:1; +/* process credentials */ + uid_t uid,euid,suid,fsuid; + gid_t gid,egid,sgid,fsgid; + int ngroups; + gid_t groups[NGROUPS]; + kernel_cap_t cap_effective, cap_inheritable, cap_permitted; + int keep_capabilities:1; + struct user_struct *user; +/* limits */ + struct rlimit rlim[RLIM_NLIMITS]; + unsigned short used_math; + char comm[16]; +/* file system info */ + int link_count; + struct tty_struct *tty; /* NULL if no tty */ + unsigned int locks; /* How many file locks are being held */ +/* ipc stuff */ + struct sem_undo *semundo; + struct sem_queue *semsleeping; +/* CPU-specific state of this task */ + struct thread_struct thread; +/* filesystem information */ + struct fs_struct *fs; +/* open file information */ + struct files_struct *files; +/* signal handlers */ + spinlock_t sigmask_lock; /* Protects signal and blocked */ + struct signal_struct *sig; + + sigset_t blocked; + struct sigpending pending; + + unsigned long sas_ss_sp; + size_t sas_ss_size; + int (*notifier)(void *priv); + void *notifier_data; + sigset_t *notifier_mask; + +/* Thread group tracking */ + u32 parent_exec_id; + u32 self_exec_id; +/* Protection of (de-)allocation: mm, files, fs, tty */ + spinlock_t alloc_lock; +#ifdef CONFIG_MOSIX + struct mosix_task mosix; +#endif CONFIG_MOSIX + void *journal_info; +}; + +/* + * Per process flags + */ +#define PF_ALIGNWARN 0x00000001 /* Print alignment warning msgs */ + /* Not implemented yet, only for 486*/ +#define PF_STARTING 0x00000002 /* being created */ +#define PF_EXITING 0x00000004 /* getting shut down */ +#define PF_FORKNOEXEC 0x00000040 /* forked but didn't exec */ +#define PF_SUPERPRIV 0x00000100 /* used super-user privileges */ +#define PF_DUMPCORE 0x00000200 /* dumped core */ +#define PF_SIGNALED 0x00000400 /* killed by a signal */ +#define PF_MEMALLOC 0x00000800 /* Allocating memory */ +#define PF_VFORK 0x00001000 /* Wake up parent in mm_release */ + +#define PF_USEDFPU 0x00100000 /* task used FPU this quantum (SMP) */ + +/* + * Ptrace flags + */ + +#define PT_PTRACED 0x00000001 +#define PT_TRACESYS 0x00000002 +#define PT_DTRACE 0x00000004 /* delayed trace (used on m68k, i386) */ +#define PT_TRACESYSGOOD 0x00000008 + +/* + * Limit the stack by to some sane default: root can always + * increase this limit if needed.. 8MB seems reasonable. + */ +#define _STK_LIM (8*1024*1024) + +#define DEF_COUNTER (10*HZ/100) /* 100 ms time slice */ +#define MAX_COUNTER (20*HZ/100) +#define DEF_NICE (0) + +/* + * INIT_TASK is used to set up the first task table, touch at + * your own risk!. Base=0, limit=0x1fffff (=2MB) + */ +#define INIT_TASK(tsk) \ +{ \ + state: 0, \ + flags: 0, \ + sigpending: 0, \ + addr_limit: KERNEL_DS, \ + exec_domain: &default_exec_domain, \ + lock_depth: -1, \ + counter: DEF_COUNTER, \ + nice: DEF_NICE, \ + policy: SCHED_OTHER, \ + mm: NULL, \ + active_mm: &init_mm, \ + cpus_allowed: -1, \ + run_list: LIST_HEAD_INIT(tsk.run_list), \ + next_task: &tsk, \ + prev_task: &tsk, \ + p_opptr: &tsk, \ + p_pptr: &tsk, \ + thread_group: LIST_HEAD_INIT(tsk.thread_group), \ + wait_chldexit: __WAIT_QUEUE_HEAD_INITIALIZER(tsk.wait_chldexit),\ + real_timer: { \ + function: it_real_fn \ + }, \ + cap_effective: CAP_INIT_EFF_SET, \ + cap_inheritable: CAP_INIT_INH_SET, \ + cap_permitted: CAP_FULL_SET, \ + keep_capabilities: 0, \ + rlim: INIT_RLIMITS, \ + user: INIT_USER, \ + comm: "swapper", \ + thread: INIT_THREAD, \ + fs: &init_fs, \ + files: &init_files, \ + sigmask_lock: SPIN_LOCK_UNLOCKED, \ + sig: &init_signals, \ + pending: { NULL, &tsk.pending.head, {{0}}}, \ + blocked: {{0}}, \ + alloc_lock: SPIN_LOCK_UNLOCKED, \ + MOSIX_INIT_TASK(tsk), \ + journal_info: NULL \ +} + + +#ifndef INIT_TASK_SIZE +# define INIT_TASK_SIZE 2048*sizeof(long) +#endif + +union task_union { + struct task_struct task; + unsigned long stack[INIT_TASK_SIZE/sizeof(long)]; +}; + +extern union task_union init_task_union; + +extern struct mm_struct init_mm; +extern struct task_struct *init_tasks[NR_CPUS]; + +/* PID hashing. (shouldnt this be dynamic?) */ +#define PIDHASH_SZ (4096 >> 2) +extern struct task_struct *pidhash[PIDHASH_SZ]; + +#define pid_hashfn(x) ((((x) >> 8) ^ (x)) & (PIDHASH_SZ - 1)) + +static inline void hash_pid(struct task_struct *p) +{ + struct task_struct **htable = &pidhash[pid_hashfn(p->pid)]; + + if((p->pidhash_next = *htable) != NULL) + (*htable)->pidhash_pprev = &p->pidhash_next; + *htable = p; + p->pidhash_pprev = htable; +} + +static inline void unhash_pid(struct task_struct *p) +{ + if(p->pidhash_next) + p->pidhash_next->pidhash_pprev = p->pidhash_pprev; + *p->pidhash_pprev = p->pidhash_next; +} + +static inline struct task_struct *find_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; +#ifdef CONFIG_MOSIX + if(p && (p->mosix.dflags & DREMOTE)) + p = NULL; +#endif CONFIG_MOSIX + return p; +} + +#ifdef CONFIG_MOSIX +static inline struct task_struct *find_any_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; + return p; +} +#endif CONFIG_MOSIX + +/* per-UID process charging. */ +extern struct user_struct * alloc_uid(uid_t); +extern void free_uid(struct user_struct *); + +#include <asm/current.h> + +extern unsigned long volatile jiffies; +extern unsigned long itimer_ticks; +extern unsigned long itimer_next; +extern struct timeval xtime; +extern void do_timer(struct pt_regs *); + +extern unsigned int * prof_buffer; +extern unsigned long prof_len; +extern unsigned long prof_shift; + +#define CURRENT_TIME (xtime.tv_sec) + +extern void FASTCALL(__wake_up(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(__wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern void FASTCALL(interruptible_sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(interruptible_sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern int FASTCALL(wake_up_process(struct task_struct * tsk)); + +#define wake_up(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_nr(x, nr) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_all(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 0) +#define wake_up_sync(x) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_sync_nr(x, nr) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible(x) __wake_up((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_nr(x, nr) __wake_up((x),TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible_all(x) __wake_up((x),TASK_INTERRUPTIBLE, 0) +#define wake_up_interruptible_sync(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_sync_nr(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, nr) +asmlinkage long sys_wait4(pid_t pid,unsigned int * stat_addr, int options, struct rusage * ru); + +extern int in_group_p(gid_t); +extern int in_egroup_p(gid_t); + +extern void proc_caches_init(void); +extern void flush_signals(struct task_struct *); +extern void flush_signal_handlers(struct task_struct *); +extern int dequeue_signal(sigset_t *, siginfo_t *); +extern void block_all_signals(int (*notifier)(void *priv), void *priv, + sigset_t *mask); +extern void unblock_all_signals(void); +extern int send_sig_info(int, struct siginfo *, struct task_struct *); +extern int force_sig_info(int, struct siginfo *, struct task_struct *); +extern int kill_pg_info(int, struct siginfo *, pid_t); +extern int kill_sl_info(int, struct siginfo *, pid_t); +extern int kill_proc_info(int, struct siginfo *, pid_t); +extern void notify_parent(struct task_struct *, int); +extern void do_notify_parent(struct task_struct *, int); +extern void force_sig(int, struct task_struct *); +extern int send_sig(int, struct task_struct *, int); +extern int kill_pg(pid_t, int, int); +extern int kill_sl(pid_t, int, int); +extern int kill_proc(pid_t, int, int); +extern int do_sigaction(int, const struct k_sigaction *, struct k_sigaction *); +extern int do_sigaltstack(const stack_t *, stack_t *, unsigned long); + +static inline int signal_pending(struct task_struct *p) +{ + return (p->sigpending != 0); +} + +/* + * Re-calculate pending state from the set of locally pending + * signals, globally pending signals, and blocked signals. + */ +static inline int has_pending_signals(sigset_t *signal, sigset_t *blocked) +{ + unsigned long ready; + long i; + + switch (_NSIG_WORDS) { + default: + for (i = _NSIG_WORDS, ready = 0; --i >= 0 ;) + ready |= signal->sig[i] &~ blocked->sig[i]; + break; + + case 4: ready = signal->sig[3] &~ blocked->sig[3]; + ready |= signal->sig[2] &~ blocked->sig[2]; + ready |= signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 2: ready = signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 1: ready = signal->sig[0] &~ blocked->sig[0]; + } + return ready != 0; +} + +/* Reevaluate whether the task has signals pending delivery. + This is required every time the blocked sigset_t changes. + All callers should have t->sigmask_lock. */ + +static inline void recalc_sigpending(struct task_struct *t) +{ +#ifdef CONFIG_MOSIX + t->mosix.ignoreoldsigs = 0; + if(t->mosix.dflags & DHEAVYSLEEP) + t->sigpending = 0; + else if(t->mosix.dflags & DFAKESIGNAL) + t->sigpending = 1; + else +#endif CONFIG_MOSIX + t->sigpending = has_pending_signals(&t->pending.signal, &t->blocked); +} + +/* True if we are on the alternate signal stack. */ + +static inline int on_sig_stack(unsigned long sp) +{ + return (sp - current->sas_ss_sp < current->sas_ss_size); +} + +static inline int sas_ss_flags(unsigned long sp) +{ + return (current->sas_ss_size == 0 ? SS_DISABLE + : on_sig_stack(sp) ? SS_ONSTACK : 0); +} + +extern int request_irq(unsigned int, + void (*handler)(int, void *, struct pt_regs *), + unsigned long, const char *, void *); +extern void free_irq(unsigned int, void *); + +/* + * This has now become a routine instead of a macro, it sets a flag if + * it returns true (to do BSD-style accounting where the process is flagged + * if it uses root privs). The implication of this is that you should do + * normal permissions checks first, and check suser() last. + * + * [Dec 1997 -- Chris Evans] + * For correctness, the above considerations need to be extended to + * fsuser(). This is done, along with moving fsuser() checks to be + * last. + * + * These will be removed, but in the mean time, when the SECURE_NOROOT + * flag is set, uids don't grant privilege. + */ +static inline int suser(void) +{ + if (!issecure(SECURE_NOROOT) && current->euid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +static inline int fsuser(void) +{ + if (!issecure(SECURE_NOROOT) && current->fsuid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * capable() checks for a particular capability. + * New privilege checks should use this interface, rather than suser() or + * fsuser(). See include/linux/capability.h for defined capabilities. + */ + +static inline int capable(int cap) +{ +#if 1 /* ok now */ + if (cap_raised(current->cap_effective, cap)) +#else + if (cap_is_fs_cap(cap) ? current->fsuid == 0 : current->euid == 0) +#endif + { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * Routines for handling mm_structs + */ +extern struct mm_struct * mm_alloc(void); + +extern struct mm_struct * start_lazy_tlb(void); +extern void end_lazy_tlb(struct mm_struct *mm); + +/* mmdrop drops the mm and the page tables */ +extern inline void FASTCALL(__mmdrop(struct mm_struct *)); +static inline void mmdrop(struct mm_struct * mm) +{ + if (atomic_dec_and_test(&mm->mm_count)) + __mmdrop(mm); +} + +/* mmput gets rid of the mappings and all user-space */ +extern void mmput(struct mm_struct *); +/* Remove the current tasks stale references to the old mm_struct */ +extern void mm_release(void); + +/* + * Routines for handling the fd arrays + */ +extern struct file ** alloc_fd_array(int); +extern int expand_fd_array(struct files_struct *, int nr); +extern void free_fd_array(struct file **, int); + +extern fd_set *alloc_fdset(int); +extern int expand_fdset(struct files_struct *, int nr); +extern void free_fdset(fd_set *, int); + +extern int copy_thread(int, unsigned long, unsigned long, unsigned long, struct task_struct *, struct pt_regs *); +extern void flush_thread(void); +extern void exit_thread(void); + +extern void exit_mm(struct task_struct *); +extern void exit_files(struct task_struct *); +extern void exit_sighand(struct task_struct *); + +extern void daemonize(void); + +extern int do_execve(char *, char **, char **, struct pt_regs *); +extern int do_fork(unsigned long, unsigned long, struct pt_regs *, unsigned long); + +extern void FASTCALL(add_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(add_wait_queue_exclusive(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(remove_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); + +#define __wait_event(wq, condition) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_UNINTERRUPTIBLE); \ + if (condition) \ + break; \ + schedule(); \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event(wq, condition) \ +do { \ + if (condition) \ + break; \ + __wait_event(wq, condition); \ +} while (0) + +#define __wait_event_interruptible(wq, condition, ret) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_INTERRUPTIBLE); \ + if (condition) \ + break; \ + if (!signal_pending(current)) { \ + schedule(); \ + continue; \ + } \ + ret = -ERESTARTSYS; \ + break; \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event_interruptible(wq, condition) \ +({ \ + int __ret = 0; \ + if (!(condition)) \ + __wait_event_interruptible(wq, condition, __ret); \ + __ret; \ +}) + +#define REMOVE_LINKS(p) do { \ + (p)->next_task->prev_task = (p)->prev_task; \ + (p)->prev_task->next_task = (p)->next_task; \ + if ((p)->p_osptr) \ + (p)->p_osptr->p_ysptr = (p)->p_ysptr; \ + if ((p)->p_ysptr) \ + (p)->p_ysptr->p_osptr = (p)->p_osptr; \ + else \ + (p)->p_pptr->p_cptr = (p)->p_osptr; \ + } while (0) + +#define SET_LINKS(p) do { \ + (p)->next_task = &init_task; \ + (p)->prev_task = init_task.prev_task; \ + init_task.prev_task->next_task = (p); \ + init_task.prev_task = (p); \ + (p)->p_ysptr = NULL; \ + if (((p)->p_osptr = (p)->p_pptr->p_cptr) != NULL) \ + (p)->p_osptr->p_ysptr = p; \ + (p)->p_pptr->p_cptr = p; \ + } while (0) + +#define for_each_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) + +#define next_thread(p) \ + list_entry((p)->thread_group.next, struct task_struct, thread_group) + +#ifdef CONFIG_MOSIX +#define for_each_local_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) \ + if(!(p->mosix.dflags & DREMOTE)) + +static inline long +LOGICAL_STATE(struct task_struct *p) +{ + int result; + unsigned long flags; + + read_lock_irqsave(&p->mosix.state_lock, flags); + result = (p->mosix.bstate == TASK_SAME) ? p->state : p->mosix.bstate; + read_unlock_irqrestore(&p->mosix.state_lock, flags); + return(result); +} + +extern void run_on(struct task_struct *); +extern void run_off(struct task_struct *); + +#define need_interim_while_asleep() mosix_need_while_asleep() +#define run_interim_while_asleep() mosix_run_while_asleep() +#endif CONFIG_MOSIX + +static inline void del_from_runqueue(struct task_struct * p) +{ + nr_running--; + p->sleep_time = jiffies; +#ifdef CONFIG_MOSIX + run_off(p); +#endif CONFIG_MOSIX + list_del(&p->run_list); + p->run_list.next = NULL; +} + +static inline int task_on_runqueue(struct task_struct *p) +{ + return (p->run_list.next != NULL); +} + +static inline void unhash_process(struct task_struct *p) +{ + if (task_on_runqueue(p)) BUG(); + write_lock_irq(&tasklist_lock); + nr_threads--; + unhash_pid(p); + REMOVE_LINKS(p); + list_del(&p->thread_group); + write_unlock_irq(&tasklist_lock); +} + +/* Protects ->fs, ->files, ->mm, and synchronises with wait4(). Nests inside tasklist_lock */ +static inline void task_lock(struct task_struct *p) +{ + spin_lock(&p->alloc_lock); +} + +static inline void task_unlock(struct task_struct *p) +{ + spin_unlock(&p->alloc_lock); +} + +/* write full pathname into buffer and return start of pathname */ +static inline char * d_path(struct dentry *dentry, struct vfsmount *vfsmnt, + char *buf, int buflen) +{ + char *res; + struct vfsmount *rootmnt; + struct dentry *root; +#if defined(CONFIG_MOSIX_DFSA) || defined(CONFIG_MOSIX_FS) + if(dentry->d_inode && dentry->d_inode->i_op && + dentry->d_inode->i_op->check_path) + dentry->d_inode->i_op->check_path(dentry); +#endif CONFIG_MOSIX_DFSA || CONFIG_MOSIX_FS + read_lock(¤t->fs->lock); + rootmnt = mntget(current->fs->rootmnt); + root = dget(current->fs->root); + read_unlock(¤t->fs->lock); + spin_lock(&dcache_lock); + res = __d_path(dentry, vfsmnt, root, rootmnt, buf, buflen); + spin_unlock(&dcache_lock); + dput(root); + mntput(rootmnt); + return res; +} + +#endif /* __KERNEL__ */ + +#endif diff --git a/sys-kernel/linux-headers/linux-headers-2.4.6-r2.ebuild b/sys-kernel/linux-headers/linux-headers-2.4.6-r2.ebuild index 6a686433b00d..55e4ed5de076 100644 --- a/sys-kernel/linux-headers/linux-headers-2.4.6-r2.ebuild +++ b/sys-kernel/linux-headers/linux-headers-2.4.6-r2.ebuild @@ -128,13 +128,13 @@ src_unpack() { cd MOSIX-${MOSV} tar xzf MOSIX-${MOSV}.tar.gz patches.${OKV} kernel.new.${OKV}.tar cd ${S} - try cat ${S2}/MOSIX-${MOSV}/patches.2.4.6 | patch -p0 - tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.2.4.6.tar + try cat ${S2}/MOSIX-${MOSV}/patches.${KV} | patch -p0 + tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.${KV}.tar fi cd ${S} echo "Applying reiserfs-NFS fix..." - try cat ${FILESDIR}/2.4.6/linux-2.4.6-reiserfs-NFS.patch | patch -N -p1 + try cat ${FILESDIR}/${PVR}/linux-${KV}-reiserfs-NFS.patch | patch -N -p1 if [ "`use lvm`" ] then @@ -146,14 +146,14 @@ src_unpack() { # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${S}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" + CFLAGS="${CFLAGS} -I${S}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" cd PATCHES try make KERNEL_VERSION=${KV} KERNEL_DIR=${S} cd ${S} # the -l option allows this patch to apply cleanly (ignore whitespace changes) try patch -l -p1 < ${S2}/LVM/${LVMV}/PATCHES/lvm-${LVMV}-${KV}.patch cd ${S}/drivers/md - try patch -p0 < ${FILESDIR}/${KV}/lvm.c.diff + try patch -p0 < ${FILESDIR}/${PVR}/lvm.c.diff fi # if [ "`use lm_sensors`" ] @@ -193,7 +193,7 @@ src_unpack() { # echo "Unpacking pcmcia-cs tools..." # cd ${S2} # unpack pcmcia-cs-${PCV}.tar.gz - # patch -p0 < ${FILESDIR}/${KV}/pcmcia-cs-${PCV}-gentoo.diff + # patch -p0 < ${FILESDIR}/${PVR}/pcmcia-cs-${PCV}-gentoo.diff # fi #JFS patch works; commented out because it's not ready for production use @@ -223,7 +223,7 @@ src_unpack() { echo echo "Fixing reject in include/linux/sched.h..." echo - cp ${FILESDIR}/${KV}/sched.h include/linux + cp ${FILESDIR}/${PVR}/sched.h include/linux fi fi @@ -271,8 +271,8 @@ src_compile() { fi # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${KS}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" - try make + CFLAGS="${CFLAGS} -I${KS}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" + try make fi # if [ "`use lm_sensors`" ] @@ -294,6 +294,7 @@ src_compile() { if [ "$PN" == "linux" ] then + cd ${KS} try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" dep try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" bzImage try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" modules @@ -326,7 +327,7 @@ src_install() { then cd ${KS2}/LVM/${LVMV}/tools - try CFLAGS="${CFLAGS} -I${KS}/include" make install -e prefix=${D} mandir=${D}/usr/share/man \ + CFLAGS="${CFLAGS} -I${KS}/include" try make install -e prefix=${D} mandir=${D}/usr/share/man \ sbindir=${D}/sbin libdir=${D}/lib #no need for a static library in /lib mv ${D}/lib/*.a ${D}/usr/lib diff --git a/sys-kernel/linux-sources/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch b/sys-kernel/linux-sources/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch new file mode 100644 index 000000000000..5f3830c8ffe1 --- /dev/null +++ b/sys-kernel/linux-sources/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch @@ -0,0 +1,647 @@ +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/inode.c linux-2.4.6-mmap/fs/nfs/inode.c +--- linux-2.4.6-file/fs/nfs/inode.c Wed Jul 4 17:30:28 2001 ++++ linux-2.4.6-mmap/fs/nfs/inode.c Wed Jul 4 20:48:01 2001 +@@ -48,6 +48,7 @@ + static void nfs_write_inode(struct inode *,int); + static void nfs_delete_inode(struct inode *); + static void nfs_put_super(struct super_block *); ++static void nfs_clear_inode(struct inode *); + static void nfs_umount_begin(struct super_block *); + static int nfs_statfs(struct super_block *, struct statfs *); + +@@ -57,6 +58,7 @@ + delete_inode: nfs_delete_inode, + put_super: nfs_put_super, + statfs: nfs_statfs, ++ clear_inode: nfs_clear_inode, + umount_begin: nfs_umount_begin, + }; + +@@ -141,6 +143,21 @@ + clear_inode(inode); + } + ++/* ++ * For the moment, the only task for the NFS clear_inode method is to ++ * release the mmap credential ++ */ ++static void ++nfs_clear_inode(struct inode *inode) ++{ ++ struct rpc_cred *cred = NFS_I(inode)->mm_cred; ++ ++ if (cred) { ++ put_rpccred(cred); ++ NFS_I(inode)->mm_cred = 0; ++ } ++} ++ + void + nfs_put_super(struct super_block *sb) + { +@@ -600,7 +617,6 @@ + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + NFS_CACHE_CTIME(inode) = fattr->ctime; + NFS_CACHE_MTIME(inode) = fattr->mtime; +- NFS_CACHE_ATIME(inode) = fattr->atime; + NFS_CACHE_ISIZE(inode) = fattr->size; + NFS_ATTRTIMEO(inode) = NFS_MINATTRTIMEO(inode); + NFS_ATTRTIMEO_UPDATE(inode) = jiffies; +@@ -794,6 +810,20 @@ + } + + /* ++ * Ensure that mmap has a recent RPC credential for use when writing out ++ * shared pages ++ */ ++static inline void ++nfs_set_mmcred(struct inode *inode, struct rpc_cred *cred) ++{ ++ struct rpc_cred *oldcred = NFS_I(inode)->mm_cred; ++ ++ NFS_I(inode)->mm_cred = get_rpccred(cred); ++ if (oldcred) ++ put_rpccred(oldcred); ++} ++ ++/* + * These are probably going to contain hooks for + * allocating and releasing RPC credentials for + * the file. I'll have to think about Tronds patch +@@ -808,20 +838,20 @@ + auth = NFS_CLIENT(inode)->cl_auth; + cred = rpcauth_lookupcred(auth, 0); + filp->private_data = cred; ++ if (filp->f_mode & FMODE_WRITE) ++ nfs_set_mmcred(inode, cred); + unlock_kernel(); + return 0; + } + + int nfs_release(struct inode *inode, struct file *filp) + { +- struct rpc_auth *auth; + struct rpc_cred *cred; + + lock_kernel(); +- auth = NFS_CLIENT(inode)->cl_auth; + cred = nfs_file_cred(filp); + if (cred) +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + unlock_kernel(); + return 0; + } +@@ -976,7 +1006,6 @@ + NFS_CACHE_CTIME(inode) = fattr->ctime; + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + +- NFS_CACHE_ATIME(inode) = fattr->atime; + inode->i_atime = nfs_time_to_secs(fattr->atime); + + NFS_CACHE_MTIME(inode) = new_mtime; +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/unlink.c linux-2.4.6-mmap/fs/nfs/unlink.c +--- linux-2.4.6-file/fs/nfs/unlink.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/fs/nfs/unlink.c Wed Jul 4 20:48:47 2001 +@@ -128,7 +128,7 @@ + dir_i = dir->d_inode; + nfs_zap_caches(dir_i); + NFS_PROTO(dir_i)->unlink_done(dir, &task->tk_msg); +- rpcauth_releasecred(task->tk_auth, data->cred); ++ put_rpccred(data->cred); + data->cred = NULL; + dput(dir); + } +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/write.c linux-2.4.6-mmap/fs/nfs/write.c +--- linux-2.4.6-file/fs/nfs/write.c Mon May 21 21:29:14 2001 ++++ linux-2.4.6-mmap/fs/nfs/write.c Wed Jul 4 20:48:24 2001 +@@ -181,7 +181,9 @@ + + + if (file) +- cred = nfs_file_cred(file); ++ cred = get_rpccred(nfs_file_cred(file)); ++ if (!cred) ++ cred = get_rpccred(NFS_I(inode)->mm_cred); + + dprintk("NFS: nfs_writepage_sync(%x/%Ld %d@%Ld)\n", + inode->i_dev, (long long)NFS_FILEID(inode), +@@ -226,6 +228,8 @@ + + io_error: + kunmap(page); ++ if (cred) ++ put_rpccred(cred); + + return written? written : result; + } +@@ -241,6 +245,9 @@ + status = (IS_ERR(req)) ? PTR_ERR(req) : 0; + if (status < 0) + goto out; ++ if (!req->wb_cred) ++ req->wb_cred = get_rpccred(NFS_I(inode)->mm_cred); ++ nfs_unlock_request(req); + nfs_release_request(req); + nfs_strategy(inode); + out: +@@ -557,13 +564,11 @@ + req->wb_bytes = count; + req->wb_file = file; + +- /* If we have a struct file, use its cached credentials +- * else cache the current process' credentials. */ ++ /* If we have a struct file, use its cached credentials */ + if (file) { + get_file(file); + req->wb_cred = nfs_file_cred(file); +- } else +- req->wb_cred = rpcauth_lookupcred(NFS_CLIENT(inode)->cl_auth, 0); ++ } + req->wb_inode = inode; + req->wb_count = 1; + +@@ -608,8 +613,8 @@ + /* Release struct file or cached credential */ + if (req->wb_file) + fput(req->wb_file); +- else +- rpcauth_releasecred(NFS_CLIENT(inode)->cl_auth, req->wb_cred); ++ else if (req->wb_cred) ++ put_rpccred(req->wb_cred); + page_cache_release(page); + nfs_page_free(req); + /* wake up anyone waiting to allocate a request */ +@@ -927,8 +932,6 @@ + if (end > rqend) + req->wb_bytes = end - req->wb_offset; + +- nfs_unlock_request(req); +- + return req; + } + +@@ -1049,6 +1052,7 @@ + goto done; + + status = 0; ++ nfs_unlock_request(req); + /* If we wrote past the end of the page. + * Call the strategy routine so it can send out a bunch + * of requests. +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs.h linux-2.4.6-mmap/include/linux/nfs_fs.h +--- linux-2.4.6-file/include/linux/nfs_fs.h Sat May 26 03:02:11 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs.h Wed Jul 4 17:32:16 2001 +@@ -63,6 +63,11 @@ + */ + #define NFS_SUPER_MAGIC 0x6969 + ++static inline struct nfs_inode_info *NFS_I(struct inode *inode) ++{ ++ return &inode->u.nfs_i; ++} ++ + #define NFS_FH(inode) (&(inode)->u.nfs_i.fh) + #define NFS_SERVER(inode) (&(inode)->i_sb->u.nfs_sb.s_server) + #define NFS_CLIENT(inode) (NFS_SERVER(inode)->client) +@@ -74,7 +79,6 @@ + #define NFS_READTIME(inode) ((inode)->u.nfs_i.read_cache_jiffies) + #define NFS_CACHE_CTIME(inode) ((inode)->u.nfs_i.read_cache_ctime) + #define NFS_CACHE_MTIME(inode) ((inode)->u.nfs_i.read_cache_mtime) +-#define NFS_CACHE_ATIME(inode) ((inode)->u.nfs_i.read_cache_atime) + #define NFS_CACHE_ISIZE(inode) ((inode)->u.nfs_i.read_cache_isize) + #define NFS_NEXTSCAN(inode) ((inode)->u.nfs_i.nextscan) + #define NFS_CACHEINV(inode) \ +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs_i.h linux-2.4.6-mmap/include/linux/nfs_fs_i.h +--- linux-2.4.6-file/include/linux/nfs_fs_i.h Thu May 31 15:27:10 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs_i.h Wed Jul 4 17:32:16 2001 +@@ -45,7 +45,6 @@ + unsigned long read_cache_jiffies; + __u64 read_cache_ctime; + __u64 read_cache_mtime; +- __u64 read_cache_atime; + __u64 read_cache_isize; + unsigned long attrtimeo; + unsigned long attrtimeo_timestamp; +@@ -73,6 +72,9 @@ + struct inode *hash_next, + *hash_prev; + unsigned long nextscan; ++ ++ /* Credentials for shared mmap */ ++ struct rpc_cred *mm_cred; + }; + + /* +diff -u --recursive --new-file linux-2.4.6-file/include/linux/sunrpc/auth.h linux-2.4.6-mmap/include/linux/sunrpc/auth.h +--- linux-2.4.6-file/include/linux/sunrpc/auth.h Fri Jun 22 19:30:22 2001 ++++ linux-2.4.6-mmap/include/linux/sunrpc/auth.h Wed Jul 4 20:51:35 2001 +@@ -14,6 +14,8 @@ + #include <linux/config.h> + #include <linux/sunrpc/sched.h> + ++#include <asm/atomic.h> ++ + /* size of the nodename buffer */ + #define UNX_MAXNODENAME 32 + +@@ -22,8 +24,10 @@ + */ + struct rpc_cred { + struct rpc_cred * cr_next; /* linked list */ ++ struct rpc_auth * cr_auth; ++ struct rpc_credops * cr_ops; + unsigned long cr_expire; /* when to gc */ +- unsigned short cr_count; /* ref count */ ++ atomic_t cr_count; /* ref count */ + unsigned short cr_flags; /* various flags */ + #ifdef RPC_DEBUG + unsigned long cr_magic; /* 0x0f4aa4f0 */ +@@ -71,6 +75,9 @@ + void (*destroy)(struct rpc_auth *); + + struct rpc_cred * (*crcreate)(int); ++}; ++ ++struct rpc_credops { + void (*crdestroy)(struct rpc_cred *); + + int (*crmatch)(struct rpc_cred *, int); +@@ -92,8 +99,7 @@ + struct rpc_cred * rpcauth_lookupcred(struct rpc_auth *, int); + struct rpc_cred * rpcauth_bindcred(struct rpc_task *); + void rpcauth_holdcred(struct rpc_task *); +-void rpcauth_releasecred(struct rpc_auth *, +- struct rpc_cred *); ++void put_rpccred(struct rpc_cred *); + void rpcauth_unbindcred(struct rpc_task *); + int rpcauth_matchcred(struct rpc_auth *, + struct rpc_cred *, int); +@@ -106,6 +112,13 @@ + void rpcauth_free_credcache(struct rpc_auth *); + void rpcauth_insert_credcache(struct rpc_auth *, + struct rpc_cred *); ++ ++static inline ++struct rpc_cred * get_rpccred(struct rpc_cred *cred) ++{ ++ atomic_inc(&cred->cr_count); ++ return cred; ++} + + #endif /* __KERNEL__ */ + #endif /* _LINUX_SUNRPC_AUTH_H */ +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth.c linux-2.4.6-mmap/net/sunrpc/auth.c +--- linux-2.4.6-file/net/sunrpc/auth.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth.c Wed Jul 4 20:51:04 2001 +@@ -81,42 +81,61 @@ + auth->au_nextgc = jiffies + (auth->au_expire >> 1); + } + ++/* ++ * Destroy an unreferenced credential ++ */ + static inline void +-rpcauth_crdestroy(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_crdestroy(struct rpc_cred *cred) + { + #ifdef RPC_DEBUG + if (cred->cr_magic != RPCAUTH_CRED_MAGIC) + BUG(); + cred->cr_magic = 0; ++ if (atomic_read(&cred->cr_count) || cred->cr_auth) ++ BUG(); + #endif +- if (auth->au_ops->crdestroy) +- auth->au_ops->crdestroy(cred); +- else +- rpc_free(cred); ++ cred->cr_ops->crdestroy(cred); + } + + /* +- * Clear the RPC credential cache ++ * Destroy a list of credentials ++ */ ++static inline ++void rpcauth_destroy_credlist(struct rpc_cred *head) ++{ ++ struct rpc_cred *cred; ++ ++ while ((cred = head) != NULL) { ++ head = cred->cr_next; ++ rpcauth_crdestroy(cred); ++ } ++} ++ ++/* ++ * Clear the RPC credential cache, and delete those credentials ++ * that are not referenced. + */ + void + rpcauth_free_credcache(struct rpc_auth *auth) + { +- struct rpc_cred **q, *cred; +- void (*destroy)(struct rpc_cred *); ++ struct rpc_cred **q, *cred, *free = NULL; + int i; + +- if (!(destroy = auth->au_ops->crdestroy)) +- destroy = (void (*)(struct rpc_cred *)) rpc_free; +- + spin_lock(&rpc_credcache_lock); + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { + *q = cred->cr_next; +- destroy(cred); ++ cred->cr_auth = NULL; ++ if (atomic_read(&cred->cr_count) == 0) { ++ cred->cr_next = free; ++ free = cred; ++ } else ++ cred->cr_next = NULL; + } + } + spin_unlock(&rpc_credcache_lock); ++ rpcauth_destroy_credlist(free); + } + + /* +@@ -133,9 +152,10 @@ + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { +- if (!cred->cr_count && ++ if (!atomic_read(&cred->cr_count) && + time_before(cred->cr_expire, jiffies)) { + *q = cred->cr_next; ++ cred->cr_auth = NULL; + cred->cr_next = free; + free = cred; + continue; +@@ -144,10 +164,7 @@ + } + } + spin_unlock(&rpc_credcache_lock); +- while ((cred = free) != NULL) { +- free = cred->cr_next; +- rpcauth_crdestroy(auth, cred); +- } ++ rpcauth_destroy_credlist(free); + auth->au_nextgc = jiffies + auth->au_expire; + } + +@@ -163,8 +180,8 @@ + spin_lock(&rpc_credcache_lock); + cred->cr_next = auth->au_credcache[nr]; + auth->au_credcache[nr] = cred; +- cred->cr_count++; +- cred->cr_expire = jiffies + auth->au_expire; ++ cred->cr_auth = auth; ++ get_rpccred(cred); + spin_unlock(&rpc_credcache_lock); + } + +@@ -187,7 +204,7 @@ + q = &auth->au_credcache[nr]; + while ((cred = *q) != NULL) { + if (!(cred->cr_flags & RPCAUTH_CRED_DEAD) && +- auth->au_ops->crmatch(cred, taskflags)) { ++ cred->cr_ops->crmatch(cred, taskflags)) { + *q = cred->cr_next; + break; + } +@@ -213,23 +230,23 @@ + * Remove cred handle from cache + */ + static void +-rpcauth_remove_credcache(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_remove_credcache(struct rpc_cred *cred) + { ++ struct rpc_auth *auth = cred->cr_auth; + struct rpc_cred **q, *cr; + int nr; + + nr = (cred->cr_uid & RPC_CREDCACHE_MASK); +- spin_lock(&rpc_credcache_lock); + q = &auth->au_credcache[nr]; + while ((cr = *q) != NULL) { + if (cred == cr) { + *q = cred->cr_next; + cred->cr_next = NULL; ++ cred->cr_auth = NULL; + break; + } + q = &cred->cr_next; + } +- spin_unlock(&rpc_credcache_lock); + } + + struct rpc_cred * +@@ -258,7 +275,7 @@ + { + dprintk("RPC: matching %s cred %d\n", + auth->au_ops->au_name, taskflags); +- return auth->au_ops->crmatch(cred, taskflags); ++ return cred->cr_ops->crmatch(cred, taskflags); + } + + void +@@ -266,26 +283,25 @@ + { + dprintk("RPC: %4d holding %s cred %p\n", + task->tk_pid, task->tk_auth->au_ops->au_name, task->tk_msg.rpc_cred); +- if (task->tk_msg.rpc_cred) { +- spin_lock(&rpc_credcache_lock); +- task->tk_msg.rpc_cred->cr_count++; +- task->tk_msg.rpc_cred->cr_expire = jiffies + task->tk_auth->au_expire; +- spin_unlock(&rpc_credcache_lock); +- } ++ if (task->tk_msg.rpc_cred) ++ get_rpccred(task->tk_msg.rpc_cred); + } + + void +-rpcauth_releasecred(struct rpc_auth *auth, struct rpc_cred *cred) ++put_rpccred(struct rpc_cred *cred) + { +- spin_lock(&rpc_credcache_lock); +- if (cred != NULL && cred->cr_count > 0) { +- if (!--cred->cr_count && (cred->cr_flags & RPCAUTH_CRED_DEAD)) { +- spin_unlock(&rpc_credcache_lock); +- rpcauth_remove_credcache(auth, cred); +- rpcauth_crdestroy(auth, cred); +- return; +- } ++ if (!atomic_dec_and_lock(&cred->cr_count, &rpc_credcache_lock)) ++ return; ++ ++ if (cred->cr_auth && cred->cr_flags & RPCAUTH_CRED_DEAD) ++ rpcauth_remove_credcache(cred); ++ ++ if (!cred->cr_auth) { ++ spin_unlock(&rpc_credcache_lock); ++ rpcauth_crdestroy(cred); ++ return; + } ++ cred->cr_expire = jiffies + cred->cr_auth->au_expire; + spin_unlock(&rpc_credcache_lock); + } + +@@ -298,7 +314,7 @@ + dprintk("RPC: %4d releasing %s cred %p\n", + task->tk_pid, auth->au_ops->au_name, cred); + +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + task->tk_msg.rpc_cred = NULL; + } + +@@ -306,10 +322,11 @@ + rpcauth_marshcred(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d marshaling %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crmarshal(task, p, ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crmarshal(task, p, + task->tk_flags & RPC_CALL_REALUID); + } + +@@ -317,20 +334,22 @@ + rpcauth_checkverf(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d validating %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crvalidate(task, p); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crvalidate(task, p); + } + + int + rpcauth_refreshcred(struct rpc_task *task) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d refreshing %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- task->tk_status = auth->au_ops->crrefresh(task); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ task->tk_status = cred->cr_ops->crrefresh(task); + return task->tk_status; + } + +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_null.c linux-2.4.6-mmap/net/sunrpc/auth_null.c +--- linux-2.4.6-file/net/sunrpc/auth_null.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_null.c Wed Jul 4 19:27:38 2001 +@@ -17,6 +17,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops null_credops; ++ + static struct rpc_auth * + nul_create(struct rpc_clnt *clnt) + { +@@ -52,9 +54,10 @@ + + if (!(cred = (struct rpc_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; +- cred->cr_count = 0; ++ atomic_set(&cred->cr_count, 0); + cred->cr_flags = RPCAUTH_CRED_UPTODATE; + cred->cr_uid = current->uid; ++ cred->cr_ops = &null_credops; + + return cred; + } +@@ -124,7 +127,11 @@ + #endif + nul_create, + nul_destroy, +- nul_create_cred, ++ nul_create_cred ++}; ++ ++static ++struct rpc_credops null_credops = { + nul_destroy_cred, + nul_match, + nul_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_unix.c linux-2.4.6-mmap/net/sunrpc/auth_unix.c +--- linux-2.4.6-file/net/sunrpc/auth_unix.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_unix.c Wed Jul 4 19:29:07 2001 +@@ -33,6 +33,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops unix_credops; ++ + static struct rpc_auth * + unx_create(struct rpc_clnt *clnt) + { +@@ -71,7 +73,7 @@ + if (!(cred = (struct unx_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 0; ++ atomic_set(&cred->uc_count, 0); + cred->uc_flags = RPCAUTH_CRED_UPTODATE; + if (flags & RPC_TASK_ROOTCREDS) { + cred->uc_uid = cred->uc_fsuid = 0; +@@ -91,6 +93,7 @@ + if (i < NFS_NGROUPS) + cred->uc_gids[i] = NOGROUP; + } ++ cred->uc_base.cr_ops = &unix_credops; + + return (struct rpc_cred *) cred; + } +@@ -106,7 +109,7 @@ + if (!(cred = (struct unx_cred *) rpc_malloc(task, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 1; ++ atomic_set(&cred->uc_count, 1); + cred->uc_flags = RPCAUTH_CRED_DEAD|RPCAUTH_CRED_UPTODATE; + cred->uc_uid = uid; + cred->uc_gid = gid; +@@ -236,7 +239,11 @@ + #endif + unx_create, + unx_destroy, +- unx_create_cred, ++ unx_create_cred ++}; ++ ++static ++struct rpc_credops unix_credops = { + unx_destroy_cred, + unx_match, + unx_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/sunrpc_syms.c linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c +--- linux-2.4.6-file/net/sunrpc/sunrpc_syms.c Fri Jan 12 00:53:02 2001 ++++ linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c Wed Jul 4 20:55:19 2001 +@@ -65,7 +65,7 @@ + EXPORT_SYMBOL(rpcauth_lookupcred); + EXPORT_SYMBOL(rpcauth_bindcred); + EXPORT_SYMBOL(rpcauth_matchcred); +-EXPORT_SYMBOL(rpcauth_releasecred); ++EXPORT_SYMBOL(put_rpccred); + + /* RPC server stuff */ + EXPORT_SYMBOL(svc_create); + +_______________________________________________ +NFS maillist - NFS@lists.sourceforge.net +http://lists.sourceforge.net/lists/listinfo/nfs + + diff --git a/sys-kernel/linux-sources/files/2.4.6-r2/lvm.c.diff b/sys-kernel/linux-sources/files/2.4.6-r2/lvm.c.diff new file mode 100644 index 000000000000..6392671b8f7e --- /dev/null +++ b/sys-kernel/linux-sources/files/2.4.6-r2/lvm.c.diff @@ -0,0 +1,20 @@ +--- lvm.c.orig Sat Apr 28 10:48:37 2001 ++++ lvm.c Sat Apr 28 11:07:07 2001 +@@ -1791,7 +1791,7 @@ + int max_hardblocksize = 0, hardblocksize; + + for (le = 0; le < lv->lv_allocated_le; le++) { +- hardblocksize = get_hardblocksize(lv->lv_current_pe[le].dev); ++ hardblocksize = get_hardsect_size(lv->lv_current_pe[le].dev); + if (hardblocksize == 0) + hardblocksize = 512; + if (hardblocksize > max_hardblocksize) +@@ -1801,7 +1801,7 @@ + if (lv->lv_access & LV_SNAPSHOT) { + for (e = 0; e < lv->lv_remap_end; e++) { + hardblocksize = +- get_hardblocksize( ++ get_hardsect_size( + lv->lv_block_exception[e].rdev_new); + if (hardblocksize == 0) + hardblocksize = 512; diff --git a/sys-kernel/linux-sources/files/2.4.6-r2/mosix.init b/sys-kernel/linux-sources/files/2.4.6-r2/mosix.init new file mode 100644 index 000000000000..f339dabb666e --- /dev/null +++ b/sys-kernel/linux-sources/files/2.4.6-r2/mosix.init @@ -0,0 +1,118 @@ +#!/bin/sh - +#RCUPDATE:2 3 4:95:This line is required for script management + +# description: MOSIX is am extension of the operating system, +# supporting scalable and transparent cluster computing. +# +# mosix Script to stop/start MOSIX +# +# Author: Amnon Shiloh + +. /etc/rc.d/config/functions + +SERVICE="mosix" +opts="start stop status restart" + +mosix_stop() { + echo 0 > /proc/mosix/admin/mospe + rm -f /var/lock/subsys/mosix +} + +mosix_start() { + if [ -s /etc/overheads -a -f /proc/mosix/admin/overheads ] + then + /bin/grep -v '^#' /etc/overheads > /proc/mosix/admin/overheads + fi + if [ -s /etc/mfscosts -a -f /proc/mosix/admin/mfscosts ] + then + /bin/grep -v '^#' /etc/mfscosts > /proc/mosix/admin/mfscosts + fi + a1= + [ -s /etc/mospe ] && a1="-p `cat /etc/mospe`" + a2= + [ -s /etc/mosgates ] && a2="-g `cat /etc/mosgates`" + /sbin/setpe -W $a1 $a2 -f /etc/mosix.map + touch /var/lock/subsys/mosix +} + +alarm() { + t=$2 + while : + do + case "$t" in 0) break ;; esac + sleep 1 + t=$(($t-1)) + done + kill -1 $1 + exit 0 +} + +# See how we were called. +start() { + eecho "Initializing MOSIX" + if [ ! -f /etc/mosix.map ] + then + echo You have no MOSIX configuration - please edit yours now: + echo -n "Editor to use [q to quit] - [vi] :- " + trap 'echo Timed-Out ; exit 1' 1 + alarm $$ 120 & + killer=$! + read editor + kill $killer + trap '' + case "$editor" in [qQ]) exit ;; + "") editor=/usr/bin/vi + [ -f $editor ] || editor=/bin/vi + ;; + esac + echo "# MOSIX CONFIGURATION" > /etc/mosix.map + echo "# ===================" >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# Each line should contain 3 fields, mapping IP addresses to MOSIX node-numbers:" >> /etc/mosix.map + echo "# 1) first MOSIX node-number in range." >> /etc/mosix.map + echo "# 2) IP address of the above node (or node-name from /etc/hosts)." >> /etc/mosix.map + echo "# 3) number of nodes in this range." >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# MOSIX-# IP number-of-nodes" >> /etc/mosix.map + echo "# ============================" >> /etc/mosix.map + $editor /etc/mosix.map + echo + while : + do + echo "If this node's standard IP address is not part of the table that" + echo "you just edited, because MOSIX uses a different or separate network," + echo "you need to type this node's MOSIX-number now." + echo + /bin/echo -n "Otherwise please press only <Enter> :- " + read me + case "$me" in "") /bin/rm -f /etc/mospe ; break ;; + [1-9] | [1-9][0-9] | [1-9][0-9][0-9] | \ + [1-9][0-9][0-9][0-9] | \ + [1-6][0-9][0-9][0-9][0-9]) + echo $me > /etc/mospe ; break ;; + *) echo Improper MOSIX number - please try again: ;; + esac + done + fi + start_mosix + eend +} + +stop() { + eecho "Stopping MOSIX" + stop_mosix + eend +} + +status() { + /sbin/setpe -r +} + +restart() { + eecho "Restarting MOSIX" + stop_mosix + start_mosix + eend +} + +doservice ${@} diff --git a/sys-kernel/linux-sources/files/2.4.6-r2/sched.h b/sys-kernel/linux-sources/files/2.4.6-r2/sched.h new file mode 100644 index 000000000000..438203f84cdd --- /dev/null +++ b/sys-kernel/linux-sources/files/2.4.6-r2/sched.h @@ -0,0 +1,997 @@ +#ifndef _LINUX_SCHED_H +#define _LINUX_SCHED_H + +#include <asm/param.h> /* for HZ */ + +extern unsigned long event; + +#include <linux/config.h> +#include <linux/binfmts.h> +#include <linux/personality.h> +#include <linux/threads.h> +#include <linux/kernel.h> +#include <linux/types.h> +#include <linux/times.h> +#include <linux/timex.h> + +#include <asm/system.h> +#include <asm/semaphore.h> +#include <asm/page.h> +#include <asm/ptrace.h> +#include <asm/mmu.h> + +#include <linux/smp.h> +#include <linux/tty.h> +#include <linux/sem.h> +#include <linux/signal.h> +#include <linux/securebits.h> + +#ifdef CONFIG_MOSIX +#include <mos/mosixtask.h> +#else +#define MOSIX_INIT_TASK(tsk) +#endif CONFIG_MOSIX + +#include <linux/fs_struct.h> + +/* + * cloning flags: + */ +#define CSIGNAL 0x000000ff /* signal mask to be sent at exit */ +#define CLONE_VM 0x00000100 /* set if VM shared between processes */ +#define CLONE_FS 0x00000200 /* set if fs info shared between processes */ +#define CLONE_FILES 0x00000400 /* set if open files shared between processes */ +#define CLONE_SIGHAND 0x00000800 /* set if signal handlers and blocked signals shared */ +#define CLONE_PID 0x00001000 /* set if pid shared */ +#define CLONE_PTRACE 0x00002000 /* set if we want to let tracing continue on the child too */ +#define CLONE_VFORK 0x00004000 /* set if the parent wants the child to wake it up on mm_release */ +#define CLONE_PARENT 0x00008000 /* set if we want to have the same parent as the cloner */ +#define CLONE_THREAD 0x00010000 /* Same thread group? */ + +#define CLONE_SIGNAL (CLONE_SIGHAND | CLONE_THREAD) + +/* + * These are the constant used to fake the fixed-point load-average + * counting. Some notes: + * - 11 bit fractions expand to 22 bits by the multiplies: this gives + * a load-average precision of 10 bits integer + 11 bits fractional + * - if you want to count load-averages more often, you need more + * precision, or rounding will get you. With 2-second counting freq, + * the EXP_n values would be 1981, 2034 and 2043 if still using only + * 11 bit fractions. + */ +extern unsigned long avenrun[]; /* Load averages */ + +#define FSHIFT 11 /* nr of bits of precision */ +#define FIXED_1 (1<<FSHIFT) /* 1.0 as fixed-point */ +#define LOAD_FREQ (5*HZ) /* 5 sec intervals */ +#define EXP_1 1884 /* 1/exp(5sec/1min) as fixed-point */ +#define EXP_5 2014 /* 1/exp(5sec/5min) */ +#define EXP_15 2037 /* 1/exp(5sec/15min) */ + +#define CALC_LOAD(load,exp,n) \ + load *= exp; \ + load += n*(FIXED_1-exp); \ + load >>= FSHIFT; + +#define CT_TO_SECS(x) ((x) / HZ) +#define CT_TO_USECS(x) (((x) % HZ) * 1000000/HZ) + +extern int nr_running, nr_threads; +extern int last_pid; + +#include <linux/fs.h> +#include <linux/time.h> +#include <linux/param.h> +#include <linux/resource.h> +#include <linux/timer.h> + +#include <asm/processor.h> + +#define TASK_RUNNING 0 +#define TASK_INTERRUPTIBLE 1 +#define TASK_UNINTERRUPTIBLE 2 +#define TASK_ZOMBIE 4 +#define TASK_STOPPED 8 +#ifdef CONFIG_MOSIX +#define TASK_SAME 16 +#endif CONFIG_MOSIX + +#define __set_task_state(tsk, state_value) \ + do { (tsk)->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_task_state(tsk, state_value) \ + set_mb((tsk)->state, (state_value)) +#else +#define set_task_state(tsk, state_value) \ + __set_task_state((tsk), (state_value)) +#endif + +#define __set_current_state(state_value) \ + do { current->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_current_state(state_value) \ + set_mb(current->state, (state_value)) +#else +#define set_current_state(state_value) \ + __set_current_state(state_value) +#endif + +/* + * Scheduling policies + */ +#define SCHED_OTHER 0 +#define SCHED_FIFO 1 +#define SCHED_RR 2 + +/* + * This is an additional bit set when we want to + * yield the CPU for one re-schedule.. + */ +#define SCHED_YIELD 0x10 + +struct sched_param { + int sched_priority; +}; + +#ifdef __KERNEL__ + +#include <linux/spinlock.h> + +/* + * This serializes "schedule()" and also protects + * the run-queue from deletions/modifications (but + * _adding_ to the beginning of the run-queue has + * a separate lock). + */ +extern rwlock_t tasklist_lock; +extern spinlock_t runqueue_lock; +extern spinlock_t mmlist_lock; + +extern void sched_init(void); +extern void init_idle(void); +extern void show_state(void); +extern void cpu_init (void); +extern void trap_init(void); +extern void update_process_times(int user); +extern void update_one_process(struct task_struct *p, unsigned long user, + unsigned long system, int cpu); + +#define MAX_SCHEDULE_TIMEOUT LONG_MAX +extern signed long FASTCALL(schedule_timeout(signed long timeout)); +asmlinkage void schedule(void); + +extern int schedule_task(struct tq_struct *task); +extern void flush_scheduled_tasks(void); +extern int start_context_thread(void); +extern int current_is_keventd(void); + +/* + * The default fd array needs to be at least BITS_PER_LONG, + * as this is the granularity returned by copy_fdset(). + */ +#define NR_OPEN_DEFAULT BITS_PER_LONG + +/* + * Open file table structure + */ +struct files_struct { + atomic_t count; + rwlock_t file_lock; /* Protects all the below members. Nests inside tsk->alloc_lock */ + int max_fds; + int max_fdset; + int next_fd; + struct file ** fd; /* current fd array */ + fd_set *close_on_exec; + fd_set *open_fds; + fd_set close_on_exec_init; + fd_set open_fds_init; + struct file * fd_array[NR_OPEN_DEFAULT]; +#ifdef CONFIG_MOSIX_DFSA + atomic_t users; /* the real count, excluding /proc grabs */ + int count_dropped_to_one; + int maxclosed; + fd_set *closed; + fd_set closed_fds_init; + int maxopened; + fd_set *opened; + fd_set opened_fds_init; + int maxmod; + fd_set *modified; + fd_set modified_fds_init; +#define INIT_FILES_DFSA , users : ATOMIC_INIT(1), \ + closed: &init_files.closed_fds_init, \ + opened: &init_files.opened_fds_init, \ + modified: &init_files.modified_fds_init, +#else +#define INIT_FILES_DFSA +#endif CONFIG_MOSIX_DFSA +}; + +#define INIT_FILES \ +{ \ + count: ATOMIC_INIT(1), \ + file_lock: RW_LOCK_UNLOCKED, \ + max_fds: NR_OPEN_DEFAULT, \ + max_fdset: __FD_SETSIZE, \ + next_fd: 0, \ + fd: &init_files.fd_array[0], \ + close_on_exec: &init_files.close_on_exec_init, \ + open_fds: &init_files.open_fds_init, \ + close_on_exec_init: { { 0, } }, \ + open_fds_init: { { 0, } }, \ + fd_array: { NULL, } \ + INIT_FILES_DFSA \ +} + +/* Maximum number of active map areas.. This is a random (large) number */ +#define MAX_MAP_COUNT (65536) + +/* Number of map areas at which the AVL tree is activated. This is arbitrary. */ +#define AVL_MIN_MAP_COUNT 32 + +struct mm_struct { + struct vm_area_struct * mmap; /* list of VMAs */ + struct vm_area_struct * mmap_avl; /* tree of VMAs */ + struct vm_area_struct * mmap_cache; /* last find_vma result */ + pgd_t * pgd; + atomic_t mm_users; /* How many users with user space? */ + atomic_t mm_count; /* How many references to "struct mm_struct" (users count as 1) */ + int map_count; /* number of VMAs */ + struct rw_semaphore mmap_sem; + spinlock_t page_table_lock; /* Protects task page tables and mm->rss */ + + struct list_head mmlist; /* List of all active mm's. These are globally strung + * together off init_mm.mmlist, and are protected + * by mmlist_lock + */ + + unsigned long start_code, end_code, start_data, end_data; + unsigned long start_brk, brk, start_stack; + unsigned long arg_start, arg_end, env_start, env_end; + unsigned long rss, total_vm, locked_vm; + unsigned long def_flags; + unsigned long cpu_vm_mask; + unsigned long swap_address; + + /* Architecture-specific MM context */ + mm_context_t context; +#ifdef CONFIG_MOSIX + atomic_t mm_kiocount; + atomic_t mm_realusers; + unsigned long last_memsort; + int used, unused, swapped; /* break-down of memory pages */ + int private_unused; /* unused but not active/inactive */ + int mark; +#define INIT_MM_MOSIX mm_kiocount: ATOMIC_INIT(0), mm_realusers: ATOMIC_INIT(1), +#else +#define INIT_MM_MOSIX +#endif CONFIG_MOSIX +}; + +extern int mmlist_nr; + +#define INIT_MM(name) \ +{ \ + mmap: &init_mmap, \ + mmap_avl: NULL, \ + mmap_cache: NULL, \ + pgd: swapper_pg_dir, \ + mm_users: ATOMIC_INIT(2), \ + mm_count: ATOMIC_INIT(1), \ + map_count: 1, \ + mmap_sem: __RWSEM_INITIALIZER(name.mmap_sem), \ + page_table_lock: SPIN_LOCK_UNLOCKED, \ + mmlist: LIST_HEAD_INIT(name.mmlist), \ + INIT_MM_MOSIX \ +} + +struct signal_struct { + atomic_t count; + struct k_sigaction action[_NSIG]; + spinlock_t siglock; +}; + + +#define INIT_SIGNALS { \ + count: ATOMIC_INIT(1), \ + action: { {{0,}}, }, \ + siglock: SPIN_LOCK_UNLOCKED \ +} + +/* + * Some day this will be a full-fledged user tracking system.. + */ +struct user_struct { + atomic_t __count; /* reference count */ + atomic_t processes; /* How many processes does this user have? */ + atomic_t files; /* How many open files does this user have? */ + + /* Hash table maintenance information */ + struct user_struct *next, **pprev; + uid_t uid; +}; + +#define get_current_user() ({ \ + struct user_struct *__user = current->user; \ + atomic_inc(&__user->__count); \ + __user; }) + +extern struct user_struct root_user; +#define INIT_USER (&root_user) + +struct task_struct { + /* + * offsets of these are hardcoded elsewhere - touch with care + */ + volatile long state; /* -1 unrunnable, 0 runnable, >0 stopped */ + unsigned long flags; /* per process flags, defined below */ + int sigpending; + mm_segment_t addr_limit; /* thread address space: + 0-0xBFFFFFFF for user-thead + 0-0xFFFFFFFF for kernel-thread + */ + struct exec_domain *exec_domain; + volatile long need_resched; + unsigned long ptrace; + + int lock_depth; /* Lock depth */ + +/* + * offset 32 begins here on 32-bit platforms. We keep + * all fields in a single cacheline that are needed for + * the goodness() loop in schedule(). + */ + long counter; + long nice; + unsigned long policy; + struct mm_struct *mm; + int has_cpu, processor; + unsigned long cpus_allowed; + /* + * (only the 'next' pointer fits into the cacheline, but + * that's just fine.) + */ + struct list_head run_list; + unsigned long sleep_time; + + struct task_struct *next_task, *prev_task; + struct mm_struct *active_mm; + +/* task state */ + struct linux_binfmt *binfmt; + int exit_code, exit_signal; + int pdeath_signal; /* The signal sent when the parent dies */ + /* ??? */ + unsigned long personality; + int dumpable:1; + int did_exec:1; + pid_t pid; + pid_t pgrp; + pid_t tty_old_pgrp; + pid_t session; + pid_t tgid; + /* boolean value for session group leader */ + int leader; + /* + * pointers to (original) parent process, youngest child, younger sibling, + * older sibling, respectively. (p->father can be replaced with + * p->p_pptr->pid) + */ + struct task_struct *p_opptr, *p_pptr, *p_cptr, *p_ysptr, *p_osptr; + struct list_head thread_group; + + /* PID hash table linkage. */ + struct task_struct *pidhash_next; + struct task_struct **pidhash_pprev; + + wait_queue_head_t wait_chldexit; /* for wait4() */ + struct semaphore *vfork_sem; /* for vfork() */ + unsigned long rt_priority; + unsigned long it_real_value, it_prof_value, it_virt_value; + unsigned long it_real_incr, it_prof_incr, it_virt_incr; + struct timer_list real_timer; + struct tms times; + unsigned long start_time; + long per_cpu_utime[NR_CPUS], per_cpu_stime[NR_CPUS]; +/* mm fault and swap info: this can arguably be seen as either mm-specific or thread-specific */ + unsigned long min_flt, maj_flt, nswap, cmin_flt, cmaj_flt, cnswap; + int swappable:1; +/* process credentials */ + uid_t uid,euid,suid,fsuid; + gid_t gid,egid,sgid,fsgid; + int ngroups; + gid_t groups[NGROUPS]; + kernel_cap_t cap_effective, cap_inheritable, cap_permitted; + int keep_capabilities:1; + struct user_struct *user; +/* limits */ + struct rlimit rlim[RLIM_NLIMITS]; + unsigned short used_math; + char comm[16]; +/* file system info */ + int link_count; + struct tty_struct *tty; /* NULL if no tty */ + unsigned int locks; /* How many file locks are being held */ +/* ipc stuff */ + struct sem_undo *semundo; + struct sem_queue *semsleeping; +/* CPU-specific state of this task */ + struct thread_struct thread; +/* filesystem information */ + struct fs_struct *fs; +/* open file information */ + struct files_struct *files; +/* signal handlers */ + spinlock_t sigmask_lock; /* Protects signal and blocked */ + struct signal_struct *sig; + + sigset_t blocked; + struct sigpending pending; + + unsigned long sas_ss_sp; + size_t sas_ss_size; + int (*notifier)(void *priv); + void *notifier_data; + sigset_t *notifier_mask; + +/* Thread group tracking */ + u32 parent_exec_id; + u32 self_exec_id; +/* Protection of (de-)allocation: mm, files, fs, tty */ + spinlock_t alloc_lock; +#ifdef CONFIG_MOSIX + struct mosix_task mosix; +#endif CONFIG_MOSIX + void *journal_info; +}; + +/* + * Per process flags + */ +#define PF_ALIGNWARN 0x00000001 /* Print alignment warning msgs */ + /* Not implemented yet, only for 486*/ +#define PF_STARTING 0x00000002 /* being created */ +#define PF_EXITING 0x00000004 /* getting shut down */ +#define PF_FORKNOEXEC 0x00000040 /* forked but didn't exec */ +#define PF_SUPERPRIV 0x00000100 /* used super-user privileges */ +#define PF_DUMPCORE 0x00000200 /* dumped core */ +#define PF_SIGNALED 0x00000400 /* killed by a signal */ +#define PF_MEMALLOC 0x00000800 /* Allocating memory */ +#define PF_VFORK 0x00001000 /* Wake up parent in mm_release */ + +#define PF_USEDFPU 0x00100000 /* task used FPU this quantum (SMP) */ + +/* + * Ptrace flags + */ + +#define PT_PTRACED 0x00000001 +#define PT_TRACESYS 0x00000002 +#define PT_DTRACE 0x00000004 /* delayed trace (used on m68k, i386) */ +#define PT_TRACESYSGOOD 0x00000008 + +/* + * Limit the stack by to some sane default: root can always + * increase this limit if needed.. 8MB seems reasonable. + */ +#define _STK_LIM (8*1024*1024) + +#define DEF_COUNTER (10*HZ/100) /* 100 ms time slice */ +#define MAX_COUNTER (20*HZ/100) +#define DEF_NICE (0) + +/* + * INIT_TASK is used to set up the first task table, touch at + * your own risk!. Base=0, limit=0x1fffff (=2MB) + */ +#define INIT_TASK(tsk) \ +{ \ + state: 0, \ + flags: 0, \ + sigpending: 0, \ + addr_limit: KERNEL_DS, \ + exec_domain: &default_exec_domain, \ + lock_depth: -1, \ + counter: DEF_COUNTER, \ + nice: DEF_NICE, \ + policy: SCHED_OTHER, \ + mm: NULL, \ + active_mm: &init_mm, \ + cpus_allowed: -1, \ + run_list: LIST_HEAD_INIT(tsk.run_list), \ + next_task: &tsk, \ + prev_task: &tsk, \ + p_opptr: &tsk, \ + p_pptr: &tsk, \ + thread_group: LIST_HEAD_INIT(tsk.thread_group), \ + wait_chldexit: __WAIT_QUEUE_HEAD_INITIALIZER(tsk.wait_chldexit),\ + real_timer: { \ + function: it_real_fn \ + }, \ + cap_effective: CAP_INIT_EFF_SET, \ + cap_inheritable: CAP_INIT_INH_SET, \ + cap_permitted: CAP_FULL_SET, \ + keep_capabilities: 0, \ + rlim: INIT_RLIMITS, \ + user: INIT_USER, \ + comm: "swapper", \ + thread: INIT_THREAD, \ + fs: &init_fs, \ + files: &init_files, \ + sigmask_lock: SPIN_LOCK_UNLOCKED, \ + sig: &init_signals, \ + pending: { NULL, &tsk.pending.head, {{0}}}, \ + blocked: {{0}}, \ + alloc_lock: SPIN_LOCK_UNLOCKED, \ + MOSIX_INIT_TASK(tsk), \ + journal_info: NULL \ +} + + +#ifndef INIT_TASK_SIZE +# define INIT_TASK_SIZE 2048*sizeof(long) +#endif + +union task_union { + struct task_struct task; + unsigned long stack[INIT_TASK_SIZE/sizeof(long)]; +}; + +extern union task_union init_task_union; + +extern struct mm_struct init_mm; +extern struct task_struct *init_tasks[NR_CPUS]; + +/* PID hashing. (shouldnt this be dynamic?) */ +#define PIDHASH_SZ (4096 >> 2) +extern struct task_struct *pidhash[PIDHASH_SZ]; + +#define pid_hashfn(x) ((((x) >> 8) ^ (x)) & (PIDHASH_SZ - 1)) + +static inline void hash_pid(struct task_struct *p) +{ + struct task_struct **htable = &pidhash[pid_hashfn(p->pid)]; + + if((p->pidhash_next = *htable) != NULL) + (*htable)->pidhash_pprev = &p->pidhash_next; + *htable = p; + p->pidhash_pprev = htable; +} + +static inline void unhash_pid(struct task_struct *p) +{ + if(p->pidhash_next) + p->pidhash_next->pidhash_pprev = p->pidhash_pprev; + *p->pidhash_pprev = p->pidhash_next; +} + +static inline struct task_struct *find_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; +#ifdef CONFIG_MOSIX + if(p && (p->mosix.dflags & DREMOTE)) + p = NULL; +#endif CONFIG_MOSIX + return p; +} + +#ifdef CONFIG_MOSIX +static inline struct task_struct *find_any_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; + return p; +} +#endif CONFIG_MOSIX + +/* per-UID process charging. */ +extern struct user_struct * alloc_uid(uid_t); +extern void free_uid(struct user_struct *); + +#include <asm/current.h> + +extern unsigned long volatile jiffies; +extern unsigned long itimer_ticks; +extern unsigned long itimer_next; +extern struct timeval xtime; +extern void do_timer(struct pt_regs *); + +extern unsigned int * prof_buffer; +extern unsigned long prof_len; +extern unsigned long prof_shift; + +#define CURRENT_TIME (xtime.tv_sec) + +extern void FASTCALL(__wake_up(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(__wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern void FASTCALL(interruptible_sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(interruptible_sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern int FASTCALL(wake_up_process(struct task_struct * tsk)); + +#define wake_up(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_nr(x, nr) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_all(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 0) +#define wake_up_sync(x) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_sync_nr(x, nr) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible(x) __wake_up((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_nr(x, nr) __wake_up((x),TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible_all(x) __wake_up((x),TASK_INTERRUPTIBLE, 0) +#define wake_up_interruptible_sync(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_sync_nr(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, nr) +asmlinkage long sys_wait4(pid_t pid,unsigned int * stat_addr, int options, struct rusage * ru); + +extern int in_group_p(gid_t); +extern int in_egroup_p(gid_t); + +extern void proc_caches_init(void); +extern void flush_signals(struct task_struct *); +extern void flush_signal_handlers(struct task_struct *); +extern int dequeue_signal(sigset_t *, siginfo_t *); +extern void block_all_signals(int (*notifier)(void *priv), void *priv, + sigset_t *mask); +extern void unblock_all_signals(void); +extern int send_sig_info(int, struct siginfo *, struct task_struct *); +extern int force_sig_info(int, struct siginfo *, struct task_struct *); +extern int kill_pg_info(int, struct siginfo *, pid_t); +extern int kill_sl_info(int, struct siginfo *, pid_t); +extern int kill_proc_info(int, struct siginfo *, pid_t); +extern void notify_parent(struct task_struct *, int); +extern void do_notify_parent(struct task_struct *, int); +extern void force_sig(int, struct task_struct *); +extern int send_sig(int, struct task_struct *, int); +extern int kill_pg(pid_t, int, int); +extern int kill_sl(pid_t, int, int); +extern int kill_proc(pid_t, int, int); +extern int do_sigaction(int, const struct k_sigaction *, struct k_sigaction *); +extern int do_sigaltstack(const stack_t *, stack_t *, unsigned long); + +static inline int signal_pending(struct task_struct *p) +{ + return (p->sigpending != 0); +} + +/* + * Re-calculate pending state from the set of locally pending + * signals, globally pending signals, and blocked signals. + */ +static inline int has_pending_signals(sigset_t *signal, sigset_t *blocked) +{ + unsigned long ready; + long i; + + switch (_NSIG_WORDS) { + default: + for (i = _NSIG_WORDS, ready = 0; --i >= 0 ;) + ready |= signal->sig[i] &~ blocked->sig[i]; + break; + + case 4: ready = signal->sig[3] &~ blocked->sig[3]; + ready |= signal->sig[2] &~ blocked->sig[2]; + ready |= signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 2: ready = signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 1: ready = signal->sig[0] &~ blocked->sig[0]; + } + return ready != 0; +} + +/* Reevaluate whether the task has signals pending delivery. + This is required every time the blocked sigset_t changes. + All callers should have t->sigmask_lock. */ + +static inline void recalc_sigpending(struct task_struct *t) +{ +#ifdef CONFIG_MOSIX + t->mosix.ignoreoldsigs = 0; + if(t->mosix.dflags & DHEAVYSLEEP) + t->sigpending = 0; + else if(t->mosix.dflags & DFAKESIGNAL) + t->sigpending = 1; + else +#endif CONFIG_MOSIX + t->sigpending = has_pending_signals(&t->pending.signal, &t->blocked); +} + +/* True if we are on the alternate signal stack. */ + +static inline int on_sig_stack(unsigned long sp) +{ + return (sp - current->sas_ss_sp < current->sas_ss_size); +} + +static inline int sas_ss_flags(unsigned long sp) +{ + return (current->sas_ss_size == 0 ? SS_DISABLE + : on_sig_stack(sp) ? SS_ONSTACK : 0); +} + +extern int request_irq(unsigned int, + void (*handler)(int, void *, struct pt_regs *), + unsigned long, const char *, void *); +extern void free_irq(unsigned int, void *); + +/* + * This has now become a routine instead of a macro, it sets a flag if + * it returns true (to do BSD-style accounting where the process is flagged + * if it uses root privs). The implication of this is that you should do + * normal permissions checks first, and check suser() last. + * + * [Dec 1997 -- Chris Evans] + * For correctness, the above considerations need to be extended to + * fsuser(). This is done, along with moving fsuser() checks to be + * last. + * + * These will be removed, but in the mean time, when the SECURE_NOROOT + * flag is set, uids don't grant privilege. + */ +static inline int suser(void) +{ + if (!issecure(SECURE_NOROOT) && current->euid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +static inline int fsuser(void) +{ + if (!issecure(SECURE_NOROOT) && current->fsuid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * capable() checks for a particular capability. + * New privilege checks should use this interface, rather than suser() or + * fsuser(). See include/linux/capability.h for defined capabilities. + */ + +static inline int capable(int cap) +{ +#if 1 /* ok now */ + if (cap_raised(current->cap_effective, cap)) +#else + if (cap_is_fs_cap(cap) ? current->fsuid == 0 : current->euid == 0) +#endif + { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * Routines for handling mm_structs + */ +extern struct mm_struct * mm_alloc(void); + +extern struct mm_struct * start_lazy_tlb(void); +extern void end_lazy_tlb(struct mm_struct *mm); + +/* mmdrop drops the mm and the page tables */ +extern inline void FASTCALL(__mmdrop(struct mm_struct *)); +static inline void mmdrop(struct mm_struct * mm) +{ + if (atomic_dec_and_test(&mm->mm_count)) + __mmdrop(mm); +} + +/* mmput gets rid of the mappings and all user-space */ +extern void mmput(struct mm_struct *); +/* Remove the current tasks stale references to the old mm_struct */ +extern void mm_release(void); + +/* + * Routines for handling the fd arrays + */ +extern struct file ** alloc_fd_array(int); +extern int expand_fd_array(struct files_struct *, int nr); +extern void free_fd_array(struct file **, int); + +extern fd_set *alloc_fdset(int); +extern int expand_fdset(struct files_struct *, int nr); +extern void free_fdset(fd_set *, int); + +extern int copy_thread(int, unsigned long, unsigned long, unsigned long, struct task_struct *, struct pt_regs *); +extern void flush_thread(void); +extern void exit_thread(void); + +extern void exit_mm(struct task_struct *); +extern void exit_files(struct task_struct *); +extern void exit_sighand(struct task_struct *); + +extern void daemonize(void); + +extern int do_execve(char *, char **, char **, struct pt_regs *); +extern int do_fork(unsigned long, unsigned long, struct pt_regs *, unsigned long); + +extern void FASTCALL(add_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(add_wait_queue_exclusive(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(remove_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); + +#define __wait_event(wq, condition) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_UNINTERRUPTIBLE); \ + if (condition) \ + break; \ + schedule(); \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event(wq, condition) \ +do { \ + if (condition) \ + break; \ + __wait_event(wq, condition); \ +} while (0) + +#define __wait_event_interruptible(wq, condition, ret) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_INTERRUPTIBLE); \ + if (condition) \ + break; \ + if (!signal_pending(current)) { \ + schedule(); \ + continue; \ + } \ + ret = -ERESTARTSYS; \ + break; \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event_interruptible(wq, condition) \ +({ \ + int __ret = 0; \ + if (!(condition)) \ + __wait_event_interruptible(wq, condition, __ret); \ + __ret; \ +}) + +#define REMOVE_LINKS(p) do { \ + (p)->next_task->prev_task = (p)->prev_task; \ + (p)->prev_task->next_task = (p)->next_task; \ + if ((p)->p_osptr) \ + (p)->p_osptr->p_ysptr = (p)->p_ysptr; \ + if ((p)->p_ysptr) \ + (p)->p_ysptr->p_osptr = (p)->p_osptr; \ + else \ + (p)->p_pptr->p_cptr = (p)->p_osptr; \ + } while (0) + +#define SET_LINKS(p) do { \ + (p)->next_task = &init_task; \ + (p)->prev_task = init_task.prev_task; \ + init_task.prev_task->next_task = (p); \ + init_task.prev_task = (p); \ + (p)->p_ysptr = NULL; \ + if (((p)->p_osptr = (p)->p_pptr->p_cptr) != NULL) \ + (p)->p_osptr->p_ysptr = p; \ + (p)->p_pptr->p_cptr = p; \ + } while (0) + +#define for_each_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) + +#define next_thread(p) \ + list_entry((p)->thread_group.next, struct task_struct, thread_group) + +#ifdef CONFIG_MOSIX +#define for_each_local_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) \ + if(!(p->mosix.dflags & DREMOTE)) + +static inline long +LOGICAL_STATE(struct task_struct *p) +{ + int result; + unsigned long flags; + + read_lock_irqsave(&p->mosix.state_lock, flags); + result = (p->mosix.bstate == TASK_SAME) ? p->state : p->mosix.bstate; + read_unlock_irqrestore(&p->mosix.state_lock, flags); + return(result); +} + +extern void run_on(struct task_struct *); +extern void run_off(struct task_struct *); + +#define need_interim_while_asleep() mosix_need_while_asleep() +#define run_interim_while_asleep() mosix_run_while_asleep() +#endif CONFIG_MOSIX + +static inline void del_from_runqueue(struct task_struct * p) +{ + nr_running--; + p->sleep_time = jiffies; +#ifdef CONFIG_MOSIX + run_off(p); +#endif CONFIG_MOSIX + list_del(&p->run_list); + p->run_list.next = NULL; +} + +static inline int task_on_runqueue(struct task_struct *p) +{ + return (p->run_list.next != NULL); +} + +static inline void unhash_process(struct task_struct *p) +{ + if (task_on_runqueue(p)) BUG(); + write_lock_irq(&tasklist_lock); + nr_threads--; + unhash_pid(p); + REMOVE_LINKS(p); + list_del(&p->thread_group); + write_unlock_irq(&tasklist_lock); +} + +/* Protects ->fs, ->files, ->mm, and synchronises with wait4(). Nests inside tasklist_lock */ +static inline void task_lock(struct task_struct *p) +{ + spin_lock(&p->alloc_lock); +} + +static inline void task_unlock(struct task_struct *p) +{ + spin_unlock(&p->alloc_lock); +} + +/* write full pathname into buffer and return start of pathname */ +static inline char * d_path(struct dentry *dentry, struct vfsmount *vfsmnt, + char *buf, int buflen) +{ + char *res; + struct vfsmount *rootmnt; + struct dentry *root; +#if defined(CONFIG_MOSIX_DFSA) || defined(CONFIG_MOSIX_FS) + if(dentry->d_inode && dentry->d_inode->i_op && + dentry->d_inode->i_op->check_path) + dentry->d_inode->i_op->check_path(dentry); +#endif CONFIG_MOSIX_DFSA || CONFIG_MOSIX_FS + read_lock(¤t->fs->lock); + rootmnt = mntget(current->fs->rootmnt); + root = dget(current->fs->root); + read_unlock(¤t->fs->lock); + spin_lock(&dcache_lock); + res = __d_path(dentry, vfsmnt, root, rootmnt, buf, buflen); + spin_unlock(&dcache_lock); + dput(root); + mntput(rootmnt); + return res; +} + +#endif /* __KERNEL__ */ + +#endif diff --git a/sys-kernel/linux-sources/linux-sources-2.4.6-r2.ebuild b/sys-kernel/linux-sources/linux-sources-2.4.6-r2.ebuild index 6a686433b00d..55e4ed5de076 100644 --- a/sys-kernel/linux-sources/linux-sources-2.4.6-r2.ebuild +++ b/sys-kernel/linux-sources/linux-sources-2.4.6-r2.ebuild @@ -128,13 +128,13 @@ src_unpack() { cd MOSIX-${MOSV} tar xzf MOSIX-${MOSV}.tar.gz patches.${OKV} kernel.new.${OKV}.tar cd ${S} - try cat ${S2}/MOSIX-${MOSV}/patches.2.4.6 | patch -p0 - tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.2.4.6.tar + try cat ${S2}/MOSIX-${MOSV}/patches.${KV} | patch -p0 + tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.${KV}.tar fi cd ${S} echo "Applying reiserfs-NFS fix..." - try cat ${FILESDIR}/2.4.6/linux-2.4.6-reiserfs-NFS.patch | patch -N -p1 + try cat ${FILESDIR}/${PVR}/linux-${KV}-reiserfs-NFS.patch | patch -N -p1 if [ "`use lvm`" ] then @@ -146,14 +146,14 @@ src_unpack() { # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${S}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" + CFLAGS="${CFLAGS} -I${S}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" cd PATCHES try make KERNEL_VERSION=${KV} KERNEL_DIR=${S} cd ${S} # the -l option allows this patch to apply cleanly (ignore whitespace changes) try patch -l -p1 < ${S2}/LVM/${LVMV}/PATCHES/lvm-${LVMV}-${KV}.patch cd ${S}/drivers/md - try patch -p0 < ${FILESDIR}/${KV}/lvm.c.diff + try patch -p0 < ${FILESDIR}/${PVR}/lvm.c.diff fi # if [ "`use lm_sensors`" ] @@ -193,7 +193,7 @@ src_unpack() { # echo "Unpacking pcmcia-cs tools..." # cd ${S2} # unpack pcmcia-cs-${PCV}.tar.gz - # patch -p0 < ${FILESDIR}/${KV}/pcmcia-cs-${PCV}-gentoo.diff + # patch -p0 < ${FILESDIR}/${PVR}/pcmcia-cs-${PCV}-gentoo.diff # fi #JFS patch works; commented out because it's not ready for production use @@ -223,7 +223,7 @@ src_unpack() { echo echo "Fixing reject in include/linux/sched.h..." echo - cp ${FILESDIR}/${KV}/sched.h include/linux + cp ${FILESDIR}/${PVR}/sched.h include/linux fi fi @@ -271,8 +271,8 @@ src_compile() { fi # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${KS}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" - try make + CFLAGS="${CFLAGS} -I${KS}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" + try make fi # if [ "`use lm_sensors`" ] @@ -294,6 +294,7 @@ src_compile() { if [ "$PN" == "linux" ] then + cd ${KS} try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" dep try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" bzImage try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" modules @@ -326,7 +327,7 @@ src_install() { then cd ${KS2}/LVM/${LVMV}/tools - try CFLAGS="${CFLAGS} -I${KS}/include" make install -e prefix=${D} mandir=${D}/usr/share/man \ + CFLAGS="${CFLAGS} -I${KS}/include" try make install -e prefix=${D} mandir=${D}/usr/share/man \ sbindir=${D}/sbin libdir=${D}/lib #no need for a static library in /lib mv ${D}/lib/*.a ${D}/usr/lib diff --git a/sys-kernel/linux/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch b/sys-kernel/linux/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch new file mode 100644 index 000000000000..5f3830c8ffe1 --- /dev/null +++ b/sys-kernel/linux/files/2.4.6-r2/linux-2.4.6-reiserfs-NFS.patch @@ -0,0 +1,647 @@ +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/inode.c linux-2.4.6-mmap/fs/nfs/inode.c +--- linux-2.4.6-file/fs/nfs/inode.c Wed Jul 4 17:30:28 2001 ++++ linux-2.4.6-mmap/fs/nfs/inode.c Wed Jul 4 20:48:01 2001 +@@ -48,6 +48,7 @@ + static void nfs_write_inode(struct inode *,int); + static void nfs_delete_inode(struct inode *); + static void nfs_put_super(struct super_block *); ++static void nfs_clear_inode(struct inode *); + static void nfs_umount_begin(struct super_block *); + static int nfs_statfs(struct super_block *, struct statfs *); + +@@ -57,6 +58,7 @@ + delete_inode: nfs_delete_inode, + put_super: nfs_put_super, + statfs: nfs_statfs, ++ clear_inode: nfs_clear_inode, + umount_begin: nfs_umount_begin, + }; + +@@ -141,6 +143,21 @@ + clear_inode(inode); + } + ++/* ++ * For the moment, the only task for the NFS clear_inode method is to ++ * release the mmap credential ++ */ ++static void ++nfs_clear_inode(struct inode *inode) ++{ ++ struct rpc_cred *cred = NFS_I(inode)->mm_cred; ++ ++ if (cred) { ++ put_rpccred(cred); ++ NFS_I(inode)->mm_cred = 0; ++ } ++} ++ + void + nfs_put_super(struct super_block *sb) + { +@@ -600,7 +617,6 @@ + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + NFS_CACHE_CTIME(inode) = fattr->ctime; + NFS_CACHE_MTIME(inode) = fattr->mtime; +- NFS_CACHE_ATIME(inode) = fattr->atime; + NFS_CACHE_ISIZE(inode) = fattr->size; + NFS_ATTRTIMEO(inode) = NFS_MINATTRTIMEO(inode); + NFS_ATTRTIMEO_UPDATE(inode) = jiffies; +@@ -794,6 +810,20 @@ + } + + /* ++ * Ensure that mmap has a recent RPC credential for use when writing out ++ * shared pages ++ */ ++static inline void ++nfs_set_mmcred(struct inode *inode, struct rpc_cred *cred) ++{ ++ struct rpc_cred *oldcred = NFS_I(inode)->mm_cred; ++ ++ NFS_I(inode)->mm_cred = get_rpccred(cred); ++ if (oldcred) ++ put_rpccred(oldcred); ++} ++ ++/* + * These are probably going to contain hooks for + * allocating and releasing RPC credentials for + * the file. I'll have to think about Tronds patch +@@ -808,20 +838,20 @@ + auth = NFS_CLIENT(inode)->cl_auth; + cred = rpcauth_lookupcred(auth, 0); + filp->private_data = cred; ++ if (filp->f_mode & FMODE_WRITE) ++ nfs_set_mmcred(inode, cred); + unlock_kernel(); + return 0; + } + + int nfs_release(struct inode *inode, struct file *filp) + { +- struct rpc_auth *auth; + struct rpc_cred *cred; + + lock_kernel(); +- auth = NFS_CLIENT(inode)->cl_auth; + cred = nfs_file_cred(filp); + if (cred) +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + unlock_kernel(); + return 0; + } +@@ -976,7 +1006,6 @@ + NFS_CACHE_CTIME(inode) = fattr->ctime; + inode->i_ctime = nfs_time_to_secs(fattr->ctime); + +- NFS_CACHE_ATIME(inode) = fattr->atime; + inode->i_atime = nfs_time_to_secs(fattr->atime); + + NFS_CACHE_MTIME(inode) = new_mtime; +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/unlink.c linux-2.4.6-mmap/fs/nfs/unlink.c +--- linux-2.4.6-file/fs/nfs/unlink.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/fs/nfs/unlink.c Wed Jul 4 20:48:47 2001 +@@ -128,7 +128,7 @@ + dir_i = dir->d_inode; + nfs_zap_caches(dir_i); + NFS_PROTO(dir_i)->unlink_done(dir, &task->tk_msg); +- rpcauth_releasecred(task->tk_auth, data->cred); ++ put_rpccred(data->cred); + data->cred = NULL; + dput(dir); + } +diff -u --recursive --new-file linux-2.4.6-file/fs/nfs/write.c linux-2.4.6-mmap/fs/nfs/write.c +--- linux-2.4.6-file/fs/nfs/write.c Mon May 21 21:29:14 2001 ++++ linux-2.4.6-mmap/fs/nfs/write.c Wed Jul 4 20:48:24 2001 +@@ -181,7 +181,9 @@ + + + if (file) +- cred = nfs_file_cred(file); ++ cred = get_rpccred(nfs_file_cred(file)); ++ if (!cred) ++ cred = get_rpccred(NFS_I(inode)->mm_cred); + + dprintk("NFS: nfs_writepage_sync(%x/%Ld %d@%Ld)\n", + inode->i_dev, (long long)NFS_FILEID(inode), +@@ -226,6 +228,8 @@ + + io_error: + kunmap(page); ++ if (cred) ++ put_rpccred(cred); + + return written? written : result; + } +@@ -241,6 +245,9 @@ + status = (IS_ERR(req)) ? PTR_ERR(req) : 0; + if (status < 0) + goto out; ++ if (!req->wb_cred) ++ req->wb_cred = get_rpccred(NFS_I(inode)->mm_cred); ++ nfs_unlock_request(req); + nfs_release_request(req); + nfs_strategy(inode); + out: +@@ -557,13 +564,11 @@ + req->wb_bytes = count; + req->wb_file = file; + +- /* If we have a struct file, use its cached credentials +- * else cache the current process' credentials. */ ++ /* If we have a struct file, use its cached credentials */ + if (file) { + get_file(file); + req->wb_cred = nfs_file_cred(file); +- } else +- req->wb_cred = rpcauth_lookupcred(NFS_CLIENT(inode)->cl_auth, 0); ++ } + req->wb_inode = inode; + req->wb_count = 1; + +@@ -608,8 +613,8 @@ + /* Release struct file or cached credential */ + if (req->wb_file) + fput(req->wb_file); +- else +- rpcauth_releasecred(NFS_CLIENT(inode)->cl_auth, req->wb_cred); ++ else if (req->wb_cred) ++ put_rpccred(req->wb_cred); + page_cache_release(page); + nfs_page_free(req); + /* wake up anyone waiting to allocate a request */ +@@ -927,8 +932,6 @@ + if (end > rqend) + req->wb_bytes = end - req->wb_offset; + +- nfs_unlock_request(req); +- + return req; + } + +@@ -1049,6 +1052,7 @@ + goto done; + + status = 0; ++ nfs_unlock_request(req); + /* If we wrote past the end of the page. + * Call the strategy routine so it can send out a bunch + * of requests. +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs.h linux-2.4.6-mmap/include/linux/nfs_fs.h +--- linux-2.4.6-file/include/linux/nfs_fs.h Sat May 26 03:02:11 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs.h Wed Jul 4 17:32:16 2001 +@@ -63,6 +63,11 @@ + */ + #define NFS_SUPER_MAGIC 0x6969 + ++static inline struct nfs_inode_info *NFS_I(struct inode *inode) ++{ ++ return &inode->u.nfs_i; ++} ++ + #define NFS_FH(inode) (&(inode)->u.nfs_i.fh) + #define NFS_SERVER(inode) (&(inode)->i_sb->u.nfs_sb.s_server) + #define NFS_CLIENT(inode) (NFS_SERVER(inode)->client) +@@ -74,7 +79,6 @@ + #define NFS_READTIME(inode) ((inode)->u.nfs_i.read_cache_jiffies) + #define NFS_CACHE_CTIME(inode) ((inode)->u.nfs_i.read_cache_ctime) + #define NFS_CACHE_MTIME(inode) ((inode)->u.nfs_i.read_cache_mtime) +-#define NFS_CACHE_ATIME(inode) ((inode)->u.nfs_i.read_cache_atime) + #define NFS_CACHE_ISIZE(inode) ((inode)->u.nfs_i.read_cache_isize) + #define NFS_NEXTSCAN(inode) ((inode)->u.nfs_i.nextscan) + #define NFS_CACHEINV(inode) \ +diff -u --recursive --new-file linux-2.4.6-file/include/linux/nfs_fs_i.h linux-2.4.6-mmap/include/linux/nfs_fs_i.h +--- linux-2.4.6-file/include/linux/nfs_fs_i.h Thu May 31 15:27:10 2001 ++++ linux-2.4.6-mmap/include/linux/nfs_fs_i.h Wed Jul 4 17:32:16 2001 +@@ -45,7 +45,6 @@ + unsigned long read_cache_jiffies; + __u64 read_cache_ctime; + __u64 read_cache_mtime; +- __u64 read_cache_atime; + __u64 read_cache_isize; + unsigned long attrtimeo; + unsigned long attrtimeo_timestamp; +@@ -73,6 +72,9 @@ + struct inode *hash_next, + *hash_prev; + unsigned long nextscan; ++ ++ /* Credentials for shared mmap */ ++ struct rpc_cred *mm_cred; + }; + + /* +diff -u --recursive --new-file linux-2.4.6-file/include/linux/sunrpc/auth.h linux-2.4.6-mmap/include/linux/sunrpc/auth.h +--- linux-2.4.6-file/include/linux/sunrpc/auth.h Fri Jun 22 19:30:22 2001 ++++ linux-2.4.6-mmap/include/linux/sunrpc/auth.h Wed Jul 4 20:51:35 2001 +@@ -14,6 +14,8 @@ + #include <linux/config.h> + #include <linux/sunrpc/sched.h> + ++#include <asm/atomic.h> ++ + /* size of the nodename buffer */ + #define UNX_MAXNODENAME 32 + +@@ -22,8 +24,10 @@ + */ + struct rpc_cred { + struct rpc_cred * cr_next; /* linked list */ ++ struct rpc_auth * cr_auth; ++ struct rpc_credops * cr_ops; + unsigned long cr_expire; /* when to gc */ +- unsigned short cr_count; /* ref count */ ++ atomic_t cr_count; /* ref count */ + unsigned short cr_flags; /* various flags */ + #ifdef RPC_DEBUG + unsigned long cr_magic; /* 0x0f4aa4f0 */ +@@ -71,6 +75,9 @@ + void (*destroy)(struct rpc_auth *); + + struct rpc_cred * (*crcreate)(int); ++}; ++ ++struct rpc_credops { + void (*crdestroy)(struct rpc_cred *); + + int (*crmatch)(struct rpc_cred *, int); +@@ -92,8 +99,7 @@ + struct rpc_cred * rpcauth_lookupcred(struct rpc_auth *, int); + struct rpc_cred * rpcauth_bindcred(struct rpc_task *); + void rpcauth_holdcred(struct rpc_task *); +-void rpcauth_releasecred(struct rpc_auth *, +- struct rpc_cred *); ++void put_rpccred(struct rpc_cred *); + void rpcauth_unbindcred(struct rpc_task *); + int rpcauth_matchcred(struct rpc_auth *, + struct rpc_cred *, int); +@@ -106,6 +112,13 @@ + void rpcauth_free_credcache(struct rpc_auth *); + void rpcauth_insert_credcache(struct rpc_auth *, + struct rpc_cred *); ++ ++static inline ++struct rpc_cred * get_rpccred(struct rpc_cred *cred) ++{ ++ atomic_inc(&cred->cr_count); ++ return cred; ++} + + #endif /* __KERNEL__ */ + #endif /* _LINUX_SUNRPC_AUTH_H */ +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth.c linux-2.4.6-mmap/net/sunrpc/auth.c +--- linux-2.4.6-file/net/sunrpc/auth.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth.c Wed Jul 4 20:51:04 2001 +@@ -81,42 +81,61 @@ + auth->au_nextgc = jiffies + (auth->au_expire >> 1); + } + ++/* ++ * Destroy an unreferenced credential ++ */ + static inline void +-rpcauth_crdestroy(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_crdestroy(struct rpc_cred *cred) + { + #ifdef RPC_DEBUG + if (cred->cr_magic != RPCAUTH_CRED_MAGIC) + BUG(); + cred->cr_magic = 0; ++ if (atomic_read(&cred->cr_count) || cred->cr_auth) ++ BUG(); + #endif +- if (auth->au_ops->crdestroy) +- auth->au_ops->crdestroy(cred); +- else +- rpc_free(cred); ++ cred->cr_ops->crdestroy(cred); + } + + /* +- * Clear the RPC credential cache ++ * Destroy a list of credentials ++ */ ++static inline ++void rpcauth_destroy_credlist(struct rpc_cred *head) ++{ ++ struct rpc_cred *cred; ++ ++ while ((cred = head) != NULL) { ++ head = cred->cr_next; ++ rpcauth_crdestroy(cred); ++ } ++} ++ ++/* ++ * Clear the RPC credential cache, and delete those credentials ++ * that are not referenced. + */ + void + rpcauth_free_credcache(struct rpc_auth *auth) + { +- struct rpc_cred **q, *cred; +- void (*destroy)(struct rpc_cred *); ++ struct rpc_cred **q, *cred, *free = NULL; + int i; + +- if (!(destroy = auth->au_ops->crdestroy)) +- destroy = (void (*)(struct rpc_cred *)) rpc_free; +- + spin_lock(&rpc_credcache_lock); + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { + *q = cred->cr_next; +- destroy(cred); ++ cred->cr_auth = NULL; ++ if (atomic_read(&cred->cr_count) == 0) { ++ cred->cr_next = free; ++ free = cred; ++ } else ++ cred->cr_next = NULL; + } + } + spin_unlock(&rpc_credcache_lock); ++ rpcauth_destroy_credlist(free); + } + + /* +@@ -133,9 +152,10 @@ + for (i = 0; i < RPC_CREDCACHE_NR; i++) { + q = &auth->au_credcache[i]; + while ((cred = *q) != NULL) { +- if (!cred->cr_count && ++ if (!atomic_read(&cred->cr_count) && + time_before(cred->cr_expire, jiffies)) { + *q = cred->cr_next; ++ cred->cr_auth = NULL; + cred->cr_next = free; + free = cred; + continue; +@@ -144,10 +164,7 @@ + } + } + spin_unlock(&rpc_credcache_lock); +- while ((cred = free) != NULL) { +- free = cred->cr_next; +- rpcauth_crdestroy(auth, cred); +- } ++ rpcauth_destroy_credlist(free); + auth->au_nextgc = jiffies + auth->au_expire; + } + +@@ -163,8 +180,8 @@ + spin_lock(&rpc_credcache_lock); + cred->cr_next = auth->au_credcache[nr]; + auth->au_credcache[nr] = cred; +- cred->cr_count++; +- cred->cr_expire = jiffies + auth->au_expire; ++ cred->cr_auth = auth; ++ get_rpccred(cred); + spin_unlock(&rpc_credcache_lock); + } + +@@ -187,7 +204,7 @@ + q = &auth->au_credcache[nr]; + while ((cred = *q) != NULL) { + if (!(cred->cr_flags & RPCAUTH_CRED_DEAD) && +- auth->au_ops->crmatch(cred, taskflags)) { ++ cred->cr_ops->crmatch(cred, taskflags)) { + *q = cred->cr_next; + break; + } +@@ -213,23 +230,23 @@ + * Remove cred handle from cache + */ + static void +-rpcauth_remove_credcache(struct rpc_auth *auth, struct rpc_cred *cred) ++rpcauth_remove_credcache(struct rpc_cred *cred) + { ++ struct rpc_auth *auth = cred->cr_auth; + struct rpc_cred **q, *cr; + int nr; + + nr = (cred->cr_uid & RPC_CREDCACHE_MASK); +- spin_lock(&rpc_credcache_lock); + q = &auth->au_credcache[nr]; + while ((cr = *q) != NULL) { + if (cred == cr) { + *q = cred->cr_next; + cred->cr_next = NULL; ++ cred->cr_auth = NULL; + break; + } + q = &cred->cr_next; + } +- spin_unlock(&rpc_credcache_lock); + } + + struct rpc_cred * +@@ -258,7 +275,7 @@ + { + dprintk("RPC: matching %s cred %d\n", + auth->au_ops->au_name, taskflags); +- return auth->au_ops->crmatch(cred, taskflags); ++ return cred->cr_ops->crmatch(cred, taskflags); + } + + void +@@ -266,26 +283,25 @@ + { + dprintk("RPC: %4d holding %s cred %p\n", + task->tk_pid, task->tk_auth->au_ops->au_name, task->tk_msg.rpc_cred); +- if (task->tk_msg.rpc_cred) { +- spin_lock(&rpc_credcache_lock); +- task->tk_msg.rpc_cred->cr_count++; +- task->tk_msg.rpc_cred->cr_expire = jiffies + task->tk_auth->au_expire; +- spin_unlock(&rpc_credcache_lock); +- } ++ if (task->tk_msg.rpc_cred) ++ get_rpccred(task->tk_msg.rpc_cred); + } + + void +-rpcauth_releasecred(struct rpc_auth *auth, struct rpc_cred *cred) ++put_rpccred(struct rpc_cred *cred) + { +- spin_lock(&rpc_credcache_lock); +- if (cred != NULL && cred->cr_count > 0) { +- if (!--cred->cr_count && (cred->cr_flags & RPCAUTH_CRED_DEAD)) { +- spin_unlock(&rpc_credcache_lock); +- rpcauth_remove_credcache(auth, cred); +- rpcauth_crdestroy(auth, cred); +- return; +- } ++ if (!atomic_dec_and_lock(&cred->cr_count, &rpc_credcache_lock)) ++ return; ++ ++ if (cred->cr_auth && cred->cr_flags & RPCAUTH_CRED_DEAD) ++ rpcauth_remove_credcache(cred); ++ ++ if (!cred->cr_auth) { ++ spin_unlock(&rpc_credcache_lock); ++ rpcauth_crdestroy(cred); ++ return; + } ++ cred->cr_expire = jiffies + cred->cr_auth->au_expire; + spin_unlock(&rpc_credcache_lock); + } + +@@ -298,7 +314,7 @@ + dprintk("RPC: %4d releasing %s cred %p\n", + task->tk_pid, auth->au_ops->au_name, cred); + +- rpcauth_releasecred(auth, cred); ++ put_rpccred(cred); + task->tk_msg.rpc_cred = NULL; + } + +@@ -306,10 +322,11 @@ + rpcauth_marshcred(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d marshaling %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crmarshal(task, p, ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crmarshal(task, p, + task->tk_flags & RPC_CALL_REALUID); + } + +@@ -317,20 +334,22 @@ + rpcauth_checkverf(struct rpc_task *task, u32 *p) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d validating %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- return auth->au_ops->crvalidate(task, p); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ return cred->cr_ops->crvalidate(task, p); + } + + int + rpcauth_refreshcred(struct rpc_task *task) + { + struct rpc_auth *auth = task->tk_auth; ++ struct rpc_cred *cred = task->tk_msg.rpc_cred; + + dprintk("RPC: %4d refreshing %s cred %p\n", +- task->tk_pid, auth->au_ops->au_name, task->tk_msg.rpc_cred); +- task->tk_status = auth->au_ops->crrefresh(task); ++ task->tk_pid, auth->au_ops->au_name, cred); ++ task->tk_status = cred->cr_ops->crrefresh(task); + return task->tk_status; + } + +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_null.c linux-2.4.6-mmap/net/sunrpc/auth_null.c +--- linux-2.4.6-file/net/sunrpc/auth_null.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_null.c Wed Jul 4 19:27:38 2001 +@@ -17,6 +17,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops null_credops; ++ + static struct rpc_auth * + nul_create(struct rpc_clnt *clnt) + { +@@ -52,9 +54,10 @@ + + if (!(cred = (struct rpc_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; +- cred->cr_count = 0; ++ atomic_set(&cred->cr_count, 0); + cred->cr_flags = RPCAUTH_CRED_UPTODATE; + cred->cr_uid = current->uid; ++ cred->cr_ops = &null_credops; + + return cred; + } +@@ -124,7 +127,11 @@ + #endif + nul_create, + nul_destroy, +- nul_create_cred, ++ nul_create_cred ++}; ++ ++static ++struct rpc_credops null_credops = { + nul_destroy_cred, + nul_match, + nul_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/auth_unix.c linux-2.4.6-mmap/net/sunrpc/auth_unix.c +--- linux-2.4.6-file/net/sunrpc/auth_unix.c Fri Feb 9 20:29:44 2001 ++++ linux-2.4.6-mmap/net/sunrpc/auth_unix.c Wed Jul 4 19:29:07 2001 +@@ -33,6 +33,8 @@ + # define RPCDBG_FACILITY RPCDBG_AUTH + #endif + ++static struct rpc_credops unix_credops; ++ + static struct rpc_auth * + unx_create(struct rpc_clnt *clnt) + { +@@ -71,7 +73,7 @@ + if (!(cred = (struct unx_cred *) rpc_allocate(flags, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 0; ++ atomic_set(&cred->uc_count, 0); + cred->uc_flags = RPCAUTH_CRED_UPTODATE; + if (flags & RPC_TASK_ROOTCREDS) { + cred->uc_uid = cred->uc_fsuid = 0; +@@ -91,6 +93,7 @@ + if (i < NFS_NGROUPS) + cred->uc_gids[i] = NOGROUP; + } ++ cred->uc_base.cr_ops = &unix_credops; + + return (struct rpc_cred *) cred; + } +@@ -106,7 +109,7 @@ + if (!(cred = (struct unx_cred *) rpc_malloc(task, sizeof(*cred)))) + return NULL; + +- cred->uc_count = 1; ++ atomic_set(&cred->uc_count, 1); + cred->uc_flags = RPCAUTH_CRED_DEAD|RPCAUTH_CRED_UPTODATE; + cred->uc_uid = uid; + cred->uc_gid = gid; +@@ -236,7 +239,11 @@ + #endif + unx_create, + unx_destroy, +- unx_create_cred, ++ unx_create_cred ++}; ++ ++static ++struct rpc_credops unix_credops = { + unx_destroy_cred, + unx_match, + unx_marshal, +diff -u --recursive --new-file linux-2.4.6-file/net/sunrpc/sunrpc_syms.c linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c +--- linux-2.4.6-file/net/sunrpc/sunrpc_syms.c Fri Jan 12 00:53:02 2001 ++++ linux-2.4.6-mmap/net/sunrpc/sunrpc_syms.c Wed Jul 4 20:55:19 2001 +@@ -65,7 +65,7 @@ + EXPORT_SYMBOL(rpcauth_lookupcred); + EXPORT_SYMBOL(rpcauth_bindcred); + EXPORT_SYMBOL(rpcauth_matchcred); +-EXPORT_SYMBOL(rpcauth_releasecred); ++EXPORT_SYMBOL(put_rpccred); + + /* RPC server stuff */ + EXPORT_SYMBOL(svc_create); + +_______________________________________________ +NFS maillist - NFS@lists.sourceforge.net +http://lists.sourceforge.net/lists/listinfo/nfs + + diff --git a/sys-kernel/linux/files/2.4.6-r2/lvm.c.diff b/sys-kernel/linux/files/2.4.6-r2/lvm.c.diff new file mode 100644 index 000000000000..6392671b8f7e --- /dev/null +++ b/sys-kernel/linux/files/2.4.6-r2/lvm.c.diff @@ -0,0 +1,20 @@ +--- lvm.c.orig Sat Apr 28 10:48:37 2001 ++++ lvm.c Sat Apr 28 11:07:07 2001 +@@ -1791,7 +1791,7 @@ + int max_hardblocksize = 0, hardblocksize; + + for (le = 0; le < lv->lv_allocated_le; le++) { +- hardblocksize = get_hardblocksize(lv->lv_current_pe[le].dev); ++ hardblocksize = get_hardsect_size(lv->lv_current_pe[le].dev); + if (hardblocksize == 0) + hardblocksize = 512; + if (hardblocksize > max_hardblocksize) +@@ -1801,7 +1801,7 @@ + if (lv->lv_access & LV_SNAPSHOT) { + for (e = 0; e < lv->lv_remap_end; e++) { + hardblocksize = +- get_hardblocksize( ++ get_hardsect_size( + lv->lv_block_exception[e].rdev_new); + if (hardblocksize == 0) + hardblocksize = 512; diff --git a/sys-kernel/linux/files/2.4.6-r2/mosix.init b/sys-kernel/linux/files/2.4.6-r2/mosix.init new file mode 100644 index 000000000000..f339dabb666e --- /dev/null +++ b/sys-kernel/linux/files/2.4.6-r2/mosix.init @@ -0,0 +1,118 @@ +#!/bin/sh - +#RCUPDATE:2 3 4:95:This line is required for script management + +# description: MOSIX is am extension of the operating system, +# supporting scalable and transparent cluster computing. +# +# mosix Script to stop/start MOSIX +# +# Author: Amnon Shiloh + +. /etc/rc.d/config/functions + +SERVICE="mosix" +opts="start stop status restart" + +mosix_stop() { + echo 0 > /proc/mosix/admin/mospe + rm -f /var/lock/subsys/mosix +} + +mosix_start() { + if [ -s /etc/overheads -a -f /proc/mosix/admin/overheads ] + then + /bin/grep -v '^#' /etc/overheads > /proc/mosix/admin/overheads + fi + if [ -s /etc/mfscosts -a -f /proc/mosix/admin/mfscosts ] + then + /bin/grep -v '^#' /etc/mfscosts > /proc/mosix/admin/mfscosts + fi + a1= + [ -s /etc/mospe ] && a1="-p `cat /etc/mospe`" + a2= + [ -s /etc/mosgates ] && a2="-g `cat /etc/mosgates`" + /sbin/setpe -W $a1 $a2 -f /etc/mosix.map + touch /var/lock/subsys/mosix +} + +alarm() { + t=$2 + while : + do + case "$t" in 0) break ;; esac + sleep 1 + t=$(($t-1)) + done + kill -1 $1 + exit 0 +} + +# See how we were called. +start() { + eecho "Initializing MOSIX" + if [ ! -f /etc/mosix.map ] + then + echo You have no MOSIX configuration - please edit yours now: + echo -n "Editor to use [q to quit] - [vi] :- " + trap 'echo Timed-Out ; exit 1' 1 + alarm $$ 120 & + killer=$! + read editor + kill $killer + trap '' + case "$editor" in [qQ]) exit ;; + "") editor=/usr/bin/vi + [ -f $editor ] || editor=/bin/vi + ;; + esac + echo "# MOSIX CONFIGURATION" > /etc/mosix.map + echo "# ===================" >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# Each line should contain 3 fields, mapping IP addresses to MOSIX node-numbers:" >> /etc/mosix.map + echo "# 1) first MOSIX node-number in range." >> /etc/mosix.map + echo "# 2) IP address of the above node (or node-name from /etc/hosts)." >> /etc/mosix.map + echo "# 3) number of nodes in this range." >> /etc/mosix.map + echo "#" >> /etc/mosix.map + echo "# MOSIX-# IP number-of-nodes" >> /etc/mosix.map + echo "# ============================" >> /etc/mosix.map + $editor /etc/mosix.map + echo + while : + do + echo "If this node's standard IP address is not part of the table that" + echo "you just edited, because MOSIX uses a different or separate network," + echo "you need to type this node's MOSIX-number now." + echo + /bin/echo -n "Otherwise please press only <Enter> :- " + read me + case "$me" in "") /bin/rm -f /etc/mospe ; break ;; + [1-9] | [1-9][0-9] | [1-9][0-9][0-9] | \ + [1-9][0-9][0-9][0-9] | \ + [1-6][0-9][0-9][0-9][0-9]) + echo $me > /etc/mospe ; break ;; + *) echo Improper MOSIX number - please try again: ;; + esac + done + fi + start_mosix + eend +} + +stop() { + eecho "Stopping MOSIX" + stop_mosix + eend +} + +status() { + /sbin/setpe -r +} + +restart() { + eecho "Restarting MOSIX" + stop_mosix + start_mosix + eend +} + +doservice ${@} diff --git a/sys-kernel/linux/files/2.4.6-r2/sched.h b/sys-kernel/linux/files/2.4.6-r2/sched.h new file mode 100644 index 000000000000..438203f84cdd --- /dev/null +++ b/sys-kernel/linux/files/2.4.6-r2/sched.h @@ -0,0 +1,997 @@ +#ifndef _LINUX_SCHED_H +#define _LINUX_SCHED_H + +#include <asm/param.h> /* for HZ */ + +extern unsigned long event; + +#include <linux/config.h> +#include <linux/binfmts.h> +#include <linux/personality.h> +#include <linux/threads.h> +#include <linux/kernel.h> +#include <linux/types.h> +#include <linux/times.h> +#include <linux/timex.h> + +#include <asm/system.h> +#include <asm/semaphore.h> +#include <asm/page.h> +#include <asm/ptrace.h> +#include <asm/mmu.h> + +#include <linux/smp.h> +#include <linux/tty.h> +#include <linux/sem.h> +#include <linux/signal.h> +#include <linux/securebits.h> + +#ifdef CONFIG_MOSIX +#include <mos/mosixtask.h> +#else +#define MOSIX_INIT_TASK(tsk) +#endif CONFIG_MOSIX + +#include <linux/fs_struct.h> + +/* + * cloning flags: + */ +#define CSIGNAL 0x000000ff /* signal mask to be sent at exit */ +#define CLONE_VM 0x00000100 /* set if VM shared between processes */ +#define CLONE_FS 0x00000200 /* set if fs info shared between processes */ +#define CLONE_FILES 0x00000400 /* set if open files shared between processes */ +#define CLONE_SIGHAND 0x00000800 /* set if signal handlers and blocked signals shared */ +#define CLONE_PID 0x00001000 /* set if pid shared */ +#define CLONE_PTRACE 0x00002000 /* set if we want to let tracing continue on the child too */ +#define CLONE_VFORK 0x00004000 /* set if the parent wants the child to wake it up on mm_release */ +#define CLONE_PARENT 0x00008000 /* set if we want to have the same parent as the cloner */ +#define CLONE_THREAD 0x00010000 /* Same thread group? */ + +#define CLONE_SIGNAL (CLONE_SIGHAND | CLONE_THREAD) + +/* + * These are the constant used to fake the fixed-point load-average + * counting. Some notes: + * - 11 bit fractions expand to 22 bits by the multiplies: this gives + * a load-average precision of 10 bits integer + 11 bits fractional + * - if you want to count load-averages more often, you need more + * precision, or rounding will get you. With 2-second counting freq, + * the EXP_n values would be 1981, 2034 and 2043 if still using only + * 11 bit fractions. + */ +extern unsigned long avenrun[]; /* Load averages */ + +#define FSHIFT 11 /* nr of bits of precision */ +#define FIXED_1 (1<<FSHIFT) /* 1.0 as fixed-point */ +#define LOAD_FREQ (5*HZ) /* 5 sec intervals */ +#define EXP_1 1884 /* 1/exp(5sec/1min) as fixed-point */ +#define EXP_5 2014 /* 1/exp(5sec/5min) */ +#define EXP_15 2037 /* 1/exp(5sec/15min) */ + +#define CALC_LOAD(load,exp,n) \ + load *= exp; \ + load += n*(FIXED_1-exp); \ + load >>= FSHIFT; + +#define CT_TO_SECS(x) ((x) / HZ) +#define CT_TO_USECS(x) (((x) % HZ) * 1000000/HZ) + +extern int nr_running, nr_threads; +extern int last_pid; + +#include <linux/fs.h> +#include <linux/time.h> +#include <linux/param.h> +#include <linux/resource.h> +#include <linux/timer.h> + +#include <asm/processor.h> + +#define TASK_RUNNING 0 +#define TASK_INTERRUPTIBLE 1 +#define TASK_UNINTERRUPTIBLE 2 +#define TASK_ZOMBIE 4 +#define TASK_STOPPED 8 +#ifdef CONFIG_MOSIX +#define TASK_SAME 16 +#endif CONFIG_MOSIX + +#define __set_task_state(tsk, state_value) \ + do { (tsk)->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_task_state(tsk, state_value) \ + set_mb((tsk)->state, (state_value)) +#else +#define set_task_state(tsk, state_value) \ + __set_task_state((tsk), (state_value)) +#endif + +#define __set_current_state(state_value) \ + do { current->state = (state_value); } while (0) +#ifdef CONFIG_SMP +#define set_current_state(state_value) \ + set_mb(current->state, (state_value)) +#else +#define set_current_state(state_value) \ + __set_current_state(state_value) +#endif + +/* + * Scheduling policies + */ +#define SCHED_OTHER 0 +#define SCHED_FIFO 1 +#define SCHED_RR 2 + +/* + * This is an additional bit set when we want to + * yield the CPU for one re-schedule.. + */ +#define SCHED_YIELD 0x10 + +struct sched_param { + int sched_priority; +}; + +#ifdef __KERNEL__ + +#include <linux/spinlock.h> + +/* + * This serializes "schedule()" and also protects + * the run-queue from deletions/modifications (but + * _adding_ to the beginning of the run-queue has + * a separate lock). + */ +extern rwlock_t tasklist_lock; +extern spinlock_t runqueue_lock; +extern spinlock_t mmlist_lock; + +extern void sched_init(void); +extern void init_idle(void); +extern void show_state(void); +extern void cpu_init (void); +extern void trap_init(void); +extern void update_process_times(int user); +extern void update_one_process(struct task_struct *p, unsigned long user, + unsigned long system, int cpu); + +#define MAX_SCHEDULE_TIMEOUT LONG_MAX +extern signed long FASTCALL(schedule_timeout(signed long timeout)); +asmlinkage void schedule(void); + +extern int schedule_task(struct tq_struct *task); +extern void flush_scheduled_tasks(void); +extern int start_context_thread(void); +extern int current_is_keventd(void); + +/* + * The default fd array needs to be at least BITS_PER_LONG, + * as this is the granularity returned by copy_fdset(). + */ +#define NR_OPEN_DEFAULT BITS_PER_LONG + +/* + * Open file table structure + */ +struct files_struct { + atomic_t count; + rwlock_t file_lock; /* Protects all the below members. Nests inside tsk->alloc_lock */ + int max_fds; + int max_fdset; + int next_fd; + struct file ** fd; /* current fd array */ + fd_set *close_on_exec; + fd_set *open_fds; + fd_set close_on_exec_init; + fd_set open_fds_init; + struct file * fd_array[NR_OPEN_DEFAULT]; +#ifdef CONFIG_MOSIX_DFSA + atomic_t users; /* the real count, excluding /proc grabs */ + int count_dropped_to_one; + int maxclosed; + fd_set *closed; + fd_set closed_fds_init; + int maxopened; + fd_set *opened; + fd_set opened_fds_init; + int maxmod; + fd_set *modified; + fd_set modified_fds_init; +#define INIT_FILES_DFSA , users : ATOMIC_INIT(1), \ + closed: &init_files.closed_fds_init, \ + opened: &init_files.opened_fds_init, \ + modified: &init_files.modified_fds_init, +#else +#define INIT_FILES_DFSA +#endif CONFIG_MOSIX_DFSA +}; + +#define INIT_FILES \ +{ \ + count: ATOMIC_INIT(1), \ + file_lock: RW_LOCK_UNLOCKED, \ + max_fds: NR_OPEN_DEFAULT, \ + max_fdset: __FD_SETSIZE, \ + next_fd: 0, \ + fd: &init_files.fd_array[0], \ + close_on_exec: &init_files.close_on_exec_init, \ + open_fds: &init_files.open_fds_init, \ + close_on_exec_init: { { 0, } }, \ + open_fds_init: { { 0, } }, \ + fd_array: { NULL, } \ + INIT_FILES_DFSA \ +} + +/* Maximum number of active map areas.. This is a random (large) number */ +#define MAX_MAP_COUNT (65536) + +/* Number of map areas at which the AVL tree is activated. This is arbitrary. */ +#define AVL_MIN_MAP_COUNT 32 + +struct mm_struct { + struct vm_area_struct * mmap; /* list of VMAs */ + struct vm_area_struct * mmap_avl; /* tree of VMAs */ + struct vm_area_struct * mmap_cache; /* last find_vma result */ + pgd_t * pgd; + atomic_t mm_users; /* How many users with user space? */ + atomic_t mm_count; /* How many references to "struct mm_struct" (users count as 1) */ + int map_count; /* number of VMAs */ + struct rw_semaphore mmap_sem; + spinlock_t page_table_lock; /* Protects task page tables and mm->rss */ + + struct list_head mmlist; /* List of all active mm's. These are globally strung + * together off init_mm.mmlist, and are protected + * by mmlist_lock + */ + + unsigned long start_code, end_code, start_data, end_data; + unsigned long start_brk, brk, start_stack; + unsigned long arg_start, arg_end, env_start, env_end; + unsigned long rss, total_vm, locked_vm; + unsigned long def_flags; + unsigned long cpu_vm_mask; + unsigned long swap_address; + + /* Architecture-specific MM context */ + mm_context_t context; +#ifdef CONFIG_MOSIX + atomic_t mm_kiocount; + atomic_t mm_realusers; + unsigned long last_memsort; + int used, unused, swapped; /* break-down of memory pages */ + int private_unused; /* unused but not active/inactive */ + int mark; +#define INIT_MM_MOSIX mm_kiocount: ATOMIC_INIT(0), mm_realusers: ATOMIC_INIT(1), +#else +#define INIT_MM_MOSIX +#endif CONFIG_MOSIX +}; + +extern int mmlist_nr; + +#define INIT_MM(name) \ +{ \ + mmap: &init_mmap, \ + mmap_avl: NULL, \ + mmap_cache: NULL, \ + pgd: swapper_pg_dir, \ + mm_users: ATOMIC_INIT(2), \ + mm_count: ATOMIC_INIT(1), \ + map_count: 1, \ + mmap_sem: __RWSEM_INITIALIZER(name.mmap_sem), \ + page_table_lock: SPIN_LOCK_UNLOCKED, \ + mmlist: LIST_HEAD_INIT(name.mmlist), \ + INIT_MM_MOSIX \ +} + +struct signal_struct { + atomic_t count; + struct k_sigaction action[_NSIG]; + spinlock_t siglock; +}; + + +#define INIT_SIGNALS { \ + count: ATOMIC_INIT(1), \ + action: { {{0,}}, }, \ + siglock: SPIN_LOCK_UNLOCKED \ +} + +/* + * Some day this will be a full-fledged user tracking system.. + */ +struct user_struct { + atomic_t __count; /* reference count */ + atomic_t processes; /* How many processes does this user have? */ + atomic_t files; /* How many open files does this user have? */ + + /* Hash table maintenance information */ + struct user_struct *next, **pprev; + uid_t uid; +}; + +#define get_current_user() ({ \ + struct user_struct *__user = current->user; \ + atomic_inc(&__user->__count); \ + __user; }) + +extern struct user_struct root_user; +#define INIT_USER (&root_user) + +struct task_struct { + /* + * offsets of these are hardcoded elsewhere - touch with care + */ + volatile long state; /* -1 unrunnable, 0 runnable, >0 stopped */ + unsigned long flags; /* per process flags, defined below */ + int sigpending; + mm_segment_t addr_limit; /* thread address space: + 0-0xBFFFFFFF for user-thead + 0-0xFFFFFFFF for kernel-thread + */ + struct exec_domain *exec_domain; + volatile long need_resched; + unsigned long ptrace; + + int lock_depth; /* Lock depth */ + +/* + * offset 32 begins here on 32-bit platforms. We keep + * all fields in a single cacheline that are needed for + * the goodness() loop in schedule(). + */ + long counter; + long nice; + unsigned long policy; + struct mm_struct *mm; + int has_cpu, processor; + unsigned long cpus_allowed; + /* + * (only the 'next' pointer fits into the cacheline, but + * that's just fine.) + */ + struct list_head run_list; + unsigned long sleep_time; + + struct task_struct *next_task, *prev_task; + struct mm_struct *active_mm; + +/* task state */ + struct linux_binfmt *binfmt; + int exit_code, exit_signal; + int pdeath_signal; /* The signal sent when the parent dies */ + /* ??? */ + unsigned long personality; + int dumpable:1; + int did_exec:1; + pid_t pid; + pid_t pgrp; + pid_t tty_old_pgrp; + pid_t session; + pid_t tgid; + /* boolean value for session group leader */ + int leader; + /* + * pointers to (original) parent process, youngest child, younger sibling, + * older sibling, respectively. (p->father can be replaced with + * p->p_pptr->pid) + */ + struct task_struct *p_opptr, *p_pptr, *p_cptr, *p_ysptr, *p_osptr; + struct list_head thread_group; + + /* PID hash table linkage. */ + struct task_struct *pidhash_next; + struct task_struct **pidhash_pprev; + + wait_queue_head_t wait_chldexit; /* for wait4() */ + struct semaphore *vfork_sem; /* for vfork() */ + unsigned long rt_priority; + unsigned long it_real_value, it_prof_value, it_virt_value; + unsigned long it_real_incr, it_prof_incr, it_virt_incr; + struct timer_list real_timer; + struct tms times; + unsigned long start_time; + long per_cpu_utime[NR_CPUS], per_cpu_stime[NR_CPUS]; +/* mm fault and swap info: this can arguably be seen as either mm-specific or thread-specific */ + unsigned long min_flt, maj_flt, nswap, cmin_flt, cmaj_flt, cnswap; + int swappable:1; +/* process credentials */ + uid_t uid,euid,suid,fsuid; + gid_t gid,egid,sgid,fsgid; + int ngroups; + gid_t groups[NGROUPS]; + kernel_cap_t cap_effective, cap_inheritable, cap_permitted; + int keep_capabilities:1; + struct user_struct *user; +/* limits */ + struct rlimit rlim[RLIM_NLIMITS]; + unsigned short used_math; + char comm[16]; +/* file system info */ + int link_count; + struct tty_struct *tty; /* NULL if no tty */ + unsigned int locks; /* How many file locks are being held */ +/* ipc stuff */ + struct sem_undo *semundo; + struct sem_queue *semsleeping; +/* CPU-specific state of this task */ + struct thread_struct thread; +/* filesystem information */ + struct fs_struct *fs; +/* open file information */ + struct files_struct *files; +/* signal handlers */ + spinlock_t sigmask_lock; /* Protects signal and blocked */ + struct signal_struct *sig; + + sigset_t blocked; + struct sigpending pending; + + unsigned long sas_ss_sp; + size_t sas_ss_size; + int (*notifier)(void *priv); + void *notifier_data; + sigset_t *notifier_mask; + +/* Thread group tracking */ + u32 parent_exec_id; + u32 self_exec_id; +/* Protection of (de-)allocation: mm, files, fs, tty */ + spinlock_t alloc_lock; +#ifdef CONFIG_MOSIX + struct mosix_task mosix; +#endif CONFIG_MOSIX + void *journal_info; +}; + +/* + * Per process flags + */ +#define PF_ALIGNWARN 0x00000001 /* Print alignment warning msgs */ + /* Not implemented yet, only for 486*/ +#define PF_STARTING 0x00000002 /* being created */ +#define PF_EXITING 0x00000004 /* getting shut down */ +#define PF_FORKNOEXEC 0x00000040 /* forked but didn't exec */ +#define PF_SUPERPRIV 0x00000100 /* used super-user privileges */ +#define PF_DUMPCORE 0x00000200 /* dumped core */ +#define PF_SIGNALED 0x00000400 /* killed by a signal */ +#define PF_MEMALLOC 0x00000800 /* Allocating memory */ +#define PF_VFORK 0x00001000 /* Wake up parent in mm_release */ + +#define PF_USEDFPU 0x00100000 /* task used FPU this quantum (SMP) */ + +/* + * Ptrace flags + */ + +#define PT_PTRACED 0x00000001 +#define PT_TRACESYS 0x00000002 +#define PT_DTRACE 0x00000004 /* delayed trace (used on m68k, i386) */ +#define PT_TRACESYSGOOD 0x00000008 + +/* + * Limit the stack by to some sane default: root can always + * increase this limit if needed.. 8MB seems reasonable. + */ +#define _STK_LIM (8*1024*1024) + +#define DEF_COUNTER (10*HZ/100) /* 100 ms time slice */ +#define MAX_COUNTER (20*HZ/100) +#define DEF_NICE (0) + +/* + * INIT_TASK is used to set up the first task table, touch at + * your own risk!. Base=0, limit=0x1fffff (=2MB) + */ +#define INIT_TASK(tsk) \ +{ \ + state: 0, \ + flags: 0, \ + sigpending: 0, \ + addr_limit: KERNEL_DS, \ + exec_domain: &default_exec_domain, \ + lock_depth: -1, \ + counter: DEF_COUNTER, \ + nice: DEF_NICE, \ + policy: SCHED_OTHER, \ + mm: NULL, \ + active_mm: &init_mm, \ + cpus_allowed: -1, \ + run_list: LIST_HEAD_INIT(tsk.run_list), \ + next_task: &tsk, \ + prev_task: &tsk, \ + p_opptr: &tsk, \ + p_pptr: &tsk, \ + thread_group: LIST_HEAD_INIT(tsk.thread_group), \ + wait_chldexit: __WAIT_QUEUE_HEAD_INITIALIZER(tsk.wait_chldexit),\ + real_timer: { \ + function: it_real_fn \ + }, \ + cap_effective: CAP_INIT_EFF_SET, \ + cap_inheritable: CAP_INIT_INH_SET, \ + cap_permitted: CAP_FULL_SET, \ + keep_capabilities: 0, \ + rlim: INIT_RLIMITS, \ + user: INIT_USER, \ + comm: "swapper", \ + thread: INIT_THREAD, \ + fs: &init_fs, \ + files: &init_files, \ + sigmask_lock: SPIN_LOCK_UNLOCKED, \ + sig: &init_signals, \ + pending: { NULL, &tsk.pending.head, {{0}}}, \ + blocked: {{0}}, \ + alloc_lock: SPIN_LOCK_UNLOCKED, \ + MOSIX_INIT_TASK(tsk), \ + journal_info: NULL \ +} + + +#ifndef INIT_TASK_SIZE +# define INIT_TASK_SIZE 2048*sizeof(long) +#endif + +union task_union { + struct task_struct task; + unsigned long stack[INIT_TASK_SIZE/sizeof(long)]; +}; + +extern union task_union init_task_union; + +extern struct mm_struct init_mm; +extern struct task_struct *init_tasks[NR_CPUS]; + +/* PID hashing. (shouldnt this be dynamic?) */ +#define PIDHASH_SZ (4096 >> 2) +extern struct task_struct *pidhash[PIDHASH_SZ]; + +#define pid_hashfn(x) ((((x) >> 8) ^ (x)) & (PIDHASH_SZ - 1)) + +static inline void hash_pid(struct task_struct *p) +{ + struct task_struct **htable = &pidhash[pid_hashfn(p->pid)]; + + if((p->pidhash_next = *htable) != NULL) + (*htable)->pidhash_pprev = &p->pidhash_next; + *htable = p; + p->pidhash_pprev = htable; +} + +static inline void unhash_pid(struct task_struct *p) +{ + if(p->pidhash_next) + p->pidhash_next->pidhash_pprev = p->pidhash_pprev; + *p->pidhash_pprev = p->pidhash_next; +} + +static inline struct task_struct *find_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; +#ifdef CONFIG_MOSIX + if(p && (p->mosix.dflags & DREMOTE)) + p = NULL; +#endif CONFIG_MOSIX + return p; +} + +#ifdef CONFIG_MOSIX +static inline struct task_struct *find_any_task_by_pid(int pid) +{ + struct task_struct *p, **htable = &pidhash[pid_hashfn(pid)]; + + for(p = *htable; p && p->pid != pid; p = p->pidhash_next) + ; + return p; +} +#endif CONFIG_MOSIX + +/* per-UID process charging. */ +extern struct user_struct * alloc_uid(uid_t); +extern void free_uid(struct user_struct *); + +#include <asm/current.h> + +extern unsigned long volatile jiffies; +extern unsigned long itimer_ticks; +extern unsigned long itimer_next; +extern struct timeval xtime; +extern void do_timer(struct pt_regs *); + +extern unsigned int * prof_buffer; +extern unsigned long prof_len; +extern unsigned long prof_shift; + +#define CURRENT_TIME (xtime.tv_sec) + +extern void FASTCALL(__wake_up(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(__wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr)); +extern void FASTCALL(sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern void FASTCALL(interruptible_sleep_on(wait_queue_head_t *q)); +extern long FASTCALL(interruptible_sleep_on_timeout(wait_queue_head_t *q, + signed long timeout)); +extern int FASTCALL(wake_up_process(struct task_struct * tsk)); + +#define wake_up(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_nr(x, nr) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_all(x) __wake_up((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 0) +#define wake_up_sync(x) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, 1) +#define wake_up_sync_nr(x, nr) __wake_up_sync((x),TASK_UNINTERRUPTIBLE | TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible(x) __wake_up((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_nr(x, nr) __wake_up((x),TASK_INTERRUPTIBLE, nr) +#define wake_up_interruptible_all(x) __wake_up((x),TASK_INTERRUPTIBLE, 0) +#define wake_up_interruptible_sync(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, 1) +#define wake_up_interruptible_sync_nr(x) __wake_up_sync((x),TASK_INTERRUPTIBLE, nr) +asmlinkage long sys_wait4(pid_t pid,unsigned int * stat_addr, int options, struct rusage * ru); + +extern int in_group_p(gid_t); +extern int in_egroup_p(gid_t); + +extern void proc_caches_init(void); +extern void flush_signals(struct task_struct *); +extern void flush_signal_handlers(struct task_struct *); +extern int dequeue_signal(sigset_t *, siginfo_t *); +extern void block_all_signals(int (*notifier)(void *priv), void *priv, + sigset_t *mask); +extern void unblock_all_signals(void); +extern int send_sig_info(int, struct siginfo *, struct task_struct *); +extern int force_sig_info(int, struct siginfo *, struct task_struct *); +extern int kill_pg_info(int, struct siginfo *, pid_t); +extern int kill_sl_info(int, struct siginfo *, pid_t); +extern int kill_proc_info(int, struct siginfo *, pid_t); +extern void notify_parent(struct task_struct *, int); +extern void do_notify_parent(struct task_struct *, int); +extern void force_sig(int, struct task_struct *); +extern int send_sig(int, struct task_struct *, int); +extern int kill_pg(pid_t, int, int); +extern int kill_sl(pid_t, int, int); +extern int kill_proc(pid_t, int, int); +extern int do_sigaction(int, const struct k_sigaction *, struct k_sigaction *); +extern int do_sigaltstack(const stack_t *, stack_t *, unsigned long); + +static inline int signal_pending(struct task_struct *p) +{ + return (p->sigpending != 0); +} + +/* + * Re-calculate pending state from the set of locally pending + * signals, globally pending signals, and blocked signals. + */ +static inline int has_pending_signals(sigset_t *signal, sigset_t *blocked) +{ + unsigned long ready; + long i; + + switch (_NSIG_WORDS) { + default: + for (i = _NSIG_WORDS, ready = 0; --i >= 0 ;) + ready |= signal->sig[i] &~ blocked->sig[i]; + break; + + case 4: ready = signal->sig[3] &~ blocked->sig[3]; + ready |= signal->sig[2] &~ blocked->sig[2]; + ready |= signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 2: ready = signal->sig[1] &~ blocked->sig[1]; + ready |= signal->sig[0] &~ blocked->sig[0]; + break; + + case 1: ready = signal->sig[0] &~ blocked->sig[0]; + } + return ready != 0; +} + +/* Reevaluate whether the task has signals pending delivery. + This is required every time the blocked sigset_t changes. + All callers should have t->sigmask_lock. */ + +static inline void recalc_sigpending(struct task_struct *t) +{ +#ifdef CONFIG_MOSIX + t->mosix.ignoreoldsigs = 0; + if(t->mosix.dflags & DHEAVYSLEEP) + t->sigpending = 0; + else if(t->mosix.dflags & DFAKESIGNAL) + t->sigpending = 1; + else +#endif CONFIG_MOSIX + t->sigpending = has_pending_signals(&t->pending.signal, &t->blocked); +} + +/* True if we are on the alternate signal stack. */ + +static inline int on_sig_stack(unsigned long sp) +{ + return (sp - current->sas_ss_sp < current->sas_ss_size); +} + +static inline int sas_ss_flags(unsigned long sp) +{ + return (current->sas_ss_size == 0 ? SS_DISABLE + : on_sig_stack(sp) ? SS_ONSTACK : 0); +} + +extern int request_irq(unsigned int, + void (*handler)(int, void *, struct pt_regs *), + unsigned long, const char *, void *); +extern void free_irq(unsigned int, void *); + +/* + * This has now become a routine instead of a macro, it sets a flag if + * it returns true (to do BSD-style accounting where the process is flagged + * if it uses root privs). The implication of this is that you should do + * normal permissions checks first, and check suser() last. + * + * [Dec 1997 -- Chris Evans] + * For correctness, the above considerations need to be extended to + * fsuser(). This is done, along with moving fsuser() checks to be + * last. + * + * These will be removed, but in the mean time, when the SECURE_NOROOT + * flag is set, uids don't grant privilege. + */ +static inline int suser(void) +{ + if (!issecure(SECURE_NOROOT) && current->euid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +static inline int fsuser(void) +{ + if (!issecure(SECURE_NOROOT) && current->fsuid == 0) { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * capable() checks for a particular capability. + * New privilege checks should use this interface, rather than suser() or + * fsuser(). See include/linux/capability.h for defined capabilities. + */ + +static inline int capable(int cap) +{ +#if 1 /* ok now */ + if (cap_raised(current->cap_effective, cap)) +#else + if (cap_is_fs_cap(cap) ? current->fsuid == 0 : current->euid == 0) +#endif + { + current->flags |= PF_SUPERPRIV; + return 1; + } + return 0; +} + +/* + * Routines for handling mm_structs + */ +extern struct mm_struct * mm_alloc(void); + +extern struct mm_struct * start_lazy_tlb(void); +extern void end_lazy_tlb(struct mm_struct *mm); + +/* mmdrop drops the mm and the page tables */ +extern inline void FASTCALL(__mmdrop(struct mm_struct *)); +static inline void mmdrop(struct mm_struct * mm) +{ + if (atomic_dec_and_test(&mm->mm_count)) + __mmdrop(mm); +} + +/* mmput gets rid of the mappings and all user-space */ +extern void mmput(struct mm_struct *); +/* Remove the current tasks stale references to the old mm_struct */ +extern void mm_release(void); + +/* + * Routines for handling the fd arrays + */ +extern struct file ** alloc_fd_array(int); +extern int expand_fd_array(struct files_struct *, int nr); +extern void free_fd_array(struct file **, int); + +extern fd_set *alloc_fdset(int); +extern int expand_fdset(struct files_struct *, int nr); +extern void free_fdset(fd_set *, int); + +extern int copy_thread(int, unsigned long, unsigned long, unsigned long, struct task_struct *, struct pt_regs *); +extern void flush_thread(void); +extern void exit_thread(void); + +extern void exit_mm(struct task_struct *); +extern void exit_files(struct task_struct *); +extern void exit_sighand(struct task_struct *); + +extern void daemonize(void); + +extern int do_execve(char *, char **, char **, struct pt_regs *); +extern int do_fork(unsigned long, unsigned long, struct pt_regs *, unsigned long); + +extern void FASTCALL(add_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(add_wait_queue_exclusive(wait_queue_head_t *q, wait_queue_t * wait)); +extern void FASTCALL(remove_wait_queue(wait_queue_head_t *q, wait_queue_t * wait)); + +#define __wait_event(wq, condition) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_UNINTERRUPTIBLE); \ + if (condition) \ + break; \ + schedule(); \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event(wq, condition) \ +do { \ + if (condition) \ + break; \ + __wait_event(wq, condition); \ +} while (0) + +#define __wait_event_interruptible(wq, condition, ret) \ +do { \ + wait_queue_t __wait; \ + init_waitqueue_entry(&__wait, current); \ + \ + add_wait_queue(&wq, &__wait); \ + for (;;) { \ + set_current_state(TASK_INTERRUPTIBLE); \ + if (condition) \ + break; \ + if (!signal_pending(current)) { \ + schedule(); \ + continue; \ + } \ + ret = -ERESTARTSYS; \ + break; \ + } \ + current->state = TASK_RUNNING; \ + remove_wait_queue(&wq, &__wait); \ +} while (0) + +#define wait_event_interruptible(wq, condition) \ +({ \ + int __ret = 0; \ + if (!(condition)) \ + __wait_event_interruptible(wq, condition, __ret); \ + __ret; \ +}) + +#define REMOVE_LINKS(p) do { \ + (p)->next_task->prev_task = (p)->prev_task; \ + (p)->prev_task->next_task = (p)->next_task; \ + if ((p)->p_osptr) \ + (p)->p_osptr->p_ysptr = (p)->p_ysptr; \ + if ((p)->p_ysptr) \ + (p)->p_ysptr->p_osptr = (p)->p_osptr; \ + else \ + (p)->p_pptr->p_cptr = (p)->p_osptr; \ + } while (0) + +#define SET_LINKS(p) do { \ + (p)->next_task = &init_task; \ + (p)->prev_task = init_task.prev_task; \ + init_task.prev_task->next_task = (p); \ + init_task.prev_task = (p); \ + (p)->p_ysptr = NULL; \ + if (((p)->p_osptr = (p)->p_pptr->p_cptr) != NULL) \ + (p)->p_osptr->p_ysptr = p; \ + (p)->p_pptr->p_cptr = p; \ + } while (0) + +#define for_each_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) + +#define next_thread(p) \ + list_entry((p)->thread_group.next, struct task_struct, thread_group) + +#ifdef CONFIG_MOSIX +#define for_each_local_task(p) \ + for (p = &init_task ; (p = p->next_task) != &init_task ; ) \ + if(!(p->mosix.dflags & DREMOTE)) + +static inline long +LOGICAL_STATE(struct task_struct *p) +{ + int result; + unsigned long flags; + + read_lock_irqsave(&p->mosix.state_lock, flags); + result = (p->mosix.bstate == TASK_SAME) ? p->state : p->mosix.bstate; + read_unlock_irqrestore(&p->mosix.state_lock, flags); + return(result); +} + +extern void run_on(struct task_struct *); +extern void run_off(struct task_struct *); + +#define need_interim_while_asleep() mosix_need_while_asleep() +#define run_interim_while_asleep() mosix_run_while_asleep() +#endif CONFIG_MOSIX + +static inline void del_from_runqueue(struct task_struct * p) +{ + nr_running--; + p->sleep_time = jiffies; +#ifdef CONFIG_MOSIX + run_off(p); +#endif CONFIG_MOSIX + list_del(&p->run_list); + p->run_list.next = NULL; +} + +static inline int task_on_runqueue(struct task_struct *p) +{ + return (p->run_list.next != NULL); +} + +static inline void unhash_process(struct task_struct *p) +{ + if (task_on_runqueue(p)) BUG(); + write_lock_irq(&tasklist_lock); + nr_threads--; + unhash_pid(p); + REMOVE_LINKS(p); + list_del(&p->thread_group); + write_unlock_irq(&tasklist_lock); +} + +/* Protects ->fs, ->files, ->mm, and synchronises with wait4(). Nests inside tasklist_lock */ +static inline void task_lock(struct task_struct *p) +{ + spin_lock(&p->alloc_lock); +} + +static inline void task_unlock(struct task_struct *p) +{ + spin_unlock(&p->alloc_lock); +} + +/* write full pathname into buffer and return start of pathname */ +static inline char * d_path(struct dentry *dentry, struct vfsmount *vfsmnt, + char *buf, int buflen) +{ + char *res; + struct vfsmount *rootmnt; + struct dentry *root; +#if defined(CONFIG_MOSIX_DFSA) || defined(CONFIG_MOSIX_FS) + if(dentry->d_inode && dentry->d_inode->i_op && + dentry->d_inode->i_op->check_path) + dentry->d_inode->i_op->check_path(dentry); +#endif CONFIG_MOSIX_DFSA || CONFIG_MOSIX_FS + read_lock(¤t->fs->lock); + rootmnt = mntget(current->fs->rootmnt); + root = dget(current->fs->root); + read_unlock(¤t->fs->lock); + spin_lock(&dcache_lock); + res = __d_path(dentry, vfsmnt, root, rootmnt, buf, buflen); + spin_unlock(&dcache_lock); + dput(root); + mntput(rootmnt); + return res; +} + +#endif /* __KERNEL__ */ + +#endif diff --git a/sys-kernel/linux/files/digest-linux-2.4.6-r2 b/sys-kernel/linux/files/digest-linux-2.4.6-r2 new file mode 100644 index 000000000000..69dfcfcd4391 --- /dev/null +++ b/sys-kernel/linux/files/digest-linux-2.4.6-r2 @@ -0,0 +1,6 @@ +MD5 eeddd92e2b9987ef9579df240ea9bd7a MOSIX-1.1.2.tar.gz +MD5 568a3673324804e0ac846969a3e311cb cloop_0.61-1.tar.gz +MD5 a3d6be1a7b31ea54d45df8d4dd82cf56 ext3-2.4-0.9.2-246.gz +MD5 3b04fb645ab68a081ab02079be937462 jfs-1.0.0-patch.tar.gz +MD5 f0fce862f60daf4cb8d7c5b589d01158 linux-2.4.6.tar.bz2 +MD5 1ce6fb607f26dc20d4c77461057acb3c lvm_0.9.1_beta7.tar.gz diff --git a/sys-kernel/linux/linux-2.4.6-r2.ebuild b/sys-kernel/linux/linux-2.4.6-r2.ebuild index 6a686433b00d..55e4ed5de076 100644 --- a/sys-kernel/linux/linux-2.4.6-r2.ebuild +++ b/sys-kernel/linux/linux-2.4.6-r2.ebuild @@ -128,13 +128,13 @@ src_unpack() { cd MOSIX-${MOSV} tar xzf MOSIX-${MOSV}.tar.gz patches.${OKV} kernel.new.${OKV}.tar cd ${S} - try cat ${S2}/MOSIX-${MOSV}/patches.2.4.6 | patch -p0 - tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.2.4.6.tar + try cat ${S2}/MOSIX-${MOSV}/patches.${KV} | patch -p0 + tar -x --no-same-owner -vf ${S2}/MOSIX-${MOSV}/kernel.new.${KV}.tar fi cd ${S} echo "Applying reiserfs-NFS fix..." - try cat ${FILESDIR}/2.4.6/linux-2.4.6-reiserfs-NFS.patch | patch -N -p1 + try cat ${FILESDIR}/${PVR}/linux-${KV}-reiserfs-NFS.patch | patch -N -p1 if [ "`use lvm`" ] then @@ -146,14 +146,14 @@ src_unpack() { # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${S}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" + CFLAGS="${CFLAGS} -I${S}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${S}" cd PATCHES try make KERNEL_VERSION=${KV} KERNEL_DIR=${S} cd ${S} # the -l option allows this patch to apply cleanly (ignore whitespace changes) try patch -l -p1 < ${S2}/LVM/${LVMV}/PATCHES/lvm-${LVMV}-${KV}.patch cd ${S}/drivers/md - try patch -p0 < ${FILESDIR}/${KV}/lvm.c.diff + try patch -p0 < ${FILESDIR}/${PVR}/lvm.c.diff fi # if [ "`use lm_sensors`" ] @@ -193,7 +193,7 @@ src_unpack() { # echo "Unpacking pcmcia-cs tools..." # cd ${S2} # unpack pcmcia-cs-${PCV}.tar.gz - # patch -p0 < ${FILESDIR}/${KV}/pcmcia-cs-${PCV}-gentoo.diff + # patch -p0 < ${FILESDIR}/${PVR}/pcmcia-cs-${PCV}-gentoo.diff # fi #JFS patch works; commented out because it's not ready for production use @@ -223,7 +223,7 @@ src_unpack() { echo echo "Fixing reject in include/linux/sched.h..." echo - cp ${FILESDIR}/${KV}/sched.h include/linux + cp ${FILESDIR}/${PVR}/sched.h include/linux fi fi @@ -271,8 +271,8 @@ src_compile() { fi # I had to hack this in so that LVM will look in the current linux # source directory instead of /usr/src/linux for stuff - pete - try CFLAGS="${CFLAGS} -I${KS}/include" ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" - try make + CFLAGS="${CFLAGS} -I${KS}/include" try ./configure --prefix=/ --mandir=/usr/share/man --with-kernel_dir="${KS}" + try make fi # if [ "`use lm_sensors`" ] @@ -294,6 +294,7 @@ src_compile() { if [ "$PN" == "linux" ] then + cd ${KS} try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" dep try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" bzImage try make HOSTCFLAGS="${LINUX_HOSTCFLAGS}" LEX="flex -l" modules @@ -326,7 +327,7 @@ src_install() { then cd ${KS2}/LVM/${LVMV}/tools - try CFLAGS="${CFLAGS} -I${KS}/include" make install -e prefix=${D} mandir=${D}/usr/share/man \ + CFLAGS="${CFLAGS} -I${KS}/include" try make install -e prefix=${D} mandir=${D}/usr/share/man \ sbindir=${D}/sbin libdir=${D}/lib #no need for a static library in /lib mv ${D}/lib/*.a ${D}/usr/lib |