169 lines
4.5 KiB
C
169 lines
4.5 KiB
C
|
#ifndef _LINUX_POLL_H
|
||
|
#define _LINUX_POLL_H
|
||
|
|
||
|
#include <asm/poll.h>
|
||
|
|
||
|
#ifdef __KERNEL__
|
||
|
|
||
|
#include <linux/compiler.h>
|
||
|
#include <linux/ktime.h>
|
||
|
#include <linux/wait.h>
|
||
|
#include <linux/string.h>
|
||
|
#include <linux/fs.h>
|
||
|
#include <linux/sysctl.h>
|
||
|
#include <asm/uaccess.h>
|
||
|
|
||
|
extern struct ctl_table epoll_table[]; /* for sysctl */
|
||
|
/* ~832 bytes of stack space used max in sys_select/sys_poll before allocating
|
||
|
additional memory. */
|
||
|
#define MAX_STACK_ALLOC 832
|
||
|
#define FRONTEND_STACK_ALLOC 256
|
||
|
#define SELECT_STACK_ALLOC FRONTEND_STACK_ALLOC
|
||
|
#define POLL_STACK_ALLOC FRONTEND_STACK_ALLOC
|
||
|
#define WQUEUES_STACK_ALLOC (MAX_STACK_ALLOC - FRONTEND_STACK_ALLOC)
|
||
|
#define N_INLINE_POLL_ENTRIES (WQUEUES_STACK_ALLOC / sizeof(struct poll_table_entry))
|
||
|
|
||
|
#define DEFAULT_POLLMASK (POLLIN | POLLOUT | POLLRDNORM | POLLWRNORM)
|
||
|
|
||
|
struct poll_table_struct;
|
||
|
|
||
|
/*
|
||
|
* structures and helpers for f_op->poll implementations
|
||
|
*/
|
||
|
typedef void (*poll_queue_proc)(struct file *, wait_queue_head_t *, struct poll_table_struct *);
|
||
|
|
||
|
/*
|
||
|
* Do not touch the structure directly, use the access functions
|
||
|
* poll_does_not_wait() and poll_requested_events() instead.
|
||
|
*/
|
||
|
typedef struct poll_table_struct {
|
||
|
poll_queue_proc _qproc;
|
||
|
unsigned long _key;
|
||
|
} poll_table;
|
||
|
|
||
|
static inline void poll_wait(struct file * filp, wait_queue_head_t * wait_address, poll_table *p)
|
||
|
{
|
||
|
if (p && p->_qproc && wait_address)
|
||
|
p->_qproc(filp, wait_address, p);
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* Return true if it is guaranteed that poll will not wait. This is the case
|
||
|
* if the poll() of another file descriptor in the set got an event, so there
|
||
|
* is no need for waiting.
|
||
|
*/
|
||
|
static inline bool poll_does_not_wait(const poll_table *p)
|
||
|
{
|
||
|
return p == NULL || p->_qproc == NULL;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* Return the set of events that the application wants to poll for.
|
||
|
* This is useful for drivers that need to know whether a DMA transfer has
|
||
|
* to be started implicitly on poll(). You typically only want to do that
|
||
|
* if the application is actually polling for POLLIN and/or POLLOUT.
|
||
|
*/
|
||
|
static inline unsigned long poll_requested_events(const poll_table *p)
|
||
|
{
|
||
|
return p ? p->_key : ~0UL;
|
||
|
}
|
||
|
|
||
|
static inline void init_poll_funcptr(poll_table *pt, poll_queue_proc qproc)
|
||
|
{
|
||
|
pt->_qproc = qproc;
|
||
|
pt->_key = ~0UL; /* all events enabled */
|
||
|
}
|
||
|
|
||
|
struct poll_table_entry {
|
||
|
struct file *filp;
|
||
|
unsigned long key;
|
||
|
wait_queue_t wait;
|
||
|
wait_queue_head_t *wait_address;
|
||
|
};
|
||
|
|
||
|
/*
|
||
|
* Structures and helpers for select/poll syscall
|
||
|
*/
|
||
|
struct poll_wqueues {
|
||
|
poll_table pt;
|
||
|
struct poll_table_page *table;
|
||
|
struct task_struct *polling_task;
|
||
|
int triggered;
|
||
|
int error;
|
||
|
int inline_index;
|
||
|
struct poll_table_entry inline_entries[N_INLINE_POLL_ENTRIES];
|
||
|
};
|
||
|
|
||
|
extern void poll_initwait(struct poll_wqueues *pwq);
|
||
|
extern void poll_freewait(struct poll_wqueues *pwq);
|
||
|
extern int poll_schedule_timeout(struct poll_wqueues *pwq, int state,
|
||
|
ktime_t *expires, unsigned long slack);
|
||
|
extern long select_estimate_accuracy(struct timespec *tv);
|
||
|
|
||
|
|
||
|
static inline int poll_schedule(struct poll_wqueues *pwq, int state)
|
||
|
{
|
||
|
return poll_schedule_timeout(pwq, state, NULL, 0);
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* Scalable version of the fd_set.
|
||
|
*/
|
||
|
|
||
|
typedef struct {
|
||
|
unsigned long *in, *out, *ex;
|
||
|
unsigned long *res_in, *res_out, *res_ex;
|
||
|
} fd_set_bits;
|
||
|
|
||
|
/*
|
||
|
* How many longwords for "nr" bits?
|
||
|
*/
|
||
|
#define FDS_BITPERLONG (8*sizeof(long))
|
||
|
#define FDS_LONGS(nr) (((nr)+FDS_BITPERLONG-1)/FDS_BITPERLONG)
|
||
|
#define FDS_BYTES(nr) (FDS_LONGS(nr)*sizeof(long))
|
||
|
|
||
|
/*
|
||
|
* We do a VERIFY_WRITE here even though we are only reading this time:
|
||
|
* we'll write to it eventually..
|
||
|
*
|
||
|
* Use "unsigned long" accesses to let user-mode fd_set's be long-aligned.
|
||
|
*/
|
||
|
static inline
|
||
|
int get_fd_set(unsigned long nr, void __user *ufdset, unsigned long *fdset)
|
||
|
{
|
||
|
nr = FDS_BYTES(nr);
|
||
|
if (ufdset)
|
||
|
return copy_from_user(fdset, ufdset, nr) ? -EFAULT : 0;
|
||
|
|
||
|
memset(fdset, 0, nr);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static inline unsigned long __must_check
|
||
|
set_fd_set(unsigned long nr, void __user *ufdset, unsigned long *fdset)
|
||
|
{
|
||
|
if (ufdset)
|
||
|
return __copy_to_user(ufdset, fdset, FDS_BYTES(nr));
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static inline
|
||
|
void zero_fd_set(unsigned long nr, unsigned long *fdset)
|
||
|
{
|
||
|
memset(fdset, 0, FDS_BYTES(nr));
|
||
|
}
|
||
|
|
||
|
#define MAX_INT64_SECONDS (((s64)(~((u64)0)>>1)/HZ)-1)
|
||
|
|
||
|
extern int do_select(int n, fd_set_bits *fds, struct timespec *end_time);
|
||
|
extern int do_sys_poll(struct pollfd __user * ufds, unsigned int nfds,
|
||
|
struct timespec *end_time);
|
||
|
extern int core_sys_select(int n, fd_set __user *inp, fd_set __user *outp,
|
||
|
fd_set __user *exp, struct timespec *end_time);
|
||
|
|
||
|
extern int poll_select_set_timeout(struct timespec *to, long sec, long nsec);
|
||
|
|
||
|
#endif /* KERNEL */
|
||
|
|
||
|
#endif /* _LINUX_POLL_H */
|