···419419 * Poll : called when someone do a select on /dev/irnet.420420 * Just check if there are new events...421421 */422422-static inline unsigned int422422+static inline __poll_t423423irnet_ctrl_poll(irnet_socket * ap,424424 struct file * file,425425 poll_table * wait)426426{427427- unsigned int mask;427427+ __poll_t mask;428428429429 DENTER(CTRL_TRACE, "(ap=0x%p)\n", ap);430430···608608/*609609 * Poll : called when someone do a select on /dev/irnet610610 */611611-static unsigned int611611+static __poll_t612612dev_irnet_poll(struct file * file,613613 poll_table * wait)614614{615615 irnet_socket * ap = file->private_data;616616- unsigned int mask;616616+ __poll_t mask;617617618618 DENTER(FS_TRACE, "(file=0x%p, ap=0x%p)\n",619619 file, ap);
···831831 * and you use a different write policy from sock_writeable()832832 * then please supply your own write_space callback.833833 */834834-unsigned int datagram_poll(struct file *file, struct socket *sock,834834+__poll_t datagram_poll(struct file *file, struct socket *sock,835835 poll_table *wait)836836{837837 struct sock *sk = sock->sk;838838- unsigned int mask;838838+ __poll_t mask;839839840840 sock_poll_wait(file, sk_sleep(sk), wait);841841 mask = 0;
···316316 int flags, int *addr_len);317317void dccp_shutdown(struct sock *sk, int how);318318int inet_dccp_listen(struct socket *sock, int backlog);319319-unsigned int dccp_poll(struct file *file, struct socket *sock,319319+__poll_t dccp_poll(struct file *file, struct socket *sock,320320 poll_table *wait);321321int dccp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len);322322void dccp_req_err(struct sock *sk, u64 seq);
+2-2
net/dccp/proto.c
···313313 * take care of normal races (between the test and the event) and we don't314314 * go look at any of the socket buffers directly.315315 */316316-unsigned int dccp_poll(struct file *file, struct socket *sock,316316+__poll_t dccp_poll(struct file *file, struct socket *sock,317317 poll_table *wait)318318{319319- unsigned int mask;319319+ __poll_t mask;320320 struct sock *sk = sock->sk;321321322322 sock_poll_wait(file, sk_sleep(sk), wait);
···493493 * take care of normal races (between the test and the event) and we don't494494 * go look at any of the socket buffers directly.495495 */496496-unsigned int tcp_poll(struct file *file, struct socket *sock, poll_table *wait)496496+__poll_t tcp_poll(struct file *file, struct socket *sock, poll_table *wait)497497{498498- unsigned int mask;498498+ __poll_t mask;499499 struct sock *sk = sock->sk;500500 const struct tcp_sock *tp = tcp_sk(sk);501501 int state;
+2-2
net/ipv4/udp.c
···25022502 * but then block when reading it. Add special case code25032503 * to work around these arguably broken applications.25042504 */25052505-unsigned int udp_poll(struct file *file, struct socket *sock, poll_table *wait)25052505+__poll_t udp_poll(struct file *file, struct socket *sock, poll_table *wait)25062506{25072507- unsigned int mask = datagram_poll(file, sock, wait);25072507+ __poll_t mask = datagram_poll(file, sock, wait);25082508 struct sock *sk = sock->sk;2509250925102510 if (!skb_queue_empty(&udp_sk(sk)->reader_queue))
···152152 * to send to a congested destination, the system call may still fail (and153153 * return ENOBUFS).154154 */155155-static unsigned int rds_poll(struct file *file, struct socket *sock,155155+static __poll_t rds_poll(struct file *file, struct socket *sock,156156 poll_table *wait)157157{158158 struct sock *sk = sock->sk;159159 struct rds_sock *rs = rds_sk_to_rs(sk);160160- unsigned int mask = 0;160160+ __poll_t mask = 0;161161 unsigned long flags;162162163163 poll_wait(file, sk_sleep(sk), wait);
+2-2
net/rfkill/core.c
···11391139 return -ENOMEM;11401140}1141114111421142-static unsigned int rfkill_fop_poll(struct file *file, poll_table *wait)11421142+static __poll_t rfkill_fop_poll(struct file *file, poll_table *wait)11431143{11441144 struct rfkill_data *data = file->private_data;11451145- unsigned int res = POLLOUT | POLLWRNORM;11451145+ __poll_t res = POLLOUT | POLLWRNORM;1146114611471147 poll_wait(file, &data->read_wait, wait);11481148
+2-2
net/rxrpc/af_rxrpc.c
···725725/*726726 * permit an RxRPC socket to be polled727727 */728728-static unsigned int rxrpc_poll(struct file *file, struct socket *sock,728728+static __poll_t rxrpc_poll(struct file *file, struct socket *sock,729729 poll_table *wait)730730{731731 struct sock *sk = sock->sk;732732 struct rxrpc_sock *rx = rxrpc_sk(sk);733733- unsigned int mask;733733+ __poll_t mask;734734735735 sock_poll_wait(file, sk_sleep(sk), wait);736736 mask = 0;
+2-2
net/sctp/socket.c
···74977497 * here, again, by modeling the current TCP/UDP code. We don't have74987498 * a good way to test with it yet.74997499 */75007500-unsigned int sctp_poll(struct file *file, struct socket *sock, poll_table *wait)75007500+__poll_t sctp_poll(struct file *file, struct socket *sock, poll_table *wait)75017501{75027502 struct sock *sk = sock->sk;75037503 struct sctp_sock *sp = sctp_sk(sk);75047504- unsigned int mask;75047504+ __poll_t mask;7505750575067506 poll_wait(file, sk_sleep(sk), wait);75077507