1714 lines
43 KiB
C
1714 lines
43 KiB
C
/*
|
|
* libwebsockets - small server side websockets and web server implementation
|
|
*
|
|
* Copyright (C) 2010-2017 Andy Green <andy@warmcat.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation:
|
|
* version 2.1 of the License.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
|
|
* MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "private-libwebsockets.h"
|
|
|
|
static int
|
|
lws_calllback_as_writeable(struct lws *wsi)
|
|
{
|
|
struct lws_context_per_thread *pt = &wsi->context->pt[(int)wsi->tsi];
|
|
int n;
|
|
|
|
lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_WRITEABLE_CB, 1);
|
|
#if defined(LWS_WITH_STATS)
|
|
if (wsi->active_writable_req_us) {
|
|
uint64_t ul = time_in_microseconds() -
|
|
wsi->active_writable_req_us;
|
|
|
|
lws_stats_atomic_bump(wsi->context, pt,
|
|
LWSSTATS_MS_WRITABLE_DELAY, ul);
|
|
lws_stats_atomic_max(wsi->context, pt,
|
|
LWSSTATS_MS_WORST_WRITABLE_DELAY, ul);
|
|
wsi->active_writable_req_us = 0;
|
|
}
|
|
#endif
|
|
|
|
switch (wsi->mode) {
|
|
case LWSCM_RAW:
|
|
n = LWS_CALLBACK_RAW_WRITEABLE;
|
|
break;
|
|
case LWSCM_RAW_FILEDESC:
|
|
n = LWS_CALLBACK_RAW_WRITEABLE_FILE;
|
|
break;
|
|
case LWSCM_WS_CLIENT:
|
|
n = LWS_CALLBACK_CLIENT_WRITEABLE;
|
|
break;
|
|
case LWSCM_WSCL_ISSUE_HTTP_BODY:
|
|
n = LWS_CALLBACK_CLIENT_HTTP_WRITEABLE;
|
|
break;
|
|
case LWSCM_WS_SERVING:
|
|
n = LWS_CALLBACK_SERVER_WRITEABLE;
|
|
break;
|
|
default:
|
|
n = LWS_CALLBACK_HTTP_WRITEABLE;
|
|
break;
|
|
}
|
|
|
|
return user_callback_handle_rxflow(wsi->protocol->callback,
|
|
wsi, (enum lws_callback_reasons) n,
|
|
wsi->user_space, NULL, 0);
|
|
}
|
|
|
|
LWS_VISIBLE int
|
|
lws_handle_POLLOUT_event(struct lws *wsi, struct lws_pollfd *pollfd)
|
|
{
|
|
int write_type = LWS_WRITE_PONG;
|
|
struct lws_tokens eff_buf;
|
|
#ifdef LWS_WITH_HTTP2
|
|
struct lws **wsi2, *wsi2a;
|
|
#endif
|
|
int ret, m, n;
|
|
|
|
wsi->leave_pollout_active = 0;
|
|
wsi->handling_pollout = 1;
|
|
/*
|
|
* if another thread wants POLLOUT on us, from here on while
|
|
* handling_pollout is set, he will only set leave_pollout_active.
|
|
* If we are going to disable POLLOUT, we will check that first.
|
|
*/
|
|
|
|
/*
|
|
* user callback is lowest priority to get these notifications
|
|
* actually, since other pending things cannot be disordered
|
|
*/
|
|
|
|
/* Priority 1: pending truncated sends are incomplete ws fragments
|
|
* If anything else sent first the protocol would be
|
|
* corrupted.
|
|
*/
|
|
if (wsi->trunc_len) {
|
|
//lwsl_notice("%s: completing partial\n", __func__);
|
|
if (lws_issue_raw(wsi, wsi->trunc_alloc + wsi->trunc_offset,
|
|
wsi->trunc_len) < 0) {
|
|
lwsl_info("%s signalling to close\n", __func__);
|
|
goto bail_die;
|
|
}
|
|
/* leave POLLOUT active either way */
|
|
goto bail_ok;
|
|
} else
|
|
if (wsi->state == LWSS_FLUSHING_STORED_SEND_BEFORE_CLOSE) {
|
|
wsi->socket_is_permanently_unusable = 1;
|
|
goto bail_die; /* retry closing now */
|
|
}
|
|
|
|
if (wsi->mode == LWSCM_WSCL_ISSUE_HTTP_BODY)
|
|
goto user_service;
|
|
|
|
#ifdef LWS_WITH_HTTP2
|
|
/*
|
|
* Priority 2: protocol packets
|
|
*/
|
|
if (wsi->upgraded_to_http2 && wsi->u.h2.h2n->pps) {
|
|
lwsl_info("servicing pps\n");
|
|
if (lws_h2_do_pps_send(wsi)) {
|
|
wsi->socket_is_permanently_unusable = 1;
|
|
goto bail_die;
|
|
}
|
|
if (wsi->u.h2.h2n->pps)
|
|
goto bail_ok;
|
|
|
|
/* we can resume whatever we were doing */
|
|
lws_rx_flow_control(wsi, LWS_RXFLOW_REASON_APPLIES_ENABLE |
|
|
LWS_RXFLOW_REASON_H2_PPS_PENDING);
|
|
|
|
goto bail_ok; /* leave POLLOUT active */
|
|
}
|
|
#endif
|
|
|
|
#ifdef LWS_WITH_CGI
|
|
if (wsi->cgi) {
|
|
/* also one shot */
|
|
if (pollfd)
|
|
if (lws_change_pollfd(wsi, LWS_POLLOUT, 0)) {
|
|
lwsl_info("failed at set pollfd\n");
|
|
return 1;
|
|
}
|
|
goto user_service_go_again;
|
|
}
|
|
#endif
|
|
|
|
/* Priority 3: pending control packets (pong or close)
|
|
*
|
|
* 3a: close notification packet requested from close api
|
|
*/
|
|
|
|
if (wsi->state == LWSS_WAITING_TO_SEND_CLOSE_NOTIFICATION) {
|
|
lwsl_debug("sending close packet\n");
|
|
wsi->waiting_to_send_close_frame = 0;
|
|
n = lws_write(wsi, &wsi->u.ws.ping_payload_buf[LWS_PRE],
|
|
wsi->u.ws.close_in_ping_buffer_len,
|
|
LWS_WRITE_CLOSE);
|
|
if (n >= 0) {
|
|
wsi->state = LWSS_AWAITING_CLOSE_ACK;
|
|
lws_set_timeout(wsi, PENDING_TIMEOUT_CLOSE_ACK, 1);
|
|
lwsl_debug("sent close indication, awaiting ack\n");
|
|
|
|
goto bail_ok;
|
|
}
|
|
|
|
goto bail_die;
|
|
}
|
|
|
|
/* else, the send failed and we should just hang up */
|
|
|
|
if ((wsi->state == LWSS_ESTABLISHED &&
|
|
wsi->u.ws.ping_pending_flag) ||
|
|
(wsi->state == LWSS_RETURNED_CLOSE_ALREADY &&
|
|
wsi->u.ws.payload_is_close)) {
|
|
|
|
if (wsi->u.ws.payload_is_close)
|
|
write_type = LWS_WRITE_CLOSE;
|
|
|
|
n = lws_write(wsi, &wsi->u.ws.ping_payload_buf[LWS_PRE],
|
|
wsi->u.ws.ping_payload_len, write_type);
|
|
if (n < 0)
|
|
goto bail_die;
|
|
|
|
/* well he is sent, mark him done */
|
|
wsi->u.ws.ping_pending_flag = 0;
|
|
if (wsi->u.ws.payload_is_close)
|
|
/* oh... a close frame was it... then we are done */
|
|
goto bail_die;
|
|
|
|
/* otherwise for PING, leave POLLOUT active either way */
|
|
goto bail_ok;
|
|
}
|
|
|
|
if (wsi->state == LWSS_ESTABLISHED &&
|
|
!wsi->socket_is_permanently_unusable &&
|
|
wsi->u.ws.send_check_ping) {
|
|
|
|
lwsl_info("issuing ping on wsi %p\n", wsi);
|
|
wsi->u.ws.send_check_ping = 0;
|
|
n = lws_write(wsi, &wsi->u.ws.ping_payload_buf[LWS_PRE],
|
|
0, LWS_WRITE_PING);
|
|
if (n < 0)
|
|
goto bail_die;
|
|
|
|
/*
|
|
* we apparently were able to send the PING in a reasonable time
|
|
* now reset the clock on our peer to be able to send the
|
|
* PONG in a reasonable time.
|
|
*/
|
|
|
|
lws_set_timeout(wsi, PENDING_TIMEOUT_WS_PONG_CHECK_GET_PONG,
|
|
wsi->context->timeout_secs);
|
|
|
|
goto bail_ok;
|
|
}
|
|
|
|
/* Priority 4: if we are closing, not allowed to send more data frags
|
|
* which means user callback or tx ext flush banned now
|
|
*/
|
|
if (wsi->state == LWSS_RETURNED_CLOSE_ALREADY)
|
|
goto user_service;
|
|
|
|
/* Priority 5: Tx path extension with more to send
|
|
*
|
|
* These are handled as new fragments each time around
|
|
* So while we must block new writeable callback to enforce
|
|
* payload ordering, but since they are always complete
|
|
* fragments control packets can interleave OK.
|
|
*/
|
|
if (wsi->state == LWSS_ESTABLISHED && wsi->u.ws.tx_draining_ext) {
|
|
lwsl_ext("SERVICING TX EXT DRAINING\n");
|
|
if (lws_write(wsi, NULL, 0, LWS_WRITE_CONTINUATION) < 0)
|
|
goto bail_die;
|
|
/* leave POLLOUT active */
|
|
goto bail_ok;
|
|
}
|
|
|
|
/* Priority 6: user can get the callback
|
|
*/
|
|
m = lws_ext_cb_active(wsi, LWS_EXT_CB_IS_WRITEABLE, NULL, 0);
|
|
if (m)
|
|
goto bail_die;
|
|
#ifndef LWS_NO_EXTENSIONS
|
|
if (!wsi->extension_data_pending)
|
|
goto user_service;
|
|
#endif
|
|
/*
|
|
* check in on the active extensions, see if they
|
|
* had pending stuff to spill... they need to get the
|
|
* first look-in otherwise sequence will be disordered
|
|
*
|
|
* NULL, zero-length eff_buf means just spill pending
|
|
*/
|
|
|
|
ret = 1;
|
|
if (wsi->mode == LWSCM_RAW || wsi->mode == LWSCM_RAW_FILEDESC)
|
|
ret = 0;
|
|
|
|
while (ret == 1) {
|
|
|
|
/* default to nobody has more to spill */
|
|
|
|
ret = 0;
|
|
eff_buf.token = NULL;
|
|
eff_buf.token_len = 0;
|
|
|
|
/* give every extension a chance to spill */
|
|
|
|
m = lws_ext_cb_active(wsi, LWS_EXT_CB_PACKET_TX_PRESEND,
|
|
&eff_buf, 0);
|
|
if (m < 0) {
|
|
lwsl_err("ext reports fatal error\n");
|
|
goto bail_die;
|
|
}
|
|
if (m)
|
|
/*
|
|
* at least one extension told us he has more
|
|
* to spill, so we will go around again after
|
|
*/
|
|
ret = 1;
|
|
|
|
/* assuming they gave us something to send, send it */
|
|
|
|
if (eff_buf.token_len) {
|
|
n = lws_issue_raw(wsi, (unsigned char *)eff_buf.token,
|
|
eff_buf.token_len);
|
|
if (n < 0) {
|
|
lwsl_info("closing from POLLOUT spill\n");
|
|
goto bail_die;
|
|
}
|
|
/*
|
|
* Keep amount spilled small to minimize chance of this
|
|
*/
|
|
if (n != eff_buf.token_len) {
|
|
lwsl_err("Unable to spill ext %d vs %d\n",
|
|
eff_buf.token_len, n);
|
|
goto bail_die;
|
|
}
|
|
} else
|
|
continue;
|
|
|
|
/* no extension has more to spill */
|
|
|
|
if (!ret)
|
|
continue;
|
|
|
|
/*
|
|
* There's more to spill from an extension, but we just sent
|
|
* something... did that leave the pipe choked?
|
|
*/
|
|
|
|
if (!lws_send_pipe_choked(wsi))
|
|
/* no we could add more */
|
|
continue;
|
|
|
|
lwsl_info("choked in POLLOUT service\n");
|
|
|
|
/*
|
|
* Yes, he's choked. Leave the POLLOUT masked on so we will
|
|
* come back here when he is unchoked. Don't call the user
|
|
* callback to enforce ordering of spilling, he'll get called
|
|
* when we come back here and there's nothing more to spill.
|
|
*/
|
|
|
|
goto bail_ok;
|
|
}
|
|
#ifndef LWS_NO_EXTENSIONS
|
|
wsi->extension_data_pending = 0;
|
|
#endif
|
|
user_service:
|
|
/* one shot */
|
|
|
|
if (wsi->parent_carries_io) {
|
|
wsi->handling_pollout = 0;
|
|
wsi->leave_pollout_active = 0;
|
|
|
|
return lws_calllback_as_writeable(wsi);
|
|
}
|
|
|
|
if (pollfd) {
|
|
int eff = wsi->leave_pollout_active;
|
|
|
|
if (!eff)
|
|
if (lws_change_pollfd(wsi, LWS_POLLOUT, 0)) {
|
|
lwsl_info("failed at set pollfd\n");
|
|
goto bail_die;
|
|
}
|
|
|
|
wsi->handling_pollout = 0;
|
|
|
|
/* cannot get leave_pollout_active set after the above */
|
|
if (!eff && wsi->leave_pollout_active)
|
|
/* got set inbetween sampling eff and clearing
|
|
* handling_pollout, force POLLOUT on */
|
|
lws_calllback_as_writeable(wsi);
|
|
|
|
wsi->leave_pollout_active = 0;
|
|
}
|
|
|
|
if (wsi->mode != LWSCM_WSCL_ISSUE_HTTP_BODY &&
|
|
!wsi->hdr_parsing_completed)
|
|
goto bail_ok;
|
|
|
|
|
|
#ifdef LWS_WITH_CGI
|
|
user_service_go_again:
|
|
#endif
|
|
|
|
#ifdef LWS_WITH_HTTP2
|
|
/*
|
|
* we are the 'network wsi' for potentially many muxed child wsi with
|
|
* no network connection of their own, who have to use us for all their
|
|
* network actions. So we use a round-robin scheme to share out the
|
|
* POLLOUT notifications to our children.
|
|
*
|
|
* But because any child could exhaust the socket's ability to take
|
|
* writes, we can only let one child get notified each time.
|
|
*
|
|
* In addition children may be closed / deleted / added between POLLOUT
|
|
* notifications, so we can't hold pointers
|
|
*/
|
|
|
|
if (wsi->mode != LWSCM_HTTP2_SERVING) {
|
|
lwsl_info("%s: non http2\n", __func__);
|
|
goto notify;
|
|
}
|
|
|
|
wsi->u.h2.requested_POLLOUT = 0;
|
|
if (!wsi->u.h2.initialized) {
|
|
lwsl_info("pollout on uninitialized http2 conn\n");
|
|
goto bail_ok;
|
|
}
|
|
|
|
// if (SSL_want_read(wsi->ssl) || SSL_want_write(wsi->ssl)) {
|
|
// lws_callback_on_writable(wsi);
|
|
// goto bail_ok;
|
|
// }
|
|
|
|
lwsl_info("%s: %p: children waiting for POLLOUT service:\n", __func__, wsi);
|
|
wsi2a = wsi->u.h2.child_list;
|
|
while (wsi2a) {
|
|
if (wsi2a->u.h2.requested_POLLOUT)
|
|
lwsl_debug(" * %p\n", wsi2a);
|
|
else
|
|
lwsl_debug(" %p\n", wsi2a);
|
|
|
|
wsi2a = wsi2a->u.h2.sibling_list;
|
|
}
|
|
|
|
wsi2 = &wsi->u.h2.child_list;
|
|
if (!*wsi2)
|
|
goto bail_ok;
|
|
|
|
do {
|
|
struct lws *w, **wa;
|
|
|
|
wa = &(*wsi2)->u.h2.sibling_list;
|
|
if (!(*wsi2)->u.h2.requested_POLLOUT) {
|
|
lwsl_debug(" child %p doesn't want POLLOUT\n", *wsi2);
|
|
goto next_child;
|
|
}
|
|
|
|
/*
|
|
* we're going to do writable callback for this child.
|
|
* move him to be the last child
|
|
*/
|
|
|
|
lwsl_debug("servicing child %p\n", *wsi2);
|
|
|
|
w = *wsi2;
|
|
while (w) {
|
|
if (!w->u.h2.sibling_list) { /* w is the current last */
|
|
lwsl_debug("w=%p, *wsi2 = %p\n", w, *wsi2);
|
|
if (w == *wsi2) /* we are already last */
|
|
break;
|
|
w->u.h2.sibling_list = *wsi2; /* last points to us as new last */
|
|
*wsi2 = (*wsi2)->u.h2.sibling_list; /* guy pointing to us until now points to our old next */
|
|
w->u.h2.sibling_list->u.h2.sibling_list = NULL; /* we point to nothing because we are last */
|
|
w = w->u.h2.sibling_list; /* w becomes us */
|
|
break;
|
|
}
|
|
w = w->u.h2.sibling_list;
|
|
}
|
|
|
|
w->u.h2.requested_POLLOUT = 0;
|
|
lwsl_info("%s: child %p (state %d)\n", __func__, (*wsi2), (*wsi2)->state);
|
|
|
|
if (w->u.h2.pending_status_body) {
|
|
w->u.h2.send_END_STREAM = 1;
|
|
n = lws_write(w,
|
|
(uint8_t *)w->u.h2.pending_status_body + LWS_PRE,
|
|
strlen(w->u.h2.pending_status_body + LWS_PRE),
|
|
LWS_WRITE_HTTP_FINAL);
|
|
lws_free_set_NULL(w->u.h2.pending_status_body);
|
|
lws_close_free_wsi(w, LWS_CLOSE_STATUS_NOSTATUS);
|
|
wa = &wsi->u.h2.child_list;
|
|
goto next_child;
|
|
}
|
|
|
|
if (w->state == LWSS_HTTP_ISSUING_FILE) {
|
|
|
|
w->leave_pollout_active = 0;
|
|
|
|
/* >0 == completion, <0 == error
|
|
*
|
|
* We'll get a LWS_CALLBACK_HTTP_FILE_COMPLETION callback when
|
|
* it's done. That's the case even if we just completed the
|
|
* send, so wait for that.
|
|
*/
|
|
n = lws_serve_http_file_fragment(w);
|
|
lwsl_debug("lws_serve_http_file_fragment says %d\n", n);
|
|
|
|
/*
|
|
* We will often hear about out having sent the final
|
|
* DATA here... if so close the actual wsi
|
|
*/
|
|
if (n < 0 || w->u.h2.send_END_STREAM) {
|
|
lwsl_debug("Closing POLLOUT child %p\n", w);
|
|
lws_close_free_wsi(w, LWS_CLOSE_STATUS_NOSTATUS);
|
|
wa = &wsi->u.h2.child_list;
|
|
goto next_child;
|
|
}
|
|
if (n > 0)
|
|
if (lws_http_transaction_completed(w))
|
|
goto bail_die;
|
|
if (!n) {
|
|
lws_callback_on_writable(w);
|
|
(w)->u.h2.requested_POLLOUT = 1;
|
|
}
|
|
|
|
goto next_child;
|
|
}
|
|
|
|
if (lws_calllback_as_writeable(w) || w->u.h2.send_END_STREAM) {
|
|
lwsl_debug("Closing POLLOUT child\n");
|
|
lws_close_free_wsi(w, LWS_CLOSE_STATUS_NOSTATUS);
|
|
wa = &wsi->u.h2.child_list;
|
|
}
|
|
|
|
next_child:
|
|
wsi2 = wa;
|
|
} while (wsi2 && *wsi2 && !lws_send_pipe_choked(wsi));
|
|
|
|
lwsl_info("%s: %p: children waiting for POLLOUT service: %p\n", __func__, wsi, wsi->u.h2.child_list);
|
|
wsi2a = wsi->u.h2.child_list;
|
|
while (wsi2a) {
|
|
if (wsi2a->u.h2.requested_POLLOUT)
|
|
lwsl_debug(" * %p\n", wsi2a);
|
|
else
|
|
lwsl_debug(" %p\n", wsi2a);
|
|
|
|
wsi2a = wsi2a->u.h2.sibling_list;
|
|
}
|
|
|
|
|
|
wsi2a = wsi->u.h2.child_list;
|
|
while (wsi2a) {
|
|
if (wsi2a->u.h2.requested_POLLOUT) {
|
|
lws_change_pollfd(wsi, 0, LWS_POLLOUT);
|
|
break;
|
|
}
|
|
wsi2a = wsi2a->u.h2.sibling_list;
|
|
}
|
|
|
|
goto bail_ok;
|
|
|
|
|
|
notify:
|
|
#endif
|
|
wsi->leave_pollout_active = 0;
|
|
|
|
n = lws_calllback_as_writeable(wsi);
|
|
wsi->handling_pollout = 0;
|
|
|
|
if (wsi->leave_pollout_active)
|
|
lws_change_pollfd(wsi, 0, LWS_POLLOUT);
|
|
|
|
return n;
|
|
|
|
/*
|
|
* since these don't disable the POLLOUT, they are always doing the
|
|
* right thing for leave_pollout_active whether it was set or not.
|
|
*/
|
|
|
|
bail_ok:
|
|
wsi->handling_pollout = 0;
|
|
wsi->leave_pollout_active = 0;
|
|
|
|
return 0;
|
|
|
|
bail_die:
|
|
wsi->handling_pollout = 0;
|
|
wsi->leave_pollout_active = 0;
|
|
|
|
return -1;
|
|
}
|
|
|
|
int
|
|
lws_service_timeout_check(struct lws *wsi, unsigned int sec)
|
|
{
|
|
struct lws_context_per_thread *pt = &wsi->context->pt[(int)wsi->tsi];
|
|
int n = 0;
|
|
|
|
(void)n;
|
|
|
|
/*
|
|
* if extensions want in on it (eg, we are a mux parent)
|
|
* give them a chance to service child timeouts
|
|
*/
|
|
if (lws_ext_cb_active(wsi, LWS_EXT_CB_1HZ, NULL, sec) < 0)
|
|
return 0;
|
|
|
|
if (!wsi->pending_timeout)
|
|
return 0;
|
|
|
|
/*
|
|
* if we went beyond the allowed time, kill the
|
|
* connection
|
|
*/
|
|
if ((time_t)sec > wsi->pending_timeout_limit) {
|
|
|
|
if (wsi->desc.sockfd != LWS_SOCK_INVALID &&
|
|
wsi->position_in_fds_table >= 0)
|
|
n = pt->fds[wsi->position_in_fds_table].events;
|
|
|
|
lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_TIMEOUTS, 1);
|
|
|
|
/* no need to log normal idle keepalive timeout */
|
|
if (wsi->pending_timeout != PENDING_TIMEOUT_HTTP_KEEPALIVE_IDLE)
|
|
lwsl_info("wsi %p: TIMEDOUT WAITING on %d "
|
|
"(did hdr %d, ah %p, wl %d, pfd "
|
|
"events %d) %llu vs %llu\n",
|
|
(void *)wsi, wsi->pending_timeout,
|
|
wsi->hdr_parsing_completed, wsi->u.hdr.ah,
|
|
pt->ah_wait_list_length, n,
|
|
(unsigned long long)sec,
|
|
(unsigned long long)wsi->pending_timeout_limit);
|
|
|
|
/*
|
|
* Since he failed a timeout, he already had a chance to do
|
|
* something and was unable to... that includes situations like
|
|
* half closed connections. So process this "failed timeout"
|
|
* close as a violent death and don't try to do protocol
|
|
* cleanup like flush partials.
|
|
*/
|
|
wsi->socket_is_permanently_unusable = 1;
|
|
if (wsi->mode == LWSCM_WSCL_WAITING_SSL)
|
|
wsi->vhost->protocols[0].callback(wsi,
|
|
LWS_CALLBACK_CLIENT_CONNECTION_ERROR,
|
|
wsi->user_space,
|
|
(void *)"Timed out waiting SSL", 21);
|
|
|
|
lws_close_free_wsi(wsi, LWS_CLOSE_STATUS_NOSTATUS);
|
|
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int lws_rxflow_cache(struct lws *wsi, unsigned char *buf, int n, int len)
|
|
{
|
|
#if defined(LWS_WITH_HTTP2)
|
|
if (wsi->upgraded_to_http2) {
|
|
struct lws_h2_netconn *h2n = wsi->u.h2.h2n;
|
|
|
|
assert(h2n->rx_scratch);
|
|
buf += n;
|
|
len -= n;
|
|
assert ((char *)buf >= (char *)h2n->rx_scratch &&
|
|
(char *)&buf[len] <= (char *)&h2n->rx_scratch[LWS_H2_RX_SCRATCH_SIZE]);
|
|
|
|
h2n->rx_scratch_pos = ((char *)buf - (char *)h2n->rx_scratch);
|
|
h2n->rx_scratch_len = len;
|
|
|
|
lwsl_info("%s: %p: pausing h2 rx_scratch\n", __func__, wsi);
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
/* his RX is flowcontrolled, don't send remaining now */
|
|
if (wsi->rxflow_buffer) {
|
|
if (buf >= wsi->rxflow_buffer &&
|
|
&buf[len - 1] < &wsi->rxflow_buffer[wsi->rxflow_len]) {
|
|
/* rxflow while we were spilling prev rxflow */
|
|
lwsl_info("%s: staying in rxflow buf\n", __func__);
|
|
return 1;
|
|
} else {
|
|
lwsl_err("%s: conflicting rxflow buf, "
|
|
"current %p len %d, new %p len %d\n", __func__,
|
|
wsi->rxflow_buffer, wsi->rxflow_len, buf, len);
|
|
assert(0);
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
/* a new rxflow, buffer it and warn caller */
|
|
lwsl_info("%s: new rxflow input buffer len %d\n", __func__, len - n);
|
|
wsi->rxflow_buffer = lws_malloc(len - n, "rxflow buf");
|
|
if (!wsi->rxflow_buffer)
|
|
return -1;
|
|
|
|
wsi->rxflow_len = len - n;
|
|
wsi->rxflow_pos = 0;
|
|
memcpy(wsi->rxflow_buffer, buf + n, len - n);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* this is used by the platform service code to stop us waiting for network
|
|
* activity in poll() when we have something that already needs service
|
|
*/
|
|
|
|
LWS_VISIBLE LWS_EXTERN int
|
|
lws_service_adjust_timeout(struct lws_context *context, int timeout_ms, int tsi)
|
|
{
|
|
struct lws_context_per_thread *pt = &context->pt[tsi];
|
|
struct allocated_headers *ah;
|
|
|
|
/* Figure out if we really want to wait in poll()
|
|
* We only need to wait if really nothing already to do and we have
|
|
* to wait for something from network
|
|
*/
|
|
|
|
/* 1) if we know we are draining rx ext, do not wait in poll */
|
|
if (pt->rx_draining_ext_list)
|
|
return 0;
|
|
|
|
#ifdef LWS_OPENSSL_SUPPORT
|
|
/* 2) if we know we have non-network pending data, do not wait in poll */
|
|
if (lws_ssl_anybody_has_buffered_read_tsi(context, tsi)) {
|
|
lwsl_info("ssl buffered read\n");
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
/* 3) if any ah has pending rx, do not wait in poll */
|
|
ah = pt->ah_list;
|
|
while (ah) {
|
|
if (ah->rxpos != ah->rxlen) {
|
|
if (!ah->wsi) {
|
|
assert(0);
|
|
}
|
|
return 0;
|
|
}
|
|
ah = ah->next;
|
|
}
|
|
|
|
return timeout_ms;
|
|
}
|
|
|
|
/*
|
|
* guys that need POLLIN service again without waiting for network action
|
|
* can force POLLIN here if not flowcontrolled, so they will get service.
|
|
*
|
|
* Return nonzero if anybody got their POLLIN faked
|
|
*/
|
|
int
|
|
lws_service_flag_pending(struct lws_context *context, int tsi)
|
|
{
|
|
struct lws_context_per_thread *pt = &context->pt[tsi];
|
|
struct allocated_headers *ah;
|
|
#ifdef LWS_OPENSSL_SUPPORT
|
|
struct lws *wsi_next;
|
|
#endif
|
|
struct lws *wsi;
|
|
int forced = 0;
|
|
|
|
/* POLLIN faking */
|
|
|
|
/*
|
|
* 1) For all guys with already-available ext data to drain, if they are
|
|
* not flowcontrolled, fake their POLLIN status
|
|
*/
|
|
wsi = pt->rx_draining_ext_list;
|
|
while (wsi) {
|
|
pt->fds[wsi->position_in_fds_table].revents |=
|
|
pt->fds[wsi->position_in_fds_table].events & LWS_POLLIN;
|
|
if (pt->fds[wsi->position_in_fds_table].revents & LWS_POLLIN) {
|
|
forced = 1;
|
|
break;
|
|
}
|
|
wsi = wsi->u.ws.rx_draining_ext_list;
|
|
}
|
|
|
|
#ifdef LWS_OPENSSL_SUPPORT
|
|
/*
|
|
* 2) For all guys with buffered SSL read data already saved up, if they
|
|
* are not flowcontrolled, fake their POLLIN status so they'll get
|
|
* service to use up the buffered incoming data, even though their
|
|
* network socket may have nothing
|
|
*/
|
|
wsi = pt->pending_read_list;
|
|
while (wsi) {
|
|
wsi_next = wsi->pending_read_list_next;
|
|
pt->fds[wsi->position_in_fds_table].revents |=
|
|
pt->fds[wsi->position_in_fds_table].events & LWS_POLLIN;
|
|
if (pt->fds[wsi->position_in_fds_table].revents & LWS_POLLIN) {
|
|
forced = 1;
|
|
/*
|
|
* he's going to get serviced now, take him off the
|
|
* list of guys with buffered SSL. If he still has some
|
|
* at the end of the service, he'll get put back on the
|
|
* list then.
|
|
*/
|
|
lws_ssl_remove_wsi_from_buffered_list(wsi);
|
|
}
|
|
|
|
wsi = wsi_next;
|
|
}
|
|
#endif
|
|
/*
|
|
* 3) For any wsi who have an ah with pending RX who did not
|
|
* complete their current headers, and are not flowcontrolled,
|
|
* fake their POLLIN status so they will be able to drain the
|
|
* rx buffered in the ah
|
|
*/
|
|
ah = pt->ah_list;
|
|
while (ah) {
|
|
if (ah->rxpos != ah->rxlen && !ah->wsi->hdr_parsing_completed) {
|
|
pt->fds[ah->wsi->position_in_fds_table].revents |=
|
|
pt->fds[ah->wsi->position_in_fds_table].events &
|
|
LWS_POLLIN;
|
|
if (pt->fds[ah->wsi->position_in_fds_table].revents &
|
|
LWS_POLLIN) {
|
|
forced = 1;
|
|
break;
|
|
}
|
|
}
|
|
ah = ah->next;
|
|
}
|
|
|
|
return forced;
|
|
}
|
|
|
|
#ifndef LWS_NO_CLIENT
|
|
|
|
LWS_VISIBLE int
|
|
lws_http_client_read(struct lws *wsi, char **buf, int *len)
|
|
{
|
|
int rlen, n;
|
|
|
|
rlen = lws_ssl_capable_read(wsi, (unsigned char *)*buf, *len);
|
|
*len = 0;
|
|
|
|
/* allow the source to signal he has data again next time */
|
|
lws_change_pollfd(wsi, 0, LWS_POLLIN);
|
|
|
|
if (rlen == LWS_SSL_CAPABLE_ERROR) {
|
|
lwsl_notice("%s: SSL capable error\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
if (rlen == 0)
|
|
return -1;
|
|
|
|
if (rlen < 0)
|
|
return 0;
|
|
|
|
*len = rlen;
|
|
wsi->client_rx_avail = 0;
|
|
|
|
/*
|
|
* server may insist on transfer-encoding: chunked,
|
|
* so http client must deal with it
|
|
*/
|
|
spin_chunks:
|
|
while (wsi->chunked && (wsi->chunk_parser != ELCP_CONTENT) && *len) {
|
|
switch (wsi->chunk_parser) {
|
|
case ELCP_HEX:
|
|
if ((*buf)[0] == '\x0d') {
|
|
wsi->chunk_parser = ELCP_CR;
|
|
break;
|
|
}
|
|
n = char_to_hex((*buf)[0]);
|
|
if (n < 0) {
|
|
lwsl_debug("chunking failure\n");
|
|
return -1;
|
|
}
|
|
wsi->chunk_remaining <<= 4;
|
|
wsi->chunk_remaining |= n;
|
|
break;
|
|
case ELCP_CR:
|
|
if ((*buf)[0] != '\x0a') {
|
|
lwsl_debug("chunking failure\n");
|
|
return -1;
|
|
}
|
|
wsi->chunk_parser = ELCP_CONTENT;
|
|
lwsl_info("chunk %d\n", wsi->chunk_remaining);
|
|
if (wsi->chunk_remaining)
|
|
break;
|
|
lwsl_info("final chunk\n");
|
|
goto completed;
|
|
|
|
case ELCP_CONTENT:
|
|
break;
|
|
|
|
case ELCP_POST_CR:
|
|
if ((*buf)[0] != '\x0d') {
|
|
lwsl_debug("chunking failure\n");
|
|
|
|
return -1;
|
|
}
|
|
|
|
wsi->chunk_parser = ELCP_POST_LF;
|
|
break;
|
|
|
|
case ELCP_POST_LF:
|
|
if ((*buf)[0] != '\x0a')
|
|
return -1;
|
|
|
|
wsi->chunk_parser = ELCP_HEX;
|
|
wsi->chunk_remaining = 0;
|
|
break;
|
|
}
|
|
(*buf)++;
|
|
(*len)--;
|
|
}
|
|
|
|
if (wsi->chunked && !wsi->chunk_remaining)
|
|
return 0;
|
|
|
|
if (wsi->u.http.rx_content_remain &&
|
|
wsi->u.http.rx_content_remain < *len)
|
|
n = (int)wsi->u.http.rx_content_remain;
|
|
else
|
|
n = *len;
|
|
|
|
if (wsi->chunked && wsi->chunk_remaining &&
|
|
wsi->chunk_remaining < n)
|
|
n = wsi->chunk_remaining;
|
|
|
|
#ifdef LWS_WITH_HTTP_PROXY
|
|
/* hubbub */
|
|
if (wsi->perform_rewrite)
|
|
lws_rewrite_parse(wsi->rw, (unsigned char *)*buf, n);
|
|
else
|
|
#endif
|
|
if (user_callback_handle_rxflow(wsi->protocol->callback,
|
|
wsi, LWS_CALLBACK_RECEIVE_CLIENT_HTTP_READ,
|
|
wsi->user_space, *buf, n)) {
|
|
lwsl_debug("%s: LWS_CALLBACK_RECEIVE_CLIENT_HTTP_READ returned -1\n", __func__);
|
|
|
|
return -1;
|
|
}
|
|
|
|
if (wsi->chunked && wsi->chunk_remaining) {
|
|
(*buf) += n;
|
|
wsi->chunk_remaining -= n;
|
|
*len -= n;
|
|
}
|
|
|
|
if (wsi->chunked && !wsi->chunk_remaining)
|
|
wsi->chunk_parser = ELCP_POST_CR;
|
|
|
|
if (wsi->chunked && *len)
|
|
goto spin_chunks;
|
|
|
|
if (wsi->chunked)
|
|
return 0;
|
|
|
|
/* if we know the content length, decrement the content remaining */
|
|
if (wsi->u.http.rx_content_length > 0)
|
|
wsi->u.http.rx_content_remain -= n;
|
|
|
|
if (wsi->u.http.rx_content_remain || !wsi->u.http.rx_content_length)
|
|
return 0;
|
|
|
|
completed:
|
|
if (user_callback_handle_rxflow(wsi->protocol->callback,
|
|
wsi, LWS_CALLBACK_COMPLETED_CLIENT_HTTP,
|
|
wsi->user_space, NULL, 0)) {
|
|
lwsl_debug("Completed call returned -1\n");
|
|
return -1;
|
|
}
|
|
|
|
if (lws_http_transaction_completed_client(wsi)) {
|
|
lwsl_notice("%s: transaction completed says -1\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static int
|
|
lws_is_ws_with_ext(struct lws *wsi)
|
|
{
|
|
#if defined(LWS_NO_EXTENSIONS)
|
|
return 0;
|
|
#else
|
|
return wsi->state == LWSS_ESTABLISHED &&
|
|
!!wsi->count_act_ext;
|
|
#endif
|
|
}
|
|
|
|
LWS_VISIBLE int
|
|
lws_service_fd_tsi(struct lws_context *context, struct lws_pollfd *pollfd, int tsi)
|
|
{
|
|
struct lws_context_per_thread *pt = &context->pt[tsi];
|
|
lws_sockfd_type our_fd = 0, tmp_fd;
|
|
struct allocated_headers *ah;
|
|
struct lws_tokens eff_buf;
|
|
unsigned int pending = 0;
|
|
struct lws *wsi, *wsi1;
|
|
char draining_flow = 0;
|
|
int timed_out = 0;
|
|
time_t now;
|
|
int n = 0, m;
|
|
int more;
|
|
|
|
if (!context->protocol_init_done)
|
|
lws_protocol_init(context);
|
|
|
|
time(&now);
|
|
|
|
/*
|
|
* handle case that system time was uninitialized when lws started
|
|
* at boot, and got initialized a little later
|
|
*/
|
|
if (context->time_up < 1464083026 && now > 1464083026)
|
|
context->time_up = now;
|
|
|
|
/* TODO: if using libev, we should probably use timeout watchers... */
|
|
if (context->last_timeout_check_s != now) {
|
|
context->last_timeout_check_s = now;
|
|
|
|
#if defined(LWS_WITH_STATS)
|
|
if (!tsi && now - context->last_dump > 10) {
|
|
lws_stats_log_dump(context);
|
|
context->last_dump = now;
|
|
}
|
|
#endif
|
|
|
|
lws_plat_service_periodic(context);
|
|
|
|
lws_check_deferred_free(context, 0);
|
|
|
|
#if defined(LWS_WITH_PEER_LIMITS)
|
|
lws_peer_cull_peer_wait_list(context);
|
|
#endif
|
|
|
|
/* retire unused deprecated context */
|
|
#if !defined(LWS_PLAT_OPTEE) && !defined(LWS_WITH_ESP32)
|
|
#if LWS_POSIX && !defined(_WIN32)
|
|
if (context->deprecated && !context->count_wsi_allocated) {
|
|
lwsl_notice("%s: ending deprecated context\n", __func__);
|
|
kill(getpid(), SIGINT);
|
|
return 0;
|
|
}
|
|
#endif
|
|
#endif
|
|
/* global timeout check once per second */
|
|
|
|
if (pollfd)
|
|
our_fd = pollfd->fd;
|
|
|
|
/*
|
|
* Phase 1: check every wsi on the timeout check list
|
|
*/
|
|
|
|
wsi = context->pt[tsi].timeout_list;
|
|
while (wsi) {
|
|
/* we have to take copies, because he may be deleted */
|
|
wsi1 = wsi->timeout_list;
|
|
tmp_fd = wsi->desc.sockfd;
|
|
if (lws_service_timeout_check(wsi, (unsigned int)now)) {
|
|
/* he did time out... */
|
|
if (tmp_fd == our_fd)
|
|
/* it was the guy we came to service! */
|
|
timed_out = 1;
|
|
/* he's gone, no need to mark as handled */
|
|
}
|
|
wsi = wsi1;
|
|
}
|
|
|
|
/*
|
|
* Phase 2: double-check active ah timeouts independent of wsi
|
|
* timeout status
|
|
*/
|
|
|
|
ah = pt->ah_list;
|
|
while (ah) {
|
|
int len;
|
|
char buf[256];
|
|
const unsigned char *c;
|
|
|
|
if (!ah->in_use || !ah->wsi || !ah->assigned ||
|
|
(ah->wsi->vhost && now - ah->assigned <
|
|
ah->wsi->vhost->timeout_secs_ah_idle + 60)) {
|
|
ah = ah->next;
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* a single ah session somehow got held for
|
|
* an unreasonable amount of time.
|
|
*
|
|
* Dump info on the connection...
|
|
*/
|
|
wsi = ah->wsi;
|
|
buf[0] = '\0';
|
|
lws_get_peer_simple(wsi, buf, sizeof(buf));
|
|
lwsl_notice("ah excessive hold: wsi %p\n"
|
|
" peer address: %s\n"
|
|
" ah rxpos %u, rxlen %u, pos %u\n",
|
|
wsi, buf, ah->rxpos, ah->rxlen,
|
|
ah->pos);
|
|
buf[0] = '\0';
|
|
m = 0;
|
|
do {
|
|
c = lws_token_to_string(m);
|
|
if (!c)
|
|
break;
|
|
if (!(*c))
|
|
break;
|
|
|
|
len = lws_hdr_total_length(wsi, m);
|
|
if (!len || len > sizeof(buf) - 1) {
|
|
m++;
|
|
continue;
|
|
}
|
|
|
|
if (lws_hdr_copy(wsi, buf,
|
|
sizeof buf, m) > 0) {
|
|
buf[sizeof(buf) - 1] = '\0';
|
|
|
|
lwsl_notice(" %s = %s\n",
|
|
(const char *)c, buf);
|
|
}
|
|
m++;
|
|
} while (1);
|
|
|
|
/* explicitly detach the ah */
|
|
|
|
lws_header_table_force_to_detachable_state(wsi);
|
|
lws_header_table_detach(wsi, 0);
|
|
|
|
/* ... and then drop the connection */
|
|
|
|
if (wsi->desc.sockfd == our_fd)
|
|
/* it was the guy we came to service! */
|
|
timed_out = 1;
|
|
|
|
lws_close_free_wsi(wsi, LWS_CLOSE_STATUS_NOSTATUS);
|
|
|
|
ah = pt->ah_list;
|
|
}
|
|
|
|
#ifdef LWS_WITH_CGI
|
|
/*
|
|
* Phase 3: handle cgi timeouts
|
|
*/
|
|
lws_cgi_kill_terminated(pt);
|
|
#endif
|
|
#if 0
|
|
{
|
|
char s[300], *p = s;
|
|
|
|
for (n = 0; n < context->count_threads; n++)
|
|
p += sprintf(p, " %7lu (%5d), ",
|
|
context->pt[n].count_conns,
|
|
context->pt[n].fds_count);
|
|
|
|
lwsl_notice("load: %s\n", s);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* at intervals, check for ws connections needing ping-pong checks
|
|
*/
|
|
|
|
if (context->ws_ping_pong_interval &&
|
|
context->last_ws_ping_pong_check_s < now + 10) {
|
|
struct lws_vhost *vh = context->vhost_list;
|
|
context->last_ws_ping_pong_check_s = now;
|
|
|
|
while (vh) {
|
|
for (n = 0; n < vh->count_protocols; n++) {
|
|
wsi = vh->same_vh_protocol_list[n];
|
|
|
|
while (wsi) {
|
|
if (wsi->state == LWSS_ESTABLISHED &&
|
|
!wsi->socket_is_permanently_unusable &&
|
|
!wsi->u.ws.send_check_ping &&
|
|
wsi->u.ws.time_next_ping_check &&
|
|
wsi->u.ws.time_next_ping_check < now) {
|
|
|
|
lwsl_info("requesting ping-pong on wsi %p\n", wsi);
|
|
wsi->u.ws.send_check_ping = 1;
|
|
lws_set_timeout(wsi, PENDING_TIMEOUT_WS_PONG_CHECK_SEND_PING,
|
|
context->timeout_secs);
|
|
lws_callback_on_writable(wsi);
|
|
wsi->u.ws.time_next_ping_check = now +
|
|
wsi->context->ws_ping_pong_interval;
|
|
}
|
|
wsi = wsi->same_vh_protocol_next;
|
|
}
|
|
}
|
|
vh = vh->vhost_next;
|
|
}
|
|
}
|
|
|
|
|
|
/* the socket we came to service timed out, nothing to do */
|
|
if (timed_out)
|
|
return 0;
|
|
|
|
/* just here for timeout management? */
|
|
if (!pollfd)
|
|
return 0;
|
|
|
|
/* no, here to service a socket descriptor */
|
|
wsi = wsi_from_fd(context, pollfd->fd);
|
|
if (!wsi)
|
|
/* not lws connection ... leave revents alone and return */
|
|
return 0;
|
|
|
|
/*
|
|
* so that caller can tell we handled, past here we need to
|
|
* zero down pollfd->revents after handling
|
|
*/
|
|
|
|
#if LWS_POSIX
|
|
/* handle session socket closed */
|
|
|
|
if ((!(pollfd->revents & pollfd->events & LWS_POLLIN)) &&
|
|
(pollfd->revents & LWS_POLLHUP)) {
|
|
wsi->socket_is_permanently_unusable = 1;
|
|
lwsl_debug("Session Socket %p (fd=%d) dead\n",
|
|
(void *)wsi, pollfd->fd);
|
|
|
|
goto close_and_handled;
|
|
}
|
|
|
|
#ifdef _WIN32
|
|
if (pollfd->revents & LWS_POLLOUT)
|
|
wsi->sock_send_blocking = FALSE;
|
|
#endif
|
|
|
|
#endif
|
|
|
|
if ((!(pollfd->revents & pollfd->events & LWS_POLLIN)) &&
|
|
(pollfd->revents & LWS_POLLHUP)) {
|
|
lwsl_debug("pollhup\n");
|
|
wsi->socket_is_permanently_unusable = 1;
|
|
goto close_and_handled;
|
|
}
|
|
|
|
#ifdef LWS_OPENSSL_SUPPORT
|
|
if ((wsi->state == LWSS_SHUTDOWN) && lws_is_ssl(wsi) && wsi->ssl) {
|
|
n = SSL_shutdown(wsi->ssl);
|
|
lwsl_debug("SSL_shutdown=%d for fd %d\n", n, wsi->desc.sockfd);
|
|
switch (n) {
|
|
case 1:
|
|
n = shutdown(wsi->desc.sockfd, SHUT_WR);
|
|
goto close_and_handled;
|
|
|
|
case 0:
|
|
lws_change_pollfd(wsi, 0, LWS_POLLIN);
|
|
n = 0;
|
|
goto handled;
|
|
|
|
default:
|
|
n = SSL_get_error(wsi->ssl, n);
|
|
if (n != SSL_ERROR_SYSCALL && n != SSL_ERROR_SSL) {
|
|
if (SSL_want_read(wsi->ssl)) {
|
|
lwsl_debug("(wants read)\n");
|
|
lws_change_pollfd(wsi, 0, LWS_POLLIN);
|
|
n = 0;
|
|
goto handled;
|
|
}
|
|
if (SSL_want_write(wsi->ssl)) {
|
|
lwsl_debug("(wants write)\n");
|
|
lws_change_pollfd(wsi, 0, LWS_POLLOUT);
|
|
n = 0;
|
|
goto handled;
|
|
}
|
|
}
|
|
|
|
/* actual error occurred, just close the connection */
|
|
n = shutdown(wsi->desc.sockfd, SHUT_WR);
|
|
goto close_and_handled;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
/* okay, what we came here to do... */
|
|
|
|
switch (wsi->mode) {
|
|
case LWSCM_HTTP_SERVING:
|
|
case LWSCM_HTTP_CLIENT:
|
|
case LWSCM_HTTP_SERVING_ACCEPTED:
|
|
case LWSCM_SERVER_LISTENER:
|
|
case LWSCM_SSL_ACK_PENDING:
|
|
case LWSCM_SSL_ACK_PENDING_RAW:
|
|
if (wsi->state == LWSS_CLIENT_HTTP_ESTABLISHED)
|
|
goto handled;
|
|
|
|
#ifdef LWS_WITH_CGI
|
|
if (wsi->cgi && (pollfd->revents & LWS_POLLOUT)) {
|
|
n = lws_handle_POLLOUT_event(wsi, pollfd);
|
|
if (n)
|
|
goto close_and_handled;
|
|
goto handled;
|
|
}
|
|
#endif
|
|
/* fallthru */
|
|
case LWSCM_RAW:
|
|
n = lws_server_socket_service(context, wsi, pollfd);
|
|
if (n) /* closed by above */
|
|
return 1;
|
|
goto handled;
|
|
|
|
case LWSCM_RAW_FILEDESC:
|
|
|
|
if (pollfd->revents & LWS_POLLOUT) {
|
|
n = lws_calllback_as_writeable(wsi);
|
|
if (lws_change_pollfd(wsi, LWS_POLLOUT, 0)) {
|
|
lwsl_info("failed at set pollfd\n");
|
|
return 1;
|
|
}
|
|
if (n)
|
|
goto close_and_handled;
|
|
}
|
|
n = LWS_CALLBACK_RAW_RX;
|
|
if (wsi->mode == LWSCM_RAW_FILEDESC)
|
|
n = LWS_CALLBACK_RAW_RX_FILE;
|
|
|
|
if (pollfd->revents & LWS_POLLIN) {
|
|
if (user_callback_handle_rxflow(
|
|
wsi->protocol->callback,
|
|
wsi, n,
|
|
wsi->user_space, NULL, 0)) {
|
|
lwsl_debug("raw rx callback closed it\n");
|
|
goto close_and_handled;
|
|
}
|
|
}
|
|
|
|
if (pollfd->revents & LWS_POLLHUP)
|
|
goto close_and_handled;
|
|
n = 0;
|
|
goto handled;
|
|
|
|
case LWSCM_WS_SERVING:
|
|
case LWSCM_WS_CLIENT:
|
|
case LWSCM_HTTP2_SERVING:
|
|
case LWSCM_HTTP_CLIENT_ACCEPTED:
|
|
|
|
/* 1: something requested a callback when it was OK to write */
|
|
|
|
if ((pollfd->revents & LWS_POLLOUT) &&
|
|
((wsi->state == LWSS_ESTABLISHED ||
|
|
wsi->state == LWSS_HTTP2_ESTABLISHED ||
|
|
wsi->state == LWSS_HTTP2_ESTABLISHED_PRE_SETTINGS ||
|
|
wsi->state == LWSS_RETURNED_CLOSE_ALREADY ||
|
|
wsi->state == LWSS_WAITING_TO_SEND_CLOSE_NOTIFICATION ||
|
|
wsi->state == LWSS_FLUSHING_STORED_SEND_BEFORE_CLOSE)) &&
|
|
lws_handle_POLLOUT_event(wsi, pollfd)) {
|
|
if (wsi->state == LWSS_RETURNED_CLOSE_ALREADY)
|
|
wsi->state = LWSS_FLUSHING_STORED_SEND_BEFORE_CLOSE;
|
|
lwsl_info("lws_service_fd: closing\n");
|
|
goto close_and_handled;
|
|
}
|
|
|
|
if (wsi->state == LWSS_RETURNED_CLOSE_ALREADY ||
|
|
wsi->state == LWSS_WAITING_TO_SEND_CLOSE_NOTIFICATION ||
|
|
wsi->state == LWSS_AWAITING_CLOSE_ACK) {
|
|
/*
|
|
* we stopped caring about anything except control
|
|
* packets. Force flow control off, defeat tx
|
|
* draining.
|
|
*/
|
|
lws_rx_flow_control(wsi, 1);
|
|
wsi->u.ws.tx_draining_ext = 0;
|
|
}
|
|
|
|
if (wsi->u.ws.tx_draining_ext)
|
|
/* we cannot deal with new RX until the TX ext
|
|
* path has been drained. It's because new
|
|
* rx will, eg, crap on the wsi rx buf that
|
|
* may be needed to retain state.
|
|
*
|
|
* TX ext drain path MUST go through event loop
|
|
* to avoid blocking.
|
|
*/
|
|
break;
|
|
|
|
if (lws_is_flowcontrolled(wsi))
|
|
/* We cannot deal with any kind of new RX
|
|
* because we are RX-flowcontrolled.
|
|
*/
|
|
break;
|
|
|
|
#if defined(LWS_WITH_HTTP2)
|
|
if (wsi->http2_substream || wsi->upgraded_to_http2) {
|
|
wsi1 = lws_get_network_wsi(wsi);
|
|
if (wsi1 && wsi1->trunc_len)
|
|
/* We cannot deal with any kind of new RX
|
|
* because we are dealing with a partial send
|
|
* (new RX may trigger new http_action() that
|
|
* expect to be able to send)
|
|
*/
|
|
break;
|
|
}
|
|
#endif
|
|
|
|
/* 2: RX Extension needs to be drained
|
|
*/
|
|
|
|
if (wsi->state == LWSS_ESTABLISHED &&
|
|
wsi->u.ws.rx_draining_ext) {
|
|
|
|
lwsl_ext("%s: RX EXT DRAINING: Service\n", __func__);
|
|
#ifndef LWS_NO_CLIENT
|
|
if (wsi->mode == LWSCM_WS_CLIENT) {
|
|
n = lws_client_rx_sm(wsi, 0);
|
|
if (n < 0)
|
|
/* we closed wsi */
|
|
n = 0;
|
|
} else
|
|
#endif
|
|
n = lws_rx_sm(wsi, 0);
|
|
|
|
goto handled;
|
|
}
|
|
|
|
if (wsi->u.ws.rx_draining_ext)
|
|
/*
|
|
* We have RX EXT content to drain, but can't do it
|
|
* right now. That means we cannot do anything lower
|
|
* priority either.
|
|
*/
|
|
break;
|
|
|
|
/* 3: RX Flowcontrol buffer / h2 rx scratch needs to be drained
|
|
*/
|
|
|
|
if (wsi->rxflow_buffer) {
|
|
lwsl_info("draining rxflow (len %d)\n",
|
|
wsi->rxflow_len - wsi->rxflow_pos);
|
|
assert(wsi->rxflow_pos < wsi->rxflow_len);
|
|
/* well, drain it */
|
|
eff_buf.token = (char *)wsi->rxflow_buffer +
|
|
wsi->rxflow_pos;
|
|
eff_buf.token_len = wsi->rxflow_len - wsi->rxflow_pos;
|
|
draining_flow = 1;
|
|
goto drain;
|
|
}
|
|
|
|
#if defined(LWS_WITH_HTTP2)
|
|
if (wsi->upgraded_to_http2) {
|
|
struct lws_h2_netconn *h2n = wsi->u.h2.h2n;
|
|
|
|
if (h2n->rx_scratch_len) {
|
|
lwsl_info("%s: %p: resuming h2 rx_scratch pos = %d len = %d\n",
|
|
__func__, wsi, h2n->rx_scratch_pos, h2n->rx_scratch_len);
|
|
eff_buf.token = (char *)h2n->rx_scratch +
|
|
h2n->rx_scratch_pos;
|
|
eff_buf.token_len = h2n->rx_scratch_len;
|
|
|
|
h2n->rx_scratch_len = 0;
|
|
goto drain;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
/* 4: any incoming (or ah-stashed incoming rx) data ready?
|
|
* notice if rx flow going off raced poll(), rx flow wins
|
|
*/
|
|
|
|
if (!(pollfd->revents & pollfd->events & LWS_POLLIN))
|
|
break;
|
|
read:
|
|
if (lws_is_flowcontrolled(wsi)) {
|
|
lwsl_info("%s: %p should be rxflow (bm 0x%x)..\n",
|
|
__func__, wsi, wsi->rxflow_bitmap);
|
|
break;
|
|
}
|
|
|
|
/* all the union members start with hdr, so even in ws mode
|
|
* we can deal with the ah via u.hdr
|
|
*/
|
|
if (wsi->u.hdr.ah) {
|
|
lwsl_info("%s: %p: inherited ah rx\n", __func__, wsi);
|
|
eff_buf.token_len = wsi->u.hdr.ah->rxlen -
|
|
wsi->u.hdr.ah->rxpos;
|
|
eff_buf.token = (char *)wsi->u.hdr.ah->rx +
|
|
wsi->u.hdr.ah->rxpos;
|
|
} else {
|
|
if (wsi->mode != LWSCM_HTTP_CLIENT_ACCEPTED) {
|
|
/*
|
|
* extension may not consume everything (eg, pmd may be constrained
|
|
* as to what it can output...) has to go in per-wsi rx buf area.
|
|
* Otherwise in large temp serv_buf area.
|
|
*/
|
|
|
|
#if defined(LWS_WITH_HTTP2)
|
|
if (wsi->upgraded_to_http2) {
|
|
if (!wsi->u.h2.h2n->rx_scratch) {
|
|
wsi->u.h2.h2n->rx_scratch = lws_malloc(LWS_H2_RX_SCRATCH_SIZE, "h2 rx scratch");
|
|
if (!wsi->u.h2.h2n->rx_scratch)
|
|
goto close_and_handled;
|
|
}
|
|
eff_buf.token = wsi->u.h2.h2n->rx_scratch;
|
|
eff_buf.token_len = LWS_H2_RX_SCRATCH_SIZE;
|
|
} else
|
|
#endif
|
|
{
|
|
eff_buf.token = (char *)pt->serv_buf;
|
|
if (lws_is_ws_with_ext(wsi)) {
|
|
eff_buf.token_len = wsi->u.ws.rx_ubuf_alloc;
|
|
} else {
|
|
eff_buf.token_len = context->pt_serv_buf_size;
|
|
}
|
|
|
|
if ((unsigned int)eff_buf.token_len > context->pt_serv_buf_size)
|
|
eff_buf.token_len = context->pt_serv_buf_size;
|
|
}
|
|
|
|
if ((int)pending > eff_buf.token_len)
|
|
pending = eff_buf.token_len;
|
|
|
|
eff_buf.token_len = lws_ssl_capable_read(wsi,
|
|
(unsigned char *)eff_buf.token, pending ? pending :
|
|
eff_buf.token_len);
|
|
switch (eff_buf.token_len) {
|
|
case 0:
|
|
lwsl_info("%s: zero length read\n", __func__);
|
|
goto close_and_handled;
|
|
case LWS_SSL_CAPABLE_MORE_SERVICE:
|
|
lwsl_info("SSL Capable more service\n");
|
|
n = 0;
|
|
goto handled;
|
|
case LWS_SSL_CAPABLE_ERROR:
|
|
lwsl_info("Closing when error\n");
|
|
goto close_and_handled;
|
|
}
|
|
// lwsl_notice("Actual RX %d\n", eff_buf.token_len);
|
|
}
|
|
}
|
|
|
|
drain:
|
|
#ifndef LWS_NO_CLIENT
|
|
if (wsi->mode == LWSCM_HTTP_CLIENT_ACCEPTED &&
|
|
!wsi->told_user_closed) {
|
|
|
|
/*
|
|
* In SSL mode we get POLLIN notification about
|
|
* encrypted data in.
|
|
*
|
|
* But that is not necessarily related to decrypted
|
|
* data out becoming available; in may need to perform
|
|
* other in or out before that happens.
|
|
*
|
|
* simply mark ourselves as having readable data
|
|
* and turn off our POLLIN
|
|
*/
|
|
wsi->client_rx_avail = 1;
|
|
lws_change_pollfd(wsi, LWS_POLLIN, 0);
|
|
|
|
/* let user code know, he'll usually ask for writeable
|
|
* callback and drain / re-enable it there
|
|
*/
|
|
if (user_callback_handle_rxflow(
|
|
wsi->protocol->callback,
|
|
wsi, LWS_CALLBACK_RECEIVE_CLIENT_HTTP,
|
|
wsi->user_space, NULL, 0)) {
|
|
lwsl_info("RECEIVE_CLIENT_HTTP closed it\n");
|
|
goto close_and_handled;
|
|
}
|
|
|
|
n = 0;
|
|
goto handled;
|
|
}
|
|
#endif
|
|
/*
|
|
* give any active extensions a chance to munge the buffer
|
|
* before parse. We pass in a pointer to an lws_tokens struct
|
|
* prepared with the default buffer and content length that's in
|
|
* there. Rather than rewrite the default buffer, extensions
|
|
* that expect to grow the buffer can adapt .token to
|
|
* point to their own per-connection buffer in the extension
|
|
* user allocation. By default with no extensions or no
|
|
* extension callback handling, just the normal input buffer is
|
|
* used then so it is efficient.
|
|
*/
|
|
do {
|
|
more = 0;
|
|
|
|
m = lws_ext_cb_active(wsi, LWS_EXT_CB_PACKET_RX_PREPARSE,
|
|
&eff_buf, 0);
|
|
if (m < 0)
|
|
goto close_and_handled;
|
|
if (m)
|
|
more = 1;
|
|
|
|
/* service incoming data */
|
|
|
|
if (eff_buf.token_len) {
|
|
/*
|
|
* if draining from rxflow buffer, not
|
|
* critical to track what was used since at the
|
|
* use it bumps wsi->rxflow_pos. If we come
|
|
* around again it will pick up from where it
|
|
* left off.
|
|
*/
|
|
n = lws_read(wsi, (unsigned char *)eff_buf.token,
|
|
eff_buf.token_len);
|
|
if (n < 0) {
|
|
/* we closed wsi */
|
|
n = 0;
|
|
goto handled;
|
|
}
|
|
}
|
|
|
|
eff_buf.token = NULL;
|
|
eff_buf.token_len = 0;
|
|
} while (more);
|
|
|
|
if (wsi->u.hdr.ah) {
|
|
lwsl_debug("%s: %p: detaching\n", __func__, wsi);
|
|
lws_header_table_force_to_detachable_state(wsi);
|
|
/* we can run the normal ah detach flow despite
|
|
* being in ws union mode, since all union members
|
|
* start with hdr */
|
|
lws_header_table_detach(wsi, 0);
|
|
}
|
|
|
|
pending = lws_ssl_pending(wsi);
|
|
if (pending) {
|
|
if (lws_is_ws_with_ext(wsi))
|
|
pending = pending > wsi->u.ws.rx_ubuf_alloc ?
|
|
wsi->u.ws.rx_ubuf_alloc : pending;
|
|
else
|
|
pending = pending > context->pt_serv_buf_size ?
|
|
context->pt_serv_buf_size : pending;
|
|
goto read;
|
|
}
|
|
|
|
if (draining_flow && wsi->rxflow_buffer &&
|
|
wsi->rxflow_pos == wsi->rxflow_len) {
|
|
lwsl_info("%s: %p flow buf: drained\n", __func__, wsi);
|
|
lws_free_set_NULL(wsi->rxflow_buffer);
|
|
/* having drained the rxflow buffer, can rearm POLLIN */
|
|
#ifdef LWS_NO_SERVER
|
|
n =
|
|
#endif
|
|
_lws_rx_flow_control(wsi);
|
|
/* n ignored, needed for NO_SERVER case */
|
|
}
|
|
|
|
break;
|
|
#ifdef LWS_WITH_CGI
|
|
case LWSCM_CGI: /* we exist to handle a cgi's stdin/out/err data...
|
|
* do the callback on our master wsi
|
|
*/
|
|
{
|
|
struct lws_cgi_args args;
|
|
|
|
if (wsi->cgi_channel >= LWS_STDOUT &&
|
|
!(pollfd->revents & pollfd->events & LWS_POLLIN))
|
|
break;
|
|
if (wsi->cgi_channel == LWS_STDIN &&
|
|
!(pollfd->revents & pollfd->events & LWS_POLLOUT))
|
|
break;
|
|
|
|
if (wsi->cgi_channel == LWS_STDIN)
|
|
if (lws_change_pollfd(wsi, LWS_POLLOUT, 0)) {
|
|
lwsl_info("failed at set pollfd\n");
|
|
return 1;
|
|
}
|
|
|
|
args.ch = wsi->cgi_channel;
|
|
args.stdwsi = &wsi->parent->cgi->stdwsi[0];
|
|
args.hdr_state = wsi->hdr_state;
|
|
|
|
lwsl_debug("CGI LWS_STDOUT %p mode %d state %d\n",
|
|
wsi->parent, wsi->parent->mode,
|
|
wsi->parent->state);
|
|
|
|
if (user_callback_handle_rxflow(
|
|
wsi->parent->protocol->callback,
|
|
wsi->parent, LWS_CALLBACK_CGI,
|
|
wsi->parent->user_space,
|
|
(void *)&args, 0))
|
|
return 1;
|
|
|
|
break;
|
|
}
|
|
#endif
|
|
/*
|
|
* something went wrong with parsing the handshake, and
|
|
* we ended up back in the event loop without completing it
|
|
*/
|
|
case LWSCM_PRE_WS_SERVING_ACCEPT:
|
|
wsi->socket_is_permanently_unusable = 1;
|
|
goto close_and_handled;
|
|
|
|
default:
|
|
#ifdef LWS_NO_CLIENT
|
|
break;
|
|
#else
|
|
if ((pollfd->revents & LWS_POLLOUT) &&
|
|
lws_handle_POLLOUT_event(wsi, pollfd)) {
|
|
lwsl_debug("POLLOUT event closed it\n");
|
|
goto close_and_handled;
|
|
}
|
|
|
|
n = lws_client_socket_service(context, wsi, pollfd);
|
|
if (n)
|
|
return 1;
|
|
goto handled;
|
|
#endif
|
|
}
|
|
|
|
n = 0;
|
|
goto handled;
|
|
|
|
close_and_handled:
|
|
lwsl_debug("%p: Close and handled\n", wsi);
|
|
lws_close_free_wsi(wsi, LWS_CLOSE_STATUS_NOSTATUS);
|
|
/*
|
|
* pollfd may point to something else after the close
|
|
* due to pollfd swapping scheme on delete on some platforms
|
|
* we can't clear revents now because it'd be the wrong guy's revents
|
|
*/
|
|
return 1;
|
|
|
|
handled:
|
|
pollfd->revents = 0;
|
|
return n;
|
|
}
|
|
|
|
LWS_VISIBLE int
|
|
lws_service_fd(struct lws_context *context, struct lws_pollfd *pollfd)
|
|
{
|
|
return lws_service_fd_tsi(context, pollfd, 0);
|
|
}
|
|
|
|
LWS_VISIBLE int
|
|
lws_service(struct lws_context *context, int timeout_ms)
|
|
{
|
|
return lws_plat_service(context, timeout_ms);
|
|
}
|
|
|
|
LWS_VISIBLE int
|
|
lws_service_tsi(struct lws_context *context, int timeout_ms, int tsi)
|
|
{
|
|
return _lws_plat_service_tsi(context, timeout_ms, tsi);
|
|
}
|
|
|