2012-07-06 03:47:57 -04:00
|
|
|
/*
|
|
|
|
|
* include/types/connection.h
|
|
|
|
|
* This file describes the connection struct and associated constants.
|
|
|
|
|
*
|
2014-01-23 09:26:18 -05:00
|
|
|
* Copyright (C) 2000-2014 Willy Tarreau - w@1wt.eu
|
2012-07-06 03:47:57 -04:00
|
|
|
*
|
|
|
|
|
* This library is free software; you can redistribute it and/or
|
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
|
* License as published by the Free Software Foundation, version 2.1
|
|
|
|
|
* exclusively.
|
|
|
|
|
*
|
|
|
|
|
* This library is distributed in the hope that it will be useful,
|
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
|
*
|
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
|
* License along with this library; if not, write to the Free Software
|
|
|
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#ifndef _TYPES_CONNECTION_H
|
|
|
|
|
#define _TYPES_CONNECTION_H
|
|
|
|
|
|
|
|
|
|
#include <stdlib.h>
|
|
|
|
|
#include <sys/socket.h>
|
|
|
|
|
|
|
|
|
|
#include <common/config.h>
|
2017-09-21 13:40:52 -04:00
|
|
|
#include <common/ist.h>
|
2012-07-06 03:47:57 -04:00
|
|
|
|
2012-09-12 16:58:11 -04:00
|
|
|
#include <types/listener.h>
|
2012-11-11 18:42:33 -05:00
|
|
|
#include <types/obj_type.h>
|
2012-12-08 16:29:20 -05:00
|
|
|
#include <types/port_range.h>
|
2012-09-12 16:58:11 -04:00
|
|
|
#include <types/protocol.h>
|
|
|
|
|
|
2016-08-10 12:57:38 -04:00
|
|
|
#include <netinet/in_systm.h>
|
2016-06-04 10:11:10 -04:00
|
|
|
#include <netinet/ip.h>
|
|
|
|
|
#include <netinet/ip6.h>
|
|
|
|
|
|
2012-07-06 03:47:57 -04:00
|
|
|
/* referenced below */
|
2012-08-24 12:12:41 -04:00
|
|
|
struct connection;
|
2017-09-13 12:30:23 -04:00
|
|
|
struct conn_stream;
|
2012-08-24 12:12:41 -04:00
|
|
|
struct buffer;
|
2016-12-22 15:13:18 -05:00
|
|
|
struct server;
|
2012-08-24 12:12:41 -04:00
|
|
|
struct pipe;
|
2012-07-06 03:47:57 -04:00
|
|
|
|
2018-08-01 11:06:43 -04:00
|
|
|
enum sub_event_type {
|
|
|
|
|
SUB_CAN_SEND = 0x00000001, /* Schedule the tasklet when we can send more */
|
|
|
|
|
SUB_CAN_RECV = 0x00000002, /* Schedule the tasklet when we can recv more */
|
|
|
|
|
};
|
|
|
|
|
|
2018-07-17 12:46:31 -04:00
|
|
|
struct wait_list {
|
|
|
|
|
struct tasklet *task;
|
|
|
|
|
struct list list;
|
2018-08-01 11:06:43 -04:00
|
|
|
int wait_reason;
|
2018-07-17 12:46:31 -04:00
|
|
|
};
|
2017-08-24 08:31:19 -04:00
|
|
|
|
|
|
|
|
/* A connection handle is how we differenciate two connections on the lower
|
|
|
|
|
* layers. It usually is a file descriptor but can be a connection id.
|
|
|
|
|
*/
|
|
|
|
|
union conn_handle {
|
|
|
|
|
int fd; /* file descriptor, for regular sockets */
|
|
|
|
|
};
|
|
|
|
|
|
2017-09-13 12:30:23 -04:00
|
|
|
/* conn_stream flags */
|
|
|
|
|
enum {
|
|
|
|
|
CS_FL_NONE = 0x00000000, /* Just for initialization purposes */
|
|
|
|
|
CS_FL_DATA_RD_ENA = 0x00000001, /* receiving data is allowed */
|
|
|
|
|
CS_FL_DATA_WR_ENA = 0x00000002, /* sending data is desired */
|
|
|
|
|
|
2017-10-05 09:06:07 -04:00
|
|
|
CS_FL_SHRD = 0x00000010, /* read shut, draining extra data */
|
|
|
|
|
CS_FL_SHRR = 0x00000020, /* read shut, resetting extra data */
|
|
|
|
|
CS_FL_SHR = CS_FL_SHRD | CS_FL_SHRR, /* read shut status */
|
|
|
|
|
|
|
|
|
|
CS_FL_SHWN = 0x00000040, /* write shut, verbose mode */
|
|
|
|
|
CS_FL_SHWS = 0x00000080, /* write shut, silent mode */
|
|
|
|
|
CS_FL_SHW = CS_FL_SHWN | CS_FL_SHWS, /* write shut status */
|
|
|
|
|
|
|
|
|
|
|
2017-09-13 12:30:23 -04:00
|
|
|
CS_FL_ERROR = 0x00000100, /* a fatal error was reported */
|
2017-12-10 15:13:25 -05:00
|
|
|
CS_FL_RCV_MORE = 0x00000200, /* more bytes to receive but not enough room */
|
2018-03-02 06:25:45 -05:00
|
|
|
CS_FL_EOS = 0x00001000, /* End of stream delivered to data layer */
|
|
|
|
|
CS_FL_REOS = 0x00002000, /* End of stream received (buffer not empty) */
|
2017-11-27 12:41:32 -05:00
|
|
|
CS_FL_WAIT_FOR_HS = 0x00010000, /* This stream is waiting for handhskae */
|
2017-09-13 12:30:23 -04:00
|
|
|
};
|
2017-08-24 08:31:19 -04:00
|
|
|
|
2017-10-05 09:06:07 -04:00
|
|
|
/* cs_shutr() modes */
|
|
|
|
|
enum cs_shr_mode {
|
|
|
|
|
CS_SHR_DRAIN = 0, /* read shutdown, drain any extra stuff */
|
|
|
|
|
CS_SHR_RESET = 1, /* read shutdown, reset any extra stuff */
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
/* cs_shutw() modes */
|
|
|
|
|
enum cs_shw_mode {
|
|
|
|
|
CS_SHW_NORMAL = 0, /* regular write shutdown */
|
|
|
|
|
CS_SHW_SILENT = 1, /* imminent close, don't notify peer */
|
|
|
|
|
};
|
|
|
|
|
|
2014-01-23 09:26:18 -05:00
|
|
|
/* For each direction, we have a CO_FL_{SOCK,DATA}_<DIR>_ENA flag, which
|
|
|
|
|
* indicates if read or write is desired in that direction for the respective
|
|
|
|
|
* layers. The current status corresponding to the current layer being used is
|
|
|
|
|
* remembered in the CO_FL_CURR_<DIR>_ENA flag. The need to poll (ie receipt of
|
|
|
|
|
* EAGAIN) is remembered at the file descriptor level so that even when the
|
|
|
|
|
* activity is stopped and restarted, we still remember whether it was needed
|
|
|
|
|
* to poll before attempting the I/O.
|
2012-08-17 05:55:04 -04:00
|
|
|
*
|
2014-01-23 09:26:18 -05:00
|
|
|
* The CO_FL_CURR_<DIR>_ENA flag is set from the FD status in
|
|
|
|
|
* conn_refresh_polling_flags(). The FD state is updated according to these
|
|
|
|
|
* flags in conn_cond_update_polling().
|
2012-08-17 05:55:04 -04:00
|
|
|
*/
|
|
|
|
|
|
2012-07-06 03:52:14 -04:00
|
|
|
/* flags for use in connection->flags */
|
|
|
|
|
enum {
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_NONE = 0x00000000, /* Just for initialization purposes */
|
|
|
|
|
|
|
|
|
|
/* Do not change these values without updating conn_*_poll_changes() ! */
|
2012-11-05 11:52:26 -05:00
|
|
|
CO_FL_SOCK_RD_ENA = 0x00000001, /* receiving handshakes is allowed */
|
2017-09-13 12:30:23 -04:00
|
|
|
CO_FL_XPRT_RD_ENA = 0x00000002, /* receiving data is allowed */
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_CURR_RD_ENA = 0x00000004, /* receiving is currently allowed */
|
2014-01-22 13:46:33 -05:00
|
|
|
/* unused : 0x00000008 */
|
2012-11-05 11:52:26 -05:00
|
|
|
|
|
|
|
|
CO_FL_SOCK_WR_ENA = 0x00000010, /* sending handshakes is desired */
|
2017-09-13 12:30:23 -04:00
|
|
|
CO_FL_XPRT_WR_ENA = 0x00000020, /* sending data is desired */
|
2012-11-05 11:52:26 -05:00
|
|
|
CO_FL_CURR_WR_ENA = 0x00000040, /* sending is currently desired */
|
2014-01-22 13:46:33 -05:00
|
|
|
/* unused : 0x00000080 */
|
2012-10-03 14:00:18 -04:00
|
|
|
|
MAJOR: connection: add two new flags to indicate readiness of control/transport
Currently the control and transport layers of a connection are supposed
to be initialized when their respective pointers are not NULL. This will
not work anymore when we plan to reuse connections, because there is an
asymmetry between the accept() side and the connect() side :
- on accept() side, the fd is set first, then the ctrl layer then the
transport layer ; upon error, they must be undone in the reverse order,
then the FD must be closed. The FD must not be deleted if the control
layer was not yet initialized ;
- on the connect() side, the fd is set last and there is no reliable way
to know if it has been initialized or not. In practice it's initialized
to -1 first but this is hackish and supposes that local FDs only will
be used forever. Also, there are even less solutions for keeping trace
of the transport layer's state.
Also it is possible to support delayed close() when something (eg: logs)
tracks some information requiring the transport and/or control layers,
making it even more difficult to clean them.
So the proposed solution is to add two flags to the connection :
- CO_FL_CTRL_READY is set when the control layer is initialized (fd_insert)
and cleared after it's released (fd_delete).
- CO_FL_XPRT_READY is set when the control layer is initialized (xprt->init)
and cleared after it's released (xprt->close).
The functions have been adapted to rely on this and not on the pointers
anymore. conn_xprt_close() was unused and dangerous : it did not close
the control layer (eg: the socket itself) but still marks the transport
layer as closed, preventing any future call to conn_full_close() from
finishing the job.
The problem comes from conn_full_close() in fact. It needs to close the
xprt and ctrl layers independantly. After that we're still having an issue :
we don't know based on ->ctrl alone whether the fd was registered or not.
For this we use the two new flags CO_FL_XPRT_READY and CO_FL_CTRL_READY. We
now rely on this and not on conn->xprt nor conn->ctrl anymore to decide what
remains to be done on the connection.
In order not to miss some flag assignments, we introduce conn_ctrl_init()
to initialize the control layer, register the fd using fd_insert() and set
the flag, and conn_ctrl_close() which unregisters the fd and removes the
flag, but only if the transport layer was closed.
Similarly, at the transport layer, conn_xprt_init() calls ->init and sets
the flag, while conn_xprt_close() checks the flag, calls ->close and clears
the flag, regardless xprt_ctx or xprt_st. This also ensures that the ->init
and the ->close functions are called only once each and in the correct order.
Note that conn_xprt_close() does nothing if the transport layer is still
tracked.
conn_full_close() now simply calls conn_xprt_close() then conn_full_close()
in turn, which do nothing if CO_FL_XPRT_TRACKED is set.
In order to handle the error path, we also provide conn_force_close() which
ignores CO_FL_XPRT_TRACKED and closes the transport and the control layers
in turns. All relevant instances of fd_delete() have been replaced with
conn_force_close(). Now we always know what state the connection is in and
we can expect to split its initialization.
2013-10-21 10:30:56 -04:00
|
|
|
/* These flags indicate whether the Control and Transport layers are initialized */
|
|
|
|
|
CO_FL_CTRL_READY = 0x00000100, /* FD was registered, fd_delete() needed */
|
|
|
|
|
CO_FL_XPRT_READY = 0x00000200, /* xprt_init() done, xprt_close() needed */
|
|
|
|
|
|
2017-10-25 03:22:43 -04:00
|
|
|
CO_FL_WILL_UPDATE = 0x00000400, /* the conn handler will take care of updating the polling */
|
2017-04-26 10:25:12 -04:00
|
|
|
|
|
|
|
|
/* This flag is used by data layers to indicate they had to stop
|
|
|
|
|
* receiving data because a buffer was full. The connection handler
|
|
|
|
|
* clears it before first calling the I/O and data callbacks.
|
2012-08-17 05:55:04 -04:00
|
|
|
*/
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_WAIT_ROOM = 0x00000800, /* data sink is full */
|
2012-08-17 05:55:04 -04:00
|
|
|
|
2012-08-30 15:11:38 -04:00
|
|
|
/* These flags are used to report whether the from/to addresses are set or not */
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_ADDR_FROM_SET = 0x00001000, /* addr.from is set */
|
|
|
|
|
CO_FL_ADDR_TO_SET = 0x00002000, /* addr.to is set */
|
|
|
|
|
|
MEDIUM: ssl: Handle early data with OpenSSL 1.1.1
When compiled with Openssl >= 1.1.1, before attempting to do the handshake,
try to read any early data. If any early data is present, then we'll create
the session, read the data, and handle the request before we're doing the
handshake.
For this, we add a new connection flag, CO_FL_EARLY_SSL_HS, which is not
part of the CO_FL_HANDSHAKE set, allowing to proceed with a session even
before an SSL handshake is completed.
As early data do have security implication, we let the origin server know
the request comes from early data by adding the "Early-Data" header, as
specified in this draft from the HTTP working group :
https://datatracker.ietf.org/doc/html/draft-ietf-httpbis-replay
2017-09-22 12:26:28 -04:00
|
|
|
CO_FL_EARLY_SSL_HS = 0x00004000, /* We have early data pending, don't start SSL handshake yet */
|
|
|
|
|
CO_FL_EARLY_DATA = 0x00008000, /* At least some of the data are early data */
|
2017-08-30 03:59:52 -04:00
|
|
|
/* unused : 0x00010000 */
|
|
|
|
|
/* unused : 0x00020000 */
|
2012-08-20 11:30:32 -04:00
|
|
|
|
2012-08-17 05:55:04 -04:00
|
|
|
/* flags used to remember what shutdown have been performed/reported */
|
|
|
|
|
CO_FL_SOCK_RD_SH = 0x00040000, /* SOCK layer was notified about shutr/read0 */
|
|
|
|
|
CO_FL_SOCK_WR_SH = 0x00080000, /* SOCK layer asked for shutw */
|
|
|
|
|
|
BUG/MEDIUM: connection: ensure to always report the end of handshakes
Despite the previous commit working fine on all tests, it's still not
sufficient to completely address the problem. If the connection handler
is called with an event validating an L4 connection but some handshakes
remain (eg: accept-proxy), it will still wake the function up, which
will not report the activity, and will not detect a change once the
handshake it complete so it will not notify the ->wake() handler.
In fact the only reason why the ->wake() handler is still called here
is because after dropping the last handshake, we try to call ->recv()
and ->send() in turn and change the flags in order to detect a data
activity. But if for any reason the data layer is not interested in
reading nor writing, it will not get these events.
A cleaner way to address this is to call the ->wake() handler only
on definitive status changes (shut, error), on real data activity,
and on a complete connection setup, measured as CONNECTED with no
more handshake pending.
It could be argued that the handshake flags have to be made part of
the condition to set CO_FL_CONNECTED but that would currently break
a part of the health checks. Also a handshake could appear at any
moment even after a connection is established so we'd lose the
ability to detect a second end of handshake.
For now the situation around CO_FL_CONNECTED is not clean :
- session_accept() only sets CO_FL_CONNECTED if there's no pending
handshake ;
- conn_fd_handler() will set it once L4 and L6 are complete, which
will do what session_accept() above refrained from doing even if
an accept_proxy handshake is still pending ;
- ssl_sock_infocbk() and ssl_sock_handshake() consider that a
handshake performed with CO_FL_CONNECTED set is a renegociation ;
=> they should instead filter on CO_FL_WAIT_L6_CONN
- all ssl_fc_* sample fetch functions wait for CO_FL_CONNECTED before
accepting to fetch information
=> they should also get rid of any pending handshake
- smp_fetch_fc_rcvd_proxy() uses !CO_FL_CONNECTED instead of
CO_FL_ACCEPT_PROXY
- health checks (standard and tcp-checks) don't check for HANDSHAKE
and may report a successful check based on CO_FL_CONNECTED while
not yet done (eg: send buffer full on send_proxy).
This patch aims at solving some of these side effects in a backportable
way before this is reworked in depth :
- we need to call ->wake() to report connection success, measure
connection time, notify that the data layer is ready and update
the data layer after activity ; this has to be done either if
we switch from pending {L4,L6}_CONN to nothing with no handshakes
left, or if we notice some handshakes were pending and are now
done.
- we document that CO_FL_CONNECTED exactly means "L4 connection
setup confirmed at least once, L6 connection setup confirmed
at least once or not necessary, all this regardless of any
possibly remaining handshakes or future L6 negociations".
This patch also renames CO_FL_CONN_STATUS to the more explicit
CO_FL_NOTIFY_DATA, and works around the previous flags trick consiting
in setting an impossible combination of flags to notify the data layer,
by simply clearing the current flags.
This fix should be backported to 1.7, 1.6 and 1.5.
2017-03-19 02:54:28 -04:00
|
|
|
/* flags used to report connection errors or other closing conditions */
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_ERROR = 0x00100000, /* a fatal error was reported */
|
2017-08-28 09:46:01 -04:00
|
|
|
CO_FL_NOTIFY_DONE = 0x001C0000, /* any xprt shut/error flags above needs to be reported */
|
2017-09-20 11:46:46 -04:00
|
|
|
CO_FL_NOTIFY_DATA = 0x001C0000, /* any shut/error flags above needs to be reported */
|
BUG/MEDIUM: connection: ensure to always report the end of handshakes
Despite the previous commit working fine on all tests, it's still not
sufficient to completely address the problem. If the connection handler
is called with an event validating an L4 connection but some handshakes
remain (eg: accept-proxy), it will still wake the function up, which
will not report the activity, and will not detect a change once the
handshake it complete so it will not notify the ->wake() handler.
In fact the only reason why the ->wake() handler is still called here
is because after dropping the last handshake, we try to call ->recv()
and ->send() in turn and change the flags in order to detect a data
activity. But if for any reason the data layer is not interested in
reading nor writing, it will not get these events.
A cleaner way to address this is to call the ->wake() handler only
on definitive status changes (shut, error), on real data activity,
and on a complete connection setup, measured as CONNECTED with no
more handshake pending.
It could be argued that the handshake flags have to be made part of
the condition to set CO_FL_CONNECTED but that would currently break
a part of the health checks. Also a handshake could appear at any
moment even after a connection is established so we'd lose the
ability to detect a second end of handshake.
For now the situation around CO_FL_CONNECTED is not clean :
- session_accept() only sets CO_FL_CONNECTED if there's no pending
handshake ;
- conn_fd_handler() will set it once L4 and L6 are complete, which
will do what session_accept() above refrained from doing even if
an accept_proxy handshake is still pending ;
- ssl_sock_infocbk() and ssl_sock_handshake() consider that a
handshake performed with CO_FL_CONNECTED set is a renegociation ;
=> they should instead filter on CO_FL_WAIT_L6_CONN
- all ssl_fc_* sample fetch functions wait for CO_FL_CONNECTED before
accepting to fetch information
=> they should also get rid of any pending handshake
- smp_fetch_fc_rcvd_proxy() uses !CO_FL_CONNECTED instead of
CO_FL_ACCEPT_PROXY
- health checks (standard and tcp-checks) don't check for HANDSHAKE
and may report a successful check based on CO_FL_CONNECTED while
not yet done (eg: send buffer full on send_proxy).
This patch aims at solving some of these side effects in a backportable
way before this is reworked in depth :
- we need to call ->wake() to report connection success, measure
connection time, notify that the data layer is ready and update
the data layer after activity ; this has to be done either if
we switch from pending {L4,L6}_CONN to nothing with no handshakes
left, or if we notice some handshakes were pending and are now
done.
- we document that CO_FL_CONNECTED exactly means "L4 connection
setup confirmed at least once, L6 connection setup confirmed
at least once or not necessary, all this regardless of any
possibly remaining handshakes or future L6 negociations".
This patch also renames CO_FL_CONN_STATUS to the more explicit
CO_FL_NOTIFY_DATA, and works around the previous flags trick consiting
in setting an impossible combination of flags to notify the data layer,
by simply clearing the current flags.
This fix should be backported to 1.7, 1.6 and 1.5.
2017-03-19 02:54:28 -04:00
|
|
|
|
|
|
|
|
/* flags used to report connection status updates */
|
|
|
|
|
CO_FL_CONNECTED = 0x00200000, /* L4+L6 now ready ; extra handshakes may or may not exist */
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_WAIT_L4_CONN = 0x00400000, /* waiting for L4 to be connected */
|
|
|
|
|
CO_FL_WAIT_L6_CONN = 0x00800000, /* waiting for L6 to be connected (eg: SSL) */
|
2012-09-01 11:26:16 -04:00
|
|
|
|
2012-10-03 14:00:18 -04:00
|
|
|
/*** All the flags below are used for connection handshakes. Any new
|
|
|
|
|
* handshake should be added after this point, and CO_FL_HANDSHAKE
|
|
|
|
|
* should be updated.
|
2012-09-01 11:26:16 -04:00
|
|
|
*/
|
2013-10-24 16:01:26 -04:00
|
|
|
CO_FL_SEND_PROXY = 0x01000000, /* send a valid PROXY protocol header */
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_SSL_WAIT_HS = 0x02000000, /* wait for an SSL handshake to complete */
|
2012-10-04 17:55:57 -04:00
|
|
|
CO_FL_ACCEPT_PROXY = 0x04000000, /* receive a valid PROXY protocol header */
|
2016-06-04 10:11:10 -04:00
|
|
|
CO_FL_ACCEPT_CIP = 0x08000000, /* receive a valid NetScaler Client IP header */
|
2012-10-03 14:00:18 -04:00
|
|
|
|
|
|
|
|
/* below we have all handshake flags grouped into one */
|
2016-06-04 10:11:10 -04:00
|
|
|
CO_FL_HANDSHAKE = CO_FL_SEND_PROXY | CO_FL_SSL_WAIT_HS | CO_FL_ACCEPT_PROXY | CO_FL_ACCEPT_CIP,
|
2012-10-03 14:00:18 -04:00
|
|
|
|
|
|
|
|
/* when any of these flags is set, polling is defined by socket-layer
|
|
|
|
|
* operations, as opposed to data-layer. Transport is explicitly not
|
|
|
|
|
* mentionned here to avoid any confusion, since it can be the same
|
|
|
|
|
* as DATA or SOCK on some implementations.
|
2012-09-01 11:26:16 -04:00
|
|
|
*/
|
2012-10-03 14:00:18 -04:00
|
|
|
CO_FL_POLL_SOCK = CO_FL_HANDSHAKE | CO_FL_WAIT_L4_CONN | CO_FL_WAIT_L6_CONN,
|
2012-10-12 11:50:05 -04:00
|
|
|
|
2015-08-04 13:24:13 -04:00
|
|
|
/* This connection may not be shared between clients */
|
|
|
|
|
CO_FL_PRIVATE = 0x10000000,
|
|
|
|
|
|
2017-01-05 09:11:44 -05:00
|
|
|
/* This flag is used to know that a PROXY protocol header was sent by the client */
|
|
|
|
|
CO_FL_RCVD_PROXY = 0x20000000,
|
|
|
|
|
|
|
|
|
|
/* unused : 0x40000000 */
|
MAJOR: connection: add two new flags to indicate readiness of control/transport
Currently the control and transport layers of a connection are supposed
to be initialized when their respective pointers are not NULL. This will
not work anymore when we plan to reuse connections, because there is an
asymmetry between the accept() side and the connect() side :
- on accept() side, the fd is set first, then the ctrl layer then the
transport layer ; upon error, they must be undone in the reverse order,
then the FD must be closed. The FD must not be deleted if the control
layer was not yet initialized ;
- on the connect() side, the fd is set last and there is no reliable way
to know if it has been initialized or not. In practice it's initialized
to -1 first but this is hackish and supposes that local FDs only will
be used forever. Also, there are even less solutions for keeping trace
of the transport layer's state.
Also it is possible to support delayed close() when something (eg: logs)
tracks some information requiring the transport and/or control layers,
making it even more difficult to clean them.
So the proposed solution is to add two flags to the connection :
- CO_FL_CTRL_READY is set when the control layer is initialized (fd_insert)
and cleared after it's released (fd_delete).
- CO_FL_XPRT_READY is set when the control layer is initialized (xprt->init)
and cleared after it's released (xprt->close).
The functions have been adapted to rely on this and not on the pointers
anymore. conn_xprt_close() was unused and dangerous : it did not close
the control layer (eg: the socket itself) but still marks the transport
layer as closed, preventing any future call to conn_full_close() from
finishing the job.
The problem comes from conn_full_close() in fact. It needs to close the
xprt and ctrl layers independantly. After that we're still having an issue :
we don't know based on ->ctrl alone whether the fd was registered or not.
For this we use the two new flags CO_FL_XPRT_READY and CO_FL_CTRL_READY. We
now rely on this and not on conn->xprt nor conn->ctrl anymore to decide what
remains to be done on the connection.
In order not to miss some flag assignments, we introduce conn_ctrl_init()
to initialize the control layer, register the fd using fd_insert() and set
the flag, and conn_ctrl_close() which unregisters the fd and removes the
flag, but only if the transport layer was closed.
Similarly, at the transport layer, conn_xprt_init() calls ->init and sets
the flag, while conn_xprt_close() checks the flag, calls ->close and clears
the flag, regardless xprt_ctx or xprt_st. This also ensures that the ->init
and the ->close functions are called only once each and in the correct order.
Note that conn_xprt_close() does nothing if the transport layer is still
tracked.
conn_full_close() now simply calls conn_xprt_close() then conn_full_close()
in turn, which do nothing if CO_FL_XPRT_TRACKED is set.
In order to handle the error path, we also provide conn_force_close() which
ignores CO_FL_XPRT_TRACKED and closes the transport and the control layers
in turns. All relevant instances of fd_delete() have been replaced with
conn_force_close(). Now we always know what state the connection is in and
we can expect to split its initialization.
2013-10-21 10:30:56 -04:00
|
|
|
|
2012-10-12 11:50:05 -04:00
|
|
|
/* This last flag indicates that the transport layer is used (for instance
|
|
|
|
|
* by logs) and must not be cleared yet. The last call to conn_xprt_close()
|
|
|
|
|
* must be done after clearing this flag.
|
|
|
|
|
*/
|
|
|
|
|
CO_FL_XPRT_TRACKED = 0x80000000,
|
2012-07-06 03:52:14 -04:00
|
|
|
};
|
|
|
|
|
|
2012-11-30 11:33:05 -05:00
|
|
|
|
|
|
|
|
/* possible connection error codes */
|
|
|
|
|
enum {
|
|
|
|
|
CO_ER_NONE, /* no error */
|
2014-01-24 10:06:50 -05:00
|
|
|
|
|
|
|
|
CO_ER_CONF_FDLIM, /* reached process' configured FD limitation */
|
|
|
|
|
CO_ER_PROC_FDLIM, /* reached process' FD limitation */
|
|
|
|
|
CO_ER_SYS_FDLIM, /* reached system's FD limitation */
|
|
|
|
|
CO_ER_SYS_MEMLIM, /* reached system buffers limitation */
|
|
|
|
|
CO_ER_NOPROTO, /* protocol not supported */
|
|
|
|
|
CO_ER_SOCK_ERR, /* other socket error */
|
|
|
|
|
|
|
|
|
|
CO_ER_PORT_RANGE, /* source port range exhausted */
|
|
|
|
|
CO_ER_CANT_BIND, /* can't bind to source address */
|
|
|
|
|
CO_ER_FREE_PORTS, /* no more free ports on the system */
|
|
|
|
|
CO_ER_ADDR_INUSE, /* local address already in use */
|
|
|
|
|
|
2012-12-03 09:41:18 -05:00
|
|
|
CO_ER_PRX_EMPTY, /* nothing received in PROXY protocol header */
|
|
|
|
|
CO_ER_PRX_ABORT, /* client abort during PROXY protocol header */
|
2012-12-03 09:35:00 -05:00
|
|
|
CO_ER_PRX_TIMEOUT, /* timeout while waiting for a PROXY header */
|
2012-12-03 09:41:18 -05:00
|
|
|
CO_ER_PRX_TRUNCATED, /* truncated PROXY protocol header */
|
|
|
|
|
CO_ER_PRX_NOT_HDR, /* not a PROXY protocol header */
|
|
|
|
|
CO_ER_PRX_BAD_HDR, /* bad PROXY protocol header */
|
|
|
|
|
CO_ER_PRX_BAD_PROTO, /* unsupported protocol in PROXY header */
|
|
|
|
|
|
2016-06-04 10:11:10 -04:00
|
|
|
CO_ER_CIP_EMPTY, /* nothing received in NetScaler Client IP header */
|
|
|
|
|
CO_ER_CIP_ABORT, /* client abort during NetScaler Client IP header */
|
|
|
|
|
CO_ER_CIP_TIMEOUT, /* timeout while waiting for a NetScaler Client IP header */
|
|
|
|
|
CO_ER_CIP_TRUNCATED, /* truncated NetScaler Client IP header */
|
|
|
|
|
CO_ER_CIP_BAD_MAGIC, /* bad magic number in NetScaler Client IP header */
|
|
|
|
|
CO_ER_CIP_BAD_PROTO, /* unsupported protocol in NetScaler Client IP header */
|
|
|
|
|
|
2012-12-03 10:32:10 -05:00
|
|
|
CO_ER_SSL_EMPTY, /* client closed during SSL handshake */
|
|
|
|
|
CO_ER_SSL_ABORT, /* client abort during SSL handshake */
|
2012-12-03 09:35:00 -05:00
|
|
|
CO_ER_SSL_TIMEOUT, /* timeout during SSL handshake */
|
2012-12-03 10:32:10 -05:00
|
|
|
CO_ER_SSL_TOO_MANY, /* too many SSL connections */
|
|
|
|
|
CO_ER_SSL_NO_MEM, /* no more memory to allocate an SSL connection */
|
|
|
|
|
CO_ER_SSL_RENEG, /* forbidden client renegociation */
|
|
|
|
|
CO_ER_SSL_CA_FAIL, /* client cert verification failed in the CA chain */
|
|
|
|
|
CO_ER_SSL_CRT_FAIL, /* client cert verification failed on the certificate */
|
2017-07-26 14:09:56 -04:00
|
|
|
CO_ER_SSL_MISMATCH, /* Server presented an SSL certificate different from the configured one */
|
|
|
|
|
CO_ER_SSL_MISMATCH_SNI, /* Server presented an SSL certificate different from the expected one */
|
2012-12-03 10:32:10 -05:00
|
|
|
CO_ER_SSL_HANDSHAKE, /* SSL error during handshake */
|
2014-04-25 12:54:29 -04:00
|
|
|
CO_ER_SSL_HANDSHAKE_HB, /* SSL error during handshake with heartbeat present */
|
2014-04-25 14:02:39 -04:00
|
|
|
CO_ER_SSL_KILLED_HB, /* Stopped a TLSv1 heartbeat attack (CVE-2014-0160) */
|
|
|
|
|
CO_ER_SSL_NO_TARGET, /* unknown target (not client nor server) */
|
2017-11-03 11:27:47 -04:00
|
|
|
CO_ER_SSL_EARLY_FAILED, /* Server refused early data */
|
2012-11-30 11:33:05 -05:00
|
|
|
};
|
|
|
|
|
|
2012-12-08 16:29:20 -05:00
|
|
|
/* source address settings for outgoing connections */
|
|
|
|
|
enum {
|
|
|
|
|
/* Tproxy exclusive values from 0 to 7 */
|
|
|
|
|
CO_SRC_TPROXY_ADDR = 0x0001, /* bind to this non-local address when connecting */
|
|
|
|
|
CO_SRC_TPROXY_CIP = 0x0002, /* bind to the client's IP address when connecting */
|
|
|
|
|
CO_SRC_TPROXY_CLI = 0x0003, /* bind to the client's IP+port when connecting */
|
|
|
|
|
CO_SRC_TPROXY_DYN = 0x0004, /* bind to a dynamically computed non-local address */
|
|
|
|
|
CO_SRC_TPROXY_MASK = 0x0007, /* bind to a non-local address when connecting */
|
|
|
|
|
|
|
|
|
|
CO_SRC_BIND = 0x0008, /* bind to a specific source address when connecting */
|
|
|
|
|
};
|
|
|
|
|
|
2018-06-19 00:23:38 -04:00
|
|
|
/* flags that can be passed to xprt->rcv_buf() and mux->rcv_buf() */
|
|
|
|
|
enum {
|
|
|
|
|
CO_RFL_BUF_WET = 0x0001, /* Buffer still has some output data present */
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
/* flags that can be passed to xprt->snd_buf() and mux->snd_buf() */
|
2014-02-01 19:51:17 -05:00
|
|
|
enum {
|
|
|
|
|
CO_SFL_MSG_MORE = 0x0001, /* More data to come afterwards */
|
2014-02-01 20:00:24 -05:00
|
|
|
CO_SFL_STREAMER = 0x0002, /* Producer is continuously streaming data */
|
2014-02-01 19:51:17 -05:00
|
|
|
};
|
2012-12-08 16:29:20 -05:00
|
|
|
|
2016-12-22 14:25:26 -05:00
|
|
|
/* known transport layers (for ease of lookup) */
|
|
|
|
|
enum {
|
|
|
|
|
XPRT_RAW = 0,
|
|
|
|
|
XPRT_SSL = 1,
|
|
|
|
|
XPRT_ENTRIES /* must be last one */
|
|
|
|
|
};
|
|
|
|
|
|
2017-12-20 10:14:44 -05:00
|
|
|
/* MUX-specific flags */
|
|
|
|
|
enum {
|
|
|
|
|
MX_FL_NONE = 0x00000000,
|
|
|
|
|
MX_FL_CLEAN_ABRT = 0x00000001, /* abort is clearly reported as an error */
|
|
|
|
|
};
|
|
|
|
|
|
REORG: connection: rename the data layer the "transport layer"
While working on the changes required to make the health checks use the
new connections, it started to become obvious that some naming was not
logical at all in the connections. Specifically, it is not logical to
call the "data layer" the layer which is in charge for all the handshake
and which does not yet provide a data layer once established until a
session has allocated all the required buffers.
In fact, it's more a transport layer, which makes much more sense. The
transport layer offers a medium on which data can transit, and it offers
the functions to move these data when the upper layer requests this. And
it is the upper layer which iterates over the transport layer's functions
to move data which should be called the data layer.
The use case where it's obvious is with embryonic sessions : an incoming
SSL connection is accepted. Only the connection is allocated, not the
buffers nor stream interface, etc... The connection handles the SSL
handshake by itself. Once this handshake is complete, we can't use the
data functions because the buffers and stream interface are not there
yet. Hence we have to first call a specific function to complete the
session initialization, after which we'll be able to use the data
functions. This clearly proves that SSL here is only a transport layer
and that the stream interface constitutes the data layer.
A similar change will be performed to rename app_cb => data, but the
two could not be in the same commit for obvious reasons.
2012-10-02 18:19:48 -04:00
|
|
|
/* xprt_ops describes transport-layer operations for a connection. They
|
|
|
|
|
* generally run over a socket-based control layer, but not always. Some
|
|
|
|
|
* of them are used for data transfer with the upper layer (rcv_*, snd_*)
|
|
|
|
|
* and the other ones are used to setup and release the transport layer.
|
2012-08-24 12:12:41 -04:00
|
|
|
*/
|
REORG: connection: rename the data layer the "transport layer"
While working on the changes required to make the health checks use the
new connections, it started to become obvious that some naming was not
logical at all in the connections. Specifically, it is not logical to
call the "data layer" the layer which is in charge for all the handshake
and which does not yet provide a data layer once established until a
session has allocated all the required buffers.
In fact, it's more a transport layer, which makes much more sense. The
transport layer offers a medium on which data can transit, and it offers
the functions to move these data when the upper layer requests this. And
it is the upper layer which iterates over the transport layer's functions
to move data which should be called the data layer.
The use case where it's obvious is with embryonic sessions : an incoming
SSL connection is accepted. Only the connection is allocated, not the
buffers nor stream interface, etc... The connection handles the SSL
handshake by itself. Once this handshake is complete, we can't use the
data functions because the buffers and stream interface are not there
yet. Hence we have to first call a specific function to complete the
session initialization, after which we'll be able to use the data
functions. This clearly proves that SSL here is only a transport layer
and that the stream interface constitutes the data layer.
A similar change will be performed to rename app_cb => data, but the
two could not be in the same commit for obvious reasons.
2012-10-02 18:19:48 -04:00
|
|
|
struct xprt_ops {
|
2018-06-19 00:15:17 -04:00
|
|
|
size_t (*rcv_buf)(struct connection *conn, struct buffer *buf, size_t count, int flags); /* recv callback */
|
2018-06-14 12:31:46 -04:00
|
|
|
size_t (*snd_buf)(struct connection *conn, const struct buffer *buf, size_t count, int flags); /* send callback */
|
2012-08-24 12:12:41 -04:00
|
|
|
int (*rcv_pipe)(struct connection *conn, struct pipe *pipe, unsigned int count); /* recv-to-pipe callback */
|
|
|
|
|
int (*snd_pipe)(struct connection *conn, struct pipe *pipe); /* send-to-pipe callback */
|
|
|
|
|
void (*shutr)(struct connection *, int); /* shutr function */
|
|
|
|
|
void (*shutw)(struct connection *, int); /* shutw function */
|
REORG: connection: rename the data layer the "transport layer"
While working on the changes required to make the health checks use the
new connections, it started to become obvious that some naming was not
logical at all in the connections. Specifically, it is not logical to
call the "data layer" the layer which is in charge for all the handshake
and which does not yet provide a data layer once established until a
session has allocated all the required buffers.
In fact, it's more a transport layer, which makes much more sense. The
transport layer offers a medium on which data can transit, and it offers
the functions to move these data when the upper layer requests this. And
it is the upper layer which iterates over the transport layer's functions
to move data which should be called the data layer.
The use case where it's obvious is with embryonic sessions : an incoming
SSL connection is accepted. Only the connection is allocated, not the
buffers nor stream interface, etc... The connection handles the SSL
handshake by itself. Once this handshake is complete, we can't use the
data functions because the buffers and stream interface are not there
yet. Hence we have to first call a specific function to complete the
session initialization, after which we'll be able to use the data
functions. This clearly proves that SSL here is only a transport layer
and that the stream interface constitutes the data layer.
A similar change will be performed to rename app_cb => data, but the
two could not be in the same commit for obvious reasons.
2012-10-02 18:19:48 -04:00
|
|
|
void (*close)(struct connection *); /* close the transport layer */
|
|
|
|
|
int (*init)(struct connection *conn); /* initialize the transport layer */
|
2016-12-22 11:19:24 -05:00
|
|
|
int (*prepare_bind_conf)(struct bind_conf *conf); /* prepare a whole bind_conf */
|
2016-12-22 11:30:20 -05:00
|
|
|
void (*destroy_bind_conf)(struct bind_conf *conf); /* destroy a whole bind_conf */
|
2016-12-22 15:13:18 -05:00
|
|
|
int (*prepare_srv)(struct server *srv); /* prepare a server context */
|
|
|
|
|
void (*destroy_srv)(struct server *srv); /* destroy a server context */
|
2016-12-04 12:42:09 -05:00
|
|
|
int (*get_alpn)(const struct connection *conn, const char **str, int *len); /* get application layer name */
|
2016-11-24 10:58:12 -05:00
|
|
|
char name[8]; /* transport layer name, zero-terminated */
|
2018-07-17 12:46:31 -04:00
|
|
|
int (*subscribe)(struct connection *conn, int event_type, void *param); /* Subscribe to events, such as "being able to send" */
|
2012-08-24 12:12:41 -04:00
|
|
|
};
|
|
|
|
|
|
MEDIUM: connection: start to introduce a mux layer between xprt and data
For HTTP/2 and QUIC, we'll need to deal with multiplexed streams inside
a connection. After quite a long brainstorming, it appears that the
connection interface to the existing streams is appropriate just like
the connection interface to the lower layers. In fact we need to have
the mux layer in the middle of the connection, between the transport
and the data layer.
A mux can exist on two directions/sides. On the inbound direction, it
instanciates new streams from incoming connections, while on the outbound
direction it muxes streams into outgoing connections. The difference is
visible on the mux->init() call : in one case, an upper context is already
known (outgoing connection), and in the other case, the upper context is
not yet known (incoming connection) and will have to be allocated by the
mux. The session doesn't have to create the new streams anymore, as this
is performed by the mux itself.
This patch introduces this and creates a pass-through mux called
"mux_pt" which is used for all new connections and which only
calls the data layer's recv,send,wake() calls. One incoming stream
is immediately created when init() is called on the inbound direction.
There should not be any visible impact.
Note that the connection's mux is purposely not set until the session
is completed so that we don't accidently run with the wrong mux. This
must not cause any issue as the xprt_done_cb function is always called
prior to using mux's recv/send functions.
2017-08-28 04:53:00 -04:00
|
|
|
/* mux_ops describes the mux operations, which are to be performed at the
|
|
|
|
|
* connection level after data are exchanged with the transport layer in order
|
|
|
|
|
* to propagate them to streams. The <init> function will automatically be
|
|
|
|
|
* called once the mux is instanciated by the connection's owner at the end
|
|
|
|
|
* of a transport handshake, when it is about to transfer data and the data
|
|
|
|
|
* layer is not ready yet.
|
|
|
|
|
*/
|
|
|
|
|
struct mux_ops {
|
|
|
|
|
int (*init)(struct connection *conn); /* early initialization */
|
|
|
|
|
void (*recv)(struct connection *conn); /* mux-layer recv callback */
|
|
|
|
|
int (*wake)(struct connection *conn); /* mux-layer callback to report activity, mandatory */
|
2017-09-13 12:30:23 -04:00
|
|
|
void (*update_poll)(struct conn_stream *cs); /* commit cs flags to mux/conn */
|
2018-06-19 00:15:17 -04:00
|
|
|
size_t (*rcv_buf)(struct conn_stream *cs, struct buffer *buf, size_t count, int flags); /* Called from the upper layer to get data */
|
2018-07-27 05:59:41 -04:00
|
|
|
size_t (*snd_buf)(struct conn_stream *cs, struct buffer *buf, size_t count, int flags); /* Called from the upper layer to send data */
|
2017-09-13 12:30:23 -04:00
|
|
|
int (*rcv_pipe)(struct conn_stream *cs, struct pipe *pipe, unsigned int count); /* recv-to-pipe callback */
|
|
|
|
|
int (*snd_pipe)(struct conn_stream *cs, struct pipe *pipe); /* send-to-pipe callback */
|
2017-10-05 09:25:48 -04:00
|
|
|
void (*shutr)(struct conn_stream *cs, enum cs_shr_mode); /* shutr function */
|
|
|
|
|
void (*shutw)(struct conn_stream *cs, enum cs_shw_mode); /* shutw function */
|
2017-09-13 12:30:23 -04:00
|
|
|
|
|
|
|
|
struct conn_stream *(*attach)(struct connection *); /* Create and attach a conn_stream to an outgoing connection */
|
|
|
|
|
void (*detach)(struct conn_stream *); /* Detach a conn_stream from an outgoing connection, when the request is done */
|
2018-07-13 05:56:34 -04:00
|
|
|
void (*show_fd)(struct buffer *, struct connection *); /* append some data about connection into chunk for "show fd" */
|
2018-07-17 12:46:31 -04:00
|
|
|
int (*subscribe)(struct conn_stream *cs, int event_type, void *param); /* Subscribe to events, such as "being able to send" */
|
2017-12-20 10:14:44 -05:00
|
|
|
unsigned int flags; /* some flags characterizing the mux's capabilities (MX_FL_*) */
|
MEDIUM: connection: start to introduce a mux layer between xprt and data
For HTTP/2 and QUIC, we'll need to deal with multiplexed streams inside
a connection. After quite a long brainstorming, it appears that the
connection interface to the existing streams is appropriate just like
the connection interface to the lower layers. In fact we need to have
the mux layer in the middle of the connection, between the transport
and the data layer.
A mux can exist on two directions/sides. On the inbound direction, it
instanciates new streams from incoming connections, while on the outbound
direction it muxes streams into outgoing connections. The difference is
visible on the mux->init() call : in one case, an upper context is already
known (outgoing connection), and in the other case, the upper context is
not yet known (incoming connection) and will have to be allocated by the
mux. The session doesn't have to create the new streams anymore, as this
is performed by the mux itself.
This patch introduces this and creates a pass-through mux called
"mux_pt" which is used for all new connections and which only
calls the data layer's recv,send,wake() calls. One incoming stream
is immediately created when init() is called on the inbound direction.
There should not be any visible impact.
Note that the connection's mux is purposely not set until the session
is completed so that we don't accidently run with the wrong mux. This
must not cause any issue as the xprt_done_cb function is always called
prior to using mux's recv/send functions.
2017-08-28 04:53:00 -04:00
|
|
|
char name[8]; /* mux layer name, zero-terminated */
|
|
|
|
|
};
|
|
|
|
|
|
2012-10-02 18:41:04 -04:00
|
|
|
/* data_cb describes the data layer's recv and send callbacks which are called
|
REORG: connection: rename the data layer the "transport layer"
While working on the changes required to make the health checks use the
new connections, it started to become obvious that some naming was not
logical at all in the connections. Specifically, it is not logical to
call the "data layer" the layer which is in charge for all the handshake
and which does not yet provide a data layer once established until a
session has allocated all the required buffers.
In fact, it's more a transport layer, which makes much more sense. The
transport layer offers a medium on which data can transit, and it offers
the functions to move these data when the upper layer requests this. And
it is the upper layer which iterates over the transport layer's functions
to move data which should be called the data layer.
The use case where it's obvious is with embryonic sessions : an incoming
SSL connection is accepted. Only the connection is allocated, not the
buffers nor stream interface, etc... The connection handles the SSL
handshake by itself. Once this handshake is complete, we can't use the
data functions because the buffers and stream interface are not there
yet. Hence we have to first call a specific function to complete the
session initialization, after which we'll be able to use the data
functions. This clearly proves that SSL here is only a transport layer
and that the stream interface constitutes the data layer.
A similar change will be performed to rename app_cb => data, but the
two could not be in the same commit for obvious reasons.
2012-10-02 18:19:48 -04:00
|
|
|
* when I/O activity was detected after the transport layer is ready. These
|
|
|
|
|
* callbacks are supposed to make use of the xprt_ops above to exchange data
|
2012-10-02 14:07:22 -04:00
|
|
|
* from/to buffers and pipes. The <wake> callback is used to report activity
|
|
|
|
|
* at the transport layer, which can be a connection opening/close, or any
|
2017-08-28 09:46:01 -04:00
|
|
|
* data movement. It may abort a connection by returning < 0.
|
2012-08-24 12:12:41 -04:00
|
|
|
*/
|
2012-10-02 18:41:04 -04:00
|
|
|
struct data_cb {
|
2017-09-13 12:30:23 -04:00
|
|
|
void (*recv)(struct conn_stream *cs); /* data-layer recv callback */
|
|
|
|
|
int (*wake)(struct conn_stream *cs); /* data-layer callback to report activity */
|
2018-07-17 12:46:31 -04:00
|
|
|
int (*subscribe)(struct conn_stream *cs, int event_type, void *param); /* Subscribe to events, such as "being able to send" */
|
2016-11-24 10:58:12 -05:00
|
|
|
char name[8]; /* data layer name, zero-terminated */
|
2012-08-24 12:12:41 -04:00
|
|
|
};
|
|
|
|
|
|
2016-07-04 17:51:33 -04:00
|
|
|
struct my_tcphdr {
|
2016-11-20 05:42:38 -05:00
|
|
|
uint16_t source;
|
|
|
|
|
uint16_t dest;
|
2016-07-04 17:51:33 -04:00
|
|
|
};
|
|
|
|
|
|
2012-12-08 16:29:20 -05:00
|
|
|
/* a connection source profile defines all the parameters needed to properly
|
|
|
|
|
* bind an outgoing connection for a server or proxy.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
struct conn_src {
|
|
|
|
|
unsigned int opts; /* CO_SRC_* */
|
|
|
|
|
int iface_len; /* bind interface name length */
|
|
|
|
|
char *iface_name; /* bind interface name or NULL */
|
|
|
|
|
struct port_range *sport_range; /* optional per-server TCP source ports */
|
|
|
|
|
struct sockaddr_storage source_addr; /* the address to which we want to bind for connect() */
|
2015-08-20 13:35:14 -04:00
|
|
|
#if defined(CONFIG_HAP_TRANSPARENT)
|
2012-12-08 16:29:20 -05:00
|
|
|
struct sockaddr_storage tproxy_addr; /* non-local address we want to bind to for connect() */
|
|
|
|
|
char *bind_hdr_name; /* bind to this header name if defined */
|
|
|
|
|
int bind_hdr_len; /* length of the name of the header above */
|
|
|
|
|
int bind_hdr_occ; /* occurrence number of header above: >0 = from first, <0 = from end, 0=disabled */
|
|
|
|
|
#endif
|
|
|
|
|
};
|
|
|
|
|
|
2017-09-13 12:30:23 -04:00
|
|
|
/*
|
|
|
|
|
* This structure describes the elements of a connection relevant to a stream
|
|
|
|
|
*/
|
|
|
|
|
struct conn_stream {
|
|
|
|
|
enum obj_type obj_type; /* differentiates connection from applet context */
|
2018-03-02 04:43:58 -05:00
|
|
|
/* 3 bytes hole here */
|
2018-07-17 12:46:31 -04:00
|
|
|
unsigned int flags; /* CS_FL_* */
|
2017-09-13 12:30:23 -04:00
|
|
|
struct connection *conn; /* xprt-level connection */
|
2018-07-17 12:49:38 -04:00
|
|
|
struct wait_list wait_list; /* We're in a wait list for send */
|
2018-07-17 12:46:31 -04:00
|
|
|
struct list send_wait_list; /* list of tasks to wake when we're ready to send */
|
2017-09-13 12:30:23 -04:00
|
|
|
void *data; /* pointer to upper layer's entity (eg: stream interface) */
|
|
|
|
|
const struct data_cb *data_cb; /* data layer callbacks. Must be set before xprt->init() */
|
|
|
|
|
void *ctx; /* mux-specific context */
|
|
|
|
|
};
|
|
|
|
|
|
2012-07-06 03:47:57 -04:00
|
|
|
/* This structure describes a connection with its methods and data.
|
|
|
|
|
* A connection may be performed to proxy or server via a local or remote
|
|
|
|
|
* socket, and can also be made to an internal applet. It can support
|
2013-09-29 03:06:42 -04:00
|
|
|
* several transport schemes (raw, ssl, ...). It can support several
|
2012-07-06 03:47:57 -04:00
|
|
|
* connection control schemes, generally a protocol for socket-oriented
|
2017-08-28 09:46:01 -04:00
|
|
|
* connections, but other methods for applets. The xprt_done_cb() callback
|
|
|
|
|
* is called once the transport layer initialization is done (success or
|
|
|
|
|
* failure). It may return < 0 to report an error and require an abort of the
|
|
|
|
|
* connection being instanciated. It must be removed once done.
|
2012-07-06 03:47:57 -04:00
|
|
|
*/
|
|
|
|
|
struct connection {
|
2018-07-17 12:46:31 -04:00
|
|
|
/* first cache line */
|
2013-09-29 03:06:42 -04:00
|
|
|
enum obj_type obj_type; /* differentiates connection from applet context */
|
2013-12-06 15:09:57 -05:00
|
|
|
unsigned char err_code; /* CO_ER_* */
|
|
|
|
|
signed short send_proxy_ofs; /* <0 = offset to (re)send from the end, >0 = send all */
|
2013-10-24 15:10:08 -04:00
|
|
|
unsigned int flags; /* CO_FL_* */
|
2012-08-24 12:12:41 -04:00
|
|
|
const struct protocol *ctrl; /* operations at the socket layer */
|
2012-10-13 08:33:58 -04:00
|
|
|
const struct xprt_ops *xprt; /* operations at the transport layer */
|
MEDIUM: connection: start to introduce a mux layer between xprt and data
For HTTP/2 and QUIC, we'll need to deal with multiplexed streams inside
a connection. After quite a long brainstorming, it appears that the
connection interface to the existing streams is appropriate just like
the connection interface to the lower layers. In fact we need to have
the mux layer in the middle of the connection, between the transport
and the data layer.
A mux can exist on two directions/sides. On the inbound direction, it
instanciates new streams from incoming connections, while on the outbound
direction it muxes streams into outgoing connections. The difference is
visible on the mux->init() call : in one case, an upper context is already
known (outgoing connection), and in the other case, the upper context is
not yet known (incoming connection) and will have to be allocated by the
mux. The session doesn't have to create the new streams anymore, as this
is performed by the mux itself.
This patch introduces this and creates a pass-through mux called
"mux_pt" which is used for all new connections and which only
calls the data layer's recv,send,wake() calls. One incoming stream
is immediately created when init() is called on the inbound direction.
There should not be any visible impact.
Note that the connection's mux is purposely not set until the session
is completed so that we don't accidently run with the wrong mux. This
must not cause any issue as the xprt_done_cb function is always called
prior to using mux's recv/send functions.
2017-08-28 04:53:00 -04:00
|
|
|
const struct mux_ops *mux; /* mux layer opreations. Must be set before xprt->init() */
|
2012-10-13 08:33:58 -04:00
|
|
|
void *xprt_ctx; /* general purpose pointer, initialized to NULL */
|
MEDIUM: connection: start to introduce a mux layer between xprt and data
For HTTP/2 and QUIC, we'll need to deal with multiplexed streams inside
a connection. After quite a long brainstorming, it appears that the
connection interface to the existing streams is appropriate just like
the connection interface to the lower layers. In fact we need to have
the mux layer in the middle of the connection, between the transport
and the data layer.
A mux can exist on two directions/sides. On the inbound direction, it
instanciates new streams from incoming connections, while on the outbound
direction it muxes streams into outgoing connections. The difference is
visible on the mux->init() call : in one case, an upper context is already
known (outgoing connection), and in the other case, the upper context is
not yet known (incoming connection) and will have to be allocated by the
mux. The session doesn't have to create the new streams anymore, as this
is performed by the mux itself.
This patch introduces this and creates a pass-through mux called
"mux_pt" which is used for all new connections and which only
calls the data layer's recv,send,wake() calls. One incoming stream
is immediately created when init() is called on the inbound direction.
There should not be any visible impact.
Note that the connection's mux is purposely not set until the session
is completed so that we don't accidently run with the wrong mux. This
must not cause any issue as the xprt_done_cb function is always called
prior to using mux's recv/send functions.
2017-08-28 04:53:00 -04:00
|
|
|
void *mux_ctx; /* mux-specific context, initialized to NULL */
|
2017-09-13 12:30:23 -04:00
|
|
|
void *owner; /* pointer to the owner session for incoming connections, or NULL */
|
2018-07-17 12:46:31 -04:00
|
|
|
enum obj_type *target; /* the target to connect to (server, proxy, applet, ...) */
|
|
|
|
|
|
|
|
|
|
/* second cache line */
|
|
|
|
|
struct list send_wait_list; /* list of tasks to wake when we're ready to send */
|
|
|
|
|
struct list list; /* attach point to various connection lists (idle, ...) */
|
2013-10-24 15:10:08 -04:00
|
|
|
int xprt_st; /* transport layer state, initialized to zero */
|
MEDIUM: connection: start to introduce a mux layer between xprt and data
For HTTP/2 and QUIC, we'll need to deal with multiplexed streams inside
a connection. After quite a long brainstorming, it appears that the
connection interface to the existing streams is appropriate just like
the connection interface to the lower layers. In fact we need to have
the mux layer in the middle of the connection, between the transport
and the data layer.
A mux can exist on two directions/sides. On the inbound direction, it
instanciates new streams from incoming connections, while on the outbound
direction it muxes streams into outgoing connections. The difference is
visible on the mux->init() call : in one case, an upper context is already
known (outgoing connection), and in the other case, the upper context is
not yet known (incoming connection) and will have to be allocated by the
mux. The session doesn't have to create the new streams anymore, as this
is performed by the mux itself.
This patch introduces this and creates a pass-through mux called
"mux_pt" which is used for all new connections and which only
calls the data layer's recv,send,wake() calls. One incoming stream
is immediately created when init() is called on the inbound direction.
There should not be any visible impact.
Note that the connection's mux is purposely not set until the session
is completed so that we don't accidently run with the wrong mux. This
must not cause any issue as the xprt_done_cb function is always called
prior to using mux's recv/send functions.
2017-08-28 04:53:00 -04:00
|
|
|
int tmp_early_data; /* 1st byte of early data, if any */
|
2017-11-23 12:21:29 -05:00
|
|
|
int sent_early_data; /* Amount of early data we sent so far */
|
2017-08-24 08:31:19 -04:00
|
|
|
union conn_handle handle; /* connection handle at the socket layer */
|
2018-07-17 12:46:31 -04:00
|
|
|
const struct netns_entry *proxy_netns;
|
2017-08-28 09:46:01 -04:00
|
|
|
int (*xprt_done_cb)(struct connection *conn); /* callback to notify of end of handshake */
|
2018-07-17 12:46:31 -04:00
|
|
|
|
|
|
|
|
/* third cache line and beyond */
|
2017-10-08 05:16:46 -04:00
|
|
|
void (*destroy_cb)(struct connection *conn); /* callback to notify of imminent death of the connection */
|
2012-08-30 15:11:38 -04:00
|
|
|
struct {
|
|
|
|
|
struct sockaddr_storage from; /* client address, or address to spoof when connecting to the server */
|
2012-09-27 16:14:33 -04:00
|
|
|
struct sockaddr_storage to; /* address reached by the client, or address to connect to */
|
2012-08-30 15:11:38 -04:00
|
|
|
} addr; /* addresses of the remote side, client for producer and server for consumer */
|
2012-07-06 03:47:57 -04:00
|
|
|
};
|
|
|
|
|
|
2018-04-10 08:33:41 -04:00
|
|
|
/* PROTO token registration */
|
|
|
|
|
enum proto_proxy_mode {
|
|
|
|
|
PROTO_MODE_NONE = 0,
|
|
|
|
|
PROTO_MODE_TCP = 1 << 0, // must not be changed!
|
|
|
|
|
PROTO_MODE_HTTP = 1 << 1, // must not be changed!
|
|
|
|
|
PROTO_MODE_ANY = PROTO_MODE_TCP | PROTO_MODE_HTTP,
|
2017-09-21 13:40:52 -04:00
|
|
|
};
|
|
|
|
|
|
2018-04-10 08:33:41 -04:00
|
|
|
enum proto_proxy_side {
|
|
|
|
|
PROTO_SIDE_NONE = 0,
|
|
|
|
|
PROTO_SIDE_FE = 1, // same as PR_CAP_FE
|
|
|
|
|
PROTO_SIDE_BE = 2, // same as PR_CAP_BE
|
|
|
|
|
PROTO_SIDE_BOTH = PROTO_SIDE_FE | PROTO_SIDE_BE,
|
2018-03-06 08:43:47 -05:00
|
|
|
};
|
|
|
|
|
|
2018-04-10 08:33:41 -04:00
|
|
|
struct mux_proto_list {
|
2017-09-21 13:40:52 -04:00
|
|
|
const struct ist token; /* token name and length. Empty is catch-all */
|
2018-04-10 08:33:41 -04:00
|
|
|
enum proto_proxy_mode mode;
|
|
|
|
|
enum proto_proxy_side side;
|
2017-09-21 13:40:52 -04:00
|
|
|
const struct mux_ops *mux;
|
|
|
|
|
struct list list;
|
|
|
|
|
};
|
|
|
|
|
|
2014-05-08 23:42:08 -04:00
|
|
|
/* proxy protocol v2 definitions */
|
2014-06-14 02:28:06 -04:00
|
|
|
#define PP2_SIGNATURE "\x0D\x0A\x0D\x0A\x00\x0D\x0A\x51\x55\x49\x54\x0A"
|
|
|
|
|
#define PP2_SIGNATURE_LEN 12
|
|
|
|
|
#define PP2_HEADER_LEN 16
|
|
|
|
|
|
|
|
|
|
/* ver_cmd byte */
|
|
|
|
|
#define PP2_CMD_LOCAL 0x00
|
|
|
|
|
#define PP2_CMD_PROXY 0x01
|
|
|
|
|
#define PP2_CMD_MASK 0x0F
|
|
|
|
|
|
|
|
|
|
#define PP2_VERSION 0x20
|
|
|
|
|
#define PP2_VERSION_MASK 0xF0
|
|
|
|
|
|
|
|
|
|
/* fam byte */
|
|
|
|
|
#define PP2_TRANS_UNSPEC 0x00
|
|
|
|
|
#define PP2_TRANS_STREAM 0x01
|
|
|
|
|
#define PP2_TRANS_DGRAM 0x02
|
|
|
|
|
#define PP2_TRANS_MASK 0x0F
|
|
|
|
|
|
|
|
|
|
#define PP2_FAM_UNSPEC 0x00
|
|
|
|
|
#define PP2_FAM_INET 0x10
|
|
|
|
|
#define PP2_FAM_INET6 0x20
|
|
|
|
|
#define PP2_FAM_UNIX 0x30
|
|
|
|
|
#define PP2_FAM_MASK 0xF0
|
|
|
|
|
|
|
|
|
|
#define PP2_ADDR_LEN_UNSPEC (0)
|
|
|
|
|
#define PP2_ADDR_LEN_INET (4 + 4 + 2 + 2)
|
|
|
|
|
#define PP2_ADDR_LEN_INET6 (16 + 16 + 2 + 2)
|
|
|
|
|
#define PP2_ADDR_LEN_UNIX (108 + 108)
|
|
|
|
|
|
|
|
|
|
#define PP2_HDR_LEN_UNSPEC (PP2_HEADER_LEN + PP2_ADDR_LEN_UNSPEC)
|
|
|
|
|
#define PP2_HDR_LEN_INET (PP2_HEADER_LEN + PP2_ADDR_LEN_INET)
|
|
|
|
|
#define PP2_HDR_LEN_INET6 (PP2_HEADER_LEN + PP2_ADDR_LEN_INET6)
|
|
|
|
|
#define PP2_HDR_LEN_UNIX (PP2_HEADER_LEN + PP2_ADDR_LEN_UNIX)
|
2014-05-08 23:42:08 -04:00
|
|
|
|
|
|
|
|
struct proxy_hdr_v2 {
|
|
|
|
|
uint8_t sig[12]; /* hex 0D 0A 0D 0A 00 0D 0A 51 55 49 54 0A */
|
2014-06-14 02:28:06 -04:00
|
|
|
uint8_t ver_cmd; /* protocol version and command */
|
2014-05-08 23:42:08 -04:00
|
|
|
uint8_t fam; /* protocol family and transport */
|
|
|
|
|
uint16_t len; /* number of following bytes part of the header */
|
2014-06-14 02:28:06 -04:00
|
|
|
union {
|
|
|
|
|
struct { /* for TCP/UDP over IPv4, len = 12 */
|
|
|
|
|
uint32_t src_addr;
|
|
|
|
|
uint32_t dst_addr;
|
|
|
|
|
uint16_t src_port;
|
|
|
|
|
uint16_t dst_port;
|
|
|
|
|
} ip4;
|
|
|
|
|
struct { /* for TCP/UDP over IPv6, len = 36 */
|
|
|
|
|
uint8_t src_addr[16];
|
|
|
|
|
uint8_t dst_addr[16];
|
|
|
|
|
uint16_t src_port;
|
|
|
|
|
uint16_t dst_port;
|
|
|
|
|
} ip6;
|
|
|
|
|
struct { /* for AF_UNIX sockets, len = 216 */
|
|
|
|
|
uint8_t src_addr[108];
|
|
|
|
|
uint8_t dst_addr[108];
|
|
|
|
|
} unx;
|
|
|
|
|
} addr;
|
2014-05-08 23:42:08 -04:00
|
|
|
};
|
|
|
|
|
|
2017-10-13 06:15:28 -04:00
|
|
|
#define PP2_TYPE_ALPN 0x01
|
|
|
|
|
#define PP2_TYPE_AUTHORITY 0x02
|
|
|
|
|
#define PP2_TYPE_CRC32C 0x03
|
|
|
|
|
#define PP2_TYPE_NOOP 0x04
|
|
|
|
|
#define PP2_TYPE_SSL 0x20
|
|
|
|
|
#define PP2_SUBTYPE_SSL_VERSION 0x21
|
|
|
|
|
#define PP2_SUBTYPE_SSL_CN 0x22
|
|
|
|
|
#define PP2_SUBTYPE_SSL_CIPHER 0x23
|
|
|
|
|
#define PP2_SUBTYPE_SSL_SIG_ALG 0x24
|
|
|
|
|
#define PP2_SUBTYPE_SSL_KEY_ALG 0x25
|
|
|
|
|
#define PP2_TYPE_NETNS 0x30
|
2014-05-08 23:42:08 -04:00
|
|
|
|
2014-11-17 09:11:45 -05:00
|
|
|
#define TLV_HEADER_SIZE 3
|
2014-05-08 23:42:08 -04:00
|
|
|
struct tlv {
|
|
|
|
|
uint8_t type;
|
|
|
|
|
uint8_t length_hi;
|
|
|
|
|
uint8_t length_lo;
|
|
|
|
|
uint8_t value[0];
|
|
|
|
|
}__attribute__((packed));
|
|
|
|
|
|
|
|
|
|
struct tlv_ssl {
|
|
|
|
|
struct tlv tlv;
|
|
|
|
|
uint8_t client;
|
|
|
|
|
uint32_t verify;
|
|
|
|
|
uint8_t sub_tlv[0];
|
|
|
|
|
}__attribute__((packed));
|
|
|
|
|
|
2014-07-30 10:39:13 -04:00
|
|
|
#define PP2_CLIENT_SSL 0x01
|
|
|
|
|
#define PP2_CLIENT_CERT_CONN 0x02
|
|
|
|
|
#define PP2_CLIENT_CERT_SESS 0x04
|
2014-05-08 23:42:08 -04:00
|
|
|
|
2017-04-05 16:24:59 -04:00
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Linux seems to be able to send 253 fds per sendmsg(), not sure
|
|
|
|
|
* about the other OSes.
|
|
|
|
|
*/
|
|
|
|
|
/* Max number of file descriptors we send in one sendmsg() */
|
|
|
|
|
#define MAX_SEND_FD 253
|
|
|
|
|
|
2012-07-06 03:47:57 -04:00
|
|
|
#endif /* _TYPES_CONNECTION_H */
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Local variables:
|
|
|
|
|
* c-indent-level: 8
|
|
|
|
|
* c-basic-offset: 8
|
|
|
|
|
* End:
|
|
|
|
|
*/
|