mirror of
https://github.com/apache/httpd.git
synced 2025-08-07 04:02:58 +03:00
mod_proxy_http: follow up to r1869216.
Let's call stream_reqbody() for all rb_methods, no RB_SPOOL_CL special case. This both simplifies code and allows to keep EOS into the input_brigade until it's sent, and thus detect whether we already fetched the whole body if/when proxy_http_handler() re-enters for different balancer members. git-svn-id: https://svn.apache.org/repos/asf/httpd/httpd/trunk@1869222 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
@@ -82,6 +82,30 @@ AP_DECLARE(void) ap_get_mime_headers(request_rec *r);
|
||||
AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r,
|
||||
apr_bucket_brigade *bb);
|
||||
|
||||
/**
|
||||
* @struct ap_mime_headers_ctx
|
||||
* @brief Context for ap_get_mime_headers_ex()
|
||||
*/
|
||||
typedef struct ap_mime_headers_ctx ap_mime_headers_ctx_t;
|
||||
struct ap_mime_headers_ctx {
|
||||
int limit_req_fields;
|
||||
int limit_req_fieldsize;
|
||||
apr_table_t *headers;
|
||||
apr_table_t *notes;
|
||||
apr_bucket_brigade *bb;
|
||||
unsigned int strict:1,
|
||||
compress:1;
|
||||
};
|
||||
|
||||
/**
|
||||
* Generic version of ap_get_mime_headers_core() that takes a filter as
|
||||
* parameter and options regarding limits to apply.
|
||||
* @param f The filter to read from
|
||||
* @param ctx The context/options (@see ap_mime_headers_ctx)
|
||||
*/
|
||||
AP_DECLARE(int) ap_get_mime_headers_ex(request_rec *r, ap_filter_t *f,
|
||||
ap_mime_headers_ctx_t *ctx);
|
||||
|
||||
/* Finish up stuff after a request */
|
||||
|
||||
/**
|
||||
|
@@ -303,16 +303,18 @@ static int stream_reqbody_read(proxy_http_req_t *req, apr_bucket_brigade *bb,
|
||||
return OK;
|
||||
}
|
||||
|
||||
static int stream_reqbody(proxy_http_req_t *req, rb_methods rb_method)
|
||||
static int stream_reqbody(proxy_http_req_t *req)
|
||||
{
|
||||
request_rec *r = req->r;
|
||||
int seen_eos = 0, rv = OK;
|
||||
apr_size_t hdr_len;
|
||||
char chunk_hdr[20]; /* must be here due to transient bucket. */
|
||||
conn_rec *origin = req->origin;
|
||||
proxy_conn_rec *p_conn = req->backend;
|
||||
apr_bucket_alloc_t *bucket_alloc = req->bucket_alloc;
|
||||
apr_bucket_brigade *header_brigade = req->header_brigade;
|
||||
apr_bucket_brigade *input_brigade = req->input_brigade;
|
||||
rb_methods rb_method = req->rb_method;
|
||||
apr_off_t bytes, bytes_streamed = 0;
|
||||
apr_bucket *e;
|
||||
|
||||
@@ -326,7 +328,7 @@ static int stream_reqbody(proxy_http_req_t *req, rb_methods rb_method)
|
||||
}
|
||||
|
||||
if (!APR_BRIGADE_EMPTY(input_brigade)) {
|
||||
/* If this brigade contains EOS, either stop or remove it. */
|
||||
/* If this brigade contains EOS, remove it and be done. */
|
||||
if (APR_BUCKET_IS_EOS(APR_BRIGADE_LAST(input_brigade))) {
|
||||
seen_eos = 1;
|
||||
|
||||
@@ -368,7 +370,8 @@ static int stream_reqbody(proxy_http_req_t *req, rb_methods rb_method)
|
||||
APR_BRIGADE_INSERT_TAIL(input_brigade, e);
|
||||
}
|
||||
}
|
||||
else if (bytes_streamed > req->cl_val) {
|
||||
else if (rb_method == RB_STREAM_CL
|
||||
&& bytes_streamed > req->cl_val) {
|
||||
/* C-L < bytes streamed?!?
|
||||
* We will error out after the body is completely
|
||||
* consumed, but we can't stream more bytes at the
|
||||
@@ -400,7 +403,7 @@ static int stream_reqbody(proxy_http_req_t *req, rb_methods rb_method)
|
||||
APR_BRIGADE_PREPEND(input_brigade, header_brigade);
|
||||
|
||||
/* Flush here on EOS because we won't stream_reqbody_read() again */
|
||||
rv = ap_proxy_pass_brigade(bucket_alloc, r, p_conn, req->origin,
|
||||
rv = ap_proxy_pass_brigade(bucket_alloc, r, p_conn, origin,
|
||||
input_brigade, seen_eos);
|
||||
if (rv != OK) {
|
||||
return rv;
|
||||
@@ -462,10 +465,6 @@ static int spool_reqbody_cl(proxy_http_req_t *req, apr_off_t *bytes_spooled)
|
||||
/* If this brigade contains EOS, either stop or remove it. */
|
||||
if (APR_BUCKET_IS_EOS(APR_BRIGADE_LAST(input_brigade))) {
|
||||
seen_eos = 1;
|
||||
|
||||
/* We can't pass this EOS to the output_filters. */
|
||||
e = APR_BRIGADE_LAST(input_brigade);
|
||||
apr_bucket_delete(e);
|
||||
}
|
||||
|
||||
apr_brigade_length(input_brigade, 1, &bytes);
|
||||
@@ -859,35 +858,21 @@ static int ap_proxy_http_request(proxy_http_req_t *req)
|
||||
{
|
||||
int rv;
|
||||
request_rec *r = req->r;
|
||||
apr_bucket_alloc_t *bucket_alloc = req->bucket_alloc;
|
||||
apr_bucket_brigade *header_brigade = req->header_brigade;
|
||||
apr_bucket_brigade *input_brigade = req->input_brigade;
|
||||
|
||||
/* send the request header/body, if any. */
|
||||
switch (req->rb_method) {
|
||||
case RB_SPOOL_CL:
|
||||
case RB_STREAM_CL:
|
||||
case RB_STREAM_CHUNKED:
|
||||
if (req->do_100_continue) {
|
||||
rv = ap_proxy_pass_brigade(bucket_alloc, r, req->backend,
|
||||
req->origin, header_brigade, 1);
|
||||
rv = ap_proxy_pass_brigade(req->bucket_alloc, r, req->backend,
|
||||
req->origin, req->header_brigade, 1);
|
||||
}
|
||||
else {
|
||||
rv = stream_reqbody(req, req->rb_method);
|
||||
rv = stream_reqbody(req);
|
||||
}
|
||||
break;
|
||||
|
||||
case RB_SPOOL_CL:
|
||||
/* Prefetch has built the header and spooled the whole body;
|
||||
* if we don't expect 100-continue we can flush both all at once,
|
||||
* otherwise flush the header only.
|
||||
*/
|
||||
if (!req->do_100_continue) {
|
||||
APR_BRIGADE_CONCAT(header_brigade, input_brigade);
|
||||
}
|
||||
rv = ap_proxy_pass_brigade(bucket_alloc, r, req->backend,
|
||||
req->origin, header_brigade, 1);
|
||||
break;
|
||||
|
||||
default:
|
||||
/* shouldn't be possible */
|
||||
rv = HTTP_INTERNAL_SERVER_ERROR;
|
||||
@@ -1590,15 +1575,10 @@ int ap_proxy_http_process_response(proxy_http_req_t *req)
|
||||
|
||||
/* Send the request body (fully). */
|
||||
switch(req->rb_method) {
|
||||
case RB_SPOOL_CL:
|
||||
case RB_STREAM_CL:
|
||||
case RB_STREAM_CHUNKED:
|
||||
status = stream_reqbody(req, req->rb_method);
|
||||
break;
|
||||
case RB_SPOOL_CL:
|
||||
/* Prefetch has spooled the whole body, flush it. */
|
||||
status = ap_proxy_pass_brigade(req->bucket_alloc, r,
|
||||
backend, origin,
|
||||
req->input_brigade, 1);
|
||||
status = stream_reqbody(req);
|
||||
break;
|
||||
default:
|
||||
/* Shouldn't happen */
|
||||
|
@@ -294,6 +294,223 @@ static int proxy_wstunnel_canon(request_rec *r, char *url)
|
||||
return OK;
|
||||
}
|
||||
|
||||
static request_rec *make_resp(conn_rec *c, request_rec *r)
|
||||
{
|
||||
apr_pool_t *pool;
|
||||
request_rec *rp;
|
||||
|
||||
apr_pool_create(&pool, c->pool);
|
||||
|
||||
rp = apr_pcalloc(pool, sizeof(*r));
|
||||
|
||||
rp->pool = pool;
|
||||
rp->status = HTTP_OK;
|
||||
|
||||
rp->headers_in = apr_table_make(pool, 50);
|
||||
rp->trailers_in = apr_table_make(pool, 5);
|
||||
|
||||
rp->subprocess_env = apr_table_make(pool, 50);
|
||||
rp->headers_out = apr_table_make(pool, 12);
|
||||
rp->trailers_out = apr_table_make(pool, 5);
|
||||
rp->err_headers_out = apr_table_make(pool, 5);
|
||||
rp->notes = apr_table_make(pool, 5);
|
||||
|
||||
rp->server = r->server;
|
||||
rp->log = r->log;
|
||||
rp->proxyreq = r->proxyreq;
|
||||
rp->request_time = r->request_time;
|
||||
rp->connection = c;
|
||||
rp->output_filters = c->output_filters;
|
||||
rp->input_filters = c->input_filters;
|
||||
rp->proto_output_filters = c->output_filters;
|
||||
rp->proto_input_filters = c->input_filters;
|
||||
rp->useragent_ip = c->client_ip;
|
||||
rp->useragent_addr = c->client_addr;
|
||||
|
||||
rp->request_config = ap_create_request_config(pool);
|
||||
|
||||
return rp;
|
||||
}
|
||||
|
||||
static int proxy_wstunnel_handle_http_response(request_rec *r,
|
||||
proxy_conn_rec *backend,
|
||||
proxy_server_conf *sconf,
|
||||
apr_bucket_brigade *bb)
|
||||
{
|
||||
conn_rec *origin = backend->connection;
|
||||
proxy_worker *worker = backend->worker;
|
||||
char fixed_buffer[HUGE_STRING_LEN];
|
||||
char *buffer = fixed_buffer;
|
||||
int size = HUGE_STRING_LEN;
|
||||
ap_mime_headers_ctx_t ctx;
|
||||
request_rec *resp;
|
||||
apr_status_t rv;
|
||||
apr_size_t len;
|
||||
int rc;
|
||||
|
||||
/* Only use dynamically sized buffer if user specifies ResponseFieldSize */
|
||||
if (worker->s->response_field_size_set) {
|
||||
size = worker->s->response_field_size;
|
||||
if (size > HUGE_STRING_LEN) {
|
||||
buffer = apr_palloc(r->pool, size);
|
||||
}
|
||||
}
|
||||
|
||||
resp = make_resp(origin, r);
|
||||
|
||||
rv = ap_rgetline(&buffer, size, &len, resp, 0, bb);
|
||||
apr_brigade_cleanup(bb);
|
||||
|
||||
if (rv != APR_SUCCESS || !apr_date_checkmask(buffer, "HTTP/#.# ### *")) {
|
||||
return HTTP_BAD_GATEWAY;
|
||||
}
|
||||
|
||||
r->status = atoi(&buffer[9]);
|
||||
if (!ap_is_HTTP_VALID_RESPONSE(r->status)) {
|
||||
return HTTP_BAD_GATEWAY;
|
||||
}
|
||||
r->status_line = apr_pstrdup(r->pool, &buffer[9]);
|
||||
|
||||
memset(&ctx, 0, sizeof(ctx));
|
||||
ctx.bb = bb;
|
||||
ctx.headers = r->headers_out;
|
||||
ctx.limit_req_fieldsize = size;
|
||||
rc = ap_get_mime_headers_ex(r, origin->input_filters, &ctx);
|
||||
apr_brigade_cleanup(bb);
|
||||
if (rc != OK) {
|
||||
r->status = HTTP_OK;
|
||||
r->status_line = NULL;
|
||||
apr_table_clear(r->headers_out);
|
||||
return rc;
|
||||
}
|
||||
|
||||
#if 0
|
||||
if (r->status != HTTP_SWITCHING_PROTOCOLS) {
|
||||
conn_rec *c = r->connection;
|
||||
apr_read_type_e block = APR_NONBLOCK_READ;
|
||||
apr_bucket_brigade *pass_bb = apr_brigade_create(r->pool,
|
||||
c->bucket_alloc);
|
||||
int finish = 0;
|
||||
|
||||
r->sent_bodyct = 1;
|
||||
do {
|
||||
apr_bucket *e;
|
||||
apr_off_t readbytes = 0;
|
||||
|
||||
rv = ap_get_brigade(origin->input_filters, bb,
|
||||
AP_MODE_READBYTES, block,
|
||||
sconf->io_buffer_size);
|
||||
|
||||
/* ap_get_brigade will return success with an empty brigade
|
||||
* for a non-blocking read which would block: */
|
||||
if (block == APR_NONBLOCK_READ
|
||||
&& (APR_STATUS_IS_EAGAIN(rv)
|
||||
|| (rv == APR_SUCCESS && APR_BRIGADE_EMPTY(bb)))) {
|
||||
/* flush to the client and switch to blocking mode */
|
||||
e = apr_bucket_flush_create(c->bucket_alloc);
|
||||
APR_BRIGADE_INSERT_TAIL(bb, e);
|
||||
if (ap_pass_brigade(r->output_filters, bb)
|
||||
|| c->aborted) {
|
||||
finish = 1;
|
||||
rc = DONE;
|
||||
}
|
||||
apr_brigade_cleanup(bb);
|
||||
block = APR_BLOCK_READ;
|
||||
continue;
|
||||
}
|
||||
if (rv == APR_EOF) {
|
||||
break;
|
||||
}
|
||||
if (rv != APR_SUCCESS) {
|
||||
if (rv == APR_ENOSPC) {
|
||||
ap_log_rerror(APLOG_MARK, APLOG_ERR, rv, r, APLOGNO(02475)
|
||||
"Response chunk/line was too large to parse");
|
||||
}
|
||||
else if (rv == APR_ENOTIMPL) {
|
||||
ap_log_rerror(APLOG_MARK, APLOG_ERR, rv, r, APLOGNO(02476)
|
||||
"Response Transfer-Encoding was not recognised");
|
||||
}
|
||||
else {
|
||||
ap_log_rerror(APLOG_MARK, APLOG_ERR, rv, r, APLOGNO(01110)
|
||||
"Network error reading response");
|
||||
}
|
||||
|
||||
/* In this case, we are in real trouble because
|
||||
* our backend bailed on us. Given we're half way
|
||||
* through a response, our only option is to
|
||||
* disconnect the client too.
|
||||
*/
|
||||
e = ap_bucket_error_create(HTTP_BAD_GATEWAY, NULL,
|
||||
r->pool, c->bucket_alloc);
|
||||
APR_BRIGADE_INSERT_TAIL(bb, e);
|
||||
e = ap_bucket_eoc_create(c->bucket_alloc);
|
||||
APR_BRIGADE_INSERT_TAIL(bb, e);
|
||||
ap_pass_brigade(r->output_filters, bb);
|
||||
apr_brigade_cleanup(bb);
|
||||
rc = DONE;
|
||||
break;
|
||||
}
|
||||
|
||||
/* next time try a non-blocking read */
|
||||
block = APR_NONBLOCK_READ;
|
||||
|
||||
if (!apr_is_empty_table(resp->trailers_in)) {
|
||||
apr_table_do(add_trailers, r->trailers_out,
|
||||
resp->trailers_in, NULL);
|
||||
apr_table_clear(resp->trailers_in);
|
||||
}
|
||||
|
||||
apr_brigade_length(bb, 0, &readbytes);
|
||||
backend->worker->s->read += readbytes;
|
||||
|
||||
/* sanity check */
|
||||
if (APR_BRIGADE_EMPTY(bb)) {
|
||||
break;
|
||||
}
|
||||
|
||||
/* Switch the allocator lifetime of the buckets */
|
||||
ap_proxy_buckets_lifetime_transform(r, bb, pass_bb);
|
||||
|
||||
/* found the last brigade? */
|
||||
if (APR_BUCKET_IS_EOS(APR_BRIGADE_LAST(pass_bb))) {
|
||||
/* the brigade may contain transient buckets that contain
|
||||
* data that lives only as long as the backend connection.
|
||||
* Force a setaside so these transient buckets become heap
|
||||
* buckets that live as long as the request.
|
||||
*/
|
||||
for (e = APR_BRIGADE_FIRST(pass_bb);
|
||||
e != APR_BRIGADE_SENTINEL(pass_bb);
|
||||
e = APR_BUCKET_NEXT(e)) {
|
||||
apr_bucket_setaside(e, r->pool);
|
||||
}
|
||||
/* finally it is safe to clean up the brigade from the
|
||||
* connection pool, as we have forced a setaside on all
|
||||
* buckets.
|
||||
*/
|
||||
apr_brigade_cleanup(bb);
|
||||
finish = 1;
|
||||
}
|
||||
|
||||
/* try send what we read */
|
||||
if (ap_pass_brigade(r->output_filters, pass_bb) != APR_SUCCESS
|
||||
|| c->aborted) {
|
||||
/* Ack! Phbtt! Die! User aborted! */
|
||||
finish = 1;
|
||||
rc = DONE;
|
||||
}
|
||||
|
||||
/* make sure we always clean up after ourselves */
|
||||
apr_brigade_cleanup(pass_bb);
|
||||
apr_brigade_cleanup(bb);
|
||||
} while (!finish);
|
||||
|
||||
return rc;
|
||||
}
|
||||
#endif
|
||||
|
||||
return DECLINED;
|
||||
}
|
||||
|
||||
/*
|
||||
* process the request and write the response.
|
||||
*/
|
||||
@@ -318,14 +535,13 @@ static int proxy_wstunnel_request(apr_pool_t *p, request_rec *r,
|
||||
apr_bucket_brigade *bb = apr_brigade_create(p, c->bucket_alloc);
|
||||
apr_socket_t *client_socket = ap_get_conn_socket(c);
|
||||
ws_baton_t *baton = apr_pcalloc(r->pool, sizeof(ws_baton_t));
|
||||
int status;
|
||||
proxyws_dir_conf *dconf = ap_get_module_config(r->per_dir_config, &proxy_wstunnel_module);
|
||||
const char *upgrade_method = *worker->s->upgrade ? worker->s->upgrade : "WebSocket";
|
||||
|
||||
header_brigade = apr_brigade_create(p, backconn->bucket_alloc);
|
||||
int status;
|
||||
|
||||
ap_log_rerror(APLOG_MARK, APLOG_TRACE2, 0, r, "sending request");
|
||||
|
||||
header_brigade = apr_brigade_create(p, backconn->bucket_alloc);
|
||||
rv = ap_proxy_create_hdrbrgd(p, header_brigade, r, conn,
|
||||
worker, conf, uri, url, server_portstr,
|
||||
&old_cl_val, &old_te_val);
|
||||
@@ -334,13 +550,19 @@ static int proxy_wstunnel_request(apr_pool_t *p, request_rec *r,
|
||||
}
|
||||
|
||||
if (ap_cstr_casecmp(upgrade_method, "NONE") == 0) {
|
||||
buf = apr_pstrdup(p, "Upgrade: WebSocket" CRLF "Connection: Upgrade" CRLF CRLF);
|
||||
buf = apr_pstrdup(p, "Upgrade: WebSocket" CRLF
|
||||
"Connection: Upgrade" CRLF
|
||||
CRLF);
|
||||
} else if (ap_cstr_casecmp(upgrade_method, "ANY") == 0) {
|
||||
const char *upgrade;
|
||||
upgrade = apr_table_get(r->headers_in, "Upgrade");
|
||||
buf = apr_pstrcat(p, "Upgrade: ", upgrade, CRLF "Connection: Upgrade" CRLF CRLF, NULL);
|
||||
buf = apr_pstrcat(p, "Upgrade: ", upgrade, CRLF
|
||||
"Connection: Upgrade" CRLF
|
||||
CRLF, NULL);
|
||||
} else {
|
||||
buf = apr_pstrcat(p, "Upgrade: ", upgrade_method, CRLF "Connection: Upgrade" CRLF CRLF, NULL);
|
||||
buf = apr_pstrcat(p, "Upgrade: ", upgrade_method, CRLF
|
||||
"Connection: Upgrade" CRLF
|
||||
CRLF, NULL);
|
||||
}
|
||||
ap_xlate_proto_to_ascii(buf, strlen(buf));
|
||||
e = apr_bucket_pool_create(buf, strlen(buf), p, c->bucket_alloc);
|
||||
@@ -350,7 +572,9 @@ static int proxy_wstunnel_request(apr_pool_t *p, request_rec *r,
|
||||
header_brigade, 1)) != OK)
|
||||
return rv;
|
||||
|
||||
apr_brigade_cleanup(header_brigade);
|
||||
if ((rv = proxy_wstunnel_handle_http_response(r, conn, conf,
|
||||
header_brigade)) != DECLINED)
|
||||
return rv;
|
||||
|
||||
ap_log_rerror(APLOG_MARK, APLOG_TRACE2, 0, r, "setting up poll()");
|
||||
|
||||
@@ -466,7 +690,7 @@ static int proxy_wstunnel_handler(request_rec *r, proxy_worker *worker,
|
||||
char *locurl = url;
|
||||
apr_uri_t *uri;
|
||||
int is_ssl = 0;
|
||||
const char *upgrade_method = *worker->s->upgrade ? worker->s->upgrade : "WebSocket";
|
||||
const char *upgrade_method;
|
||||
|
||||
if (ap_cstr_casecmpn(url, "wss:", 4) == 0) {
|
||||
scheme = "WSS";
|
||||
@@ -480,6 +704,7 @@ static int proxy_wstunnel_handler(request_rec *r, proxy_worker *worker,
|
||||
return DECLINED;
|
||||
}
|
||||
|
||||
upgrade_method = *worker->s->upgrade ? worker->s->upgrade : "WebSocket";
|
||||
if (ap_cstr_casecmp(upgrade_method, "NONE") != 0) {
|
||||
const char *upgrade;
|
||||
upgrade = apr_table_get(r->headers_in, "Upgrade");
|
||||
|
@@ -554,22 +554,7 @@ AP_DECLARE(apr_status_t) ap_rgetline(char **s, apr_size_t n,
|
||||
apr_size_t *read, request_rec *r,
|
||||
int flags, apr_bucket_brigade *bb)
|
||||
{
|
||||
apr_status_t rv;
|
||||
|
||||
rv = ap_fgetline_core(s, n, read, r->proto_input_filters, flags,
|
||||
bb, r->pool);
|
||||
#if APR_CHARSET_EBCDIC
|
||||
/* On EBCDIC boxes, each complete http protocol input line needs to be
|
||||
* translated into the code page used by the compiler. Since
|
||||
* ap_fgetline_core uses recursion, we do the translation in a wrapper
|
||||
* function to ensure that each input character gets translated only once.
|
||||
*/
|
||||
if (*read) {
|
||||
ap_xlate_proto_from_ascii(*s, *read);
|
||||
}
|
||||
#endif
|
||||
|
||||
return rv;
|
||||
return ap_fgetline(s, n, read, r->proto_input_filters, flags, bb, r->pool);
|
||||
}
|
||||
|
||||
AP_DECLARE(int) ap_getline(char *s, int n, request_rec *r, int flags)
|
||||
@@ -1004,23 +989,30 @@ rrl_failed:
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int table_do_fn_check_lengths(void *r_, const char *key,
|
||||
struct table_do_fn_check_lengths_baton {
|
||||
request_rec *r;
|
||||
ap_mime_headers_ctx_t *ctx;
|
||||
};
|
||||
static int table_do_fn_check_lengths(void *arg, const char *key,
|
||||
const char *value)
|
||||
{
|
||||
request_rec *r = r_;
|
||||
if (value == NULL || r->server->limit_req_fieldsize >= strlen(value) )
|
||||
struct table_do_fn_check_lengths_baton *baton = arg;
|
||||
|
||||
if (value == NULL || baton->ctx->limit_req_fieldsize >= strlen(value))
|
||||
return 1;
|
||||
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
apr_table_setn(r->notes, "error-notes",
|
||||
"Size of a request header field exceeds server limit.");
|
||||
ap_log_rerror(APLOG_MARK, APLOG_INFO, 0, r, APLOGNO(00560) "Request "
|
||||
"header exceeds LimitRequestFieldSize after merging: %.*s",
|
||||
if (baton->ctx->notes) {
|
||||
apr_table_setn(baton->ctx->notes, "error-notes",
|
||||
"Size of a header field exceeds limit.");
|
||||
}
|
||||
ap_log_rerror(APLOG_MARK, APLOG_INFO, 0, baton->r, APLOGNO(00560)
|
||||
"Header exceeds size limit after merging: %.*s",
|
||||
field_name_len(key), key);
|
||||
return 0;
|
||||
}
|
||||
|
||||
AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb)
|
||||
AP_DECLARE(int) ap_get_mime_headers_ex(request_rec *r, ap_filter_t *f,
|
||||
ap_mime_headers_ctx_t *ctx)
|
||||
{
|
||||
char *last_field = NULL;
|
||||
apr_size_t last_len = 0;
|
||||
@@ -1030,44 +1022,53 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
apr_size_t len;
|
||||
int fields_read = 0;
|
||||
char *tmp_field;
|
||||
core_server_config *conf = ap_get_core_module_config(r->server->module_config);
|
||||
int strict = (conf->http_conformance != AP_HTTP_CONFORMANCE_UNSAFE);
|
||||
apr_bucket_brigade *bb = ctx->bb;
|
||||
int rc = OK;
|
||||
|
||||
if (bb == NULL) {
|
||||
bb = apr_brigade_create(r->pool, r->connection->bucket_alloc);
|
||||
}
|
||||
if (ctx->headers == NULL) {
|
||||
ctx->headers = apr_table_make(r->pool, 25);
|
||||
}
|
||||
if (ctx->limit_req_fieldsize <= 0) {
|
||||
ctx->limit_req_fieldsize = HUGE_STRING_LEN;
|
||||
}
|
||||
|
||||
/*
|
||||
* Read header lines until we get the empty separator line, a read error,
|
||||
* the connection closes (EOF), reach the server limit, or we timeout.
|
||||
* the connection closes (EOF), reach the size limit(s), or we timeout.
|
||||
*/
|
||||
while(1) {
|
||||
apr_status_t rv;
|
||||
|
||||
field = NULL;
|
||||
rv = ap_rgetline(&field, r->server->limit_req_fieldsize + 2,
|
||||
&len, r, strict ? AP_GETLINE_CRLF : 0, bb);
|
||||
rv = ap_fgetline(&field, ctx->limit_req_fieldsize + 2, &len, f,
|
||||
ctx->strict ? AP_GETLINE_CRLF : 0, bb, r->pool);
|
||||
|
||||
if (rv != APR_SUCCESS) {
|
||||
if (APR_STATUS_IS_TIMEUP(rv)) {
|
||||
r->status = HTTP_REQUEST_TIME_OUT;
|
||||
}
|
||||
else {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
}
|
||||
|
||||
/* ap_rgetline returns APR_ENOSPC if it fills up the buffer before
|
||||
* finding the end-of-line. This is only going to happen if it
|
||||
* exceeds the configured limit for a field size.
|
||||
*/
|
||||
if (rv == APR_ENOSPC) {
|
||||
apr_table_setn(r->notes, "error-notes",
|
||||
"Size of a request header field "
|
||||
"exceeds server limit.");
|
||||
if (ctx->notes) {
|
||||
apr_table_setn(ctx->notes, "error-notes",
|
||||
"Size of a header field exceeds limit.");
|
||||
}
|
||||
ap_log_rerror(APLOG_MARK, APLOG_INFO, 0, r, APLOGNO(00561)
|
||||
"Request header exceeds LimitRequestFieldSize%s"
|
||||
"%.*s",
|
||||
"Header exceeds size limit%s%.*s",
|
||||
(field && *field) ? ": " : "",
|
||||
(field) ? field_name_len(field) : 0,
|
||||
(field) ? field : "");
|
||||
}
|
||||
return;
|
||||
if (APR_STATUS_IS_TIMEUP(rv)) {
|
||||
rc = HTTP_REQUEST_TIME_OUT;
|
||||
}
|
||||
else {
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
}
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
/* For all header values, and all obs-fold lines, the presence of
|
||||
@@ -1087,18 +1088,18 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
apr_size_t fold_len;
|
||||
|
||||
if (last_field == NULL) {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(03442)
|
||||
"Line folding encountered before first"
|
||||
" header line");
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
if (field[1] == '\0') {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(03443)
|
||||
"Empty folded line encountered");
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
/* Leading whitespace on an obs-fold line can be
|
||||
@@ -1115,19 +1116,19 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
*/
|
||||
fold_len = last_len + len + 1; /* trailing null */
|
||||
|
||||
if (fold_len >= (apr_size_t)(r->server->limit_req_fieldsize)) {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
if (fold_len >= (apr_size_t)ctx->limit_req_fieldsize) {
|
||||
if (ctx->notes) {
|
||||
/* report what we have accumulated so far before the
|
||||
* overflow (last_field) as the field with the problem
|
||||
*/
|
||||
apr_table_setn(r->notes, "error-notes",
|
||||
"Size of a request header field "
|
||||
"exceeds server limit.");
|
||||
apr_table_setn(ctx->notes, "error-notes",
|
||||
"Size of a header field exceeds limit.");
|
||||
}
|
||||
ap_log_rerror(APLOG_MARK, APLOG_INFO, 0, r, APLOGNO(00562)
|
||||
"Request header exceeds LimitRequestFieldSize "
|
||||
"after folding: %.*s",
|
||||
"Header exceeds size limit after folding: %.*s",
|
||||
field_name_len(last_field), last_field);
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
if (fold_len > alloc_len) {
|
||||
@@ -1157,46 +1158,47 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
* most recently read input line).
|
||||
*/
|
||||
|
||||
if (r->server->limit_req_fields
|
||||
&& (++fields_read > r->server->limit_req_fields)) {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
apr_table_setn(r->notes, "error-notes",
|
||||
"The number of request header fields "
|
||||
"exceeds this server's limit.");
|
||||
if (ctx->limit_req_fields
|
||||
&& (++fields_read > ctx->limit_req_fields)) {
|
||||
if (ctx->notes) {
|
||||
apr_table_setn(ctx->notes, "error-notes",
|
||||
"The number of header fields "
|
||||
"exceeds the limit.");
|
||||
}
|
||||
ap_log_rerror(APLOG_MARK, APLOG_INFO, 0, r, APLOGNO(00563)
|
||||
"Number of request headers exceeds "
|
||||
"LimitRequestFields");
|
||||
return;
|
||||
"Number of headers exceeds the limit");
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
if (!strict)
|
||||
if (!ctx->strict)
|
||||
{
|
||||
/* Not Strict ('Unsafe' mode), using the legacy parser */
|
||||
|
||||
if (!(value = strchr(last_field, ':'))) { /* Find ':' or */
|
||||
r->status = HTTP_BAD_REQUEST; /* abort bad request */
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(00564)
|
||||
"Request header field is missing ':' "
|
||||
"Header field is missing ':' "
|
||||
"separator: %.*s", (int)LOG_NAME_MAX_LEN,
|
||||
last_field);
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
if (value == last_field) {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(03453)
|
||||
"Request header field name was empty");
|
||||
return;
|
||||
"Header field name was empty");
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
*value++ = '\0'; /* NUL-terminate at colon */
|
||||
|
||||
if (strpbrk(last_field, "\t\n\v\f\r ")) {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(03452)
|
||||
"Request header field name presented"
|
||||
"Header field name presented"
|
||||
" invalid whitespace");
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
while (*value == ' ' || *value == '\t') {
|
||||
@@ -1204,11 +1206,11 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
}
|
||||
|
||||
if (strpbrk(value, "\n\v\f\r")) {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(03451)
|
||||
"Request header field value presented"
|
||||
"Header field value presented"
|
||||
" bad whitespace");
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
}
|
||||
else /* Using strict RFC7230 parsing */
|
||||
@@ -1216,11 +1218,11 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
/* Ensure valid token chars before ':' per RFC 7230 3.2.4 */
|
||||
value = (char *)ap_scan_http_token(last_field);
|
||||
if ((value == last_field) || *value != ':') {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(02426)
|
||||
"Request header field name is malformed: "
|
||||
"%.*s", (int)LOG_NAME_MAX_LEN, last_field);
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
|
||||
*value++ = '\0'; /* NUL-terminate last_field name at ':' */
|
||||
@@ -1238,15 +1240,15 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
* for specific header handler logic later in the cycle
|
||||
*/
|
||||
if (*tmp_field != '\0') {
|
||||
r->status = HTTP_BAD_REQUEST;
|
||||
ap_log_rerror(APLOG_MARK, APLOG_DEBUG, 0, r, APLOGNO(02427)
|
||||
"Request header value is malformed: "
|
||||
"%.*s", (int)LOG_NAME_MAX_LEN, value);
|
||||
return;
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
}
|
||||
|
||||
apr_table_addn(r->headers_in, last_field, value);
|
||||
apr_table_addn(ctx->headers, last_field, value);
|
||||
|
||||
/* This last_field header is now stored in headers_in,
|
||||
* resume processing of the current input line.
|
||||
@@ -1260,7 +1262,7 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
|
||||
/* Keep track of this new header line so that we can extend it across
|
||||
* any obs-fold or parse it on the next loop iteration. We referenced
|
||||
* our previously allocated buffer in r->headers_in,
|
||||
* our previously allocated buffer in ctx->headers,
|
||||
* so allocate a fresh buffer if required.
|
||||
*/
|
||||
alloc_len = 0;
|
||||
@@ -1271,18 +1273,51 @@ AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb
|
||||
/* Combine multiple message-header fields with the same
|
||||
* field-name, following RFC 2616, 4.2.
|
||||
*/
|
||||
apr_table_compress(r->headers_in, APR_OVERLAP_TABLES_MERGE);
|
||||
if (ctx->compress) {
|
||||
apr_table_compress(ctx->headers, APR_OVERLAP_TABLES_MERGE);
|
||||
}
|
||||
|
||||
/* enforce LimitRequestFieldSize for merged headers */
|
||||
apr_table_do(table_do_fn_check_lengths, r, r->headers_in, NULL);
|
||||
/* Enforce limit for merged headers */
|
||||
{
|
||||
struct table_do_fn_check_lengths_baton baton = { r, ctx };
|
||||
if (!apr_table_do(table_do_fn_check_lengths, &baton,
|
||||
ctx->headers, NULL)) {
|
||||
rc = HTTP_BAD_REQUEST;
|
||||
goto cleanup;
|
||||
}
|
||||
}
|
||||
|
||||
cleanup:
|
||||
if (bb != ctx->bb) {
|
||||
apr_brigade_destroy(bb);
|
||||
}
|
||||
return rc;
|
||||
}
|
||||
|
||||
AP_DECLARE(void) ap_get_mime_headers_core(request_rec *r, apr_bucket_brigade *bb)
|
||||
{
|
||||
core_server_config *conf = ap_get_core_module_config(r->server->module_config);
|
||||
ap_mime_headers_ctx_t ctx;
|
||||
int status;
|
||||
|
||||
memset(&ctx, 0, sizeof(ctx));
|
||||
ctx.strict = (conf->http_conformance != AP_HTTP_CONFORMANCE_UNSAFE);
|
||||
ctx.limit_req_fields = r->server->limit_req_fields;
|
||||
ctx.limit_req_fieldsize = r->server->limit_req_fieldsize;
|
||||
ctx.headers = r->headers_in;
|
||||
ctx.notes = r->notes;
|
||||
ctx.compress = 1;
|
||||
ctx.bb = bb;
|
||||
|
||||
status = ap_get_mime_headers_ex(r, r->proto_input_filters, &ctx);
|
||||
if (status != OK) {
|
||||
r->status = status;
|
||||
}
|
||||
}
|
||||
|
||||
AP_DECLARE(void) ap_get_mime_headers(request_rec *r)
|
||||
{
|
||||
apr_bucket_brigade *tmp_bb;
|
||||
tmp_bb = apr_brigade_create(r->pool, r->connection->bucket_alloc);
|
||||
ap_get_mime_headers_core(r, tmp_bb);
|
||||
apr_brigade_destroy(tmp_bb);
|
||||
ap_get_mime_headers_core(r, NULL);
|
||||
}
|
||||
|
||||
AP_DECLARE(request_rec *) ap_create_request(conn_rec *conn)
|
||||
|
Reference in New Issue
Block a user