CLEANUP: assorted typo fixes in the code, commits and doc
This commit is contained in:
parent
bd477d5f51
commit
27a6353ceb
@ -212,7 +212,7 @@ da_status_t da_atlas_compile(void *ctx, da_read_fn readfn, da_setpos_fn setposfn
|
||||
* da_getpropid on the atlas, and if generated by the search, the ID will be consistent across
|
||||
* different calls to search.
|
||||
* Properties added by a search that are neither in the compiled atlas, nor in the extra_props list
|
||||
* Are assigned an ID within the context that is not transferrable through different search results
|
||||
* Are assigned an ID within the context that is not transferable through different search results
|
||||
* within the same atlas.
|
||||
* @param atlas Atlas instance
|
||||
* @param extra_props properties
|
||||
|
@ -362,7 +362,7 @@ option set-process-time <var name>
|
||||
latency added by the SPOE processing for the last handled event or group.
|
||||
|
||||
If several events or groups are processed for the same stream, this value
|
||||
will be overrideen.
|
||||
will be overridden.
|
||||
|
||||
See also: "option set-total-time".
|
||||
|
||||
|
@ -12067,7 +12067,7 @@ redirect scheme <sch> [code <code>] <option> [{if | unless} <condition>]
|
||||
- "set-cookie-fmt <fmt>"
|
||||
It is equivaliant to the option above, except the "Set-Cookie" header
|
||||
will be filled with the result of the log-format string <fmt>
|
||||
evaluation. Be carefull to respect the "NAME[=value]" format because no
|
||||
evaluation. Be careful to respect the "NAME[=value]" format because no
|
||||
special check are performed during the configuration parsing.
|
||||
|
||||
- "clear-cookie NAME[=]"
|
||||
@ -13448,8 +13448,8 @@ stick-table type {ip | integer | string [len <length>] | binary [len <length>]}
|
||||
defined period. The factor must be greater than 0 and lower or
|
||||
equal to 1024.
|
||||
|
||||
[recv-only] indicates that we don't intent to use the table to perform
|
||||
updates on it, but thay we only plan on using the table to
|
||||
[recv-only] indicates that we don't intend to use the table to perform
|
||||
updates on it, but that we only plan on using the table to
|
||||
retrieve data from a remote peer which we are interested in.
|
||||
Indeed, the use of this keyword enables the retrieval of
|
||||
local-only values such as conn_cur that are not learned by
|
||||
@ -18547,7 +18547,7 @@ init-state { fully-up | up | down | fully-down }
|
||||
|
||||
The "init-state" option sets the initial state of the server:
|
||||
- when set to 'fully-up', the server is considered immediately available
|
||||
and can turn to the DOWN sate when ALL health checks fail.
|
||||
and can turn to the DOWN state when ALL health checks fail.
|
||||
- when set to 'up' (the default), the server is considered immediately
|
||||
available and will initiate a health check that can turn it to the DOWN
|
||||
state immediately if it fails.
|
||||
@ -19090,7 +19090,7 @@ resolve-opts <option>,<option>,...
|
||||
using an "agent-check" or through the runtime api.
|
||||
|
||||
* prevent-dup-ip
|
||||
Ensure HAProxy's default behavior is enforced on a server: prevent re-using
|
||||
Ensure HAProxy's default behavior is enforced on a server: prevent reusing
|
||||
an IP address already set to a server in the same backend and sharing the
|
||||
same fqdn.
|
||||
This is the opposite of allow-dup-ip.
|
||||
|
@ -38,7 +38,7 @@ extern int (*const pat_index_fcts[PAT_MATCH_NUM])(struct pattern_expr *, struct
|
||||
extern void (*const pat_prune_fcts[PAT_MATCH_NUM])(struct pattern_expr *);
|
||||
extern struct pattern *(*const pat_match_fcts[PAT_MATCH_NUM])(struct sample *, struct pattern_expr *, int);
|
||||
|
||||
/* This is the root of the list of all pattern_ref avalaibles. */
|
||||
/* This is the root of the list of all available pattern_ref values. */
|
||||
extern struct list pattern_reference;
|
||||
|
||||
int pattern_finalize_config(void);
|
||||
|
@ -184,7 +184,7 @@ enum parse_type_t {
|
||||
PARSE_TYPE_NONE = 0,
|
||||
PARSE_TYPE_INT,
|
||||
PARSE_TYPE_STR, /* string which is strdup() */
|
||||
PARSE_TYPE_ARRAY_SUBSTR, /* string splitted by colons into an array of substring */
|
||||
PARSE_TYPE_ARRAY_SUBSTR, /* string split by colons into an array of substring */
|
||||
PARSE_TYPE_ONOFF, /* "on" or "off" keyword */
|
||||
};
|
||||
|
||||
|
@ -461,7 +461,7 @@ void clock_entering_poll(void)
|
||||
|
||||
/* The time might have jumped either backwards or forwards during tasks
|
||||
* processing. It's easy to detect a backwards jump, but a forward jump
|
||||
* needs a marging. Here the upper limit of 2 seconds corresponds to a
|
||||
* needs a margin. Here the upper limit of 2 seconds corresponds to a
|
||||
* large margin at which the watchdog would already trigger so it looks
|
||||
* sufficient to avoid false positives most of the time. The goal here
|
||||
* is to make sure that before_poll can be trusted when entering
|
||||
|
6
src/h2.c
6
src/h2.c
@ -503,7 +503,7 @@ int h2_make_htx_request(struct http_hdr *list, struct htx *htx, unsigned int *ms
|
||||
goto fail;
|
||||
}
|
||||
|
||||
/* Check the number of blocks agains "tune.http.maxhdr" value before adding EOH block */
|
||||
/* Check the number of blocks against "tune.http.maxhdr" value before adding EOH block */
|
||||
if (htx_nbblks(htx) > global.tune.max_http_hdr)
|
||||
goto fail;
|
||||
|
||||
@ -767,7 +767,7 @@ int h2_make_htx_response(struct http_hdr *list, struct htx *htx, unsigned int *m
|
||||
*/
|
||||
}
|
||||
|
||||
/* Check the number of blocks agains "tune.http.maxhdr" value before adding EOH block */
|
||||
/* Check the number of blocks against "tune.http.maxhdr" value before adding EOH block */
|
||||
if (htx_nbblks(htx) > global.tune.max_http_hdr)
|
||||
goto fail;
|
||||
|
||||
@ -847,7 +847,7 @@ int h2_make_htx_trailers(struct http_hdr *list, struct htx *htx)
|
||||
goto fail;
|
||||
}
|
||||
|
||||
/* Check the number of blocks agains "tune.http.maxhdr" value before adding EOT block */
|
||||
/* Check the number of blocks against "tune.http.maxhdr" value before adding EOT block */
|
||||
if (htx_nbblks(htx) > global.tune.max_http_hdr)
|
||||
goto fail;
|
||||
|
||||
|
4
src/h3.c
4
src/h3.c
@ -898,7 +898,7 @@ static ssize_t h3_headers_to_htx(struct qcs *qcs, const struct buffer *buf,
|
||||
}
|
||||
}
|
||||
|
||||
/* Check the number of blocks agains "tune.http.maxhdr" value before adding EOH block */
|
||||
/* Check the number of blocks against "tune.http.maxhdr" value before adding EOH block */
|
||||
if (htx_nbblks(htx) > global.tune.max_http_hdr) {
|
||||
len = -1;
|
||||
goto out;
|
||||
@ -1086,7 +1086,7 @@ static ssize_t h3_trailers_to_htx(struct qcs *qcs, const struct buffer *buf,
|
||||
++hdr_idx;
|
||||
}
|
||||
|
||||
/* Check the number of blocks agains "tune.http.maxhdr" value before adding EOT block */
|
||||
/* Check the number of blocks against "tune.http.maxhdr" value before adding EOT block */
|
||||
if (htx_nbblks(htx) > global.tune.max_http_hdr) {
|
||||
len = -1;
|
||||
goto out;
|
||||
|
@ -599,7 +599,7 @@ static unsigned int hlua_nb_instruction = 0;
|
||||
|
||||
/* Wrapper to retrieve the number of instructions between two interrupts
|
||||
* depending on user settings and current hlua context. If not already
|
||||
* explicitly set, we compute the ideal value using hard limits releaved
|
||||
* explicitly set, we compute the ideal value using hard limits revealed
|
||||
* by Thierry Fournier's work, whose original notes may be found below:
|
||||
*
|
||||
* --
|
||||
@ -9562,7 +9562,7 @@ static void hlua_event_handler(struct hlua *hlua)
|
||||
hlua_timer_init(&hlua->timer, hlua_timeout_task);
|
||||
|
||||
/* make sure to reset the task expiry before each hlua_ctx_resume()
|
||||
* since the task is re-used for multiple cb function calls
|
||||
* since the task is reused for multiple cb function calls
|
||||
* We couldn't risk to have t->expire pointing to a past date because
|
||||
* it was set during last function invocation but was never reset since
|
||||
* (ie: E_AGAIN)
|
||||
|
@ -137,7 +137,7 @@ static int compute_ideal_maxconn()
|
||||
* extremely large and watchdog will kill the process, when it will try
|
||||
* to loop over the fdtab (see fd_reregister_all). Please note, that
|
||||
* fd_hard_limit is taken in account implicitly via 'ideal_maxconn'
|
||||
* value in all global.maxconn adjustements, when global.rlimit_memmax
|
||||
* value in all global.maxconn adjustments, when global.rlimit_memmax
|
||||
* is set:
|
||||
*
|
||||
* MIN(global.maxconn, capped by global.rlimit_memmax, ideal_maxconn);
|
||||
|
@ -443,7 +443,7 @@ static void fcgi_trace(enum trace_level level, uint64_t mask, const struct trace
|
||||
* a bit complex due to some API limits for now. The rules are the following :
|
||||
* - if an error or a shutdown was detected on the connection,
|
||||
we must not attempt to receive
|
||||
* - if we're subscribed for receving, no need to try again
|
||||
* - if we're subscribed for receiving, no need to try again
|
||||
* - if the demux buf failed to be allocated, we must not try to receive and
|
||||
* we know there is nothing pending (we'll be woken up once allocated)
|
||||
* - if the demux buf is full, we will not be able to receive.
|
||||
|
@ -2731,7 +2731,7 @@ static size_t h1_make_eoh(struct h1s *h1s, struct h1m *h1m, struct htx *htx, siz
|
||||
else {
|
||||
/* Both headers are missing */
|
||||
if (h1m->flags & H1_MF_RESP) {
|
||||
/* It is a esponse: Switch to unknown xfer length */
|
||||
/* It is a response: Switch to unknown xfer length */
|
||||
h1m->flags &= ~(H1_MF_XFER_LEN|H1_MF_XFER_ENC|H1_MF_CLEN|H1_MF_CHNK);
|
||||
h1s->flags &= ~(H1S_F_HAVE_CLEN|H1S_F_HAVE_CHNK);
|
||||
TRACE_STATE("Switch response to unknown XFER length", H1_EV_TX_DATA|H1_EV_TX_HDRS, h1c->conn, h1s);
|
||||
|
@ -951,7 +951,7 @@ h2c_is_dead(const struct h2c *h2c)
|
||||
* a bit complex due to some API limits for now. The rules are the following :
|
||||
* - if an error or a shutdown was detected on the connection, we must not
|
||||
* attempt to receive
|
||||
* - if we're subscribed for receving, no need to try again
|
||||
* - if we're subscribed for receiving, no need to try again
|
||||
* - if the demux buf failed to be allocated, we must not try to receive and
|
||||
* we know there is nothing pending (we'll be woken up once allocated)
|
||||
* - if the demux buf is full, we will not be able to receive.
|
||||
|
@ -2108,7 +2108,7 @@ int qcc_recv_stop_sending(struct qcc *qcc, uint64_t id, uint64_t err)
|
||||
* has already been closed locally. This is useful to not emit multiple
|
||||
* RESET_STREAM for a single stream. This is functional if stream is
|
||||
* locally closed due to all data transmitted, but in this case the RFC
|
||||
* advices to use an explicit RESET_STREAM.
|
||||
* advises to use an explicit RESET_STREAM.
|
||||
*/
|
||||
if (qcs_is_close_local(qcs)) {
|
||||
TRACE_STATE("ignoring STOP_SENDING", QMUX_EV_QCC_RECV|QMUX_EV_QCS_RECV, qcc->conn, qcs);
|
||||
|
@ -137,7 +137,7 @@ int const pat_match_types[PAT_MATCH_NUM] = {
|
||||
static THREAD_LOCAL struct pattern static_pattern;
|
||||
static THREAD_LOCAL struct sample_data static_sample_data;
|
||||
|
||||
/* This is the root of the list of all pattern_ref avalaibles. */
|
||||
/* This is the root of the list of all available pattern_ref values. */
|
||||
struct list pattern_reference = LIST_HEAD_INIT(pattern_reference);
|
||||
|
||||
static THREAD_LOCAL struct lru64_head *pat_lru_tree;
|
||||
|
@ -1931,7 +1931,7 @@ void qc_bind_tid_commit(struct quic_conn *qc, struct listener *new_li)
|
||||
conn_id = eb64_entry(node, struct quic_connection_id, seq_num);
|
||||
|
||||
/* Rebinding is considered done when CID points to the new
|
||||
* thread. quic-conn instance cannot be derefence after it.
|
||||
* thread. quic-conn instance cannot be dereferenced after it.
|
||||
*/
|
||||
HA_ATOMIC_STORE(&conn_id->tid, new_tid);
|
||||
qc = NULL;
|
||||
|
@ -1645,7 +1645,7 @@ static inline int quic_padding_check(const unsigned char *pos,
|
||||
|
||||
/* Validate the token, retry or not (provided by NEW_TOKEN) parsed into
|
||||
* <pkt> RX packet from <dgram> datagram.
|
||||
* Return 1 if succeded, 0 if not.
|
||||
* Return 1 if succeeded, 0 if not.
|
||||
*/
|
||||
static inline int quic_token_validate(struct quic_rx_packet *pkt,
|
||||
struct quic_dgram *dgram,
|
||||
|
@ -244,7 +244,7 @@ static int quic_conn_enc_level_init(struct quic_conn *qc,
|
||||
}
|
||||
|
||||
/* Ensure early-data encryption is not inserted at the end of this ->qel_list
|
||||
* list. This would perturbate the sender during handshakes. This latter adds
|
||||
* list. This would disrupt the sender during handshakes. This latter adds
|
||||
* PADDING frames to datagrams from the last encryption level in this list,
|
||||
* for datagram with at least an ack-eliciting Initial packet inside.
|
||||
* But a QUIC server has nothing to send from this early-data encryption
|
||||
|
@ -1312,7 +1312,7 @@ struct sink *sink_new_from_srv(struct server *srv, const char *from)
|
||||
/* directly create a sink of BUF type, and use UNSPEC log format to
|
||||
* inherit from caller fmt in sink_write()
|
||||
*
|
||||
* sink_name must be unique to prevent existing sink from being re-used
|
||||
* sink_name must be unique to prevent existing sink from being reused
|
||||
*/
|
||||
sink = sink_new_buf(sink_name, trash.area, LOG_FORMAT_UNSPEC, bufsize);
|
||||
ha_free(&sink_name); // no longer needed
|
||||
|
@ -482,7 +482,7 @@ int sock_get_old_sockets(const char *unixsocket)
|
||||
|
||||
/* dst_fd is always open in the worker process context because
|
||||
* it's inherited from the master via -x cmd option. It's closed
|
||||
* futher in main (after bind_listeners()) and not here for the
|
||||
* further in main (after bind_listeners()) and not here for the
|
||||
* simplicity. In main(), after bind_listeners(), it's safe just
|
||||
* to loop over all workers list, launched before this reload and
|
||||
* to close its ipc_fd[0], thus we also close this fd. If we
|
||||
|
@ -4807,7 +4807,7 @@ int ckch_conf_parse(char **args, int cur_arg, struct ckch_conf *f, int *found, c
|
||||
int n = 0;
|
||||
char *b, *e;
|
||||
|
||||
/* split a string into substring splitted by colons */
|
||||
/* split a string into substring split by colons */
|
||||
|
||||
e = b = args[cur_arg + 1];
|
||||
do {
|
||||
|
16
src/stream.c
16
src/stream.c
@ -81,8 +81,8 @@ static void strm_trace(enum trace_level level, uint64_t mask,
|
||||
/* The event representation is split like this :
|
||||
* strm - stream
|
||||
* sc - stream connector
|
||||
* http - http analyzis
|
||||
* tcp - tcp analyzis
|
||||
* http - http analysis
|
||||
* tcp - tcp analysis
|
||||
*
|
||||
* STRM_EV_* macros are defined in <proto/stream.h>
|
||||
*/
|
||||
@ -96,13 +96,13 @@ static const struct trace_event strm_trace_events[] = {
|
||||
{ .mask = STRM_EV_CS_ST, .name = "sc_state", .desc = "processing connector states" },
|
||||
|
||||
{ .mask = STRM_EV_HTTP_ANA, .name = "http_ana", .desc = "HTTP analyzers" },
|
||||
{ .mask = STRM_EV_HTTP_ERR, .name = "http_err", .desc = "error during HTTP analyzis" },
|
||||
{ .mask = STRM_EV_HTTP_ERR, .name = "http_err", .desc = "error during HTTP analysis" },
|
||||
|
||||
{ .mask = STRM_EV_TCP_ANA, .name = "tcp_ana", .desc = "TCP analyzers" },
|
||||
{ .mask = STRM_EV_TCP_ERR, .name = "tcp_err", .desc = "error during TCP analyzis" },
|
||||
{ .mask = STRM_EV_TCP_ERR, .name = "tcp_err", .desc = "error during TCP analysis" },
|
||||
|
||||
{ .mask = STRM_EV_FLT_ANA, .name = "flt_ana", .desc = "Filter analyzers" },
|
||||
{ .mask = STRM_EV_FLT_ERR, .name = "flt_err", .desc = "error during filter analyzis" },
|
||||
{ .mask = STRM_EV_FLT_ERR, .name = "flt_err", .desc = "error during filter analysis" },
|
||||
{}
|
||||
};
|
||||
|
||||
@ -1148,7 +1148,7 @@ static int process_switching_rules(struct stream *s, struct channel *req, int an
|
||||
|
||||
}
|
||||
|
||||
/* Se the max connection retries for the stream. may be overwriten later */
|
||||
/* Se the max connection retries for the stream. may be overwritten later */
|
||||
s->max_retries = s->be->conn_retries;
|
||||
|
||||
/* we don't want to run the TCP or HTTP filters again if the backend has not changed */
|
||||
@ -2767,7 +2767,7 @@ void sess_change_server(struct stream *strm, struct server *newsrv)
|
||||
* and that removal will be done under isolation anyway. Thus by
|
||||
* decrementing served after detaching from the list, we're
|
||||
* guaranteeing that served==0 implies that no stream is in the
|
||||
* list anymore, which is a sufficient guarantee for tha goal.
|
||||
* list anymore, which is a sufficient guarantee for that goal.
|
||||
*/
|
||||
_HA_ATOMIC_DEC(&oldsrv->proxy->served);
|
||||
stream_del_srv_conn(strm);
|
||||
@ -3716,7 +3716,7 @@ static void __strm_dump_to_buffer(struct buffer *buf, const struct show_sess_ctx
|
||||
}
|
||||
}
|
||||
|
||||
/* Context-less function to append a complet dump of a stream state onto the
|
||||
/* Context-less function to append a complete dump of a stream state onto the
|
||||
* buffer. It relies on __strm_dump_to_buffer.
|
||||
*/
|
||||
void strm_dump_to_buffer(struct buffer *buf, const struct stream *strm, const char *pfx, uint32_t anon_key)
|
||||
|
@ -7106,7 +7106,7 @@ int clean_env(void)
|
||||
/* Restores **environ from backup created by backup_env(). Must be always
|
||||
* preceded by clean_env() in order to properly restore the process environment.
|
||||
* global init_env ptr array must be freed by the upper layers.
|
||||
* Returns 0 on sucess and -1 in case if the process has run out of memory. If
|
||||
* Returns 0 on success and -1 in case if the process has run out of memory. If
|
||||
* setenv() fails with EINVAL or the parsed string doesn't contain '=' (the
|
||||
* latter is mandatory format for strings kept in **environ), emits warning and
|
||||
* continues. This allows to terminate the process at the startup stage, if it
|
||||
|
Loading…
x
Reference in New Issue
Block a user