- Add new patches (see https://www.haproxy.org/bugs/bugs-1.8.8.html) - Raise patch-level to 03 Signed-off-by: Christian Lachner <gladiac@gmail.com>lilik-openwrt-22.03
@ -0,0 +1,31 @@ | |||||
commit 80e179128cfd78d95cdebf7195fd21299e7931b6 | |||||
Author: Willy Tarreau <w@1wt.eu> | |||||
Date: Sat Apr 28 07:18:15 2018 +0200 | |||||
BUG/MINOR: config: disable http-reuse on TCP proxies | |||||
Louis Chanouha reported an inappropriate warning when http-reuse is | |||||
present in a defaults section while a TCP proxy accidently inherits | |||||
it and finds a conflict with other options like the use of the PROXY | |||||
protocol. To fix this patch removes the http-reuse option for TCP | |||||
proxies. | |||||
This fix needs to be backported to 1.8, 1.7 and possibly 1.6. | |||||
(cherry picked from commit 46deab6e64bfda7211b7c3199ad01f136141c86f) | |||||
Signed-off-by: Christopher Faulet <cfaulet@haproxy.com> | |||||
diff --git a/src/cfgparse.c b/src/cfgparse.c | |||||
index 5a460381..63d2de58 100644 | |||||
--- a/src/cfgparse.c | |||||
+++ b/src/cfgparse.c | |||||
@@ -8702,6 +8702,9 @@ out_uri_auth_compat: | |||||
} | |||||
#endif | |||||
+ if ((curproxy->mode != PR_MODE_HTTP) && (curproxy->options & PR_O_REUSE_MASK) != PR_O_REUSE_NEVR) | |||||
+ curproxy->options &= ~PR_O_REUSE_MASK; | |||||
+ | |||||
if ((curproxy->options & PR_O_REUSE_MASK) != PR_O_REUSE_NEVR) { | |||||
if ((curproxy->conn_src.opts & CO_SRC_TPROXY_MASK) == CO_SRC_TPROXY_CLI || | |||||
(curproxy->conn_src.opts & CO_SRC_TPROXY_MASK) == CO_SRC_TPROXY_CIP || |
@ -0,0 +1,42 @@ | |||||
commit edb5a1efd22eb9918574d962640cd2ae3bb45ad3 | |||||
Author: Christopher Faulet <cfaulet@haproxy.com> | |||||
Date: Wed May 2 12:12:45 2018 +0200 | |||||
BUG/MINOR: checks: Fix check->health computation for flapping servers | |||||
This patch fixes an old bug introduced in the commit 7b1d47ce ("MAJOR: checks: | |||||
move health checks changes to set_server_check_status()"). When a DOWN server is | |||||
flapping, everytime a check succeds, check->health is incremented. But when a | |||||
check fails, it is decremented only when it is higher than the rise value. So if | |||||
only one check succeds for a DOWN server, check->health will remain set to 1 for | |||||
all subsequent failing checks. | |||||
So, at first glance, it seems not that terrible because the server remains | |||||
DOWN. But it is reported in the transitional state "DOWN server, going up". And | |||||
it will remain in this state until it is UP again. And there is also an | |||||
insidious side effect. If a DOWN server is flapping time to time, It will end to | |||||
be considered UP after a uniq successful check, , regardless the rise threshold, | |||||
because check->health will be increased slowly and never decreased. | |||||
To fix the bug, we just need to reset check->health to 0 when a check fails for | |||||
a DOWN server. To do so, we just need to relax the condition to handle a failure | |||||
in the function set_server_check_status. | |||||
This patch must be backported to haproxy 1.5 and newer. | |||||
(cherry picked from commit b119a79fc336f2b6074de1c3113b1682c717985c) | |||||
Signed-off-by: Willy Tarreau <w@1wt.eu> | |||||
diff --git a/src/checks.c b/src/checks.c | |||||
index 80a9c70d..d07a82f8 100644 | |||||
--- a/src/checks.c | |||||
+++ b/src/checks.c | |||||
@@ -243,7 +243,7 @@ static void set_server_check_status(struct check *check, short status, const cha | |||||
*/ | |||||
if ((!(check->state & CHK_ST_AGENT) || | |||||
(check->status >= HCHK_STATUS_L57DATA)) && | |||||
- (check->health >= check->rise)) { | |||||
+ (check->health > 0)) { | |||||
HA_ATOMIC_ADD(&s->counters.failed_checks, 1); | |||||
report = 1; | |||||
check->health--; |
@ -0,0 +1,48 @@ | |||||
commit 830324444e57c042666b17ac4584352cca85dafd | |||||
Author: Christopher Faulet <cfaulet@haproxy.com> | |||||
Date: Wed May 2 16:58:40 2018 +0200 | |||||
BUG/MEDIUM: threads: Fix the sync point for more than 32 threads | |||||
In the sync point, to know if a thread has requested a synchronization, we call | |||||
the function thread_need_sync(). It should return 1 if yes, otherwise it should | |||||
return 0. It is intended to return a signed integer. | |||||
But internally, instead of returning 0 or 1, it returns 0 or tid_bit | |||||
(threads_want_sync & tid_bit). So, tid_bit is casted in integer. For the first | |||||
32 threads, it's ok, because we always check if thread_need_sync() returns | |||||
something else than 0. But this is a problem if HAProxy is started with more | |||||
than 32 threads, because for threads 33 to 64 (so for tid 32 to 63), their | |||||
tid_bit casted to integer are evaluated to 0. So the sync point does not work for | |||||
more than 32 threads. | |||||
Now, the function thread_need_sync() respects its contract, returning 0 or | |||||
1. the function thread_no_sync() has also been updated to avoid any ambiguities. | |||||
This patch must be backported in HAProxy 1.8. | |||||
(cherry picked from commit 148b16e1ceb819dfcef4c45828121d9cd7474b35) | |||||
Signed-off-by: Willy Tarreau <w@1wt.eu> | |||||
diff --git a/src/hathreads.c b/src/hathreads.c | |||||
index daf226ce..944a0d5b 100644 | |||||
--- a/src/hathreads.c | |||||
+++ b/src/hathreads.c | |||||
@@ -85,7 +85,7 @@ void thread_want_sync() | |||||
/* Returns 1 if no thread has requested a sync. Otherwise, it returns 0. */ | |||||
int thread_no_sync() | |||||
{ | |||||
- return (threads_want_sync == 0); | |||||
+ return (threads_want_sync == 0UL); | |||||
} | |||||
/* Returns 1 if the current thread has requested a sync. Otherwise, it returns | |||||
@@ -93,7 +93,7 @@ int thread_no_sync() | |||||
*/ | |||||
int thread_need_sync() | |||||
{ | |||||
- return (threads_want_sync & tid_bit); | |||||
+ return ((threads_want_sync & tid_bit) != 0UL); | |||||
} | |||||
/* Thread barrier. Synchronizes all threads at the barrier referenced by |
@ -0,0 +1,31 @@ | |||||
commit 335bc7b74eee84f0a3bcb615cadd23fe01d1336c | |||||
Author: PiBa-NL <PiBa.NL.dev@gmail.com> | |||||
Date: Wed May 2 22:27:14 2018 +0200 | |||||
BUG/MINOR: lua: Put tasks to sleep when waiting for data | |||||
If a lua socket is waiting for data it currently spins at 100% cpu usage. | |||||
This because the TICK_ETERNITY returned by the socket is ignored when | |||||
setting the 'expire' time of the task. | |||||
Fixed by removing the check for yields that return TICK_ETERNITY. | |||||
This should be backported to at least 1.8. | |||||
(cherry picked from commit fe971b35aeca9994f3823112c783aa796e74075a) | |||||
Signed-off-by: Willy Tarreau <w@1wt.eu> | |||||
diff --git a/src/hlua.c b/src/hlua.c | |||||
index bd0b87e3..0100e7cf 100644 | |||||
--- a/src/hlua.c | |||||
+++ b/src/hlua.c | |||||
@@ -5536,8 +5536,7 @@ static struct task *hlua_process_task(struct task *task) | |||||
case HLUA_E_AGAIN: /* co process or timeout wake me later. */ | |||||
notification_gc(&hlua->com); | |||||
- if (hlua->wake_time != TICK_ETERNITY) | |||||
- task->expire = hlua->wake_time; | |||||
+ task->expire = hlua->wake_time; | |||||
break; | |||||
/* finished with error. */ |
@ -0,0 +1,252 @@ | |||||
commit 016feef5483397491af3242162934d9e9dbc6263 | |||||
Author: Patrick Hemmer <haproxy@stormcloud9.net> | |||||
Date: Tue May 1 21:30:41 2018 -0400 | |||||
DOC/MINOR: clean up LUA documentation re: servers & array/table. | |||||
* A few typos | |||||
* Fix definitions of values which are tables, not arrays. | |||||
* Consistent US English naming for "server" instead of "serveur". | |||||
[tfo: should be backported to 1.6 and higher] | |||||
(cherry picked from commit c6a1d711a4d47d68611aa28adecdadba96221bde) | |||||
Signed-off-by: Willy Tarreau <w@1wt.eu> | |||||
diff --git a/doc/lua-api/index.rst b/doc/lua-api/index.rst | |||||
index e7aa425d..2d210945 100644 | |||||
--- a/doc/lua-api/index.rst | |||||
+++ b/doc/lua-api/index.rst | |||||
@@ -169,9 +169,9 @@ Core class | |||||
**context**: task, action, sample-fetch, converter | |||||
- This attribute is an array of declared proxies (frontend and backends). Each | |||||
- proxy give an access to his list of listeners and servers. Each entry is of | |||||
- type :ref:`proxy_class` | |||||
+ This attribute is a table of declared proxies (frontend and backends). Each | |||||
+ proxy give an access to his list of listeners and servers. The table is | |||||
+ indexed by proxy name, and each entry is of type :ref:`proxy_class`. | |||||
Warning, if you are declared frontend and backend with the same name, only one | |||||
of these are listed. | |||||
@@ -183,12 +183,9 @@ Core class | |||||
**context**: task, action, sample-fetch, converter | |||||
- This attribute is an array of declared proxies with backend capability. Each | |||||
- proxy give an access to his list of listeners and servers. Each entry is of | |||||
- type :ref:`proxy_class` | |||||
- | |||||
- Warning, if you are declared frontend and backend with the same name, only one | |||||
- of these are listed. | |||||
+ This attribute is a table of declared proxies with backend capability. Each | |||||
+ proxy give an access to his list of listeners and servers. The table is | |||||
+ indexed by the backend name, and each entry is of type :ref:`proxy_class`. | |||||
:see: :js:attr:`core.proxies` | |||||
:see: :js:attr:`core.frontends` | |||||
@@ -197,12 +194,9 @@ Core class | |||||
**context**: task, action, sample-fetch, converter | |||||
- This attribute is an array of declared proxies with frontend capability. Each | |||||
- proxy give an access to his list of listeners and servers. Each entry is of | |||||
- type :ref:`proxy_class` | |||||
- | |||||
- Warning, if you are declared frontend and backend with the same name, only one | |||||
- of these are listed. | |||||
+ This attribute is a table of declared proxies with frontend capability. Each | |||||
+ proxy give an access to his list of listeners and servers. The table is | |||||
+ indexed by the frontend name, and each entry is of type :ref:`proxy_class`. | |||||
:see: :js:attr:`core.proxies` | |||||
:see: :js:attr:`core.backends` | |||||
@@ -336,7 +330,7 @@ Core class | |||||
Lua execution or resume, so two consecutive call to the function "now" will | |||||
probably returns the same result. | |||||
- :returns: an array which contains two entries "sec" and "usec". "sec" | |||||
+ :returns: a table which contains two entries "sec" and "usec". "sec" | |||||
contains the current at the epoch format, and "usec" contains the | |||||
current microseconds. | |||||
@@ -439,9 +433,12 @@ Core class | |||||
**context**: body, init, task, action, sample-fetch, converter | |||||
- proxies is an array containing the list of all proxies declared in the | |||||
- configuration file. Each entry of the proxies array is an object of type | |||||
- :ref:`proxy_class` | |||||
+ proxies is a table containing the list of all proxies declared in the | |||||
+ configuration file. The table is indexed by the proxy name, and each entry | |||||
+ of the proxies table is an object of type :ref:`proxy_class`. | |||||
+ | |||||
+ Warning, if you have declared a frontend and backend with the same name, only | |||||
+ one of these are listed. | |||||
.. js:function:: core.register_action(name, actions, func [, nb_args]) | |||||
@@ -852,13 +849,14 @@ Proxy class | |||||
.. js:attribute:: Proxy.servers | |||||
- Contain an array with the attached servers. Each server entry is an object of | |||||
- type :ref:`server_class`. | |||||
+ Contain a table with the attached servers. The table is indexed by server | |||||
+ name, and each server entry is an object of type :ref:`server_class`. | |||||
.. js:attribute:: Proxy.listeners | |||||
- Contain an array with the attached listeners. Each listeners entry is an | |||||
- object of type :ref:`listener_class`. | |||||
+ Contain a table with the attached listeners. The table is indexed by listener | |||||
+ name, and each each listeners entry is an object of type | |||||
+ :ref:`listener_class`. | |||||
.. js:function:: Proxy.pause(px) | |||||
@@ -908,21 +906,25 @@ Proxy class | |||||
.. js:function:: Proxy.get_stats(px) | |||||
- Returns an array containg the proxy statistics. The statistics returned are | |||||
+ Returns a table containg the proxy statistics. The statistics returned are | |||||
not the same if the proxy is frontend or a backend. | |||||
:param class_proxy px: A :ref:`proxy_class` which indicates the manipulated | |||||
proxy. | |||||
- :returns: a key/value array containing stats | |||||
+ :returns: a key/value table containing stats | |||||
.. _server_class: | |||||
Server class | |||||
============ | |||||
+.. js:class:: Server | |||||
+ | |||||
+ This class provides a way for manipulating servers and retrieving information. | |||||
+ | |||||
.. js:function:: Server.is_draining(sv) | |||||
- Return true if the server is currently draining stiky connections. | |||||
+ Return true if the server is currently draining sticky connections. | |||||
:param class_server sv: A :ref:`server_class` which indicates the manipulated | |||||
server. | |||||
@@ -930,7 +932,7 @@ Server class | |||||
.. js:function:: Server.set_weight(sv, weight) | |||||
- Dynamically change the weight of the serveur. See the management socket | |||||
+ Dynamically change the weight of the server. See the management socket | |||||
documentation for more information about the format of the string. | |||||
:param class_server sv: A :ref:`server_class` which indicates the manipulated | |||||
@@ -939,7 +941,7 @@ Server class | |||||
.. js:function:: Server.get_weight(sv) | |||||
- This function returns an integer representing the serveur weight. | |||||
+ This function returns an integer representing the server weight. | |||||
:param class_server sv: A :ref:`server_class` which indicates the manipulated | |||||
server. | |||||
@@ -947,16 +949,16 @@ Server class | |||||
.. js:function:: Server.set_addr(sv, addr) | |||||
- Dynamically change the address of the serveur. See the management socket | |||||
+ Dynamically change the address of the server. See the management socket | |||||
documentation for more information about the format of the string. | |||||
:param class_server sv: A :ref:`server_class` which indicates the manipulated | |||||
server. | |||||
- :param string weight: A string describing the server address. | |||||
+ :param string addr: A string describing the server address. | |||||
.. js:function:: Server.get_addr(sv) | |||||
- Returns a string describing the address of the serveur. | |||||
+ Returns a string describing the address of the server. | |||||
:param class_server sv: A :ref:`server_class` which indicates the manipulated | |||||
server. | |||||
@@ -968,7 +970,7 @@ Server class | |||||
:param class_server sv: A :ref:`server_class` which indicates the manipulated | |||||
server. | |||||
- :returns: a key/value array containing stats | |||||
+ :returns: a key/value table containing stats | |||||
.. js:function:: Server.shut_sess(sv) | |||||
@@ -1085,7 +1087,7 @@ Listener class | |||||
:param class_listener ls: A :ref:`listener_class` which indicates the | |||||
manipulated listener. | |||||
- :returns: a key/value array containing stats | |||||
+ :returns: a key/value table containing stats | |||||
.. _concat_class: | |||||
@@ -1169,7 +1171,7 @@ Fetches class | |||||
usage. they are the chapters 7.3.2 to 7.3.6. | |||||
**warning** some sample fetches are not available in some context. These | |||||
- limitations are specified in this documentation when theire useful. | |||||
+ limitations are specified in this documentation when they're useful. | |||||
:see: :js:attr:`TXN.f` | |||||
:see: :js:attr:`TXN.sf` | |||||
@@ -1345,13 +1347,13 @@ HTTP class | |||||
.. js:function:: HTTP.req_get_headers(http) | |||||
- Returns an array containing all the request headers. | |||||
+ Returns a table containing all the request headers. | |||||
:param class_http http: The related http object. | |||||
- :returns: array of headers. | |||||
+ :returns: table of headers. | |||||
:see: :js:func:`HTTP.res_get_headers` | |||||
- This is the form of the returned array: | |||||
+ This is the form of the returned table: | |||||
.. code-block:: lua | |||||
@@ -1366,13 +1368,13 @@ HTTP class | |||||
.. js:function:: HTTP.res_get_headers(http) | |||||
- Returns an array containing all the response headers. | |||||
+ Returns a table containing all the response headers. | |||||
:param class_http http: The related http object. | |||||
- :returns: array of headers. | |||||
+ :returns: table of headers. | |||||
:see: :js:func:`HTTP.req_get_headers` | |||||
- This is the form of the returned array: | |||||
+ This is the form of the returned table: | |||||
.. code-block:: lua | |||||
@@ -2210,12 +2212,12 @@ AppletHTTP class | |||||
.. js:attribute:: AppletHTTP.headers | |||||
- :returns: array | |||||
+ :returns: table | |||||
- The attribute headers returns an array containing the HTTP | |||||
+ The attribute headers returns a table containing the HTTP | |||||
headers. The header names are always in lower case. As the header name can be | |||||
encountered more than once in each request, the value is indexed with 0 as | |||||
- first index value. The array have this form: | |||||
+ first index value. The table have this form: | |||||
.. code-block:: lua | |||||
@ -0,0 +1,31 @@ | |||||
commit b2219ae216a141acdf0e2a3f67d2c85aee2a2bc2 | |||||
Author: Dragan Dosen <ddosen@haproxy.com> | |||||
Date: Fri May 4 16:27:15 2018 +0200 | |||||
BUG/MINOR: map: correctly track reference to the last ref_elt being dumped | |||||
The bug was introduced in the commit 8d85aa4 ("BUG/MAJOR: map: fix | |||||
segfault during 'show map/acl' on cli"). | |||||
This patch should be backported to 1.8, 1.7 and 1.6. | |||||
(cherry picked from commit 336a11f75571ad46f74a7c6247c13ed44f95da93) | |||||
Signed-off-by: Willy Tarreau <w@1wt.eu> | |||||
diff --git a/src/map.c b/src/map.c | |||||
index f40e4394..a9a1e53c 100644 | |||||
--- a/src/map.c | |||||
+++ b/src/map.c | |||||
@@ -307,9 +307,9 @@ static int cli_io_handler_pat_list(struct appctx *appctx) | |||||
* reference to the last ref_elt being dumped. | |||||
*/ | |||||
if (appctx->st2 == STAT_ST_LIST) { | |||||
- if (!LIST_ISEMPTY(&appctx->ctx.sess.bref.users)) { | |||||
- LIST_DEL(&appctx->ctx.sess.bref.users); | |||||
- LIST_INIT(&appctx->ctx.sess.bref.users); | |||||
+ if (!LIST_ISEMPTY(&appctx->ctx.map.bref.users)) { | |||||
+ LIST_DEL(&appctx->ctx.map.bref.users); | |||||
+ LIST_INIT(&appctx->ctx.map.bref.users); | |||||
} | |||||
} | |||||
return 1; |
@ -0,0 +1,128 @@ | |||||
commit a0f0db361978154474d76028183647d5991f3b5c | |||||
Author: Olivier Houchard <ohouchard@haproxy.com> | |||||
Date: Fri May 4 15:46:16 2018 +0200 | |||||
BUG/MEDIUM: task: Don't free a task that is about to be run. | |||||
While running a task, we may try to delete and free a task that is about to | |||||
be run, because it's part of the local tasks list, or because rq_next points | |||||
to it. | |||||
So flag any task that is in the local tasks list to be deleted, instead of | |||||
run, by setting t->process to NULL, and re-make rq_next a global, | |||||
thread-local variable, that is modified if we attempt to delete that task. | |||||
Many thanks to PiBa-NL for reporting this and analysing the problem. | |||||
This should be backported to 1.8. | |||||
(cherry picked from commit 9b36cb4a414c22e13d344afbbe70684e9f2f1d49) | |||||
Signed-off-by: Willy Tarreau <w@1wt.eu> | |||||
diff --git a/include/proto/task.h b/include/proto/task.h | |||||
index cbc1a907..c1c4c07e 100644 | |||||
--- a/include/proto/task.h | |||||
+++ b/include/proto/task.h | |||||
@@ -90,6 +90,8 @@ extern unsigned int nb_tasks_cur; | |||||
extern unsigned int niced_tasks; /* number of niced tasks in the run queue */ | |||||
extern struct pool_head *pool_head_task; | |||||
extern struct pool_head *pool_head_notification; | |||||
+extern THREAD_LOCAL struct task *curr_task; /* task currently running or NULL */ | |||||
+extern THREAD_LOCAL struct eb32sc_node *rq_next; /* Next task to be potentially run */ | |||||
__decl_hathreads(extern HA_SPINLOCK_T rq_lock); /* spin lock related to run queue */ | |||||
__decl_hathreads(extern HA_SPINLOCK_T wq_lock); /* spin lock related to wait queue */ | |||||
@@ -177,8 +179,11 @@ static inline struct task *__task_unlink_rq(struct task *t) | |||||
static inline struct task *task_unlink_rq(struct task *t) | |||||
{ | |||||
HA_SPIN_LOCK(TASK_RQ_LOCK, &rq_lock); | |||||
- if (likely(task_in_rq(t))) | |||||
+ if (likely(task_in_rq(t))) { | |||||
+ if (&t->rq == rq_next) | |||||
+ rq_next = eb32sc_next(rq_next, tid_bit); | |||||
__task_unlink_rq(t); | |||||
+ } | |||||
HA_SPIN_UNLOCK(TASK_RQ_LOCK, &rq_lock); | |||||
return t; | |||||
} | |||||
@@ -230,7 +235,7 @@ static inline struct task *task_new(unsigned long thread_mask) | |||||
* Free a task. Its context must have been freed since it will be lost. | |||||
* The task count is decremented. | |||||
*/ | |||||
-static inline void task_free(struct task *t) | |||||
+static inline void __task_free(struct task *t) | |||||
{ | |||||
pool_free(pool_head_task, t); | |||||
if (unlikely(stopping)) | |||||
@@ -238,6 +243,18 @@ static inline void task_free(struct task *t) | |||||
HA_ATOMIC_SUB(&nb_tasks, 1); | |||||
} | |||||
+static inline void task_free(struct task *t) | |||||
+{ | |||||
+ /* There's no need to protect t->state with a lock, as the task | |||||
+ * has to run on the current thread. | |||||
+ */ | |||||
+ if (t == curr_task || !(t->state & TASK_RUNNING)) | |||||
+ __task_free(t); | |||||
+ else | |||||
+ t->process = NULL; | |||||
+} | |||||
+ | |||||
+ | |||||
/* Place <task> into the wait queue, where it may already be. If the expiration | |||||
* timer is infinite, do nothing and rely on wake_expired_task to clean up. | |||||
*/ | |||||
diff --git a/src/task.c b/src/task.c | |||||
index fd9acf66..3d021bb4 100644 | |||||
--- a/src/task.c | |||||
+++ b/src/task.c | |||||
@@ -39,6 +39,7 @@ unsigned int nb_tasks_cur = 0; /* copy of the tasks count */ | |||||
unsigned int niced_tasks = 0; /* number of niced tasks in the run queue */ | |||||
THREAD_LOCAL struct task *curr_task = NULL; /* task currently running or NULL */ | |||||
+THREAD_LOCAL struct eb32sc_node *rq_next = NULL; /* Next task to be potentially run */ | |||||
__decl_hathreads(HA_SPINLOCK_T __attribute__((aligned(64))) rq_lock); /* spin lock related to run queue */ | |||||
__decl_hathreads(HA_SPINLOCK_T __attribute__((aligned(64))) wq_lock); /* spin lock related to wait queue */ | |||||
@@ -186,7 +187,6 @@ void process_runnable_tasks() | |||||
struct task *t; | |||||
int i; | |||||
int max_processed; | |||||
- struct eb32sc_node *rq_next; | |||||
struct task *local_tasks[16]; | |||||
int local_tasks_count; | |||||
int final_tasks_count; | |||||
@@ -227,8 +227,14 @@ void process_runnable_tasks() | |||||
*/ | |||||
if (likely(t->process == process_stream)) | |||||
t = process_stream(t); | |||||
- else | |||||
- t = t->process(t); | |||||
+ else { | |||||
+ if (t->process != NULL) | |||||
+ t = t->process(t); | |||||
+ else { | |||||
+ __task_free(t); | |||||
+ t = NULL; | |||||
+ } | |||||
+ } | |||||
curr_task = NULL; | |||||
if (likely(t != NULL)) { | |||||
@@ -309,8 +315,14 @@ void process_runnable_tasks() | |||||
curr_task = t; | |||||
if (likely(t->process == process_stream)) | |||||
t = process_stream(t); | |||||
- else | |||||
- t = t->process(t); | |||||
+ else { | |||||
+ if (t->process != NULL) | |||||
+ t = t->process(t); | |||||
+ else { | |||||
+ __task_free(t); | |||||
+ t = NULL; | |||||
+ } | |||||
+ } | |||||
curr_task = NULL; | |||||
if (t) | |||||
local_tasks[final_tasks_count++] = t; |
@ -0,0 +1,48 @@ | |||||
commit 52ec3578c3ddc688ae14da3cd3e7e351494603d8 | |||||
Author: PiBa-NL <PiBa.NL.dev@gmail.com> | |||||
Date: Sat May 5 23:51:42 2018 +0200 | |||||
BUG/MINOR: lua: schedule socket task upon lua connect() | |||||
The parameters like server-address, port and timeout should be set before | |||||
process_stream task is called to avoid the stream being 'closed' before it | |||||
got initialized properly. This is most clearly visible when running with | |||||
tune.lua.forced-yield=1.. So scheduling the task should not be done when | |||||
creating the lua socket, but when connect is called. The error | |||||
"socket: not yet initialised, you can't set timeouts." would then appear. | |||||
Below code for example also shows this issue, as the sleep will | |||||
yield the lua code: | |||||
local con = core.tcp() | |||||
core.sleep(1) | |||||
con:settimeout(10) | |||||
(cherry picked from commit 706d5ee0c366787536213ccd6dea264d20b76a22) | |||||
[wt: must be backported to 1.7 and 1.6 as well with a different patch, | |||||
see https://www.mail-archive.com/haproxy@formilux.org/msg29924.html] | |||||
Signed-off-by: Willy Tarreau <w@1wt.eu> | |||||
diff --git a/src/hlua.c b/src/hlua.c | |||||
index 0100e7cf..5cc918c9 100644 | |||||
--- a/src/hlua.c | |||||
+++ b/src/hlua.c | |||||
@@ -2415,6 +2415,10 @@ __LJMP static int hlua_socket_connect(struct lua_State *L) | |||||
WILL_LJMP(luaL_error(L, "out of memory")); | |||||
} | |||||
xref_unlock(&socket->xref, peer); | |||||
+ | |||||
+ task_wakeup(s->task, TASK_WOKEN_INIT); | |||||
+ /* Return yield waiting for connection. */ | |||||
+ | |||||
WILL_LJMP(hlua_yieldk(L, 0, 0, hlua_socket_connect_yield, TICK_ETERNITY, 0)); | |||||
return 0; | |||||
@@ -2566,8 +2570,6 @@ __LJMP static int hlua_socket_new(lua_State *L) | |||||
strm->flags |= SF_DIRECT | SF_ASSIGNED | SF_ADDR_SET | SF_BE_ASSIGNED; | |||||
strm->target = &socket_tcp.obj_type; | |||||
- task_wakeup(strm->task, TASK_WOKEN_INIT); | |||||
- /* Return yield waiting for connection. */ | |||||
return 1; | |||||
out_fail_stream: |