[master] 3e2e3fb Name variables which point to struct vdp_entry_s "vdpe" to make life easier for patch in the pipeline.
Poul-Henning Kamp
phk at FreeBSD.org
Mon Sep 25 18:20:06 UTC 2017
commit 3e2e3fbda283893b4a15e1984fe95a31a594f7b9
Author: Poul-Henning Kamp <phk at FreeBSD.org>
Date: Mon Sep 25 17:43:15 2017 +0000
Name variables which point to struct vdp_entry_s "vdpe" to make
life easier for patch in the pipeline.
diff --git a/bin/varnishd/cache/cache.h b/bin/varnishd/cache/cache.h
index b6e6825..c228b69 100644
--- a/bin/varnishd/cache/cache.h
+++ b/bin/varnishd/cache/cache.h
@@ -574,9 +574,9 @@ struct req {
struct objcore *stale_oc;
/* Deliver pipeline */
- struct vdp_entry_s vdp;
- struct vdp_entry *vdp_nxt;
- unsigned vdp_retval;
+ struct vdp_entry_s vdpe;
+ struct vdp_entry *vdpe_nxt;
+ unsigned vdpe_retval;
/* Delivery mode */
unsigned res_mode;
diff --git a/bin/varnishd/cache/cache_deliver_proc.c b/bin/varnishd/cache/cache_deliver_proc.c
index 5d3b23c..ad8f46f 100644
--- a/bin/varnishd/cache/cache_deliver_proc.c
+++ b/bin/varnishd/cache/cache_deliver_proc.c
@@ -39,7 +39,7 @@
* This function picks and calls the next delivery processor from the
* list. The return value is the return value of the delivery
* processor. Upon seeing a non-zero return value, that lowest value
- * observed is latched in req->vdp_retval and all subsequent calls to
+ * observed is latched in req->vdpe_retval and all subsequent calls to
* VDP_bytes will return that value directly without calling the next
* processor.
*
@@ -52,63 +52,63 @@ int
VDP_bytes(struct req *req, enum vdp_action act, const void *ptr, ssize_t len)
{
int retval;
- struct vdp_entry *vdp;
+ struct vdp_entry *vdpe;
CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
assert(act == VDP_NULL || act == VDP_FLUSH);
- if (req->vdp_retval)
- return (req->vdp_retval);
- vdp = req->vdp_nxt;
- CHECK_OBJ_NOTNULL(vdp, VDP_ENTRY_MAGIC);
- req->vdp_nxt = VTAILQ_NEXT(vdp, list);
+ if (req->vdpe_retval)
+ return (req->vdpe_retval);
+ vdpe = req->vdpe_nxt;
+ CHECK_OBJ_NOTNULL(vdpe, VDP_ENTRY_MAGIC);
+ req->vdpe_nxt = VTAILQ_NEXT(vdpe, list);
assert(act > VDP_NULL || len > 0);
/* Call the present layer, while pointing to the next layer down */
- retval = vdp->func(req, act, &vdp->priv, ptr, len);
- if (retval && (req->vdp_retval == 0 || retval < req->vdp_retval))
- req->vdp_retval = retval; /* Latch error value */
- req->vdp_nxt = vdp;
- return (req->vdp_retval);
+ retval = vdpe->func(req, act, &vdpe->priv, ptr, len);
+ if (retval && (req->vdpe_retval == 0 || retval < req->vdpe_retval))
+ req->vdpe_retval = retval; /* Latch error value */
+ req->vdpe_nxt = vdpe;
+ return (req->vdpe_retval);
}
void
VDP_push(struct req *req, vdp_bytes *func, void *priv, int bottom,
const char *id)
{
- struct vdp_entry *vdp;
+ struct vdp_entry *vdpe;
CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
AN(func);
- vdp = WS_Alloc(req->ws, sizeof *vdp);
- if (vdp == NULL)
+ vdpe = WS_Alloc(req->ws, sizeof *vdpe);
+ if (vdpe == NULL)
return;
- INIT_OBJ(vdp, VDP_ENTRY_MAGIC);
- vdp->func = func;
- vdp->priv = priv;
- vdp->id = id;
+ INIT_OBJ(vdpe, VDP_ENTRY_MAGIC);
+ vdpe->func = func;
+ vdpe->priv = priv;
+ vdpe->id = id;
if (bottom)
- VTAILQ_INSERT_TAIL(&req->vdp, vdp, list);
+ VTAILQ_INSERT_TAIL(&req->vdpe, vdpe, list);
else
- VTAILQ_INSERT_HEAD(&req->vdp, vdp, list);
- req->vdp_nxt = VTAILQ_FIRST(&req->vdp);
+ VTAILQ_INSERT_HEAD(&req->vdpe, vdpe, list);
+ req->vdpe_nxt = VTAILQ_FIRST(&req->vdpe);
- AZ(vdp->func(req, VDP_INIT, &vdp->priv, NULL, 0));
+ AZ(vdpe->func(req, VDP_INIT, &vdpe->priv, NULL, 0));
}
static void
vdp_pop(struct req *req, vdp_bytes *func)
{
- struct vdp_entry *vdp;
+ struct vdp_entry *vdpe;
CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
- vdp = VTAILQ_FIRST(&req->vdp);
- CHECK_OBJ_NOTNULL(vdp, VDP_ENTRY_MAGIC);
- assert(vdp->func == func);
- VTAILQ_REMOVE(&req->vdp, vdp, list);
- AZ(vdp->func(req, VDP_FINI, &vdp->priv, NULL, 0));
- AZ(vdp->priv);
- req->vdp_nxt = VTAILQ_FIRST(&req->vdp);
+ vdpe = VTAILQ_FIRST(&req->vdpe);
+ CHECK_OBJ_NOTNULL(vdpe, VDP_ENTRY_MAGIC);
+ assert(vdpe->func == func);
+ VTAILQ_REMOVE(&req->vdpe, vdpe, list);
+ AZ(vdpe->func(req, VDP_FINI, &vdpe->priv, NULL, 0));
+ AZ(vdpe->priv);
+ req->vdpe_nxt = VTAILQ_FIRST(&req->vdpe);
}
void
@@ -116,8 +116,8 @@ VDP_close(struct req *req)
{
CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
- while (!VTAILQ_EMPTY(&req->vdp))
- vdp_pop(req, VTAILQ_FIRST(&req->vdp)->func);
+ while (!VTAILQ_EMPTY(&req->vdpe))
+ vdp_pop(req, VTAILQ_FIRST(&req->vdpe)->func);
}
/*--------------------------------------------------------------------*/
diff --git a/bin/varnishd/cache/cache_gzip.c b/bin/varnishd/cache/cache_gzip.c
index 4370866..4c24e4b 100644
--- a/bin/varnishd/cache/cache_gzip.c
+++ b/bin/varnishd/cache/cache_gzip.c
@@ -327,7 +327,7 @@ VDP_gunzip(struct req *req, enum vdp_action act, void **priv,
* If the size is non-zero AND we are the top
* VDP (ie: no ESI), we know what size the output will be.
*/
- if (u != 0 && VTAILQ_FIRST(&req->vdp)->func == VDP_gunzip)
+ if (u != 0 && VTAILQ_FIRST(&req->vdpe)->func == VDP_gunzip)
req->resp_len = u;
return (0);
@@ -355,7 +355,7 @@ VDP_gunzip(struct req *req, enum vdp_action act, void **priv,
return (-1);
if (vg->m_len == vg->m_sz || vr != VGZ_OK) {
if (VDP_bytes(req, VDP_FLUSH, vg->m_buf, vg->m_len))
- return (req->vdp_retval);
+ return (req->vdpe_retval);
vg->m_len = 0;
VGZ_Obuf(vg, vg->m_buf, vg->m_sz);
}
diff --git a/bin/varnishd/cache/cache_req.c b/bin/varnishd/cache/cache_req.c
index 2d06616..56ed74c 100644
--- a/bin/varnishd/cache/cache_req.c
+++ b/bin/varnishd/cache/cache_req.c
@@ -132,8 +132,8 @@ Req_New(const struct worker *wrk, struct sess *sp)
req->t_prev = NAN;
req->t_req = NAN;
- req->vdp_nxt = 0;
- VTAILQ_INIT(&req->vdp);
+ req->vdpe_nxt = 0;
+ VTAILQ_INIT(&req->vdpe);
return (req);
}
diff --git a/bin/varnishd/cache/cache_req_fsm.c b/bin/varnishd/cache/cache_req_fsm.c
index 9e1fbe5..a0caba3 100644
--- a/bin/varnishd/cache/cache_req_fsm.c
+++ b/bin/varnishd/cache/cache_req_fsm.c
@@ -793,7 +793,7 @@ cnt_recv(struct worker *wrk, struct req *req)
req->director_hint = VCL_DefaultDirector(req->vcl);
AN(req->director_hint);
- req->vdp_retval = 0;
+ req->vdpe_retval = 0;
req->d_ttl = -1;
req->disable_esi = 0;
req->hash_always_miss = 0;
diff --git a/bin/varnishd/http1/cache_http1_deliver.c b/bin/varnishd/http1/cache_http1_deliver.c
index e6a6845..b335471 100644
--- a/bin/varnishd/http1/cache_http1_deliver.c
+++ b/bin/varnishd/http1/cache_http1_deliver.c
@@ -46,7 +46,7 @@ v1d_bytes(struct req *req, enum vdp_action act, void **priv,
if (act == VDP_INIT || act == VDP_FINI)
return (0);
- AZ(req->vdp_nxt); /* always at the bottom of the pile */
+ AZ(req->vdpe_nxt); /* always at the bottom of the pile */
if (len > 0)
wl = V1L_Write(req->wrk, ptr, len);
More information about the varnish-commit
mailing list