[master] 3e2e3fb Name variables which point to struct vdp_entry_s "vdpe" to make life easier for patch in the pipeline.

Poul-Henning Kamp phk at FreeBSD.org
Mon Sep 25 18:20:06 UTC 2017


commit 3e2e3fbda283893b4a15e1984fe95a31a594f7b9
Author: Poul-Henning Kamp <phk at FreeBSD.org>
Date:   Mon Sep 25 17:43:15 2017 +0000

    Name variables which point to struct vdp_entry_s "vdpe" to make
    life easier for patch in the pipeline.

diff --git a/bin/varnishd/cache/cache.h b/bin/varnishd/cache/cache.h
index b6e6825..c228b69 100644
--- a/bin/varnishd/cache/cache.h
+++ b/bin/varnishd/cache/cache.h
@@ -574,9 +574,9 @@ struct req {
 	struct objcore		*stale_oc;
 
 	/* Deliver pipeline */
-	struct vdp_entry_s	vdp;
-	struct vdp_entry	*vdp_nxt;
-	unsigned		vdp_retval;
+	struct vdp_entry_s	vdpe;
+	struct vdp_entry	*vdpe_nxt;
+	unsigned		vdpe_retval;
 
 	/* Delivery mode */
 	unsigned		res_mode;
diff --git a/bin/varnishd/cache/cache_deliver_proc.c b/bin/varnishd/cache/cache_deliver_proc.c
index 5d3b23c..ad8f46f 100644
--- a/bin/varnishd/cache/cache_deliver_proc.c
+++ b/bin/varnishd/cache/cache_deliver_proc.c
@@ -39,7 +39,7 @@
  * This function picks and calls the next delivery processor from the
  * list. The return value is the return value of the delivery
  * processor. Upon seeing a non-zero return value, that lowest value
- * observed is latched in req->vdp_retval and all subsequent calls to
+ * observed is latched in req->vdpe_retval and all subsequent calls to
  * VDP_bytes will return that value directly without calling the next
  * processor.
  *
@@ -52,63 +52,63 @@ int
 VDP_bytes(struct req *req, enum vdp_action act, const void *ptr, ssize_t len)
 {
 	int retval;
-	struct vdp_entry *vdp;
+	struct vdp_entry *vdpe;
 
 	CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
 	assert(act == VDP_NULL || act == VDP_FLUSH);
-	if (req->vdp_retval)
-		return (req->vdp_retval);
-	vdp = req->vdp_nxt;
-	CHECK_OBJ_NOTNULL(vdp, VDP_ENTRY_MAGIC);
-	req->vdp_nxt = VTAILQ_NEXT(vdp, list);
+	if (req->vdpe_retval)
+		return (req->vdpe_retval);
+	vdpe = req->vdpe_nxt;
+	CHECK_OBJ_NOTNULL(vdpe, VDP_ENTRY_MAGIC);
+	req->vdpe_nxt = VTAILQ_NEXT(vdpe, list);
 
 	assert(act > VDP_NULL || len > 0);
 	/* Call the present layer, while pointing to the next layer down */
-	retval = vdp->func(req, act, &vdp->priv, ptr, len);
-	if (retval && (req->vdp_retval == 0 || retval < req->vdp_retval))
-		req->vdp_retval = retval; /* Latch error value */
-	req->vdp_nxt = vdp;
-	return (req->vdp_retval);
+	retval = vdpe->func(req, act, &vdpe->priv, ptr, len);
+	if (retval && (req->vdpe_retval == 0 || retval < req->vdpe_retval))
+		req->vdpe_retval = retval; /* Latch error value */
+	req->vdpe_nxt = vdpe;
+	return (req->vdpe_retval);
 }
 
 void
 VDP_push(struct req *req, vdp_bytes *func, void *priv, int bottom,
     const char *id)
 {
-	struct vdp_entry *vdp;
+	struct vdp_entry *vdpe;
 
 	CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
 	AN(func);
 
-	vdp = WS_Alloc(req->ws, sizeof *vdp);
-	if (vdp == NULL)
+	vdpe = WS_Alloc(req->ws, sizeof *vdpe);
+	if (vdpe == NULL)
 		return;
-	INIT_OBJ(vdp, VDP_ENTRY_MAGIC);
-	vdp->func = func;
-	vdp->priv = priv;
-	vdp->id = id;
+	INIT_OBJ(vdpe, VDP_ENTRY_MAGIC);
+	vdpe->func = func;
+	vdpe->priv = priv;
+	vdpe->id = id;
 	if (bottom)
-		VTAILQ_INSERT_TAIL(&req->vdp, vdp, list);
+		VTAILQ_INSERT_TAIL(&req->vdpe, vdpe, list);
 	else
-		VTAILQ_INSERT_HEAD(&req->vdp, vdp, list);
-	req->vdp_nxt = VTAILQ_FIRST(&req->vdp);
+		VTAILQ_INSERT_HEAD(&req->vdpe, vdpe, list);
+	req->vdpe_nxt = VTAILQ_FIRST(&req->vdpe);
 
-	AZ(vdp->func(req, VDP_INIT, &vdp->priv, NULL, 0));
+	AZ(vdpe->func(req, VDP_INIT, &vdpe->priv, NULL, 0));
 }
 
 static void
 vdp_pop(struct req *req, vdp_bytes *func)
 {
-	struct vdp_entry *vdp;
+	struct vdp_entry *vdpe;
 	CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
 
-	vdp = VTAILQ_FIRST(&req->vdp);
-	CHECK_OBJ_NOTNULL(vdp, VDP_ENTRY_MAGIC);
-	assert(vdp->func == func);
-	VTAILQ_REMOVE(&req->vdp, vdp, list);
-	AZ(vdp->func(req, VDP_FINI, &vdp->priv, NULL, 0));
-	AZ(vdp->priv);
-	req->vdp_nxt = VTAILQ_FIRST(&req->vdp);
+	vdpe = VTAILQ_FIRST(&req->vdpe);
+	CHECK_OBJ_NOTNULL(vdpe, VDP_ENTRY_MAGIC);
+	assert(vdpe->func == func);
+	VTAILQ_REMOVE(&req->vdpe, vdpe, list);
+	AZ(vdpe->func(req, VDP_FINI, &vdpe->priv, NULL, 0));
+	AZ(vdpe->priv);
+	req->vdpe_nxt = VTAILQ_FIRST(&req->vdpe);
 }
 
 void
@@ -116,8 +116,8 @@ VDP_close(struct req *req)
 {
 
 	CHECK_OBJ_NOTNULL(req, REQ_MAGIC);
-	while (!VTAILQ_EMPTY(&req->vdp))
-		vdp_pop(req, VTAILQ_FIRST(&req->vdp)->func);
+	while (!VTAILQ_EMPTY(&req->vdpe))
+		vdp_pop(req, VTAILQ_FIRST(&req->vdpe)->func);
 }
 
 /*--------------------------------------------------------------------*/
diff --git a/bin/varnishd/cache/cache_gzip.c b/bin/varnishd/cache/cache_gzip.c
index 4370866..4c24e4b 100644
--- a/bin/varnishd/cache/cache_gzip.c
+++ b/bin/varnishd/cache/cache_gzip.c
@@ -327,7 +327,7 @@ VDP_gunzip(struct req *req, enum vdp_action act, void **priv,
 		 * If the size is non-zero AND we are the top
 		 * VDP (ie: no ESI), we know what size the output will be.
 		 */
-		if (u != 0 && VTAILQ_FIRST(&req->vdp)->func == VDP_gunzip)
+		if (u != 0 && VTAILQ_FIRST(&req->vdpe)->func == VDP_gunzip)
 			req->resp_len = u;
 
 		return (0);
@@ -355,7 +355,7 @@ VDP_gunzip(struct req *req, enum vdp_action act, void **priv,
 			return (-1);
 		if (vg->m_len == vg->m_sz || vr != VGZ_OK) {
 			if (VDP_bytes(req, VDP_FLUSH, vg->m_buf, vg->m_len))
-				return (req->vdp_retval);
+				return (req->vdpe_retval);
 			vg->m_len = 0;
 			VGZ_Obuf(vg, vg->m_buf, vg->m_sz);
 		}
diff --git a/bin/varnishd/cache/cache_req.c b/bin/varnishd/cache/cache_req.c
index 2d06616..56ed74c 100644
--- a/bin/varnishd/cache/cache_req.c
+++ b/bin/varnishd/cache/cache_req.c
@@ -132,8 +132,8 @@ Req_New(const struct worker *wrk, struct sess *sp)
 	req->t_prev = NAN;
 	req->t_req = NAN;
 
-	req->vdp_nxt = 0;
-	VTAILQ_INIT(&req->vdp);
+	req->vdpe_nxt = 0;
+	VTAILQ_INIT(&req->vdpe);
 
 	return (req);
 }
diff --git a/bin/varnishd/cache/cache_req_fsm.c b/bin/varnishd/cache/cache_req_fsm.c
index 9e1fbe5..a0caba3 100644
--- a/bin/varnishd/cache/cache_req_fsm.c
+++ b/bin/varnishd/cache/cache_req_fsm.c
@@ -793,7 +793,7 @@ cnt_recv(struct worker *wrk, struct req *req)
 	req->director_hint = VCL_DefaultDirector(req->vcl);
 	AN(req->director_hint);
 
-	req->vdp_retval = 0;
+	req->vdpe_retval = 0;
 	req->d_ttl = -1;
 	req->disable_esi = 0;
 	req->hash_always_miss = 0;
diff --git a/bin/varnishd/http1/cache_http1_deliver.c b/bin/varnishd/http1/cache_http1_deliver.c
index e6a6845..b335471 100644
--- a/bin/varnishd/http1/cache_http1_deliver.c
+++ b/bin/varnishd/http1/cache_http1_deliver.c
@@ -46,7 +46,7 @@ v1d_bytes(struct req *req, enum vdp_action act, void **priv,
 	if (act == VDP_INIT || act == VDP_FINI)
 		return (0);
 
-	AZ(req->vdp_nxt);		/* always at the bottom of the pile */
+	AZ(req->vdpe_nxt);		/* always at the bottom of the pile */
 
 	if (len > 0)
 		wl = V1L_Write(req->wrk, ptr, len);


More information about the varnish-commit mailing list