From mboxrd@z Thu Jan 1 00:00:00 1970 From: Matthias Fischer To: development@lists.ipfire.org Subject: [PATCH] squid 4.5: latest patches (01-09) Date: Mon, 04 Feb 2019 18:54:35 +0100 Message-ID: <20190204175435.8456-1-matthias.fischer@ipfire.org> MIME-Version: 1.0 Content-Type: multipart/mixed; boundary="===============3786999699197001518==" List-Id: --===============3786999699197001518== Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable For details see: http://www.squid-cache.org/Versions/v4/changesets/ Best, Matthias Signed-off-by: Matthias Fischer --- lfs/squid | 11 +- ...ile_errors_with_-O3_optimization_288.patch | 107 ++++++++ ...GoIntoBackground_fork_call_fails_344.patch | 29 +++ ...associated_with_auto-consumption_348.patch | 118 +++++++++ ...ds_that_define_OPENSSL_NO_ENGINE_349.patch | 22 ++ ...disker-to-worker_queue_overflows_353.patch | 92 +++++++ ..._when_reserving_a_new_cache_slot_350.patch | 246 ++++++++++++++++++ ...opped_some_of_the_write_requests_352.patch | 164 ++++++++++++ ...call_setsid_in_--foreground_mode_354.patch | 36 +++ ...ect_IPv6_loopback_binding_errors_355.patch | 54 ++++ 10 files changed, 878 insertions(+), 1 deletion(-) create mode 100644 src/patches/squid/01_Bug_4875_pt2_GCC-8_compile_errors_wi= th_-O3_optimization_288.patch create mode 100644 src/patches/squid/02_Bug_4856_Exit_when_GoIntoBackground_= fork_call_fails_344.patch create mode 100644 src/patches/squid/03_Fix_BodyPipe_Sink_memory_leaks_assoc= iated_with_auto-consumption_348.patch create mode 100644 src/patches/squid/04_Fix_OpenSSL_builds_that_define_OPENS= SL_NO_ENGINE_349.patch create mode 100644 src/patches/squid/05_Fixed_disker-to-worker_queue_overflo= ws_353.patch create mode 100644 src/patches/squid/06_Initialize_StoreMapSlice_when_reserv= ing_a_new_cache_slot_350.patch create mode 100644 src/patches/squid/07_Fail_Rock_swapout_if_the_disk_droppe= d_some_of_the_write_requests_352.patch create mode 100644 src/patches/squid/08_Bug_4914_Do_not_call_setsid_in_--for= eground_mode_354.patch create mode 100644 src/patches/squid/09_Bug_4915_Detect_IPv6_loopback_bindin= g_errors_355.patch diff --git a/lfs/squid b/lfs/squid index 6033ab394..6c4706240 100644 --- a/lfs/squid +++ b/lfs/squid @@ -72,6 +72,15 @@ $(subst %,%_MD5,$(objects)) : $(TARGET) : $(patsubst %,$(DIR_DL)/%,$(objects)) @$(PREBUILD) @rm -rf $(DIR_APP) && cd $(DIR_SRC) && tar xaf $(DIR_DL)/$(DL_FILE) + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/01_Bug_4875_pt2= _GCC-8_compile_errors_with_-O3_optimization_288.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/02_Bug_4856_Exi= t_when_GoIntoBackground_fork_call_fails_344.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/03_Fix_BodyPipe= _Sink_memory_leaks_associated_with_auto-consumption_348.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/04_Fix_OpenSSL_= builds_that_define_OPENSSL_NO_ENGINE_349.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/05_Fixed_disker= -to-worker_queue_overflows_353.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/06_Initialize_S= toreMapSlice_when_reserving_a_new_cache_slot_350.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/07_Fail_Rock_sw= apout_if_the_disk_dropped_some_of_the_write_requests_352.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/08_Bug_4914_Do_= not_call_setsid_in_--foreground_mode_354.patch + cd $(DIR_APP) && patch -Np1 -i $(DIR_SRC)/src/patches/squid/09_Bug_4915_Det= ect_IPv6_loopback_binding_errors_355.patch cd $(DIR_APP) && patch -Np0 -i $(DIR_SRC)/src/patches/squid/squid-4.5-fix-m= ax-file-descriptors.patch =20 cd $(DIR_APP) && autoreconf -vfi @@ -91,7 +100,7 @@ $(TARGET) : $(patsubst %,$(DIR_DL)/%,$(objects)) --disable-kqueue \ --disable-esi \ --disable-arch-native \ - --enable-ipv6 \ + --disable-ipv6 \ --enable-poll \ --enable-ident-lookups \ --enable-storeio=3Daufs,diskd,ufs \ diff --git a/src/patches/squid/01_Bug_4875_pt2_GCC-8_compile_errors_with_-O3_= optimization_288.patch b/src/patches/squid/01_Bug_4875_pt2_GCC-8_compile_erro= rs_with_-O3_optimization_288.patch new file mode 100644 index 000000000..1e3b731a0 --- /dev/null +++ b/src/patches/squid/01_Bug_4875_pt2_GCC-8_compile_errors_with_-O3_optimiz= ation_288.patch @@ -0,0 +1,107 @@ +commit d6adda4644e614dcaa6831abc5efffb2af51a7c8 +Author: Amos Jeffries +Date: 2019-01-06 13:22:19 +0000 + + Bug 4875 pt2: GCC-8 compile errors with -O3 optimization (#288) + =20 + GCC-8 warnings exposed at -O3 optimization causes its + own static analyzer to detect optimized code is eliding + initialization on paths that do not use the + configuration variables. + =20 + Refactor the parseTimeLine() API to return the parsed + values so that there is no need to initialize anything prior + to parsing. + +diff --git a/src/cache_cf.cc b/src/cache_cf.cc +index c4eb955..a45ea2e 100644 +--- a/src/cache_cf.cc ++++ b/src/cache_cf.cc +@@ -160,7 +160,6 @@ static bool setLogformat(CustomLog *cl, const char *name= , const bool dieWhenMiss + static void configDoConfigure(void); + static void parse_refreshpattern(RefreshPattern **); + static uint64_t parseTimeUnits(const char *unit, bool allowMsec); +-static void parseTimeLine(time_msec_t * tptr, const char *units, bool allow= Msec, bool expectMoreArguments); + static void parse_u_short(unsigned short * var); + static void parse_string(char **); + static void default_all(void); +@@ -1034,19 +1033,19 @@ parse_obsolete(const char *name) +=20 + /* Parse a time specification from the config file. Store the + * result in 'tptr', after converting it to 'units' */ +-static void +-parseTimeLine(time_msec_t * tptr, const char *units, bool allowMsec, bool= expectMoreArguments =3D false) ++static time_msec_t ++parseTimeLine(const char *units, bool allowMsec, bool expectMoreArguments= =3D false) + { + time_msec_t u =3D parseTimeUnits(units, allowMsec); + if (u =3D=3D 0) { + self_destruct(); +- return; ++ return 0; + } +=20 +- char *token =3D ConfigParser::NextToken();; ++ char *token =3D ConfigParser::NextToken(); + if (!token) { + self_destruct(); +- return; ++ return 0; + } +=20 + double d =3D xatof(token); +@@ -1059,7 +1058,7 @@ parseTimeLine(time_msec_t * tptr, const char *units, = bool allowMsec, bool expe +=20 + } else if (!expectMoreArguments) { + self_destruct(); +- return; ++ return 0; +=20 + } else { + token =3D NULL; // show default units if dying below +@@ -1068,13 +1067,14 @@ parseTimeLine(time_msec_t * tptr, const char *units,= bool allowMsec, bool expe + } else + token =3D NULL; // show default units if dying below. +=20 +- *tptr =3D static_cast(m * d); ++ const auto result =3D static_cast(m * d); +=20 +- if (static_cast(*tptr) * 2 !=3D m * d * 2) { ++ if (static_cast(result) * 2 !=3D m * d * 2) { + debugs(3, DBG_CRITICAL, "FATAL: Invalid value '" << + d << " " << (token ? token : units) << ": integer overflow (= time_msec_t)."); + self_destruct(); + } ++ return result; + } +=20 + static uint64_t +@@ -2918,8 +2918,7 @@ dump_time_t(StoreEntry * entry, const char *name, time= _t var) + void + parse_time_t(time_t * var) + { +- time_msec_t tval; +- parseTimeLine(&tval, T_SECOND_STR, false); ++ time_msec_t tval =3D parseTimeLine(T_SECOND_STR, false); + *var =3D static_cast(tval/1000); + } +=20 +@@ -2941,7 +2940,7 @@ dump_time_msec(StoreEntry * entry, const char *name, t= ime_msec_t var) + void + parse_time_msec(time_msec_t * var) + { +- parseTimeLine(var, T_SECOND_STR, true); ++ *var =3D parseTimeLine(T_SECOND_STR, true); + } +=20 + static void +@@ -4814,8 +4813,7 @@ static void free_ftp_epsv(acl_access **ftp_epsv) + static void + parse_UrlHelperTimeout(SquidConfig::UrlHelperTimeout *config) + { +- time_msec_t tval; +- parseTimeLine(&tval, T_SECOND_STR, false, true); ++ const auto tval =3D parseTimeLine(T_SECOND_STR, false, true); + Config.Timeout.urlRewrite =3D static_cast(tval/1000); +=20 + char *key, *value; diff --git a/src/patches/squid/02_Bug_4856_Exit_when_GoIntoBackground_fork_ca= ll_fails_344.patch b/src/patches/squid/02_Bug_4856_Exit_when_GoIntoBackground= _fork_call_fails_344.patch new file mode 100644 index 000000000..3159b9d1c --- /dev/null +++ b/src/patches/squid/02_Bug_4856_Exit_when_GoIntoBackground_fork_call_fail= s_344.patch @@ -0,0 +1,29 @@ +commit 6aa83b8d51dbb66bf5383f18615a59d8ff0a10a4 +Author: Marcos Mello +Date: 2019-01-08 10:14:09 +0000 + + Bug 4856: Exit when GoIntoBackground() fork() call fails (#344) + =20 + Ignoring the (implicit/default) command-line instruction to background + Squid is wrong in general and confuses at least some daemon managers. + +diff --git a/src/main.cc b/src/main.cc +index eeb8f72..7f5ec80 100644 +--- a/src/main.cc ++++ b/src/main.cc +@@ -1865,13 +1865,12 @@ GoIntoBackground() + pid_t pid; + if ((pid =3D fork()) < 0) { + int xerrno =3D errno; +- syslog(LOG_ALERT, "fork failed: %s", xstrerr(xerrno)); +- // continue anyway, mimicking --foreground mode (XXX?) ++ throw TexcHere(ToSBuf("failed to fork(2) the master process: ", xst= rerr(xerrno))); + } else if (pid > 0) { + // parent + exit(EXIT_SUCCESS); + } +- // child, running as a background daemon (or a failed-to-fork parent) ++ // child, running as a background daemon + } +=20 + static void diff --git a/src/patches/squid/03_Fix_BodyPipe_Sink_memory_leaks_associated_w= ith_auto-consumption_348.patch b/src/patches/squid/03_Fix_BodyPipe_Sink_memor= y_leaks_associated_with_auto-consumption_348.patch new file mode 100644 index 000000000..efb0d4c69 --- /dev/null +++ b/src/patches/squid/03_Fix_BodyPipe_Sink_memory_leaks_associated_with_aut= o-consumption_348.patch @@ -0,0 +1,118 @@ +commit dd9d539a35275071d7718cfbdbc3203d7aa05139 +Author: Alex Rousskov +Date: 2019-01-08 15:14:18 +0000 + + Fix BodyPipe/Sink memory leaks associated with auto-consumption (#348) + =20 + Auto-consumption happens (and could probably leak memory) in many cases, + but this leak was exposed by an eCAP service that blocked or replaced + virgin messages. + =20 + The BodySink job termination algorithm relies on body production + notifications. A BodySink job created after the body production had + ended can never stop and, hence, leaks (leaking the associated BodyPipe + object with it). Such a job is also useless: If production is over, + there is no need to free space for more body data! This change avoids + creating such leaking and useless jobs. + +diff --git a/src/BodyPipe.cc b/src/BodyPipe.cc +index 0fdf0dc..59477bd 100644 +--- a/src/BodyPipe.cc ++++ b/src/BodyPipe.cc +@@ -286,8 +286,7 @@ BodyPipe::expectNoConsumption() + abortedConsumption =3D true; +=20 + // in case somebody enabled auto-consumption before regular one abo= rted +- if (mustAutoConsume) +- startAutoConsumption(); ++ startAutoConsumptionIfNeeded(); + } + } +=20 +@@ -313,23 +312,28 @@ BodyPipe::consume(size_t size) + postConsume(size); + } +=20 +-// In the AutoConsumption mode the consumer has gone but the producer cont= inues +-// producing data. We are using a BodySink BodyConsumer which just discards= the produced data. + void + BodyPipe::enableAutoConsumption() + { + mustAutoConsume =3D true; + debugs(91,5, HERE << "enabled auto consumption" << status()); +- if (!theConsumer && theBuf.hasContent()) +- startAutoConsumption(); ++ startAutoConsumptionIfNeeded(); + } +=20 +-// start auto consumption by creating body sink ++/// Check the current need and, if needed, start auto consumption. In auto ++/// consumption mode, the consumer is gone, but the producer continues to ++/// produce data. We use a BodySink BodyConsumer to discard that data. + void +-BodyPipe::startAutoConsumption() ++BodyPipe::startAutoConsumptionIfNeeded() + { +- Must(mustAutoConsume); +- Must(!theConsumer); ++ const auto startNow =3D ++ mustAutoConsume && // was enabled ++ !theConsumer && // has not started yet ++ theProducer.valid() && // still useful (and will eventually stop) ++ theBuf.hasContent(); // has something to consume right now ++ if (!startNow) ++ return; ++ + theConsumer =3D new BodySink(this); + debugs(91,7, HERE << "starting auto consumption" << status()); + scheduleBodyDataNotification(); +@@ -393,15 +397,14 @@ BodyPipe::postAppend(size_t size) + thePutSize +=3D size; + debugs(91,7, HERE << "added " << size << " bytes" << status()); +=20 +- if (mustAutoConsume && !theConsumer && size > 0) +- startAutoConsumption(); ++ if (!mayNeedMoreData()) ++ clearProducer(true); // reached end-of-body +=20 + // We should not consume here even if mustAutoConsume because the + // caller may not be ready for the data to be consumed during this call. + scheduleBodyDataNotification(); +=20 +- if (!mayNeedMoreData()) +- clearProducer(true); // reached end-of-body ++ startAutoConsumptionIfNeeded(); + } +=20 + void +diff --git a/src/BodyPipe.h b/src/BodyPipe.h +index 2bf3220..7061c78 100644 +--- a/src/BodyPipe.h ++++ b/src/BodyPipe.h +@@ -131,7 +131,7 @@ public: + bool exhausted() const; // saw eof/abort and all data consumed + bool stillConsuming(const Consumer::Pointer &consumer) const { return t= heConsumer =3D=3D consumer; } +=20 +- // start or continue consuming when there is no consumer ++ /// start or continue consuming when producing without consumer + void enableAutoConsumption(); +=20 + const MemBuf &buf() const { return theBuf; } +@@ -151,7 +151,7 @@ protected: + void postConsume(size_t size); + void postAppend(size_t size); +=20 +- void startAutoConsumption(); // delayed start of enabled consumption ++ void startAutoConsumptionIfNeeded(); +=20 + private: + int64_t theBodySize; // expected total content length, if known +@@ -163,7 +163,7 @@ private: +=20 + MemBuf theBuf; // produced but not yet consumed content, if any +=20 +- bool mustAutoConsume; // consume when there is no consumer ++ bool mustAutoConsume; ///< keep theBuf empty when producing without con= sumer + bool abortedConsumption; ///< called BodyProducer::noteBodyConsumerAbor= ted + bool isCheckedOut; // to keep track of checkout violations + }; diff --git a/src/patches/squid/04_Fix_OpenSSL_builds_that_define_OPENSSL_NO_E= NGINE_349.patch b/src/patches/squid/04_Fix_OpenSSL_builds_that_define_OPENSSL= _NO_ENGINE_349.patch new file mode 100644 index 000000000..c868d585c --- /dev/null +++ b/src/patches/squid/04_Fix_OpenSSL_builds_that_define_OPENSSL_NO_ENGINE_3= 49.patch @@ -0,0 +1,22 @@ +commit 186fc1d7c3bad845882d8e3497af9c3a42d75e8f +Author: Rosen Penev +Date: 2019-01-09 17:42:21 +0000 + + Fix OpenSSL builds that define OPENSSL_NO_ENGINE (#349) + =20 + Even with ENGINE support disabled, OpenSSL provides the openssl/engine.h + header. We have to use OPENSSL_NO_ENGINE to detect ENGINE support. + +diff --git a/src/ssl/support.cc b/src/ssl/support.cc +index f6d4ce7..e350a73 100644 +--- a/src/ssl/support.cc ++++ b/src/ssl/support.cc +@@ -485,7 +485,7 @@ Ssl::Initialize(void) +=20 + SQUID_OPENSSL_init_ssl(); +=20 +-#if HAVE_OPENSSL_ENGINE_H ++#if !defined(OPENSSL_NO_ENGINE) + if (::Config.SSL.ssl_engine) { + ENGINE_load_builtin_engines(); + ENGINE *e; diff --git a/src/patches/squid/05_Fixed_disker-to-worker_queue_overflows_353.= patch b/src/patches/squid/05_Fixed_disker-to-worker_queue_overflows_353.patch new file mode 100644 index 000000000..1b656549f --- /dev/null +++ b/src/patches/squid/05_Fixed_disker-to-worker_queue_overflows_353.patch @@ -0,0 +1,92 @@ +commit 4ce035593f5257ee2deb6ca85c7b9a0c4b850f70 +Author: Eduard Bagdasaryan +Date: 2019-01-12 10:05:30 +0000 + + Fixed disker-to-worker queue overflows (#353) + =20 + Before this fix, Squid sometimes logged the following error: + =20 + BUG: Worker I/O pop queue for ... overflow: ... + =20 + The bug could result in truncated hit responses, reduced hit ratio, and, + combined with buggy lost I/O handling code (GitHub PR #352), even cache + corruption. + =20 + The bug could be triggered by the following sequence of events: + =20 + * Disker dequeues one I/O request from the worker push queue. + * Worker pushes more I/O requests to that disker, reaching 1024 requests + in its push queue (QueueCapacity or just "N" below). No overflow here! + * Worker process is suspended (or is just too busy to pop I/O results). + * Disker satisfies all 1+N requests, adding each to the worker pop queue + and overflows that queue when adding the last processed request. + =20 + This fix limits worker push so that the sum of all pending requests + never exceeds (pop) queue capacity. This approach will continue to work + even if diskers are enhanced to dequeue multiple requests for seek + optimization and/or priority-based scheduling. + =20 + Pop queue and push queue can still accommodate N requests each. The fix + appears to reduce supported disker "concurrency" levels from 2N down to + N pending I/O requests, reducing queue memory utilization. However, the + actual reduction is from N+1 to N: Since a worker pops all its satisfied + requests before queuing a new one, there could never be more than N+1 + pending requests (N in the push queue and 1 worked on by the disker). + =20 + We left the BUG reporting and handling intact. There are no known bugs + in that code now. If the bug never surfaces again, it can be replaced + with code that translates low-level queue overflow exception into a + user-friendly TextException. + +diff --git a/src/DiskIO/IpcIo/IpcIoFile.cc b/src/DiskIO/IpcIo/IpcIoFile.cc +index 735cf0f..95d6d8b 100644 +--- a/src/DiskIO/IpcIo/IpcIoFile.cc ++++ b/src/DiskIO/IpcIo/IpcIoFile.cc +@@ -364,6 +364,10 @@ IpcIoFile::push(IpcIoPendingRequest *const pending) +=20 + debugs(47, 7, HERE << "pushing " << SipcIo(KidIdentifier, ipcIo, di= skId)); +=20 ++ // protect DiskerHandleRequest() from pop queue overflow ++ if (pendingRequests() >=3D QueueCapacity) ++ throw Ipc::OneToOneUniQueue::Full(); ++ + if (queue->push(diskId, ipcIo)) + Notify(diskId); // must notify disker + trackPendingRequest(ipcIo.requestId, pending); +@@ -879,10 +883,8 @@ IpcIoFile::DiskerHandleRequest(const int workerId, IpcI= oMsg &ipcIo) + if (queue->push(workerId, ipcIo)) + Notify(workerId); // must notify worker + } catch (const Queue::Full &) { +- // The worker queue should not overflow because the worker should p= op() +- // before push()ing and because if disker pops N requests at a time, +- // we should make sure the worker pop() queue length is the worker +- // push queue length plus N+1. XXX: implement the N+1 difference. ++ // The worker pop queue should not overflow because the worker can ++ // push only if pendingRequests() is less than QueueCapacity. + debugs(47, DBG_IMPORTANT, "BUG: Worker I/O pop queue for " << + DbName << " overflow: " << + SipcIo(workerId, ipcIo, KidIdentifier)); // TODO: report que= ue len +diff --git a/src/DiskIO/IpcIo/IpcIoFile.h b/src/DiskIO/IpcIo/IpcIoFile.h +index f89197e..53a3749 100644 +--- a/src/DiskIO/IpcIo/IpcIoFile.h ++++ b/src/DiskIO/IpcIo/IpcIoFile.h +@@ -55,6 +55,8 @@ public: +=20 + class IpcIoPendingRequest; +=20 ++/// In a worker process, represents a single (remote) cache_dir disker file. ++/// In a disker process, used as a bunch of static methods handling that fi= le. + class IpcIoFile: public DiskFile + { + CBDATA_CLASS(IpcIoFile); +@@ -98,6 +100,10 @@ private: + void push(IpcIoPendingRequest *const pending); + IpcIoPendingRequest *dequeueRequest(const unsigned int requestId); +=20 ++ /// the total number of I/O requests in push queue and pop queue ++ /// (but no, the implementation does not add push and pop queue sizes) ++ size_t pendingRequests() const { return olderRequests->size() + newerRe= quests->size(); } ++ + static void Notify(const int peerId); +=20 + static void OpenTimeout(void *const param); diff --git a/src/patches/squid/06_Initialize_StoreMapSlice_when_reserving_a_n= ew_cache_slot_350.patch b/src/patches/squid/06_Initialize_StoreMapSlice_when_= reserving_a_new_cache_slot_350.patch new file mode 100644 index 000000000..5f04bffce --- /dev/null +++ b/src/patches/squid/06_Initialize_StoreMapSlice_when_reserving_a_new_cach= e_slot_350.patch @@ -0,0 +1,246 @@ +commit 8e022a61922053764fd1627ba4b65a1617a1bd8a +Author: Eduard Bagdasaryan +Date: 2019-01-23 03:51:12 +0000 + + Initialize StoreMapSlice when reserving a new cache slot (#350) + =20 + Rock sets the StoreMapSlice::next field when sending a slice to disk. To + avoid writing slice A twice, Rock allocates a new slice B to prime + A.next right before writing A. Scheduling A's writing and, sometimes, + lack of data to fill B create a gap between B's allocation and B's + writing (which sets B.next). During that time, A.next points to B, but + B.next is untouched. + =20 + If writing slice A or swapout in general fails, the chain of failed + entry slices (now containing both A and B) is freed. If untouched B.next + contains garbage, then freeChainAt() adds "random" slices after B to the + free slice pool. Subsequent swapouts use those incorrectly freed slices, + effectively overwriting portions of random cache entries, corrupting the + cache. + =20 + How did B.next get dirty in the first place? freeChainAt() cleans the + slices it frees, but Rock also makes direct noteFreeMapSlice() calls. + Shared memory cache may have avoided this corruption because it makes no + such calls. + =20 + Ipc::StoreMap::prepFreeSlice() now clears allocated slices. Long-term, + we may be able to move free slice management into StoreMap to automate + this cleanup. + =20 + Also simplified and polished slot allocation code a little, removing the + Rock::IoState::reserveSlotForWriting() middleman. This change also + improves the symmetry between Rock and shared memory cache code. + +diff --git a/src/MemStore.cc b/src/MemStore.cc +index 1b24469..1ff48a0 100644 +--- a/src/MemStore.cc ++++ b/src/MemStore.cc +@@ -772,13 +772,15 @@ MemStore::reserveSapForWriting(Ipc::Mem::PageId &page) + { + Ipc::Mem::PageId slot; + if (freeSlots->pop(slot)) { +- debugs(20, 5, "got a previously free slot: " << slot); ++ const auto slotId =3D slot.number - 1; ++ debugs(20, 5, "got a previously free slot: " << slotId); +=20 + if (Ipc::Mem::GetPage(Ipc::Mem::PageId::cachePage, page)) { + debugs(20, 5, "and got a previously free page: " << page); +- return slot.number - 1; ++ map->prepFreeSlice(slotId); ++ return slotId; + } else { +- debugs(20, 3, "but there is no free page, returning " << slot); ++ debugs(20, 3, "but there is no free page, returning " << slotId= ); + freeSlots->push(slot); + } + } +@@ -791,8 +793,10 @@ MemStore::reserveSapForWriting(Ipc::Mem::PageId &page) + assert(!waitingFor); // noteFreeMapSlice() should have cleared it + assert(slot.set()); + assert(page.set()); +- debugs(20, 5, "got previously busy " << slot << " and " << page); +- return slot.number - 1; ++ const auto slotId =3D slot.number - 1; ++ map->prepFreeSlice(slotId); ++ debugs(20, 5, "got previously busy " << slotId << " and " << page); ++ return slotId; + } + assert(waitingFor.slot =3D=3D &slot && waitingFor.page =3D=3D &page); + waitingFor.slot =3D NULL; +diff --git a/src/fs/rock/RockIoState.cc b/src/fs/rock/RockIoState.cc +index bba24ae..50b6e6d 100644 +--- a/src/fs/rock/RockIoState.cc ++++ b/src/fs/rock/RockIoState.cc +@@ -190,7 +190,7 @@ Rock::IoState::tryWrite(char const *buf, size_t size, of= f_t coreOff) + // allocate the first slice during the first write + if (!coreOff) { + assert(sidCurrent < 0); +- sidCurrent =3D reserveSlotForWriting(); // throws on failures ++ sidCurrent =3D dir->reserveSlotForWriting(); // throws on failures + assert(sidCurrent >=3D 0); + writeAnchor().start =3D sidCurrent; + } +@@ -207,7 +207,7 @@ Rock::IoState::tryWrite(char const *buf, size_t size, of= f_t coreOff) + // We do not write a full buffer without overflow because + // we would not yet know what to set the nextSlot to. + if (overflow) { +- const SlotId sidNext =3D reserveSlotForWriting(); // throws ++ const auto sidNext =3D dir->reserveSlotForWriting(); // throws + assert(sidNext >=3D 0); + writeToDisk(sidNext); + } else if (Store::Root().transientReaders(*e)) { +@@ -306,21 +306,6 @@ Rock::IoState::writeBufToDisk(const SlotId sidNext, con= st bool eof, const bool l + theFile->write(r); + } +=20 +-/// finds and returns a free db slot to fill or throws +-Rock::SlotId +-Rock::IoState::reserveSlotForWriting() +-{ +- Ipc::Mem::PageId pageId; +- if (dir->useFreeSlot(pageId)) +- return pageId.number-1; +- +- // This may happen when the number of available db slots is close to the +- // number of concurrent requests reading or writing those slots, which = may +- // happen when the db is "small" compared to the request traffic OR whe= n we +- // are rebuilding and have not loaded "many" entries or empty slots yet. +- throw TexcHere("ran out of free db slots"); +-} +- + void + Rock::IoState::finishedWriting(const int errFlag) + { +diff --git a/src/fs/rock/RockIoState.h b/src/fs/rock/RockIoState.h +index 3c21355..b8f2061 100644 +--- a/src/fs/rock/RockIoState.h ++++ b/src/fs/rock/RockIoState.h +@@ -66,7 +66,6 @@ private: + size_t writeToBuffer(char const *buf, size_t size); + void writeToDisk(const SlotId nextSlot); + void writeBufToDisk(const SlotId nextSlot, const bool eof, const bool l= astWrite); +- SlotId reserveSlotForWriting(); +=20 + void callBack(int errflag); +=20 +diff --git a/src/fs/rock/RockSwapDir.cc b/src/fs/rock/RockSwapDir.cc +index 06fb763..f170bb6 100644 +--- a/src/fs/rock/RockSwapDir.cc ++++ b/src/fs/rock/RockSwapDir.cc +@@ -699,12 +699,16 @@ Rock::SwapDir::diskOffsetLimit() const + return diskOffset(map->sliceLimit()); + } +=20 +-bool +-Rock::SwapDir::useFreeSlot(Ipc::Mem::PageId &pageId) ++Rock::SlotId ++Rock::SwapDir::reserveSlotForWriting() + { ++ Ipc::Mem::PageId pageId; ++ + if (freeSlots->pop(pageId)) { +- debugs(47, 5, "got a previously free slot: " << pageId); +- return true; ++ const auto slotId =3D pageId.number - 1; ++ debugs(47, 5, "got a previously free slot: " << slotId); ++ map->prepFreeSlice(slotId); ++ return slotId; + } +=20 + // catch free slots delivered to noteFreeMapSlice() +@@ -713,14 +717,20 @@ Rock::SwapDir::useFreeSlot(Ipc::Mem::PageId &pageId) + if (map->purgeOne()) { + assert(!waitingForPage); // noteFreeMapSlice() should have cleared = it + assert(pageId.set()); +- debugs(47, 5, "got a previously busy slot: " << pageId); +- return true; ++ const auto slotId =3D pageId.number - 1; ++ debugs(47, 5, "got a previously busy slot: " << slotId); ++ map->prepFreeSlice(slotId); ++ return slotId; + } + assert(waitingForPage =3D=3D &pageId); + waitingForPage =3D NULL; +=20 ++ // This may happen when the number of available db slots is close to the ++ // number of concurrent requests reading or writing those slots, which = may ++ // happen when the db is "small" compared to the request traffic OR whe= n we ++ // are rebuilding and have not loaded "many" entries or empty slots yet. + debugs(47, 3, "cannot get a slot; entries: " << map->entryCount()); +- return false; ++ throw TexcHere("ran out of free db slots"); + } +=20 + bool +diff --git a/src/fs/rock/RockSwapDir.h b/src/fs/rock/RockSwapDir.h +index 879bd45..aa88283 100644 +--- a/src/fs/rock/RockSwapDir.h ++++ b/src/fs/rock/RockSwapDir.h +@@ -62,10 +62,12 @@ public: + int64_t slotLimitAbsolute() const; ///< Rock store implementation limit + int64_t slotLimitActual() const; ///< total number of slots in this db +=20 +- /// removes a slot from a list of free slots or returns false +- bool useFreeSlot(Ipc::Mem::PageId &pageId); + /// whether the given slot ID may point to a slot in this db + bool validSlotId(const SlotId slotId) const; ++ ++ /// finds and returns a free db slot to fill or throws ++ SlotId reserveSlotForWriting(); ++ + /// purges one or more entries to make full() false and free some slots + void purgeSome(); +=20 +diff --git a/src/ipc/StoreMap.cc b/src/ipc/StoreMap.cc +index ba1d4df..e778bba 100644 +--- a/src/ipc/StoreMap.cc ++++ b/src/ipc/StoreMap.cc +@@ -346,8 +346,7 @@ Ipc::StoreMap::freeChainAt(SliceId sliceId, const SliceI= d splicingPoint) + while (sliceId >=3D 0) { + Slice &slice =3D sliceAt(sliceId); + const SliceId nextId =3D slice.next; +- slice.size =3D 0; +- slice.next =3D -1; ++ slice.clear(); + if (cleaner) + cleaner->noteFreeMapSlice(sliceId); // might change slice state + if (sliceId =3D=3D splicingPoint) { +@@ -360,6 +359,14 @@ Ipc::StoreMap::freeChainAt(SliceId sliceId, const Slice= Id splicingPoint) + debugs(54, 7, "freed chain #" << chainId << " in " << path); + } +=20 ++void ++Ipc::StoreMap::prepFreeSlice(const SliceId sliceId) ++{ ++ // TODO: Move freeSlots here, along with reserveSlotForWriting() logic. ++ assert(validSlice(sliceId)); ++ sliceAt(sliceId).clear(); ++} ++ + Ipc::StoreMap::SliceId + Ipc::StoreMap::sliceContaining(const sfileno fileno, const uint64_t bytesNe= eded) const + { +diff --git a/src/ipc/StoreMap.h b/src/ipc/StoreMap.h +index 5d44379..8a4909e 100644 +--- a/src/ipc/StoreMap.h ++++ b/src/ipc/StoreMap.h +@@ -42,6 +42,9 @@ public: + return *this; + } +=20 ++ /// restore default-constructed state ++ void clear() { size =3D 0; next =3D -1; } ++ + std::atomic size; ///< slice contents size + std::atomic next; ///< ID of the next entry slice + }; +@@ -292,6 +295,9 @@ public: + /// readable anchor for the entry created by openForReading() + const Anchor &readableEntry(const AnchorId anchorId) const; +=20 ++ /// prepare a chain-unaffiliated slice for being added to an entry chain ++ void prepFreeSlice(const SliceId sliceId); ++ + /// Returns the ID of the entry slice containing n-th byte or + /// a negative ID if the entry does not store that many bytes (yet). + /// Requires a read lock. diff --git a/src/patches/squid/07_Fail_Rock_swapout_if_the_disk_dropped_some_= of_the_write_requests_352.patch b/src/patches/squid/07_Fail_Rock_swapout_if_t= he_disk_dropped_some_of_the_write_requests_352.patch new file mode 100644 index 000000000..14ce96a33 --- /dev/null +++ b/src/patches/squid/07_Fail_Rock_swapout_if_the_disk_dropped_some_of_the_= write_requests_352.patch @@ -0,0 +1,164 @@ +commit a374216b3262b6b40dc19b8ea5423cd2f27f94a3 +Author: Eduard Bagdasaryan +Date: 2019-01-23 04:30:40 +0000 + + Fail Rock swapout if the disk dropped some of the write requests (#352) + =20 + Detecting dropped writes earlier is more than a TODO: If the last entry + write was successful, the whole entry becomes available for hits + immediately. IpcIoFile::checkTimeouts() that runs every 7 seconds + (IpcIoFile::Timeout) would eventually notify Rock about the timeout, + allowing Rock to release the failed entry, but that notification may + be too late. + =20 + The precise outcome of hitting an entry with a missing on-disk slice is + unknown (because the bug was detected by temporary hit validation code + that turned such hits into misses), but SWAPFAIL is the best we could + hope for. + +diff --git a/src/fs/rock/RockSwapDir.cc b/src/fs/rock/RockSwapDir.cc +index f170bb6..b70dc8b 100644 +--- a/src/fs/rock/RockSwapDir.cc ++++ b/src/fs/rock/RockSwapDir.cc +@@ -858,42 +858,59 @@ Rock::SwapDir::writeCompleted(int errflag, size_t, Ref= Count< ::WriteRequest> r) +=20 + debugs(79, 7, "errflag=3D" << errflag << " rlen=3D" << request->len << = " eof=3D" << request->eof); +=20 +- // TODO: Fail if disk dropped one of the previous write requests. +- +- if (errflag =3D=3D DISK_OK) { +- // do not increment sio.offset_ because we do it in sio->write() +- +- // finalize the shared slice info after writing slice contents to d= isk +- Ipc::StoreMap::Slice &slice =3D +- map->writeableSlice(sio.swap_filen, request->sidCurrent); +- slice.size =3D request->len - sizeof(DbCellHeader); +- slice.next =3D request->sidNext; +- +- if (request->eof) { +- assert(sio.e); +- assert(sio.writeableAnchor_); +- if (sio.touchingStoreEntry()) { +- sio.e->swap_file_sz =3D sio.writeableAnchor_->basics.swap_f= ile_sz =3D +- sio.offset_; +- map->switchWritingToReading(sio.swap_filen); +- // sio.e keeps the (now read) lock on the anchor +- } +- sio.writeableAnchor_ =3D NULL; +- sio.splicingPoint =3D request->sidCurrent; +- sio.finishedWriting(errflag); +- } +- } else { +- noteFreeMapSlice(request->sidNext); +- +- writeError(sio); +- sio.finishedWriting(errflag); +- // and wait for the finalizeSwapoutFailure() call to close the map = entry +- } ++ if (errflag !=3D DISK_OK) ++ handleWriteCompletionProblem(errflag, *request); ++ else if (droppedEarlierRequest(*request)) ++ handleWriteCompletionProblem(DISK_ERROR, *request); ++ else ++ handleWriteCompletionSuccess(*request); +=20 + if (sio.touchingStoreEntry()) + CollapsedForwarding::Broadcast(*sio.e); + } +=20 ++/// code shared by writeCompleted() success handling cases ++void ++Rock::SwapDir::handleWriteCompletionSuccess(const WriteRequest &request) ++{ ++ auto &sio =3D *(request.sio); ++ sio.splicingPoint =3D request.sidCurrent; ++ // do not increment sio.offset_ because we do it in sio->write() ++ ++ // finalize the shared slice info after writing slice contents to disk ++ Ipc::StoreMap::Slice &slice =3D ++ map->writeableSlice(sio.swap_filen, request.sidCurrent); ++ slice.size =3D request.len - sizeof(DbCellHeader); ++ slice.next =3D request.sidNext; ++ ++ if (request.eof) { ++ assert(sio.e); ++ assert(sio.writeableAnchor_); ++ if (sio.touchingStoreEntry()) { ++ sio.e->swap_file_sz =3D sio.writeableAnchor_->basics.swap_file_= sz =3D ++ sio.offset_; ++ ++ map->switchWritingToReading(sio.swap_filen); ++ // sio.e keeps the (now read) lock on the anchor ++ } ++ sio.writeableAnchor_ =3D NULL; ++ sio.finishedWriting(DISK_OK); ++ } ++} ++ ++/// code shared by writeCompleted() error handling cases ++void ++Rock::SwapDir::handleWriteCompletionProblem(const int errflag, const WriteR= equest &request) ++{ ++ auto &sio =3D *request.sio; ++ ++ noteFreeMapSlice(request.sidNext); ++ ++ writeError(sio); ++ sio.finishedWriting(errflag); ++ // and hope that Core will call disconnect() to close the map entry ++} ++ + void + Rock::SwapDir::writeError(StoreIOState &sio) + { +@@ -908,6 +925,23 @@ Rock::SwapDir::writeError(StoreIOState &sio) + // All callers must also call IoState callback, to propagate the error. + } +=20 ++/// whether the disk has dropped at least one of the previous write requests ++bool ++Rock::SwapDir::droppedEarlierRequest(const WriteRequest &request) const ++{ ++ const auto &sio =3D *request.sio; ++ assert(sio.writeableAnchor_); ++ const Ipc::StoreMapSliceId expectedSliceId =3D sio.splicingPoint < 0 ? ++ sio.writeableAnchor_->start : ++ map->writeableSlice(sio.swap_filen, sio.splicingPoint).next; ++ if (expectedSliceId !=3D request.sidCurrent) { ++ debugs(79, 3, "yes; expected " << expectedSliceId << ", but got " <= < request.sidCurrent); ++ return true; ++ } ++ ++ return false; ++} ++ + void + Rock::SwapDir::updateHeaders(StoreEntry *updatedE) + { +diff --git a/src/fs/rock/RockSwapDir.h b/src/fs/rock/RockSwapDir.h +index aa88283..0748337 100644 +--- a/src/fs/rock/RockSwapDir.h ++++ b/src/fs/rock/RockSwapDir.h +@@ -138,6 +138,9 @@ protected: +=20 + private: + void createError(const char *const msg); ++ void handleWriteCompletionSuccess(const WriteRequest &request); ++ void handleWriteCompletionProblem(const int errflag, const WriteRequest= &request); ++ bool droppedEarlierRequest(const WriteRequest &request) const; +=20 + DiskIOStrategy *io; + RefCount theFile; ///< cache storage for this cache_dir +diff --git a/src/fs/rock/forward.h b/src/fs/rock/forward.h +index ba64d03..4e68a27 100644 +--- a/src/fs/rock/forward.h ++++ b/src/fs/rock/forward.h +@@ -40,6 +40,8 @@ class HeaderUpdater; +=20 + class DbCellHeader; +=20 ++class WriteRequest; ++ + } +=20 + #endif /* SQUID_FS_ROCK_FORWARD_H */ diff --git a/src/patches/squid/08_Bug_4914_Do_not_call_setsid_in_--foreground= _mode_354.patch b/src/patches/squid/08_Bug_4914_Do_not_call_setsid_in_--foreg= round_mode_354.patch new file mode 100644 index 000000000..65bf2e648 --- /dev/null +++ b/src/patches/squid/08_Bug_4914_Do_not_call_setsid_in_--foreground_mode_3= 54.patch @@ -0,0 +1,36 @@ +commit 6076ca90a8101888483b769edd2f9aeef0da8638 +Author: Marcos Mello +Date: 2019-01-23 15:18:43 +0000 + + Bug 4914: Do not call setsid() in --foreground mode (#354) + =20 + Squid executed in --foreground is always a process group leader. Thus, + setsid(2) is unnecessary and always fails (with EPERM) for such Squids. + +diff --git a/src/main.cc b/src/main.cc +index 7f5ec80..1e8e035 100644 +--- a/src/main.cc ++++ b/src/main.cc +@@ -1871,6 +1871,7 @@ GoIntoBackground() + exit(EXIT_SUCCESS); + } + // child, running as a background daemon ++ Must(setsid() > 0); // ought to succeed after fork() + } +=20 + static void +@@ -1916,14 +1917,6 @@ watch_child(const CommandLine &masterCommand) + if (!opt_foreground) + GoIntoBackground(); +=20 +- // TODO: Fails with --foreground if the calling process is process group +- // leader, which is always (?) the case. Should probably moved to +- // GoIntoBackground and executed only after successfully forking +- if (setsid() < 0) { +- int xerrno =3D errno; +- syslog(LOG_ALERT, "setsid failed: %s", xstrerr(xerrno)); +- } +- + closelog(); +=20 + #ifdef TIOCNOTTY diff --git a/src/patches/squid/09_Bug_4915_Detect_IPv6_loopback_binding_error= s_355.patch b/src/patches/squid/09_Bug_4915_Detect_IPv6_loopback_binding_erro= rs_355.patch new file mode 100644 index 000000000..b6fed3336 --- /dev/null +++ b/src/patches/squid/09_Bug_4915_Detect_IPv6_loopback_binding_errors_355.p= atch @@ -0,0 +1,54 @@ +commit 568e66b7c64e0d7873b9575b0dc60886805a2fc6 (refs/remotes/origin/v4) +Author: Amos Jeffries +Date: 2019-01-24 16:43:47 +0000 + + Bug 4915: Detect IPv6 loopback binding errors (#355) + =20 + Systems which have been partially 'IPv6 disabled' may allow + sockets to be opened and used but missing the IPv6 loopback + address. + =20 + Implement the outstanding TODO to detect such failures and + disable IPv6 support properly within Squid when they are found. + =20 + This should fix bug 4915 auth_param helper startup and similar + external_acl_type helper issues. For security such helpers are + not permitted to use the machine default IP address which is + globally accessible. + +diff --git a/src/ip/tools.cc b/src/ip/tools.cc +index c7a8ac8..1ccf64f 100644 +--- a/src/ip/tools.cc ++++ b/src/ip/tools.cc +@@ -10,6 +10,7 @@ +=20 + #include "squid.h" + #include "Debug.h" ++#include "ip/Address.h" + #include "ip/tools.h" +=20 + #if HAVE_UNISTD_H +@@ -55,8 +56,21 @@ Ip::ProbeTransport() + debugs(3, 2, "Missing RFC 3493 compliance - attempting split IPv4 and I= Pv6 stacks ..."); + EnableIpv6 |=3D IPV6_SPECIAL_SPLITSTACK; + #endif +- // TODO: attempt to use the socket to connect somewhere ? +- // needs to be safe to contact and with guaranteed working IPv6 at the= other end. ++ ++ // Test for IPv6 loopback/localhost address binding ++ Ip::Address ip; ++ ip.setLocalhost(); ++ if (ip.isIPv6()) { // paranoid; always succeeds if we got this far ++ struct sockaddr_in6 sin; ++ ip.getSockAddr(sin); ++ if (bind(s, reinterpret_cast(&sin), sizeof(sin))= !=3D 0) { ++ debugs(3, DBG_CRITICAL, "WARNING: BCP 177 violation. Detected n= on-functional IPv6 loopback."); ++ EnableIpv6 =3D IPV6_OFF; ++ } else { ++ debugs(3, 2, "Detected functional IPv6 loopback ..."); ++ } ++ } ++ + close(s); +=20 + #if USE_IPV6 --=20 2.18.0 --===============3786999699197001518==--