diff options
-rw-r--r-- | .travis.yml | 1 | ||||
-rw-r--r-- | CMakeLists.txt | 16 | ||||
-rw-r--r-- | debian/apt-transport-https.README | 2 | ||||
-rw-r--r-- | debian/apt-transport-https.docs | 1 | ||||
-rw-r--r-- | debian/apt-transport-https.install | 3 | ||||
-rw-r--r-- | debian/control | 20 | ||||
-rw-r--r-- | methods/CMakeLists.txt | 20 | ||||
-rw-r--r-- | methods/curl.cc | 548 | ||||
-rw-r--r-- | methods/curl.h | 88 | ||||
-rw-r--r-- | po/CMakeLists.txt | 8 | ||||
-rwxr-xr-x | test/integration/test-apt-update-failure-propagation | 6 |
11 files changed, 3 insertions, 710 deletions
diff --git a/.travis.yml b/.travis.yml index 0685cdb76..7ef7b36c9 100644 --- a/.travis.yml +++ b/.travis.yml @@ -6,7 +6,6 @@ services: env: - USER=travis CMAKE_FLAGS= - USER=root CMAKE_FLAGS=-DWITH_DOC=OFF - - USER=travis CMAKE_FLAGS="-DWITH_DOC=OFF -DFORCE_CURL=ON" install: - sed -i -e "s#1000#$(id -u)#g" Dockerfile - docker build --tag=apt-ci . diff --git a/CMakeLists.txt b/CMakeLists.txt index e5a78706b..04a8be22a 100644 --- a/CMakeLists.txt +++ b/CMakeLists.txt @@ -75,22 +75,6 @@ if (BERKELEY_DB_FOUND) set(HAVE_BDB 1) endif() - -# apt-transport-https dependencies -option(WITH_CURL "Build curl-based methods" ON) -option(FORCE_CURL "Use curl-based methods" OFF) -if (FORCE_CURL AND NOT WITH_CURL) - message(FATAL_ERROR "Cannot set FORCE_CURL=ON with WITH_CURL=OFF") -endif() -if (WITH_CURL) - find_package(CURL REQUIRED) - if (CURL_FOUND) - set(HAVE_CURL 1) - endif() -else() - set(HAVE_CURL 0) -endif() - find_package(GnuTLS REQUIRED) if (GNUTLS_FOUND) set(HAVE_GNUTLS 1) diff --git a/debian/apt-transport-https.README b/debian/apt-transport-https.README deleted file mode 100644 index 0c470a1a2..000000000 --- a/debian/apt-transport-https.README +++ /dev/null @@ -1,2 +0,0 @@ -See APT::https group of options in apt.conf(5) to configure how APT will handle -https:// URIs. diff --git a/debian/apt-transport-https.docs b/debian/apt-transport-https.docs deleted file mode 100644 index 906c1db69..000000000 --- a/debian/apt-transport-https.docs +++ /dev/null @@ -1 +0,0 @@ -debian/apt-transport-https.README diff --git a/debian/apt-transport-https.install b/debian/apt-transport-https.install deleted file mode 100644 index 4849f1484..000000000 --- a/debian/apt-transport-https.install +++ /dev/null @@ -1,3 +0,0 @@ -usr/lib/apt/methods/curl -usr/lib/apt/methods/curl+http -usr/lib/apt/methods/curl+https diff --git a/debian/control b/debian/control index 86acab587..22567e193 100644 --- a/debian/control +++ b/debian/control @@ -37,7 +37,7 @@ Depends: adduser, ${misc:Depends}, ${shlibs:Depends} Replaces: apt-utils (<< 1.3~exp2~), apt-transport-https (<< 1.5~alpha4~) -Breaks: apt-utils (<< 1.3~exp2~), apt-transport-https (<< 1.5~alpha4~) +Breaks: apt-utils (<< 1.3~exp2~), apt-transport-https Provides: apt-transport-https (= ${binary:Version}) Recommends: ca-certificates Suggests: apt-doc, @@ -146,21 +146,3 @@ Description: package management related utility programs * apt-ftparchive is used to create Packages and other index files needed to publish an archive of Debian packages * apt-sortpkgs is a Packages/Sources file normalizer. - -Package: apt-transport-https -Architecture: any -Depends: ${misc:Depends}, ${shlibs:Depends}, apt (>= 1.5~alpha4~) -Recommends: ca-certificates -Priority: optional -Description: Transitional package: curl-https download transport for APT - This package enables the usage of 'deb https://foo distro main' lines - in the /etc/apt/sources.list so that all package managers using the - libapt-pkg library can access metadata and packages available in sources - accessible over https (Hypertext Transfer Protocol Secure). - . - This transport supports server as well as client authentication - with certificates. - . - This package is no longer needed, https support was integrated into apt, - and is only provided as a fallback if some bugs are found in apt's native - https support. diff --git a/methods/CMakeLists.txt b/methods/CMakeLists.txt index 3ae3f9963..72f07e87e 100644 --- a/methods/CMakeLists.txt +++ b/methods/CMakeLists.txt @@ -6,18 +6,12 @@ add_executable(gpgv gpgv.cc) add_executable(cdrom cdrom.cc) add_executable(http http.cc http_main.cc rfc2553emu.cc connect.cc basehttp.cc) add_executable(mirror mirror.cc http.cc rfc2553emu.cc connect.cc basehttp.cc) -if (HAVE_CURL) - add_executable(curl curl.cc basehttp.cc) -endif() add_executable(ftp ftp.cc rfc2553emu.cc connect.cc) add_executable(rred rred.cc) add_executable(rsh rsh.cc) target_compile_definitions(http PRIVATE ${GNUTLS_DEFINITIONS}) target_include_directories(http PRIVATE ${GNUTLS_INCLUDE_DIR}) -if (HAVE_CURL) -target_include_directories(curl PRIVATE ${CURL_INCLUDE_DIRS}) -endif() # Link the executables against the libraries target_link_libraries(file apt-pkg) @@ -27,9 +21,6 @@ target_link_libraries(gpgv apt-pkg) target_link_libraries(cdrom apt-pkg) target_link_libraries(http apt-pkg ${GNUTLS_LIBRARIES}) target_link_libraries(mirror apt-pkg ${RESOLV_LIBRARIES} ${GNUTLS_LIBRARIES}) -if (HAVE_CURL) - target_link_libraries(curl apt-pkg ${CURL_LIBRARIES}) -endif() target_link_libraries(ftp apt-pkg ${GNUTLS_LIBRARIES}) target_link_libraries(rred apt-pkg) target_link_libraries(rsh apt-pkg) @@ -41,14 +32,5 @@ install(TARGETS file copy store gpgv cdrom http ftp rred rsh mirror add_slaves(${CMAKE_INSTALL_LIBEXECDIR}/apt/methods store gzip lzma bzip2 xz) add_slaves(${CMAKE_INSTALL_LIBEXECDIR}/apt/methods rsh ssh) -set(curl_slaves curl+https curl+http) -if (FORCE_CURL) - set(curl_slaves ${curl_slaves} https) -else() - add_slaves(${CMAKE_INSTALL_LIBEXECDIR}/apt/methods http https) -endif() -if (HAVE_CURL) - install(TARGETS curl RUNTIME DESTINATION ${CMAKE_INSTALL_LIBEXECDIR}/apt/methods) - add_slaves(${CMAKE_INSTALL_LIBEXECDIR}/apt/methods curl ${curl_slaves}) -endif() +add_slaves(${CMAKE_INSTALL_LIBEXECDIR}/apt/methods http https) diff --git a/methods/curl.cc b/methods/curl.cc deleted file mode 100644 index 8e06d858d..000000000 --- a/methods/curl.cc +++ /dev/null @@ -1,548 +0,0 @@ -//-*- mode: cpp; mode: fold -*- -// Description /*{{{*/ -// $Id: http.cc,v 1.59 2004/05/08 19:42:35 mdz Exp $ -/* ###################################################################### - - HTTPS Acquire Method - This is the HTTPS acquire method for APT. - - It uses libcurl - - ##################################################################### */ - /*}}}*/ -// Include Files /*{{{*/ -#include <config.h> - -#include <apt-pkg/configuration.h> -#include <apt-pkg/error.h> -#include <apt-pkg/fileutl.h> -#include <apt-pkg/hashes.h> -#include <apt-pkg/macros.h> -#include <apt-pkg/netrc.h> -#include <apt-pkg/proxy.h> -#include <apt-pkg/strutl.h> - -#include <ctype.h> -#include <stdio.h> -#include <stdlib.h> -#include <sys/stat.h> -#include <sys/time.h> -#include <unistd.h> - -#include <array> -#include <iostream> -#include <sstream> - -#include "curl.h" - -#include <apti18n.h> - /*}}}*/ -using namespace std; - -struct APT_HIDDEN CURLUserPointer { - HttpsMethod * const https; - HttpsMethod::FetchResult * const Res; - HttpsMethod::FetchItem const * const Itm; - RequestState * const Req; - CURLUserPointer(HttpsMethod * const https, HttpsMethod::FetchResult * const Res, - HttpsMethod::FetchItem const * const Itm, RequestState * const Req) : https(https), Res(Res), Itm(Itm), Req(Req) {} -}; - -size_t -HttpsMethod::parse_header(void *buffer, size_t size, size_t nmemb, void *userp) -{ - size_t len = size * nmemb; - CURLUserPointer *me = static_cast<CURLUserPointer *>(userp); - std::string line((char*) buffer, len); - for (--len; len > 0; --len) - if (isspace_ascii(line[len]) == 0) - { - ++len; - break; - } - line.erase(len); - - if (line.empty() == true) - { - if (me->Req->File.Open(me->Itm->DestFile, FileFd::WriteAny) == false) - return ERROR_NOT_FROM_SERVER; - - me->Req->JunkSize = 0; - if (me->Req->Result != 416 && me->Req->StartPos != 0) - ; - else if (me->Req->Result == 416) - { - bool partialHit = false; - if (me->Itm->ExpectedHashes.usable() == true) - { - Hashes resultHashes(me->Itm->ExpectedHashes); - FileFd file(me->Itm->DestFile, FileFd::ReadOnly); - me->Req->TotalFileSize = file.FileSize(); - me->Req->Date = file.ModificationTime(); - resultHashes.AddFD(file); - HashStringList const hashList = resultHashes.GetHashStringList(); - partialHit = (me->Itm->ExpectedHashes == hashList); - } - else if (me->Req->Result == 416 && me->Req->TotalFileSize == me->Req->File.FileSize()) - partialHit = true; - - if (partialHit == true) - { - me->Req->Result = 200; - me->Req->StartPos = me->Req->TotalFileSize; - // the actual size is not important for https as curl will deal with it - // by itself and e.g. doesn't bother us with transport-encoding… - me->Req->JunkSize = std::numeric_limits<unsigned long long>::max(); - } - else - me->Req->StartPos = 0; - } - else - me->Req->StartPos = 0; - - me->Res->LastModified = me->Req->Date; - me->Res->Size = me->Req->TotalFileSize; - me->Res->ResumePoint = me->Req->StartPos; - - // we expect valid data, so tell our caller we get the file now - if (me->Req->Result >= 200 && me->Req->Result < 300) - { - if (me->Res->Size != 0 && me->Res->Size > me->Res->ResumePoint) - me->https->URIStart(*me->Res); - if (me->Req->AddPartialFileToHashes(me->Req->File) == false) - return 0; - } - else - me->Req->JunkSize = std::numeric_limits<decltype(me->Req->JunkSize)>::max(); - } - else if (me->Req->HeaderLine(line) == false) - return 0; - - return size*nmemb; -} - -size_t -HttpsMethod::write_data(void *buffer, size_t size, size_t nmemb, void *userp) -{ - CURLUserPointer *me = static_cast<CURLUserPointer *>(userp); - size_t buffer_size = size * nmemb; - // we don't need to count the junk here, just drop anything we get as - // we don't always know how long it would be, e.g. in chunked encoding. - if (me->Req->JunkSize != 0) - return buffer_size; - - if(me->Req->File.Write(buffer, buffer_size) != true) - return 0; - - if(me->https->Queue->MaximumSize > 0) - { - unsigned long long const TotalWritten = me->Req->File.Tell(); - if (TotalWritten > me->https->Queue->MaximumSize) - { - me->https->SetFailReason("MaximumSizeExceeded"); - _error->Error(_("File has unexpected size (%llu != %llu). Mirror sync in progress?"), - TotalWritten, me->https->Queue->MaximumSize); - return 0; - } - } - - if (me->https->Server->GetHashes()->Add((unsigned char const * const)buffer, buffer_size) == false) - return 0; - - return buffer_size; -} - -// HttpsServerState::HttpsServerState - Constructor /*{{{*/ -HttpsServerState::HttpsServerState(URI Srv,HttpsMethod * Owner) : ServerState(Srv, Owner), Hash(NULL) -{ - TimeOut = Owner->ConfigFindI("Timeout", TimeOut); - Reset(); -} - /*}}}*/ -bool HttpsServerState::InitHashes(HashStringList const &ExpectedHashes) /*{{{*/ -{ - delete Hash; - Hash = new Hashes(ExpectedHashes); - return true; -} - /*}}}*/ -APT_PURE Hashes * HttpsServerState::GetHashes() /*{{{*/ -{ - return Hash; -} - /*}}}*/ - -bool HttpsMethod::SetupProxy() /*{{{*/ -{ - URI ServerName = Queue->Uri; - - // Determine the proxy setting - AutoDetectProxy(ServerName); - - // Curl should never read proxy settings from the environment, as - // we determine which proxy to use. Do this for consistency among - // methods and prevent an environment variable overriding a - // no-proxy ("DIRECT") setting in apt.conf. - curl_easy_setopt(curl, CURLOPT_PROXY, ""); - - // Determine the proxy setting - try https first, fallback to http and use env at last - string UseProxy = ConfigFind("Proxy::" + ServerName.Host, ""); - if (UseProxy.empty() == true) - UseProxy = ConfigFind("Proxy", ""); - // User wants to use NO proxy, so nothing to setup - if (UseProxy == "DIRECT") - return true; - - // Parse no_proxy, a comma (,) separated list of domains we don't want to use - // a proxy for so we stop right here if it is in the list - if (getenv("no_proxy") != 0 && CheckDomainList(ServerName.Host,getenv("no_proxy")) == true) - return true; - - if (UseProxy.empty() == true) - { - const char* result = nullptr; - if (std::find(methodNames.begin(), methodNames.end(), "https") != methodNames.end()) - result = getenv("https_proxy"); - // FIXME: Fall back to http_proxy is to remain compatible with - // existing setups and behaviour of apt.conf. This should be - // deprecated in the future (including apt.conf). Most other - // programs do not fall back to http proxy settings and neither - // should Apt. - if (result == nullptr && std::find(methodNames.begin(), methodNames.end(), "http") != methodNames.end()) - result = getenv("http_proxy"); - UseProxy = result == nullptr ? "" : result; - } - - // Determine what host and port to use based on the proxy settings - if (UseProxy.empty() == false) - { - Proxy = UseProxy; - AddProxyAuth(Proxy, ServerName); - - if (Proxy.Access == "socks5h") - curl_easy_setopt(curl, CURLOPT_PROXYTYPE, CURLPROXY_SOCKS5_HOSTNAME); - else if (Proxy.Access == "socks5") - curl_easy_setopt(curl, CURLOPT_PROXYTYPE, CURLPROXY_SOCKS5); - else if (Proxy.Access == "socks4a") - curl_easy_setopt(curl, CURLOPT_PROXYTYPE, CURLPROXY_SOCKS4A); - else if (Proxy.Access == "socks") - curl_easy_setopt(curl, CURLOPT_PROXYTYPE, CURLPROXY_SOCKS4); - else if (Proxy.Access == "http" || Proxy.Access == "https") - curl_easy_setopt(curl, CURLOPT_PROXYTYPE, CURLPROXY_HTTP); - else - return false; - - if (Proxy.Port != 1) - curl_easy_setopt(curl, CURLOPT_PROXYPORT, Proxy.Port); - curl_easy_setopt(curl, CURLOPT_PROXY, Proxy.Host.c_str()); - if (Proxy.User.empty() == false || Proxy.Password.empty() == false) - { - curl_easy_setopt(curl, CURLOPT_PROXYUSERNAME, Proxy.User.c_str()); - curl_easy_setopt(curl, CURLOPT_PROXYPASSWORD, Proxy.Password.c_str()); - } - } - return true; -} /*}}}*/ -// HttpsMethod::Fetch - Fetch an item /*{{{*/ -// --------------------------------------------------------------------- -/* This adds an item to the pipeline. We keep the pipeline at a fixed - depth. */ -bool HttpsMethod::Fetch(FetchItem *Itm) -{ - struct stat SBuf; - struct curl_slist *headers=NULL; - char curl_errorstr[CURL_ERROR_SIZE]; - URI Uri = Itm->Uri; - setPostfixForMethodNames(Uri.Host.c_str()); - AllowRedirect = ConfigFindB("AllowRedirect", true); - Debug = DebugEnabled(); - - // TODO: - // - http::Pipeline-Depth - // - error checking/reporting - // - more debug options? (CURLOPT_DEBUGFUNCTION?) - { - auto const plus = Binary.find('+'); - if (plus != std::string::npos) - Uri.Access = Binary.substr(plus + 1); - } - - curl_easy_reset(curl); - if (SetupProxy() == false) - return _error->Error("Unsupported proxy configured: %s", URI::SiteOnly(Proxy).c_str()); - - MaybeAddAuthTo(Uri); - if (Server == nullptr || Server->Comp(Itm->Uri) == false) - Server = CreateServerState(Itm->Uri); - - // The "+" is encoded as a workaround for a amazon S3 bug - // see LP bugs #1003633 and #1086997. (taken from http method) - Uri.Path = QuoteString(Uri.Path, "+~ "); - - FetchResult Res; - RequestState Req(this, Server.get()); - CURLUserPointer userp(this, &Res, Itm, &Req); - // callbacks - curl_easy_setopt(curl, CURLOPT_URL, static_cast<string>(Uri).c_str()); - curl_easy_setopt(curl, CURLOPT_HEADERFUNCTION, parse_header); - curl_easy_setopt(curl, CURLOPT_WRITEHEADER, &userp); - curl_easy_setopt(curl, CURLOPT_WRITEFUNCTION, write_data); - curl_easy_setopt(curl, CURLOPT_WRITEDATA, &userp); - // options - curl_easy_setopt(curl, CURLOPT_NOPROGRESS, true); - curl_easy_setopt(curl, CURLOPT_FILETIME, true); - curl_easy_setopt(curl, CURLOPT_FOLLOWLOCATION, 0); - - if (std::find(methodNames.begin(), methodNames.end(), "https") != methodNames.end()) - { - curl_easy_setopt(curl, CURLOPT_PROTOCOLS, CURLPROTO_HTTPS); - curl_easy_setopt(curl, CURLOPT_REDIR_PROTOCOLS, CURLPROTO_HTTPS); - - // File containing the list of trusted CA. - std::string const cainfo = ConfigFind("CaInfo", ""); - if(cainfo.empty() == false) - curl_easy_setopt(curl, CURLOPT_CAINFO, cainfo.c_str()); - // Check server certificate against previous CA list ... - curl_easy_setopt(curl, CURLOPT_SSL_VERIFYPEER, ConfigFindB("Verify-Peer", true) ? 1 : 0); - // ... and hostname against cert CN or subjectAltName - curl_easy_setopt(curl, CURLOPT_SSL_VERIFYHOST, ConfigFindB("Verify-Host", true) ? 2 : 0); - // Also enforce issuer of server certificate using its cert - std::string const issuercert = ConfigFind("IssuerCert", ""); - if(issuercert.empty() == false) - curl_easy_setopt(curl, CURLOPT_ISSUERCERT, issuercert.c_str()); - // For client authentication, certificate file ... - std::string const pem = ConfigFind("SslCert", ""); - if(pem.empty() == false) - curl_easy_setopt(curl, CURLOPT_SSLCERT, pem.c_str()); - // ... and associated key. - std::string const key = ConfigFind("SslKey", ""); - if(key.empty() == false) - curl_easy_setopt(curl, CURLOPT_SSLKEY, key.c_str()); - // Allow forcing SSL version to SSLv3 or TLSv1 - long final_version = CURL_SSLVERSION_DEFAULT; - std::string const sslversion = ConfigFind("SslForceVersion", ""); - if(sslversion == "TLSv1") - final_version = CURL_SSLVERSION_TLSv1; - else if(sslversion == "TLSv1.0") - final_version = CURL_SSLVERSION_TLSv1_0; - else if(sslversion == "TLSv1.1") - final_version = CURL_SSLVERSION_TLSv1_1; - else if(sslversion == "TLSv1.2") - final_version = CURL_SSLVERSION_TLSv1_2; - else if(sslversion == "SSLv3") - final_version = CURL_SSLVERSION_SSLv3; - curl_easy_setopt(curl, CURLOPT_SSLVERSION, final_version); - // CRL file - std::string const crlfile = ConfigFind("CrlFile", ""); - if(crlfile.empty() == false) - curl_easy_setopt(curl, CURLOPT_CRLFILE, crlfile.c_str()); - } - else - { - curl_easy_setopt(curl, CURLOPT_PROTOCOLS, CURLPROTO_HTTP); - curl_easy_setopt(curl, CURLOPT_REDIR_PROTOCOLS, CURLPROTO_HTTP); - } - // cache-control - if(ConfigFindB("No-Cache", false) == false) - { - // cache enabled - if (ConfigFindB("No-Store", false) == true) - headers = curl_slist_append(headers,"Cache-Control: no-store"); - std::string ss; - strprintf(ss, "Cache-Control: max-age=%u", ConfigFindI("Max-Age", 0)); - headers = curl_slist_append(headers, ss.c_str()); - } else { - // cache disabled by user - headers = curl_slist_append(headers, "Cache-Control: no-cache"); - headers = curl_slist_append(headers, "Pragma: no-cache"); - } - curl_easy_setopt(curl, CURLOPT_HTTPHEADER, headers); - // speed limit - int const dlLimit = ConfigFindI("Dl-Limit", 0) * 1024; - if (dlLimit > 0) - curl_easy_setopt(curl, CURLOPT_MAX_RECV_SPEED_LARGE, dlLimit); - - // set header - curl_easy_setopt(curl, CURLOPT_USERAGENT, ConfigFind("User-Agent", "Debian APT-CURL/1.0 (" PACKAGE_VERSION ")").c_str()); - - // set timeout - int const timeout = ConfigFindI("Timeout", 120); - curl_easy_setopt(curl, CURLOPT_CONNECTTIMEOUT, timeout); - //set really low lowspeed timeout (see #497983) - curl_easy_setopt(curl, CURLOPT_LOW_SPEED_LIMIT, DL_MIN_SPEED); - curl_easy_setopt(curl, CURLOPT_LOW_SPEED_TIME, timeout); - - if(_config->FindB("Acquire::ForceIPv4", false) == true) - curl_easy_setopt(curl, CURLOPT_IPRESOLVE, CURL_IPRESOLVE_V4); - else if(_config->FindB("Acquire::ForceIPv6", false) == true) - curl_easy_setopt(curl, CURLOPT_IPRESOLVE, CURL_IPRESOLVE_V6); - - // debug - if (Debug == true) - curl_easy_setopt(curl, CURLOPT_VERBOSE, true); - - // error handling - curl_errorstr[0] = '\0'; - curl_easy_setopt(curl, CURLOPT_ERRORBUFFER, curl_errorstr); - - // If we ask for uncompressed files servers might respond with content- - // negotiation which lets us end up with compressed files we do not support, - // see 657029, 657560 and co, so if we have no extension on the request - // ask for text only. As a sidenote: If there is nothing to negotate servers - // seem to be nice and ignore it. - if (ConfigFindB("SendAccept", true)) - { - size_t const filepos = Itm->Uri.find_last_of('/'); - string const file = Itm->Uri.substr(filepos + 1); - if (flExtension(file) == file) - headers = curl_slist_append(headers, "Accept: text/*"); - } - - // if we have the file send an if-range query with a range header - if (Server->RangesAllowed && stat(Itm->DestFile.c_str(),&SBuf) >= 0 && SBuf.st_size > 0) - { - std::string Buf; - strprintf(Buf, "Range: bytes=%lli-", (long long) SBuf.st_size); - headers = curl_slist_append(headers, Buf.c_str()); - strprintf(Buf, "If-Range: %s", TimeRFC1123(SBuf.st_mtime, false).c_str()); - headers = curl_slist_append(headers, Buf.c_str()); - } - else if(Itm->LastModified > 0) - { - curl_easy_setopt(curl, CURLOPT_TIMECONDITION, CURL_TIMECOND_IFMODSINCE); - curl_easy_setopt(curl, CURLOPT_TIMEVALUE, Itm->LastModified); - } - - if (Server->InitHashes(Itm->ExpectedHashes) == false) - return false; - - // keep apt updated - Res.Filename = Itm->DestFile; - - // get it! - CURLcode success = curl_easy_perform(curl); - - // If the server returns 200 OK but the If-Modified-Since condition is not - // met, CURLINFO_CONDITION_UNMET will be set to 1 - long curl_condition_unmet = 0; - curl_easy_getinfo(curl, CURLINFO_CONDITION_UNMET, &curl_condition_unmet); - if (curl_condition_unmet == 1) - Req.Result = 304; - - Req.File.Close(); - curl_slist_free_all(headers); - - // cleanup - if (success != CURLE_OK) - { -#pragma GCC diagnostic push -#pragma GCC diagnostic ignored "-Wswitch" - switch (success) - { - case CURLE_COULDNT_RESOLVE_PROXY: - case CURLE_COULDNT_RESOLVE_HOST: - SetFailReason("ResolveFailure"); - break; - case CURLE_COULDNT_CONNECT: - SetFailReason("ConnectionRefused"); - break; - case CURLE_OPERATION_TIMEDOUT: - SetFailReason("Timeout"); - break; - } -#pragma GCC diagnostic pop - // only take curls technical errors if we haven't our own - // (e.g. for the maximum size limit we have and curls can be confusing) - if (_error->PendingError() == false) - _error->Error("%s", curl_errorstr); - else - _error->Warning("curl: %s", curl_errorstr); - return false; - } - - switch (DealWithHeaders(Res, Req)) - { - case BaseHttpMethod::IMS_HIT: - URIDone(Res); - break; - - case BaseHttpMethod::ERROR_WITH_CONTENT_PAGE: - // unlink, no need keep 401/404 page content in partial/ - RemoveFile(Binary.c_str(), Req.File.Name()); - // Fall through. - case BaseHttpMethod::ERROR_UNRECOVERABLE: - case BaseHttpMethod::ERROR_NOT_FROM_SERVER: - return false; - - case BaseHttpMethod::TRY_AGAIN_OR_REDIRECT: - Redirect(NextURI); - break; - - case BaseHttpMethod::FILE_IS_OPEN: - struct stat resultStat; - if (unlikely(stat(Req.File.Name().c_str(), &resultStat) != 0)) - { - _error->Errno("stat", "Unable to access file %s", Req.File.Name().c_str()); - return false; - } - Res.Size = resultStat.st_size; - - // Timestamp - curl_easy_getinfo(curl, CURLINFO_FILETIME, &Res.LastModified); - if (Res.LastModified != -1) - { - struct timeval times[2]; - times[0].tv_sec = Res.LastModified; - times[1].tv_sec = Res.LastModified; - times[0].tv_usec = times[1].tv_usec = 0; - utimes(Req.File.Name().c_str(), times); - } - else - Res.LastModified = resultStat.st_mtime; - - // take hashes - Res.TakeHashes(*(Server->GetHashes())); - - // keep apt updated - URIDone(Res); - break; - } - return true; -} - /*}}}*/ -std::unique_ptr<ServerState> HttpsMethod::CreateServerState(URI const &uri)/*{{{*/ -{ - return std::unique_ptr<ServerState>(new HttpsServerState(uri, this)); -} - /*}}}*/ -HttpsMethod::HttpsMethod(std::string &&pProg) : BaseHttpMethod(std::move(pProg),"1.2",Pipeline | SendConfig)/*{{{*/ -{ - auto addName = std::inserter(methodNames, methodNames.begin()); - addName = "http"; - auto const plus = Binary.find('+'); - if (plus != std::string::npos) - { - addName = Binary.substr(plus + 1); - auto base = Binary.substr(0, plus); - if (base != "https") - addName = base; - } - if (std::find(methodNames.begin(), methodNames.end(), "https") != methodNames.end()) - curl_global_init(CURL_GLOBAL_SSL); - else - curl_global_init(CURL_GLOBAL_NOTHING); - curl = curl_easy_init(); -} - /*}}}*/ -HttpsMethod::~HttpsMethod() /*{{{*/ -{ - curl_easy_cleanup(curl); -} - /*}}}*/ -int main(int, const char *argv[]) /*{{{*/ -{ - std::string Binary = flNotDir(argv[0]); - if (Binary.find('+') == std::string::npos && Binary != "https") - Binary.append("+https"); - return HttpsMethod(std::move(Binary)).Run(); -} - /*}}}*/ diff --git a/methods/curl.h b/methods/curl.h deleted file mode 100644 index d352dc52a..000000000 --- a/methods/curl.h +++ /dev/null @@ -1,88 +0,0 @@ -// -*- mode: cpp; mode: fold -*- -// Description /*{{{*/// $Id: http.h,v 1.12 2002/04/18 05:09:38 jgg Exp $ -// $Id: http.h,v 1.12 2002/04/18 05:09:38 jgg Exp $ -/* ###################################################################### - - HTTP Acquire Method - This is the HTTP acquire method for APT. - - ##################################################################### */ - /*}}}*/ - -#ifndef APT_HTTPS_H -#define APT_HTTPS_H - -#include <iostream> -#include <memory> -#include <string> -#include <curl/curl.h> -#include <stddef.h> - -#include "basehttp.h" - -using std::cout; -using std::endl; - -class Hashes; -class HttpsMethod; -class FileFd; - -class HttpsServerState : public ServerState -{ - Hashes * Hash; - - protected: - virtual bool ReadHeaderLines(std::string &/*Data*/) APT_OVERRIDE { return false; } - virtual bool LoadNextResponse(bool const /*ToFile*/, RequestState &/*Req*/) APT_OVERRIDE { return false; } - - public: - virtual bool WriteResponse(std::string const &/*Data*/) APT_OVERRIDE { return false; } - - /** \brief Transfer the data from the socket */ - virtual bool RunData(RequestState &) APT_OVERRIDE { return false; } - virtual bool RunDataToDevNull(RequestState &) APT_OVERRIDE { return false; } - - virtual bool Open() APT_OVERRIDE { return false; } - virtual bool IsOpen() APT_OVERRIDE { return false; } - virtual bool Close() APT_OVERRIDE { return false; } - virtual bool InitHashes(HashStringList const &ExpectedHashes) APT_OVERRIDE; - virtual Hashes * GetHashes() APT_OVERRIDE; - virtual bool Die(RequestState &/*Req*/) APT_OVERRIDE { return false; } - virtual bool Flush(FileFd * const /*File*/) APT_OVERRIDE { return false; } - virtual bool Go(bool /*ToFile*/, RequestState &/*Req*/) APT_OVERRIDE { return false; } - - HttpsServerState(URI Srv, HttpsMethod *Owner); - virtual ~HttpsServerState() {Close();}; -}; - -class HttpsMethod : public BaseHttpMethod -{ - // minimum speed in bytes/se that triggers download timeout handling - static const int DL_MIN_SPEED = 10; - - virtual bool Fetch(FetchItem *) APT_OVERRIDE; - - static size_t parse_header(void *buffer, size_t size, size_t nmemb, void *userp); - static size_t write_data(void *buffer, size_t size, size_t nmemb, void *userp); - static int progress_callback(void *clientp, double dltotal, double dlnow, - double ultotal, double ulnow); - bool SetupProxy(); - CURL *curl; - - // Used by BaseHttpMethods unused by https - virtual void SendReq(FetchItem *) APT_OVERRIDE { exit(42); } - virtual void RotateDNS() APT_OVERRIDE { exit(42); } - - public: - - virtual std::unique_ptr<ServerState> CreateServerState(URI const &uri) APT_OVERRIDE; - using pkgAcqMethod::FetchResult; - using pkgAcqMethod::FetchItem; - - explicit HttpsMethod(std::string &&pProg); - virtual ~HttpsMethod(); -}; - -#include <apt-pkg/strutl.h> -URI Proxy; - -#endif diff --git a/po/CMakeLists.txt b/po/CMakeLists.txt index 743b2f0eb..57015eca7 100644 --- a/po/CMakeLists.txt +++ b/po/CMakeLists.txt @@ -10,18 +10,12 @@ apt_add_translation_domain( EXCLUDE_LANGUAGES ${languages_excluded} ) -if (HAVE_CURL) - set(curl_methods curl) -else() - set(curl_methods) -endif() - apt_add_translation_domain( DOMAIN apt TARGETS apt apt-cache apt-get apt-config apt-cdrom apt-helper apt-mark apt-private # Methods - file copy store gpgv cdrom http ${curl_methods} ftp rred rsh mirror + file copy store gpgv cdrom http ftp rred rsh mirror SCRIPTS ../dselect/install ../dselect/update EXCLUDE_LANGUAGES ${languages_excluded} ) diff --git a/test/integration/test-apt-update-failure-propagation b/test/integration/test-apt-update-failure-propagation index 20d941f7d..5e8239f19 100755 --- a/test/integration/test-apt-update-failure-propagation +++ b/test/integration/test-apt-update-failure-propagation @@ -85,13 +85,7 @@ for FILE in rootdir/etc/apt/sources.list.d/*-stable-* ; do sed -i -e "s#:${APTHTTPSPORT}/#:666/#" "$FILE" done testwarning aptget update -o Dir::Bin::Methods::https="${OLDMETHODS}/https" -if grep -q FORCE_CURL:BOOL=OFF $PROJECT_BINARY_DIR/CMakeCache.txt; then testequalor2 "W: Failed to fetch https://localhost:666/dists/stable/InRelease Failed to connect to localhost port 666: Connection refused W: Some index files failed to download. They have been ignored, or old ones used instead." "W: Failed to fetch https://localhost:666/dists/stable/InRelease Could not connect to localhost:666 (127.0.0.1). - connect (111: Connection refused) W: Some index files failed to download. They have been ignored, or old ones used instead." tail -n 2 rootdir/tmp/testwarning.output -else -testequalor2 "W: Failed to fetch https://localhost:666/dists/stable/InRelease Failed to connect to localhost port 666: Connection refused -W: Some index files failed to download. They have been ignored, or old ones used instead." "W: Failed to fetch https://localhost:666/dists/stable/InRelease couldn't connect to host -W: Some index files failed to download. They have been ignored, or old ones used instead." tail -n 2 rootdir/tmp/testwarning.output -fi posttest |