• We value your experience with Plesk during 2024
    Plesk strives to perform even better in 2025. To help us improve further, please answer a few questions about your experience with Plesk Obsidian 2024.
    Please take this short survey:

    https://pt-research.typeform.com/to/AmZvSXkx
  • The Horde webmail has been deprecated. Its complete removal is scheduled for April 2025. For details and recommended actions, see the Feature and Deprecation Plan.
  • We’re working on enhancing the Monitoring feature in Plesk, and we could really use your expertise! If you’re open to sharing your experiences with server and website monitoring or providing feedback, we’d love to have a one-hour online meeting with you.

Resolved Unable to fetch URL: wget -q

Ankebut

Regular Pleskian
Hello,
when i am run a sheduled task, i receive always a error below, in command line sa it me that is support it. I would be very happy if someone would help me

root@x ~ # curl --version | grep Protocols

Protocols: dict file ftp ftps gopher http https imap imaps ldap ldaps pop3 pop3s rtmp rtsp scp sftp smb smbs smtp smtps telnet tftp



Task "wget -q https://doitauto.de/backend/SwpAfterbuyFeeds/Run/pass/19691969/type/import_articles/feedID/10" completed with error in 0 seconds, output:

curl: (1) Protocol "wget -q https" not supported or disabled in libcurl
Unable to fetch URL: wget -q https://doitauto.de/backend/SwpAfterbuyFeeds/Run/pass/19691969/type/import_articles/feedID/10
 
Hello,

could you please how this task looks in plesk UI ?
Or you can execute something like
crontab -l -u YOUR_USER
and past output here (without real username, of course)

btw, are you running this task under some subscription schedule tasks or it is global (server) task ?

is it critical for you to use wget ? Probably it will work for you to use curl instead, something like
curl https://doitauto.de/backend/SwpAfterbuyFeeds/Run/pass/19691969/type/import_articles/feedID/10 -O > /dev/null 2>&1
 
-O tels curl that it must save downloaded file with it's original name (by default it output downloaded content to stdout)

"> /dev/null 2>&1" - this is output redirect (download progress, errors) to the /dev/null - it reduces the log files
you can omit this part. in this case curl messages will be saved to the task log.
 
Back
Top