• If you are still using CentOS 7.9, it's time to convert to Alma 8 with the free centos2alma tool by Plesk or Plesk Migrator. Please let us know your experiences or concerns in this thread:
    CentOS2Alma discussion
  • Inviting everyone to the UX test of a new security feature in the WP Toolkit
    For WordPress site owners, threats posed by hackers are ever-present. Because of this, we are developing a new security feature for the WP Toolkit. If the topic of WordPress website security is relevant to you, we would be grateful if you could share your experience and help us test the usability of this feature. We invite you to join us for a 1-hour online session via Google Meet. Select a convenient meeting time with our friendly UX staff here.

Resolved Unable to fetch URL: wget -q

Ankebut

Regular Pleskian
Hello,
when i am run a sheduled task, i receive always a error below, in command line sa it me that is support it. I would be very happy if someone would help me

root@x ~ # curl --version | grep Protocols

Protocols: dict file ftp ftps gopher http https imap imaps ldap ldaps pop3 pop3s rtmp rtsp scp sftp smb smbs smtp smtps telnet tftp



Task "wget -q https://doitauto.de/backend/SwpAfterbuyFeeds/Run/pass/19691969/type/import_articles/feedID/10" completed with error in 0 seconds, output:

curl: (1) Protocol "wget -q https" not supported or disabled in libcurl
Unable to fetch URL: wget -q https://doitauto.de/backend/SwpAfterbuyFeeds/Run/pass/19691969/type/import_articles/feedID/10
 
Hello,

could you please how this task looks in plesk UI ?
Or you can execute something like
crontab -l -u YOUR_USER
and past output here (without real username, of course)

btw, are you running this task under some subscription schedule tasks or it is global (server) task ?

is it critical for you to use wget ? Probably it will work for you to use curl instead, something like
curl https://doitauto.de/backend/SwpAfterbuyFeeds/Run/pass/19691969/type/import_articles/feedID/10 -O > /dev/null 2>&1
 
-O tels curl that it must save downloaded file with it's original name (by default it output downloaded content to stdout)

"> /dev/null 2>&1" - this is output redirect (download progress, errors) to the /dev/null - it reduces the log files
you can omit this part. in this case curl messages will be saved to the task log.
 
Back
Top