Questions tagged [wget]
GNU Wget is a free software package for retrieving files using HTTP, HTTPS and FTP, the most widely-used Internet protocols. It is a non-interactive command line tool, so it may easily be called from scripts, Cron jobs, terminals without X Window System (X11) support, etc.
917 questions
1
vote
1
answer
120
views
Why do wget and curl download a specific file only partially?
I want to download a specific WordPress sitemap, but wget and curl both fail to download the latest sitemap fully. If I open the sitemap in a browser, that sitemap simply has more links in it, then ...
1
vote
1
answer
192
views
Skip previously downloaded files in wget or alternatives
I used wget to download all media from a website. I would like to download new uploads from the site, but I also want to be able to delete unneeded files to avoid clutter and save space.
I've tried ...
4
votes
1
answer
725
views
wget throws "Is a directory" error
I need to download about 10,000 files from a HTTP server. I am using wget. It works. This command created a subfolder 192.168.1.173+8080/files with about 4,000 files, after which I believe the HTTP ...
1
vote
1
answer
81
views
How to explore folders of a web directory in the terminal?
I am analyzing some data using my university SSH server. Some new data was shared with me in a web format (share.servername.com/url/XYZ/abc/…).
The page includes multiple folders and subfolders. I ...
1
vote
0
answers
32
views
How does WGET handle a large number of concurrent FTP downloads where the .listing file always has the exact same name? No conflicts?
I download several dozen files from an FTP server as they're updated.
Each file is retrieved with a separate WGET command which is called again every few minutes. Since all these WGET calls always ...
0
votes
0
answers
85
views
wget converting links not working for all links when trying to archive a wordpress blog
I'm trying to archive a wordpress website using wget.
Here's the command line I'm using:
wget --mirror --page-requisites --convert-links -k -e robots=off https://url.info
I tried different variations ...
0
votes
1
answer
286
views
wget - how to download all files of specific type from a website?
I'm trying to batch download hundreds of .mp3 files from a website.
Entering this:
wget -A mp3 http://traffic.libsyn.com/gardnermuseum/
Gives this:
--2025-03-30 11:39:34-- http://traffic.libsyn.com/...
0
votes
1
answer
186
views
How to find wget download location for Amazon Corretto in PowerShell 5?
I would like to run a package for Python, py-causal, that uses a Java library, Tetrad. I am stuck at getting Java on Windows 10.
The Tetrad repo has a link to this Amazon page, where I ended up on the ...
3
votes
1
answer
899
views
Catching download filename of wget or cURL command?
I've been trying to figure out how to "catch" the filename of a downloaded file via wget or curl commands in a shell script.
What I'd like (and what seems impossible) is something like this:
...
1
vote
0
answers
186
views
How to wait a server for SSL handshake using Curl?
I need get a website using Curl but the server does not respond the SSL connection and Curl says an error on the 2 minutes but the server respond in the minute ~5. I try to use the timeout argument ...
1
vote
2
answers
268
views
Use the page title for HTML file name when downloaded using wget
I can download a single stand alone HTML file with:
wget https://www.bbc.co.uk/news/articles/c99rgj0xkryo
but wget will save the file as index.html rather than Nation falls silent as King leads ...
1
vote
1
answer
122
views
How to archive a website with occasional PHP errors?
I am trying to archive a website that will soon vanish. I tried wget and httrack.
The problem is that the website returns PHP errors (database connection error) from time to time and the downloaded ...
0
votes
1
answer
213
views
How to mirror a google.sites site with wget and convert links into local link references?
I am trying to mirror/download onto my PC a sites.google website I have using wget. I can get it to download no problems but when I click on index.html it starts with the local folder and then links ...
1
vote
0
answers
310
views
Is there a possibility for wget to handle long(er) paths?
I am using wget for a local backup of a remote ftp/s directory using the following syntax:
wget --no-dns-cache -nv -r -c -l inf -P/local_dir -nH --cut-dirs=1 --restrict-file-names=windows --progress=...
1
vote
1
answer
219
views
How to download a remote directory list and not its contents?
From a remote directory I have to automatically download a file of which I know part of the name, only (e.g. "file-vers.1.2.3.zip", where the version number is the unknown part).
Since ...