curl vs Wget
Please let me know if you have other thoughts or comments on this document.
File issues or pull-requests if you find problems or have improvements.
What both commands do
- both are command line tools that can download contents from FTP, HTTP(S)
- both can send HTTP POST requests
- both support HTTP cookies
- both support HSTS and HTTP proxy
- both are designed to work without user interaction
- both are fully open source and free software
- both projects started in 1996 (under other names)
- both are portable and run on many operating systems
How they differ
library: curl is powered by libcurl - a cross-platform library with a stable API that can be used by each and everyone. This difference is major since it creates a completely different attitude on how to do things internally. It is also slightly harder to make a library than a "mere" command line tool.
pipes: curl works more like the traditional Unix cat command, it sends more stuff to stdout, and reads more from stdin in a "everything is a pipe" manner. Wget is more like cp, using the same analogue.
Single shot: curl is basically made to do single-shot transfers of data. It transfers just the URLs that the user specifies, and does not contain any recursive downloading logic nor any sort of HTML parser.
More protocols: curl supports FTP(S), GOPHER(S), HTTP(S), SCP, SFTP, TFTP, TELNET, DICT, LDAP(S), MQTT, FILE, POP3(S), IMAP(S), SMB(S), SMTP(S), RTMP and RTSP. Wget supports HTTP(S) and FTP.
More portable: curl builds and runs on lots of more platforms than wget. For example: OS/400, TPF and other more "exotic" platforms that aren't straight-forward Unix clones. curl requires but a C89 compiler.
More SSL libraries and SSL support: curl can be built with one out of thirteen (13!) different SSL/TLS libraries, and it offers more control and wider support for protocol details.
HTTP auth: curl supports more HTTP authentication methods, especially over HTTP proxies: Basic, Digest, NTLM and Negotiate
SOCKS: curl supports SOCKS4 and SOCKS5 for proxy access. With local or proxy based name resolving.
curl supports HTTPS proxy, that is HTTPS to the proxy. wget does not.
Bidirectional: curl offers upload and sending capabilities. Wget only offers plain HTTP POST support.
HTTP multipart/form-data sending, which allows users to do HTTP "upload" and in general emulate browsers and do HTTP automation to a wider extent.
curl supports gzip, brotli, zstd and deflate Content-Encoding and does automatic decompression.
curl offers and performs decompression of Transfer-Encoded HTTP, wget doesn't.
curl supports HTTP/2, HTTP/3 and alt-svc
curl does dual-stack (IPv4 + IPv6) connects using Happy Eyeballs
curl can do many transfers in parallel (
Much more developer activity. While this can be debated, I consider three metrics here: mailing list activity, source code commit frequency and release frequency. Anyone following these two projects can see that the curl project has a lot higher pace in all these areas, and it has been so for 15+ years. Compare on openhub.
curl comes pre-installed on macOS and Windows 10. Wget does not.
Wget is command line only. There's no library.
Recursive!: Wget's major strong side compared to curl is its ability to download recursively, or even just download everything that is referred to from a remote resource, be it a HTML page or a FTP directory listing.
GPL: Wget is GPL v3. curl is MIT licensed.
GNU: Wget is part of the GNU project and all copyrights are assigned to FSF. The curl project is entirely stand-alone and independent with no organization parenting at all with almost all copyrights owned by Daniel.
Wget requires no extra options to simply download a remote URL to a local file, while curl requires -o or -O.
Wget supports only GnuTLS or OpenSSL for SSL/TLS support.
Wget supports only Basic auth as the only auth type over HTTP proxy.
Wget has no SOCKS support.
Its ability to recover from a prematurely broken transfer and continue downloading has no counterpart in curl.
Wget still supports metalink, curl dropped that support due to security concerns
Wget enables more features by default: cookies, redirect-following, time stamping from the remote resource etc. With curl most of those features need to be explicitly enabled.
There's a 'wget' in BusyBox, there's no curl there (it is not the actual wget, just a stripped down clone with the same name).
Wget can be typed in using only the left hand on a qwerty keyboard!
Wget requires a C99 compiler and also relies on gnulib.
When to use which
Primarily: use the one that gets the job done for you.
Wget has (recursive) downloading powers that curl does not feature and it also handle download retries over unreliable connections possibly slightly more effective.
For just about everything else, curl is probably the more suitable tool.
In recent years, wget2 is worked on to become the replacement for wget. This comparison will eventually get wget2 details as well.
For a stricter feature by feature comparison (that also compares other similar tools), see the curl comparison table
Feedback and improvements by: Micah Cowan, Olemis Lang
Updated: June 18, 2022 12:25 (Central European, Stockholm Sweden)