Age | Commit message (Collapse) | Author |
|
USE_WINDOWS_SSPI on Windows, and then libcurl will be built to use the native
way to do NTLM. SSPI also allows libcurl to pass on the current user and its
password in the request.
|
|
referencing the proxy name pointer.
|
|
|
|
that feature 64 bit 'long'.
Some systems have 64 bit time_t and deal with years beyond 2038. However, even
some of the systems with 64 bit time_t returns -1 for dates beyond 03:14:07
UTC, January 19, 2038. (Such as AIX 5100-06)
|
|
file got a Last-Modified: header written to the data stream, corrupting the
actual data. This was because some conditions from the previous FTP code was
not properly brought into the new FTP code. I fixed and I added test case 520
to verify. (This bug was introduced in 7.13.1)
|
|
on the remote side. This then converts the operation to an ordinary STOR
upload. This was requested/pointed out by Ignacio Vazquez-Abrams.
It also proved (and I fixed) a bug in the newly rewritten ftp code (and
present in the 7.13.1 release) when trying to resume an upload and the servers
returns an error to the SIZE command. libcurl then loops and sends SIZE
commands infinitely.
|
|
|
|
used, it slows initialization too much reading an infinitely long file!
|
|
the cookie "engine" without having to provide an empty or non-existing file.
|
|
|
|
|
|
|
|
server.
|
|
|
|
|
|
got no notification, no mail, no nothing.
You didn't even bother to mail us when you went public with this. Cool.
NTLM buffer overflow fix, as reported here:
http://www.securityfocus.com/archive/1/391042
|
|
requested data from a host and then followed a redirect to another
host. libcurl then didn't use the proxy-auth properly in the second request,
due to the host-only check for original host name wrongly being extended to
the proxy auth as well. Added test case 233 to verify the flaw and that the
fix removed the problem.
|
|
Bug report #1124588 by David
|
|
that picks NTLM. Thanks to David Byron letting me test NTLM against his
servers, I could quickly repeat and fix the problem. It turned out to be:
When libcurl POSTs without knowing/using an authentication and it gets back a
list of types from which it picks NTLM, it needs to either continue sending
its data if it keeps the connection alive, or not send the data but close the
connection. Then do the first step in the NTLM auth. libcurl didn't send the
data nor close the connection but simply read the response-body and then sent
the first negotiation step. Which then failed miserably of course. The fixed
version forces a connection if there is more than 2000 bytes left to send.
|
|
do pretransfer stuff like Curl_pretransfer().
|
|
|
|
week day names and month names and servers don't like that.
|
|
|
|
|
|
at fixing this issue.
|
|
|
|
|
|
Don't free 'lud_dn' twice in case curl_unescape()
fails.
|
|
|
|
For ftp only?
|
|
|
|
The tag 'before_ftp_statemachine' was set just before this commit in case
of future need.
|
|
"Each command line must be terminated by a CRLF".
|
|
|
|
CURLE_OK or CURLE_OUT_OF_MEMORY.
Add typecast in hostares.c.
|
|
|
|
operation to the caller. Disconnecting has the disadvantage that the conn
pointer gets completely invalidated and this is not handled on lots of places
in the code.
|
|
produce a zlib error.
|
|
|
|
|
|
the buffer is already BUFSIZE +1 one big to fit the extra trailing zero. This
change is reported to fix David's weird SSL problem...
|
|
prevent memory leakage
|
|
properly, to avoid reading uninited variables when using file:// (valgrind)
|
|
|
|
gets closed just after the request has been sent failed and did not re-issue
a request on a fresh reconnect like the easy interface did. Now it does!
(define CURL_MULTIEASY, run test case 160)
|
|
that uses the multi interface to run the request. It is a great testbed for
the multi interface and I believe we shall do it this way for real in the
future when we have a successor to curl_multi_fdset().
|
|
|
|
libcurl leaked the last Location: URL.
|
|
timed out" even though the reason was different. Fixed this problem by not
setting this timeout to zero when using multi.
|
|
curl_easy_perform() invokes. It was previously unlocked at disconnect, which
could mean that it remained locked between multiple transfers. The DNS cache
may not live as long as the connection cache does, as they are separate.
To deal with the lack of DNS (host address) data availability in re-used
connections, libcurl now keeps a copy of the IP adress as a string, to be able
to show it even on subsequent requests on the same connection.
|