<feed xmlns='http://www.w3.org/2005/Atom'>
<title>curl/tests/data/test153, branch master</title>
<subtitle>cURL mirror with patches applied
</subtitle>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/'/>
<entry>
<title>digest: Do not quote algorithm in HTTP authorisation</title>
<updated>2020-02-07T21:46:21+00:00</updated>
<author>
<name>Pierre-Yves Bigourdan</name>
<email>pierre-yves.bigourdan@bbc.co.uk</email>
</author>
<published>2020-02-06T15:43:06+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=feba3f0549af546677ee2d924344234fd91b8f50'/>
<id>feba3f0549af546677ee2d924344234fd91b8f50</id>
<content type='text'>
RFC 7616 section 3.4 (The Authorization Header Field) states that "For
historical reasons, a sender MUST NOT generate the quoted string syntax
for the following parameters: algorithm, qop, and nc". This removes the
quoting for the algorithm parameter.

Reviewed-by: Steve Holme
Closes #4890
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
RFC 7616 section 3.4 (The Authorization Header Field) states that "For
historical reasons, a sender MUST NOT generate the quoted string syntax
for the following parameters: algorithm, qop, and nc". This removes the
quoting for the algorithm parameter.

Reviewed-by: Steve Holme
Closes #4890
</pre>
</div>
</content>
</entry>
<entry>
<title>curl: support parallel transfers</title>
<updated>2019-07-20T17:14:16+00:00</updated>
<author>
<name>Daniel Stenberg</name>
<email>daniel@haxx.se</email>
</author>
<published>2019-07-20T17:14:00+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=b88940850002a3f1c25bc6488b95ad30eb80d696'/>
<id>b88940850002a3f1c25bc6488b95ad30eb80d696</id>
<content type='text'>
This is done by making sure each individual transfer is first added to a
linked list as then they can be performed serially, or at will, in
parallel.

Closes #3804
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This is done by making sure each individual transfer is first added to a
linked list as then they can be performed serially, or at will, in
parallel.

Closes #3804
</pre>
</div>
</content>
</entry>
<entry>
<title>test153: fix content-length to avoid occasional hang</title>
<updated>2019-06-24T06:33:32+00:00</updated>
<author>
<name>Daniel Stenberg</name>
<email>daniel@haxx.se</email>
</author>
<published>2019-06-23T21:12:48+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=1e9769639b57ffd4f587d39efafaae6ff1d75c4a'/>
<id>1e9769639b57ffd4f587d39efafaae6ff1d75c4a</id>
<content type='text'>
Closes #4065
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Closes #4065
</pre>
</div>
</content>
</entry>
<entry>
<title>http: always send Host: header as first header</title>
<updated>2015-03-12T11:15:24+00:00</updated>
<author>
<name>Daniel Stenberg</name>
<email>daniel@haxx.se</email>
</author>
<published>2015-03-11T23:50:46+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=a5d994941c2b19a3c63402caa064c9683647de05'/>
<id>a5d994941c2b19a3c63402caa064c9683647de05</id>
<content type='text'>
...after the method line:

 "Since the Host field-value is critical information for handling a
 request, a user agent SHOULD generate Host as the first header field
 following the request-line." / RFC 7230 section 5.4

Additionally, this will also make libcurl ignore multiple specified
custom Host: headers and only use the first one. Test 1121 has been
updated accordingly

Bug: http://curl.haxx.se/bug/view.cgi?id=1491
Reported-by: Rainer Canavan
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
...after the method line:

 "Since the Host field-value is critical information for handling a
 request, a user agent SHOULD generate Host as the first header field
 following the request-line." / RFC 7230 section 5.4

Additionally, this will also make libcurl ignore multiple specified
custom Host: headers and only use the first one. Test 1121 has been
updated accordingly

Bug: http://curl.haxx.se/bug/view.cgi?id=1491
Reported-by: Rainer Canavan
</pre>
</div>
</content>
</entry>
<entry>
<title>tests: Don't run HTTP digest tests for SSPI based builds</title>
<updated>2014-11-06T20:36:08+00:00</updated>
<author>
<name>Steve Holme</name>
<email>steve_holme@hotmail.com</email>
</author>
<published>2014-11-06T19:57:16+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=b5c496f1090ab8a8c535acde33af61f197bd99ba'/>
<id>b5c496f1090ab8a8c535acde33af61f197bd99ba</id>
<content type='text'>
Added !SSPI to the features list of the HTTP digest tests, as SSPI
based builds now use the Windows SSPI messaging API rather than the
internal functions, and we can't control the random numbers that get
used as part of the digest.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Added !SSPI to the features list of the HTTP digest tests, as SSPI
based builds now use the Windows SSPI messaging API rather than the
internal functions, and we can't control the random numbers that get
used as part of the digest.
</pre>
</div>
</content>
</entry>
<entry>
<title>Added some &lt;keywords&gt; sections and use some key words more consistently.</title>
<updated>2007-10-12T01:44:22+00:00</updated>
<author>
<name>Dan Fandrich</name>
<email>dan@coneharvesters.com</email>
</author>
<published>2007-10-12T01:44:22+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=d7fbe07ee24f2bf1e0f158577a7c995426ef9958'/>
<id>d7fbe07ee24f2bf1e0f158577a7c995426ef9958</id>
<content type='text'>
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
</pre>
</div>
</content>
</entry>
<entry>
<title>Replaced 127.0.0.1 with %HOSTIP where possible</title>
<updated>2007-09-14T19:32:31+00:00</updated>
<author>
<name>Dan Fandrich</name>
<email>dan@coneharvesters.com</email>
</author>
<published>2007-09-14T19:32:31+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=05e4a3026da7c2dfbd3ed340a01d4a221a721168'/>
<id>05e4a3026da7c2dfbd3ed340a01d4a221a721168</id>
<content type='text'>
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
</pre>
</div>
</content>
</entry>
<entry>
<title>Updated the test harness to add a new "crypto" feature check and updated the</title>
<updated>2007-03-09T23:39:42+00:00</updated>
<author>
<name>Dan Fandrich</name>
<email>dan@coneharvesters.com</email>
</author>
<published>2007-03-09T23:39:42+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=e12d46ac50ddce4330d1a2e1d4a7e5e2accd280e'/>
<id>e12d46ac50ddce4330d1a2e1d4a7e5e2accd280e</id>
<content type='text'>
appropriate test case to use it.  For now, this is treated the same as the
"SSL" feature because curl doesn't list it separately.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
appropriate test case to use it.  For now, this is treated the same as the
"SSL" feature because curl doesn't list it separately.
</pre>
</div>
</content>
</entry>
<entry>
<title>Convert (most of) the test data files into genuine XML.  A handful still</title>
<updated>2007-01-23T02:25:56+00:00</updated>
<author>
<name>Dan Fandrich</name>
<email>dan@coneharvesters.com</email>
</author>
<published>2007-01-23T02:25:56+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=33bea767ebd0bbf11867fe1af31dc345e9f816af'/>
<id>33bea767ebd0bbf11867fe1af31dc345e9f816af</id>
<content type='text'>
are not, due mainly to the lack of support for XML character entities
(e.g. &amp; =&gt; &amp;amp; ).  This will make it easier to validate test files using
tools like xmllint, as well as edit and view them using XML tools.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
are not, due mainly to the lack of support for XML character entities
(e.g. &amp; =&gt; &amp;amp; ).  This will make it easier to validate test files using
tools like xmllint, as well as edit and view them using XML tools.
</pre>
</div>
</content>
</entry>
<entry>
<title>Venkat Akella found out that libcurl did not like HTTP responses that simply</title>
<updated>2006-11-25T13:32:04+00:00</updated>
<author>
<name>Daniel Stenberg</name>
<email>daniel@haxx.se</email>
</author>
<published>2006-11-25T13:32:04+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=da58d03ff7be9fc5e0219d95ce2f9932e2272473'/>
<id>da58d03ff7be9fc5e0219d95ce2f9932e2272473</id>
<content type='text'>
responded with a single status line and no headers nor body. Starting now, a
HTTP response on a persistent connection (i.e not set to be closed after the
response has been taken care of) must have Content-Length or chunked
encoding set, or libcurl will simply assume that there is no body.

To my horror I learned that we had no less than 57(!) test cases that did bad
HTTP responses like this, and even the test http server (sws) responded badly
when queried by the test system if it is the test system. So although the
actual fix for the problem was tiny, going through all the newly failing test
cases got really painful and boring.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
responded with a single status line and no headers nor body. Starting now, a
HTTP response on a persistent connection (i.e not set to be closed after the
response has been taken care of) must have Content-Length or chunked
encoding set, or libcurl will simply assume that there is no body.

To my horror I learned that we had no less than 57(!) test cases that did bad
HTTP responses like this, and even the test http server (sws) responded badly
when queried by the test system if it is the test system. So although the
actual fix for the problem was tiny, going through all the newly failing test
cases got really painful and boring.
</pre>
</div>
</content>
</entry>
</feed>
