<feed xmlns='http://www.w3.org/2005/Atom'>
<title>curl/tests/data/test273, branch gemini-meta</title>
<subtitle>cURL mirror with patches applied
</subtitle>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/'/>
<entry>
<title>http: always send Host: header as first header</title>
<updated>2015-03-12T11:15:24+00:00</updated>
<author>
<name>Daniel Stenberg</name>
<email>daniel@haxx.se</email>
</author>
<published>2015-03-11T23:50:46+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=a5d994941c2b19a3c63402caa064c9683647de05'/>
<id>a5d994941c2b19a3c63402caa064c9683647de05</id>
<content type='text'>
...after the method line:

 "Since the Host field-value is critical information for handling a
 request, a user agent SHOULD generate Host as the first header field
 following the request-line." / RFC 7230 section 5.4

Additionally, this will also make libcurl ignore multiple specified
custom Host: headers and only use the first one. Test 1121 has been
updated accordingly

Bug: http://curl.haxx.se/bug/view.cgi?id=1491
Reported-by: Rainer Canavan
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
...after the method line:

 "Since the Host field-value is critical information for handling a
 request, a user agent SHOULD generate Host as the first header field
 following the request-line." / RFC 7230 section 5.4

Additionally, this will also make libcurl ignore multiple specified
custom Host: headers and only use the first one. Test 1121 has been
updated accordingly

Bug: http://curl.haxx.se/bug/view.cgi?id=1491
Reported-by: Rainer Canavan
</pre>
</div>
</content>
</entry>
<entry>
<title>tests: Don't run HTTP digest tests for SSPI based builds</title>
<updated>2014-11-06T20:36:08+00:00</updated>
<author>
<name>Steve Holme</name>
<email>steve_holme@hotmail.com</email>
</author>
<published>2014-11-06T19:57:16+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=b5c496f1090ab8a8c535acde33af61f197bd99ba'/>
<id>b5c496f1090ab8a8c535acde33af61f197bd99ba</id>
<content type='text'>
Added !SSPI to the features list of the HTTP digest tests, as SSPI
based builds now use the Windows SSPI messaging API rather than the
internal functions, and we can't control the random numbers that get
used as part of the digest.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Added !SSPI to the features list of the HTTP digest tests, as SSPI
based builds now use the Windows SSPI messaging API rather than the
internal functions, and we can't control the random numbers that get
used as part of the digest.
</pre>
</div>
</content>
</entry>
<entry>
<title>Do not mix CRLF and LF header endings in a couple of HTTP tests</title>
<updated>2012-11-19T09:58:02+00:00</updated>
<author>
<name>Fabian Keil</name>
<email>fk@fabiankeil.de</email>
</author>
<published>2012-11-14T18:49:12+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=ab0fa55780b518502ff0712f37e8352323389cde'/>
<id>ab0fa55780b518502ff0712f37e8352323389cde</id>
<content type='text'>
Consistently use CRLF instead. The mixed endings weren't
documented so I assume they were unintentional.

This change doesn't matter for curl itself but makes using
the tests with a proxy between curl and the test server
more convenient.

Tests that consistently use no carriage returns were
left unmodified as one can easily work around this.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Consistently use CRLF instead. The mixed endings weren't
documented so I assume they were unintentional.

This change doesn't matter for curl itself but makes using
the tests with a proxy between curl and the test server
more convenient.

Tests that consistently use no carriage returns were
left unmodified as one can easily work around this.
</pre>
</div>
</content>
</entry>
<entry>
<title>Replaced 127.0.0.1 with %HOSTIP where possible</title>
<updated>2007-09-14T19:32:31+00:00</updated>
<author>
<name>Dan Fandrich</name>
<email>dan@coneharvesters.com</email>
</author>
<published>2007-09-14T19:32:31+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=05e4a3026da7c2dfbd3ed340a01d4a221a721168'/>
<id>05e4a3026da7c2dfbd3ed340a01d4a221a721168</id>
<content type='text'>
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
</pre>
</div>
</content>
</entry>
<entry>
<title>Updated the test harness to add a new "crypto" feature check and updated the</title>
<updated>2007-03-09T23:39:42+00:00</updated>
<author>
<name>Dan Fandrich</name>
<email>dan@coneharvesters.com</email>
</author>
<published>2007-03-09T23:39:42+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=e12d46ac50ddce4330d1a2e1d4a7e5e2accd280e'/>
<id>e12d46ac50ddce4330d1a2e1d4a7e5e2accd280e</id>
<content type='text'>
appropriate test case to use it.  For now, this is treated the same as the
"SSL" feature because curl doesn't list it separately.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
appropriate test case to use it.  For now, this is treated the same as the
"SSL" feature because curl doesn't list it separately.
</pre>
</div>
</content>
</entry>
<entry>
<title>Convert (most of) the test data files into genuine XML.  A handful still</title>
<updated>2007-01-23T02:25:56+00:00</updated>
<author>
<name>Dan Fandrich</name>
<email>dan@coneharvesters.com</email>
</author>
<published>2007-01-23T02:25:56+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=33bea767ebd0bbf11867fe1af31dc345e9f816af'/>
<id>33bea767ebd0bbf11867fe1af31dc345e9f816af</id>
<content type='text'>
are not, due mainly to the lack of support for XML character entities
(e.g. &amp; =&gt; &amp;amp; ).  This will make it easier to validate test files using
tools like xmllint, as well as edit and view them using XML tools.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
are not, due mainly to the lack of support for XML character entities
(e.g. &amp; =&gt; &amp;amp; ).  This will make it easier to validate test files using
tools like xmllint, as well as edit and view them using XML tools.
</pre>
</div>
</content>
</entry>
<entry>
<title>Venkat Akella found out that libcurl did not like HTTP responses that simply</title>
<updated>2006-11-25T13:32:04+00:00</updated>
<author>
<name>Daniel Stenberg</name>
<email>daniel@haxx.se</email>
</author>
<published>2006-11-25T13:32:04+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=da58d03ff7be9fc5e0219d95ce2f9932e2272473'/>
<id>da58d03ff7be9fc5e0219d95ce2f9932e2272473</id>
<content type='text'>
responded with a single status line and no headers nor body. Starting now, a
HTTP response on a persistent connection (i.e not set to be closed after the
response has been taken care of) must have Content-Length or chunked
encoding set, or libcurl will simply assume that there is no body.

To my horror I learned that we had no less than 57(!) test cases that did bad
HTTP responses like this, and even the test http server (sws) responded badly
when queried by the test system if it is the test system. So although the
actual fix for the problem was tiny, going through all the newly failing test
cases got really painful and boring.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
responded with a single status line and no headers nor body. Starting now, a
HTTP response on a persistent connection (i.e not set to be closed after the
response has been taken care of) must have Content-Length or chunked
encoding set, or libcurl will simply assume that there is no body.

To my horror I learned that we had no less than 57(!) test cases that did bad
HTTP responses like this, and even the test http server (sws) responded badly
when queried by the test system if it is the test system. So although the
actual fix for the problem was tiny, going through all the newly failing test
cases got really painful and boring.
</pre>
</div>
</content>
</entry>
<entry>
<title>Dave Dribin made libcurl understand and handle cases when the server</title>
<updated>2005-10-20T20:07:32+00:00</updated>
<author>
<name>Daniel Stenberg</name>
<email>daniel@haxx.se</email>
</author>
<published>2005-10-20T20:07:32+00:00</published>
<link rel='alternate' type='text/html' href='https://git.benburwell.com/mirrors/curl/commit/?id=be9c873a6e97423bc0b2a2dd45835c35c7d81231'/>
<id>be9c873a6e97423bc0b2a2dd45835c35c7d81231</id>
<content type='text'>
(wrongly) sends *two* WWW-Authenticate headers for Digest. While this should
never happen in a sane world, libcurl previously got into an infinite loop
when this occurred. Dave added test 273 to verify this.
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
(wrongly) sends *two* WWW-Authenticate headers for Digest. While this should
never happen in a sane world, libcurl previously got into an infinite loop
when this occurred. Dave added test 273 to verify this.
</pre>
</div>
</content>
</entry>
</feed>
