From b88940850002a3f1c25bc6488b95ad30eb80d696 Mon Sep 17 00:00:00 2001 From: Daniel Stenberg Date: Sat, 20 Jul 2019 19:14:00 +0200 Subject: curl: support parallel transfers This is done by making sure each individual transfer is first added to a linked list as then they can be performed serially, or at will, in parallel. Closes #3804 --- tests/data/test1291 | 10 ++-------- 1 file changed, 2 insertions(+), 8 deletions(-) (limited to 'tests/data/test1291') diff --git a/tests/data/test1291 b/tests/data/test1291 index 3f1575184..a2e505fc9 100644 --- a/tests/data/test1291 +++ b/tests/data/test1291 @@ -21,7 +21,7 @@ HTTP PUT none -Attempt to upload 100K files but fail immediately +Attempt to upload 1000 files but fail immediately -K log/cmd1291 --fail-early @@ -31,7 +31,7 @@ XXXXXXXx # generate the config file -perl -e 'for(1 .. 100000) { printf("upload-file=log/upload-this\nurl=htttttp://non-existing-host.haxx.se/upload/1291\n", $_);}' > log/cmd1291; +perl -e 'for(1 .. 1000) { printf("upload-file=log/upload-this\nurl=htttttp://non-existing-host.haxx.se/upload/1291\n", $_);}' > log/cmd1291; @@ -40,11 +40,5 @@ perl -e 'for(1 .. 100000) { printf("upload-file=log/upload-this\nurl=htttttp://n 1 - -# we disable valgrind here since it takes 40+ seconds even on a fairly snappy -# machine - -disable - -- cgit v1.2.3