aboutsummaryrefslogtreecommitdiff
path: root/docs
diff options
context:
space:
mode:
Diffstat (limited to 'docs')
-rw-r--r--docs/examples/fopen.c662
1 files changed, 496 insertions, 166 deletions
diff --git a/docs/examples/fopen.c b/docs/examples/fopen.c
index 8fc4b1f7c..452995bc3 100644
--- a/docs/examples/fopen.c
+++ b/docs/examples/fopen.c
@@ -1,28 +1,53 @@
/*****************************************************************************
- * _ _ ____ _
- * Project ___| | | | _ \| |
- * / __| | | | |_) | |
- * | (__| |_| | _ <| |___
- * \___|\___/|_| \_\_____|
*
- * $Id$
+ * This example source code introduces a c library buffered I/O interface to
+ * URL reads it supports fopen(), fread(), fgets(), feof(), fclose(),
+ * rewind(). Supported functions have identical prototypes to their normal c
+ * lib namesakes and are preceaded by url_ .
*
- * This example source code introduces an fopen()/fread()/fclose() emulation
- * for URL reads. Using an approach similar to this, you could replace your
- * program's fopen() with this url_fopen() and fread() with url_fread() and
- * it should be possible to read remote streams instead of (only) local files.
+ * Using this code you can replace your program's fopen() with url_fopen()
+ * and fread() with url_fread() and it become possible to read remote streams
+ * instead of (only) local files. Local files (ie those that can be directly
+ * fopened) will drop back to using the underlying clib implementations
*
- * See the main() function at the bottom that shows a tiny app in action.
+ * See the main() function at the bottom that shows an app that retrives from a
+ * specified url using fgets() and fread() and saves as two output files.
*
- * This source code is a proof of concept. It will need further attention to
- * become production-use useful and solid.
+ * Coyright (c)2003 Simtec Electronics
+ *
+ * Re-implemented by Vincent Sanders <vince@kyllikki.org> with extensive
+ * reference to original curl example code
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. The name of the author may not be used to endorse or promote products
+ * derived from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
+ * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
+ * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
+ * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
+ * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
+ * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
+ * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
* This example requires libcurl 7.9.7 or later.
*/
+
#include <stdio.h>
#include <string.h>
#include <sys/time.h>
#include <stdlib.h>
+#include <errno.h>
#include <curl/curl.h>
@@ -30,206 +55,511 @@
#error "too old libcurl version, get the latest!"
#endif
-struct data {
- int type;
- union {
- CURL *curl;
- FILE *file;
- } handle;
-
- /* This is the documented biggest possible buffer chunk we can get from
- libcurl in one single callback! */
- char buffer[CURL_MAX_WRITE_SIZE];
- char *readptr; /* read from here */
- int bytes; /* bytes available from read pointer */
+enum fcurl_type_e { CFTYPE_NONE=0, CFTYPE_FILE=1, CFTYPE_CURL=2 };
- CURLMcode m; /* stored from a previous url_fread() */
+struct fcurl_data
+{
+ enum fcurl_type_e type; /* type of handle */
+ union {
+ CURL *curl;
+ FILE *file;
+ } handle; /* handle */
+
+ char *buffer; /* buffer to store cached data*/
+ int buffer_len; /* currently allocated buffers length */
+ int buffer_pos; /* end of data in buffer*/
+ int still_running; /* Is background url fetch still in progress */
};
-typedef struct data URL_FILE;
+typedef struct fcurl_data URL_FILE;
+
+/* exported functions */
+URL_FILE *url_fopen(char *url,const char *operation);
+int url_fclose(URL_FILE *file);
+int url_feof(URL_FILE *file);
+size_t url_fread(void *ptr, size_t size, size_t nmemb, URL_FILE *file);
+char * url_fgets(char *ptr, int size, URL_FILE *file);
+void url_rewind(URL_FILE *file);
/* we use a global one for convenience */
CURLM *multi_handle;
-static
-size_t write_callback(char *buffer,
- size_t size,
- size_t nitems,
- void *userp)
+/* curl calls this routine to get more data */
+static size_t
+write_callback(char *buffer,
+ size_t size,
+ size_t nitems,
+ void *userp)
{
- URL_FILE *url = (URL_FILE *)userp;
- size *= nitems;
+ char *newbuff;
+ int rembuff;
+
+ URL_FILE *url = (URL_FILE *)userp;
+ size *= nitems;
+
+ rembuff=url->buffer_len - url->buffer_pos;//remaining space in buffer
+
+ if(size > rembuff)
+ {
+ //not enuf space in buffer
+ newbuff=realloc(url->buffer,url->buffer_len + (size - rembuff));
+ if(newbuff==NULL)
+ {
+ fprintf(stderr,"callback buffer grow failed\n");
+ size=rembuff;
+ }
+ else
+ {
+ /* realloc suceeded increase buffer size*/
+ url->buffer_len+=size - rembuff;
+ url->buffer=newbuff;
+
+ /*printf("Callback buffer grown to %d bytes\n",url->buffer_len);*/
+ }
+ }
+
+ memcpy(&url->buffer[url->buffer_pos], buffer, size);
+ url->buffer_pos += size;
- memcpy(url->readptr, buffer, size);
- url->readptr += size;
- url->bytes += size;
+ /*fprintf(stderr, "callback %d size bytes\n", size);*/
+
+ return size;
+}
- fprintf(stderr, "callback %d size bytes\n", size);
+/* use to attempt to fill the read buffer up to requested number of bytes */
+static int
+curl_fill_buffer(URL_FILE *file,int want,int waittime)
+{
+ fd_set fdread;
+ fd_set fdwrite;
+ fd_set fdexcep;
+ int maxfd;
+ struct timeval timeout;
+ int rc;
+
+ /* only attempt to fill buffer if transactions still running and buffer
+ * doesnt exceed required size already
+ */
+ if((!file->still_running) || (file->buffer_pos > want))
+ return 0;
+
+ /* attempt to fill buffer */
+ do
+ {
+ FD_ZERO(&fdread);
+ FD_ZERO(&fdwrite);
+ FD_ZERO(&fdexcep);
+
+ /* set a suitable timeout to fail on */
+ timeout.tv_sec = 60; /* 1 minute */
+ timeout.tv_usec = 0;
+
+ /* get file descriptors from the transfers */
+ curl_multi_fdset(multi_handle, &fdread, &fdwrite, &fdexcep, &maxfd);
+
+ rc = select(maxfd+1, &fdread, &fdwrite, &fdexcep, &timeout);
+
+ switch(rc) {
+ case -1:
+ /* select error */
+ break;
+
+ case 0:
+ break;
+
+ default:
+ /* timeout or readable/writable sockets */
+ /* note we *could* be more efficient and not wait for
+ * CURLM_CALL_MULTI_PERFORM to clear here and check it on re-entry
+ * but that gets messy */
+ while(curl_multi_perform(multi_handle, &file->still_running) ==
+ CURLM_CALL_MULTI_PERFORM);
+
+ break;
+ }
+ } while(file->still_running && (file->buffer_pos < want));
+ return 1;
+}
- return size;
+/* use to remove want bytes from the front of a files buffer */
+static int
+curl_use_buffer(URL_FILE *file,int want)
+{
+ /* sort out buffer */
+ if((file->buffer_pos - want) <=0)
+ {
+ /* ditch buffer - write will recreate */
+ if(file->buffer)
+ free(file->buffer);
+
+ file->buffer=NULL;
+ file->buffer_pos=0;
+ file->buffer_len=0;
+ }
+ else
+ {
+ /* move rest down make it available for later */
+ memmove(file->buffer,
+ &file->buffer[want],
+ (file->buffer_pos - want));
+
+ file->buffer_pos -= want;
+ }
+ return 0;
}
-URL_FILE *url_fopen(char *url, char *operation)
+
+
+URL_FILE *
+url_fopen(char *url,const char *operation)
{
- /* this code could check for URLs or types in the 'url' and
- basicly use the real fopen() for standard files */
+ /* this code could check for URLs or types in the 'url' and
+ basicly use the real fopen() for standard files */
+
+ URL_FILE *file;
+ (void)operation;
+
+ file = (URL_FILE *)malloc(sizeof(URL_FILE));
+ if(!file)
+ return NULL;
- URL_FILE *file;
- int still_running;
- (void)operation;
+ memset(file, 0, sizeof(URL_FILE));
+
+ if((file->handle.file=fopen(url,operation)))
+ {
+ file->type = CFTYPE_FILE; /* marked as URL */
+ }
+ else
+ {
+ file->type = CFTYPE_CURL; /* marked as URL */
+ file->handle.curl = curl_easy_init();
- file = (URL_FILE *)malloc(sizeof(URL_FILE));
- if(!file)
- return NULL;
+ curl_easy_setopt(file->handle.curl, CURLOPT_URL, url);
+ curl_easy_setopt(file->handle.curl, CURLOPT_FILE, file);
+ curl_easy_setopt(file->handle.curl, CURLOPT_VERBOSE, FALSE);
+ curl_easy_setopt(file->handle.curl, CURLOPT_WRITEFUNCTION, write_callback);
- memset(file, 0, sizeof(URL_FILE));
+ if(!multi_handle)
+ multi_handle = curl_multi_init();
- file->type = 1; /* marked as URL, use 0 for plain file */
- file->handle.curl = curl_easy_init();
+ curl_multi_add_handle(multi_handle, file->handle.curl);
- curl_easy_setopt(file->handle.curl, CURLOPT_URL, url);
- curl_easy_setopt(file->handle.curl, CURLOPT_FILE, file);
- curl_easy_setopt(file->handle.curl, CURLOPT_VERBOSE, FALSE);
- curl_easy_setopt(file->handle.curl, CURLOPT_WRITEFUNCTION, write_callback);
+ /* lets start the fetch */
+ while(curl_multi_perform(multi_handle, &file->still_running) ==
+ CURLM_CALL_MULTI_PERFORM );
- if(!multi_handle)
- multi_handle = curl_multi_init();
+ if((file->buffer_pos == 0) && (!file->still_running))
+ {
+ /* if still_running is 0 now, we should return NULL */
- curl_multi_add_handle(multi_handle, file->handle.curl);
+ /* make sure the easy handle is not in the multi handle anymore */
+ curl_multi_remove_handle(multi_handle, file->handle.curl);
- while(CURLM_CALL_MULTI_PERFORM ==
- curl_multi_perform(multi_handle, &still_running));
+ /* cleanup */
+ curl_easy_cleanup(file->handle.curl);
- /* if still_running would be 0 now, we should return NULL */
+ free(file);
- return file;
+ file = NULL;
+ }
+ }
+ return file;
}
-void url_fclose(URL_FILE *file)
+int
+url_fclose(URL_FILE *file)
{
- /* make sure the easy handle is not in the multi handle anymore */
- curl_multi_remove_handle(multi_handle, file->handle.curl);
+ int ret=0;/* default is good return */
- /* cleanup */
- curl_easy_cleanup(file->handle.curl);
-}
+ switch(file->type)
+ {
+ case CFTYPE_FILE:
+ ret=fclose(file->handle.file); /* passthrough */
+ break;
+
+ case CFTYPE_CURL:
+ /* make sure the easy handle is not in the multi handle anymore */
+ curl_multi_remove_handle(multi_handle, file->handle.curl);
+ /* cleanup */
+ curl_easy_cleanup(file->handle.curl);
+ break;
+
+ default: /* unknown or supported type - oh dear */
+ ret=EOF;
+ errno=EBADF;
+ break;
+
+ }
+ if(file->buffer)
+ free(file->buffer);/* free any allocated buffer space */
-size_t url_fread(void *ptr, size_t size, size_t nmemb, URL_FILE *file)
+ free(file);
+
+ return ret;
+}
+
+int
+url_feof(URL_FILE *file)
{
- fd_set fdread;
- fd_set fdwrite;
- fd_set fdexcep;
- int maxfd;
- struct timeval timeout;
- int rc;
- int still_running = 0;
-
- if(!file->bytes) { /* no data available at this point */
-
- file->readptr = file->buffer; /* reset read pointer */
-
- if(CURLM_CALL_MULTI_PERFORM == file->m) {
- while(CURLM_CALL_MULTI_PERFORM ==
- curl_multi_perform(multi_handle, &still_running)) {
- if(file->bytes) {
- printf("(fread) WOAH! THis happened!\n");
- break;
- }
- }
- if(!still_running) {
- printf("DONE RUNNING AROUND!\n");
- return 0;
- }
+ int ret=0;
+
+ switch(file->type)
+ {
+ case CFTYPE_FILE:
+ ret=feof(file->handle.file);
+ break;
+
+ case CFTYPE_CURL:
+ if((file->buffer_pos == 0) && (!file->still_running))
+ ret = 1;
+ break;
+ default: /* unknown or supported type - oh dear */
+ ret=-1;
+ errno=EBADF;
+ break;
}
+ return ret;
+}
- do {
+size_t
+url_fread(void *ptr, size_t size, size_t nmemb, URL_FILE *file)
+{
+ size_t want;
+
+ switch(file->type)
+ {
+ case CFTYPE_FILE:
+ want=fread(ptr,size,nmemb,file->handle.file);
+ break;
+
+ case CFTYPE_CURL:
+ want = nmemb * size;
+
+ curl_fill_buffer(file,want,1);
+
+ /* check if theres data in the buffer - if not curl_fill_buffer()
+ * either errored or EOF */
+ if(!file->buffer_pos)
+ return 0;
+
+ /* ensure only available data is considered */
+ if(file->buffer_pos < want)
+ want = file->buffer_pos;
- FD_ZERO(&fdread);
- FD_ZERO(&fdwrite);
- FD_ZERO(&fdexcep);
-
- /* set a suitable timeout to fail on */
- timeout.tv_sec = 500; /* 5 minutes */
- timeout.tv_usec = 0;
-
- /* get file descriptors from the transfers */
- curl_multi_fdset(multi_handle, &fdread, &fdwrite, &fdexcep, &maxfd);
-
- rc = select(maxfd+1, &fdread, &fdwrite, &fdexcep, &timeout);
-
- switch(rc) {
- case -1:
- /* select error */
- break;
- case 0:
- break;
- default:
- /* timeout or readable/writable sockets */
- printf("select() returned %d!\n", rc);
- do {
- file->m = curl_multi_perform(multi_handle, &still_running);
-
- if(file->bytes)
- /* we have received data, return that now */
- break;
-
- } while(CURLM_CALL_MULTI_PERFORM == file->m);
-
-
- if(!still_running)
- printf("DONE RUNNING AROUND!\n");
-
- break;
- }
- } while(still_running && (file->bytes <= 0));
- }
- else
- printf("(fread) Skip network read\n");
-
- if(file->bytes) {
- /* data already available, return that */
- int want = size * nmemb;
-
- if(file->bytes < want)
- want = file->bytes;
-
- memcpy(ptr, file->readptr, want);
- file->readptr += want;
- file->bytes -= want;
-
- printf("(fread) return %d bytes\n", want);
+ /* xfer data to caller */
+ memcpy(ptr, file->buffer, want);
+ curl_use_buffer(file,want);
+
+ want = want / size; /* number of items - nb correct op - checked
+ * with glibc code*/
+
+ /*printf("(fread) return %d bytes %d left\n", want,file->buffer_pos);*/
+ break;
+
+ default: /* unknown or supported type - oh dear */
+ want=0;
+ errno=EBADF;
+ break;
+
+ }
return want;
- }
- return 0; /* no data available to return */
}
+char *
+url_fgets(char *ptr, int size, URL_FILE *file)
+{
+ int want = size - 1;/* always need to leave room for zero termination */
+ int loop;
+
+ switch(file->type)
+ {
+ case CFTYPE_FILE:
+ ptr = fgets(ptr,size,file->handle.file);
+ break;
+
+ case CFTYPE_CURL:
+ curl_fill_buffer(file,want,1);
+
+ /* check if theres data in the buffer - if not fill either errored or
+ * EOF */
+ if(!file->buffer_pos)
+ return NULL;
+
+ /* ensure only available data is considered */
+ if(file->buffer_pos < want)
+ want = file->buffer_pos;
+
+ /*buffer contains data */
+ /* look for newline or eof */
+ for(loop=0;loop < want;loop++)
+ {
+ if(file->buffer[loop] == '\n')
+ {
+ want=loop+1;/* include newline */
+ break;
+ }
+ }
+
+ /* xfer data to caller */
+ memcpy(ptr, file->buffer, want);
+ ptr[want]=0;/* allways null terminate */
+
+ curl_use_buffer(file,want);
+
+ /*printf("(fgets) return %d bytes %d left\n", want,file->buffer_pos);*/
+ break;
+
+ default: /* unknown or supported type - oh dear */
+ ptr=NULL;
+ errno=EBADF;
+ break;
+ }
+
+ return ptr;/*success */
+}
+
+void
+url_rewind(URL_FILE *file)
+{
+ switch(file->type)
+ {
+ case CFTYPE_FILE:
+ rewind(file->handle.file); /* passthrough */
+ break;
+
+ case CFTYPE_CURL:
+ /* halt transaction */
+ curl_multi_remove_handle(multi_handle, file->handle.curl);
+
+ /* restart */
+ curl_multi_add_handle(multi_handle, file->handle.curl);
+
+ /* ditch buffer - write will recreate - resets stream pos*/
+ if(file->buffer)
+ free(file->buffer);
+
+ file->buffer=NULL;
+ file->buffer_pos=0;
+ file->buffer_len=0;
+
+ break;
-int main(int argc, char *argv[])
+ default: /* unknown or supported type - oh dear */
+ break;
+
+ }
+
+}
+
+
+/* Small main program to retrive from a url using fgets and fread saving the
+ * output to two test files (note the fgets method will corrupt binary files if
+ * they contain 0 chars */
+int
+main(int argc, char *argv[])
{
- URL_FILE *handle;
- int nread;
- char buffer[256];
+ URL_FILE *handle;
+ FILE *outf;
+
+ int nread;
+ char buffer[256];
+ char *url;
+
+ if(argc < 2)
+ {
+ url="http://192.168.7.3/testfile";/* default to testurl */
+ }
+ else
+ {
+ url=argv[1];/* use passed url */
+ }
+
+ /* copy from url line by line with fgets */
+ outf=fopen("fgets.test","w+");
+ if(!outf)
+ {
+ perror("couldnt open fgets output file\n");
+ return 1;
+ }
+
+ handle = url_fopen(url, "r");
+ if(!handle)
+ {
+ printf("couldn't url_fopen()\n");
+ fclose(outf);
+ return 2;
+ }
+
+ while(!url_feof(handle))
+ {
+ url_fgets(buffer,sizeof(buffer),handle);
+ fwrite(buffer,1,strlen(buffer),outf);
+ }
+
+ url_fclose(handle);
+
+ fclose(outf);
+
+
+ /* Copy from url with fread */
+ outf=fopen("fread.test","w+");
+ if(!outf)
+ {
+ perror("couldnt open fread output file\n");
+ return 1;
+ }
+
+ handle = url_fopen("testfile", "r");
+ if(!handle) {
+ printf("couldn't url_fopen()\n");
+ fclose(outf);
+ return 2;
+ }
+
+ do {
+ nread = url_fread(buffer, 1,sizeof(buffer), handle);
+ fwrite(buffer,1,nread,outf);
+ } while(nread);
+
+ url_fclose(handle);
+
+ fclose(outf);
+
+
+ /* Test rewind */
+ outf=fopen("rewind.test","w+");
+ if(!outf)
+ {
+ perror("couldnt open fread output file\n");
+ return 1;
+ }
+
+ handle = url_fopen("testfile", "r");
+ if(!handle) {
+ printf("couldn't url_fopen()\n");
+ fclose(outf);
+ return 2;
+ }
+
+ nread = url_fread(buffer, 1,sizeof(buffer), handle);
+ fwrite(buffer,1,nread,outf);
+ url_rewind(handle);
- (void)argc;
- (void)argv;
+ buffer[0]='\n';
+ fwrite(buffer,1,1,outf);
- handle = url_fopen("http://curl.haxx.se/", "r");
+ nread = url_fread(buffer, 1,sizeof(buffer), handle);
+ fwrite(buffer,1,nread,outf);
- if(!handle) {
- printf("couldn't url_fopen()\n");
- }
- do {
- nread = url_fread(buffer, sizeof(buffer), 1, handle);
+ url_fclose(handle);
- printf("We got: %d bytes\n", nread);
- } while(nread);
+ fclose(outf);
- url_fclose(handle);
- return 0;
+ return 0;/* all done */
}