mirror of
https://github.com/curl/curl.git
synced 2026-04-20 04:11:16 +03:00
Before this patch curl used the C preprocessor to override standard memory allocation symbols: malloc, calloc, strdup, realloc, free. The goal of these is to replace them with curl's debug wrappers in `CURLDEBUG` builds, another was to replace them with the wrappers calling user-defined allocators in libcurl. This solution needed a bunch of workarounds to avoid breaking external headers: it relied on include order to do the overriding last. For "unity" builds it needed to reset overrides before external includes. Also in test apps, which are always built as single source files. It also needed the `(symbol)` trick to avoid overrides in some places. This would still not fix cases where the standard symbols were macros. It was also fragile and difficult to figure out which was the actual function behind an alloc or free call in a specific piece of code. This in turn caused bugs where the wrong allocator was accidentally called. To avoid these problems, this patch replaces this solution with `curlx_`-prefixed allocator macros, and mapping them _once_ to either the libcurl wrappers, the debug wrappers or the standard ones, matching the rest of the code in libtests. This concludes the long journey to avoid redefining standard functions in the curl codebase. Note: I did not update `packages/OS400/*.c` sources. They did not `#include` `curl_setup.h`, `curl_memory.h` or `memdebug.h`, meaning the overrides were never applied to them. This may or may not have been correct. For now I suppressed the direct use of standard allocators via a local `.checksrc`. Probably they (except for `curlcl.c`) should be updated to include `curl_setup.h` and use the `curlx_` macros. This patch changes mappings in two places: - `lib/curl_threads.c` in libtests: Before this patch it mapped to libcurl allocators. After, it maps to standard allocators, like the rest of libtests code. - `units`: before this patch it mapped to standard allocators. After, it maps to libcurl allocators. Also: - drop all position-dependent `curl_memory.h` and `memdebug.h` includes, and delete the now unnecessary headers. - rename `Curl_tcsdup` macro to `curlx_tcsdup` and define like the other allocators. - map `curlx_strdup()` to `_strdup()` on Windows (was: `strdup()`). To fix warnings silenced via `_CRT_NONSTDC_NO_DEPRECATE`. - multibyte: map `curlx_convert_*()` to `_strdup()` on Windows (was: `strdup()`). - src: do not reuse the `strdup` name for the local replacement. - lib509: call `_strdup()` on Windows (was: `strdup()`). - test1132: delete test obsoleted by this patch. - CHECKSRC.md: update text for `SNPRINTF`. - checksrc: ban standard allocator symbols. Follow-up tob12da22db1#18866 Follow-up todb98daab05#18844 Follow-up to4deea9396b#18814 Follow-up to9678ff5b1b#18776 Follow-up to10bac43b87#18774 Follow-up to20142f5d06#18634 Follow-up tobf7375ecc5#18503 Follow-up to9863599d69#18502 Follow-up to3bb5e58c10#17827 Closes #19626
620 lines
15 KiB
C
620 lines
15 KiB
C
/***************************************************************************
|
|
* _ _ ____ _
|
|
* Project ___| | | | _ \| |
|
|
* / __| | | | |_) | |
|
|
* | (__| |_| | _ <| |___
|
|
* \___|\___/|_| \_\_____|
|
|
*
|
|
* Copyright (C) Daniel Stenberg, <daniel@haxx.se>, et al.
|
|
*
|
|
* This software is licensed as described in the file COPYING, which
|
|
* you should have received as part of this distribution. The terms
|
|
* are also available at https://curl.se/docs/copyright.html.
|
|
*
|
|
* You may opt to use, copy, modify, merge, publish, distribute and/or sell
|
|
* copies of the Software, and permit persons to whom the Software is
|
|
* furnished to do so, under the terms of the COPYING file.
|
|
*
|
|
* This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
|
|
* KIND, either express or implied.
|
|
*
|
|
* SPDX-License-Identifier: curl
|
|
*
|
|
***************************************************************************/
|
|
|
|
#include "curl_setup.h"
|
|
#include "bufq.h"
|
|
|
|
static bool chunk_is_empty(const struct buf_chunk *chunk)
|
|
{
|
|
return chunk->r_offset >= chunk->w_offset;
|
|
}
|
|
|
|
static bool chunk_is_full(const struct buf_chunk *chunk)
|
|
{
|
|
return chunk->w_offset >= chunk->dlen;
|
|
}
|
|
|
|
static size_t chunk_len(const struct buf_chunk *chunk)
|
|
{
|
|
return chunk->w_offset - chunk->r_offset;
|
|
}
|
|
|
|
static void chunk_reset(struct buf_chunk *chunk)
|
|
{
|
|
chunk->next = NULL;
|
|
chunk->r_offset = chunk->w_offset = 0;
|
|
}
|
|
|
|
static size_t chunk_append(struct buf_chunk *chunk,
|
|
const uint8_t *buf, size_t len)
|
|
{
|
|
uint8_t *p = &chunk->x.data[chunk->w_offset];
|
|
size_t n = chunk->dlen - chunk->w_offset;
|
|
DEBUGASSERT(chunk->dlen >= chunk->w_offset);
|
|
if(n) {
|
|
n = CURLMIN(n, len);
|
|
memcpy(p, buf, n);
|
|
chunk->w_offset += n;
|
|
}
|
|
return n;
|
|
}
|
|
|
|
static size_t chunk_read(struct buf_chunk *chunk,
|
|
uint8_t *buf, size_t len)
|
|
{
|
|
uint8_t *p = &chunk->x.data[chunk->r_offset];
|
|
size_t n = chunk->w_offset - chunk->r_offset;
|
|
DEBUGASSERT(chunk->w_offset >= chunk->r_offset);
|
|
if(!n) {
|
|
return 0;
|
|
}
|
|
else if(n <= len) {
|
|
memcpy(buf, p, n);
|
|
chunk->r_offset = chunk->w_offset = 0;
|
|
return n;
|
|
}
|
|
else {
|
|
memcpy(buf, p, len);
|
|
chunk->r_offset += len;
|
|
return len;
|
|
}
|
|
}
|
|
|
|
static CURLcode chunk_slurpn(struct buf_chunk *chunk, size_t max_len,
|
|
Curl_bufq_reader *reader,
|
|
void *reader_ctx, size_t *pnread)
|
|
{
|
|
uint8_t *p = &chunk->x.data[chunk->w_offset];
|
|
size_t n = chunk->dlen - chunk->w_offset; /* free amount */
|
|
CURLcode result;
|
|
|
|
*pnread = 0;
|
|
DEBUGASSERT(chunk->dlen >= chunk->w_offset);
|
|
if(!n)
|
|
return CURLE_AGAIN;
|
|
if(max_len && n > max_len)
|
|
n = max_len;
|
|
result = reader(reader_ctx, p, n, pnread);
|
|
if(!result) {
|
|
DEBUGASSERT(*pnread <= n);
|
|
chunk->w_offset += *pnread;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
static void chunk_peek(const struct buf_chunk *chunk,
|
|
const uint8_t **pbuf, size_t *plen)
|
|
{
|
|
DEBUGASSERT(chunk->w_offset >= chunk->r_offset);
|
|
*pbuf = &chunk->x.data[chunk->r_offset];
|
|
*plen = chunk->w_offset - chunk->r_offset;
|
|
}
|
|
|
|
static void chunk_peek_at(const struct buf_chunk *chunk, size_t offset,
|
|
const uint8_t **pbuf, size_t *plen)
|
|
{
|
|
offset += chunk->r_offset;
|
|
DEBUGASSERT(chunk->w_offset >= offset);
|
|
*pbuf = &chunk->x.data[offset];
|
|
*plen = chunk->w_offset - offset;
|
|
}
|
|
|
|
static size_t chunk_skip(struct buf_chunk *chunk, size_t amount)
|
|
{
|
|
size_t n = chunk->w_offset - chunk->r_offset;
|
|
DEBUGASSERT(chunk->w_offset >= chunk->r_offset);
|
|
if(n) {
|
|
n = CURLMIN(n, amount);
|
|
chunk->r_offset += n;
|
|
if(chunk->r_offset == chunk->w_offset)
|
|
chunk->r_offset = chunk->w_offset = 0;
|
|
}
|
|
return n;
|
|
}
|
|
|
|
static void chunk_list_free(struct buf_chunk **anchor)
|
|
{
|
|
struct buf_chunk *chunk;
|
|
while(*anchor) {
|
|
chunk = *anchor;
|
|
*anchor = chunk->next;
|
|
curlx_free(chunk);
|
|
}
|
|
}
|
|
|
|
|
|
void Curl_bufcp_init(struct bufc_pool *pool,
|
|
size_t chunk_size, size_t spare_max)
|
|
{
|
|
DEBUGASSERT(chunk_size > 0);
|
|
DEBUGASSERT(spare_max > 0);
|
|
memset(pool, 0, sizeof(*pool));
|
|
pool->chunk_size = chunk_size;
|
|
pool->spare_max = spare_max;
|
|
}
|
|
|
|
static CURLcode bufcp_take(struct bufc_pool *pool,
|
|
struct buf_chunk **pchunk)
|
|
{
|
|
struct buf_chunk *chunk = NULL;
|
|
|
|
if(pool->spare) {
|
|
chunk = pool->spare;
|
|
pool->spare = chunk->next;
|
|
--pool->spare_count;
|
|
chunk_reset(chunk);
|
|
*pchunk = chunk;
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/* Check for integer overflow before allocation */
|
|
if(pool->chunk_size > SIZE_MAX - sizeof(*chunk)) {
|
|
*pchunk = NULL;
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
chunk = curlx_calloc(1, sizeof(*chunk) + pool->chunk_size);
|
|
if(!chunk) {
|
|
*pchunk = NULL;
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
chunk->dlen = pool->chunk_size;
|
|
*pchunk = chunk;
|
|
return CURLE_OK;
|
|
}
|
|
|
|
static void bufcp_put(struct bufc_pool *pool,
|
|
struct buf_chunk *chunk)
|
|
{
|
|
if(pool->spare_count >= pool->spare_max) {
|
|
curlx_free(chunk);
|
|
}
|
|
else {
|
|
chunk_reset(chunk);
|
|
chunk->next = pool->spare;
|
|
pool->spare = chunk;
|
|
++pool->spare_count;
|
|
}
|
|
}
|
|
|
|
void Curl_bufcp_free(struct bufc_pool *pool)
|
|
{
|
|
chunk_list_free(&pool->spare);
|
|
pool->spare_count = 0;
|
|
}
|
|
|
|
static void bufq_init(struct bufq *q, struct bufc_pool *pool,
|
|
size_t chunk_size, size_t max_chunks, int opts)
|
|
{
|
|
DEBUGASSERT(chunk_size > 0);
|
|
DEBUGASSERT(max_chunks > 0);
|
|
memset(q, 0, sizeof(*q));
|
|
q->chunk_size = chunk_size;
|
|
q->max_chunks = max_chunks;
|
|
q->pool = pool;
|
|
q->opts = opts;
|
|
}
|
|
|
|
void Curl_bufq_init2(struct bufq *q, size_t chunk_size, size_t max_chunks,
|
|
int opts)
|
|
{
|
|
bufq_init(q, NULL, chunk_size, max_chunks, opts);
|
|
}
|
|
|
|
void Curl_bufq_init(struct bufq *q, size_t chunk_size, size_t max_chunks)
|
|
{
|
|
bufq_init(q, NULL, chunk_size, max_chunks, BUFQ_OPT_NONE);
|
|
}
|
|
|
|
void Curl_bufq_initp(struct bufq *q, struct bufc_pool *pool,
|
|
size_t max_chunks, int opts)
|
|
{
|
|
bufq_init(q, pool, pool->chunk_size, max_chunks, opts);
|
|
}
|
|
|
|
void Curl_bufq_free(struct bufq *q)
|
|
{
|
|
chunk_list_free(&q->head);
|
|
chunk_list_free(&q->spare);
|
|
q->tail = NULL;
|
|
q->chunk_count = 0;
|
|
}
|
|
|
|
void Curl_bufq_reset(struct bufq *q)
|
|
{
|
|
struct buf_chunk *chunk;
|
|
while(q->head) {
|
|
chunk = q->head;
|
|
q->head = chunk->next;
|
|
chunk->next = q->spare;
|
|
q->spare = chunk;
|
|
}
|
|
q->tail = NULL;
|
|
}
|
|
|
|
size_t Curl_bufq_len(const struct bufq *q)
|
|
{
|
|
const struct buf_chunk *chunk = q->head;
|
|
size_t len = 0;
|
|
while(chunk) {
|
|
len += chunk_len(chunk);
|
|
chunk = chunk->next;
|
|
}
|
|
return len;
|
|
}
|
|
|
|
bool Curl_bufq_is_empty(const struct bufq *q)
|
|
{
|
|
return !q->head || chunk_is_empty(q->head);
|
|
}
|
|
|
|
bool Curl_bufq_is_full(const struct bufq *q)
|
|
{
|
|
if(!q->tail || q->spare)
|
|
return FALSE;
|
|
if(q->chunk_count < q->max_chunks)
|
|
return FALSE;
|
|
if(q->chunk_count > q->max_chunks)
|
|
return TRUE;
|
|
/* we have no spares and cannot make more, is the tail full? */
|
|
return chunk_is_full(q->tail);
|
|
}
|
|
|
|
static struct buf_chunk *get_spare(struct bufq *q)
|
|
{
|
|
struct buf_chunk *chunk = NULL;
|
|
|
|
if(q->spare) {
|
|
chunk = q->spare;
|
|
q->spare = chunk->next;
|
|
chunk_reset(chunk);
|
|
return chunk;
|
|
}
|
|
|
|
if(q->chunk_count >= q->max_chunks && (!(q->opts & BUFQ_OPT_SOFT_LIMIT)))
|
|
return NULL;
|
|
|
|
if(q->pool) {
|
|
if(bufcp_take(q->pool, &chunk))
|
|
return NULL;
|
|
++q->chunk_count;
|
|
return chunk;
|
|
}
|
|
else {
|
|
/* Check for integer overflow before allocation */
|
|
if(q->chunk_size > SIZE_MAX - sizeof(*chunk)) {
|
|
return NULL;
|
|
}
|
|
|
|
chunk = curlx_calloc(1, sizeof(*chunk) + q->chunk_size);
|
|
if(!chunk)
|
|
return NULL;
|
|
chunk->dlen = q->chunk_size;
|
|
++q->chunk_count;
|
|
return chunk;
|
|
}
|
|
}
|
|
|
|
static void prune_head(struct bufq *q)
|
|
{
|
|
struct buf_chunk *chunk;
|
|
|
|
while(q->head && chunk_is_empty(q->head)) {
|
|
chunk = q->head;
|
|
q->head = chunk->next;
|
|
if(q->tail == chunk)
|
|
q->tail = q->head;
|
|
if(q->pool) {
|
|
bufcp_put(q->pool, chunk);
|
|
--q->chunk_count;
|
|
}
|
|
else if((q->chunk_count > q->max_chunks) ||
|
|
(q->opts & BUFQ_OPT_NO_SPARES)) {
|
|
/* SOFT_LIMIT allowed us more than max. free spares until
|
|
* we are at max again. Or free them if we are configured
|
|
* to not use spares. */
|
|
curlx_free(chunk);
|
|
--q->chunk_count;
|
|
}
|
|
else {
|
|
chunk->next = q->spare;
|
|
q->spare = chunk;
|
|
}
|
|
}
|
|
}
|
|
|
|
static struct buf_chunk *get_non_full_tail(struct bufq *q)
|
|
{
|
|
struct buf_chunk *chunk;
|
|
|
|
if(q->tail && !chunk_is_full(q->tail))
|
|
return q->tail;
|
|
chunk = get_spare(q);
|
|
if(chunk) {
|
|
/* new tail, and possibly new head */
|
|
if(q->tail) {
|
|
q->tail->next = chunk;
|
|
q->tail = chunk;
|
|
}
|
|
else {
|
|
DEBUGASSERT(!q->head);
|
|
q->head = q->tail = chunk;
|
|
}
|
|
}
|
|
return chunk;
|
|
}
|
|
|
|
CURLcode Curl_bufq_write(struct bufq *q,
|
|
const uint8_t *buf, size_t len,
|
|
size_t *pnwritten)
|
|
{
|
|
struct buf_chunk *tail;
|
|
size_t n;
|
|
|
|
DEBUGASSERT(q->max_chunks > 0);
|
|
*pnwritten = 0;
|
|
while(len) {
|
|
tail = get_non_full_tail(q);
|
|
if(!tail) {
|
|
if((q->chunk_count < q->max_chunks) || (q->opts & BUFQ_OPT_SOFT_LIMIT))
|
|
/* should have gotten a tail, but did not */
|
|
return CURLE_OUT_OF_MEMORY;
|
|
break;
|
|
}
|
|
n = chunk_append(tail, buf, len);
|
|
if(!n)
|
|
break;
|
|
*pnwritten += n;
|
|
buf += n;
|
|
len -= n;
|
|
}
|
|
return (!*pnwritten && len) ? CURLE_AGAIN : CURLE_OK;
|
|
}
|
|
|
|
CURLcode Curl_bufq_cwrite(struct bufq *q,
|
|
const char *buf, size_t len,
|
|
size_t *pnwritten)
|
|
{
|
|
return Curl_bufq_write(q, (const uint8_t *)buf, len, pnwritten);
|
|
}
|
|
|
|
CURLcode Curl_bufq_read(struct bufq *q, uint8_t *buf, size_t len,
|
|
size_t *pnread)
|
|
{
|
|
*pnread = 0;
|
|
while(len && q->head) {
|
|
size_t n = chunk_read(q->head, buf, len);
|
|
if(n) {
|
|
*pnread += n;
|
|
buf += n;
|
|
len -= n;
|
|
}
|
|
prune_head(q);
|
|
}
|
|
return (!*pnread) ? CURLE_AGAIN : CURLE_OK;
|
|
}
|
|
|
|
CURLcode Curl_bufq_cread(struct bufq *q, char *buf, size_t len,
|
|
size_t *pnread)
|
|
{
|
|
return Curl_bufq_read(q, (uint8_t *)buf, len, pnread);
|
|
}
|
|
|
|
bool Curl_bufq_peek(struct bufq *q,
|
|
const uint8_t **pbuf, size_t *plen)
|
|
{
|
|
if(q->head && chunk_is_empty(q->head)) {
|
|
prune_head(q);
|
|
}
|
|
if(q->head && !chunk_is_empty(q->head)) {
|
|
chunk_peek(q->head, pbuf, plen);
|
|
return TRUE;
|
|
}
|
|
*pbuf = NULL;
|
|
*plen = 0;
|
|
return FALSE;
|
|
}
|
|
|
|
bool Curl_bufq_peek_at(struct bufq *q, size_t offset,
|
|
const uint8_t **pbuf, size_t *plen)
|
|
{
|
|
struct buf_chunk *c = q->head;
|
|
size_t clen;
|
|
|
|
while(c) {
|
|
clen = chunk_len(c);
|
|
if(!clen)
|
|
break;
|
|
if(offset >= clen) {
|
|
offset -= clen;
|
|
c = c->next;
|
|
continue;
|
|
}
|
|
chunk_peek_at(c, offset, pbuf, plen);
|
|
return TRUE;
|
|
}
|
|
*pbuf = NULL;
|
|
*plen = 0;
|
|
return FALSE;
|
|
}
|
|
|
|
void Curl_bufq_skip(struct bufq *q, size_t amount)
|
|
{
|
|
size_t n;
|
|
|
|
while(amount && q->head) {
|
|
n = chunk_skip(q->head, amount);
|
|
amount -= n;
|
|
prune_head(q);
|
|
}
|
|
}
|
|
|
|
CURLcode Curl_bufq_pass(struct bufq *q, Curl_bufq_writer *writer,
|
|
void *writer_ctx, size_t *pwritten)
|
|
{
|
|
const uint8_t *buf;
|
|
size_t blen;
|
|
CURLcode result = CURLE_OK;
|
|
|
|
*pwritten = 0;
|
|
while(Curl_bufq_peek(q, &buf, &blen)) {
|
|
size_t chunk_written;
|
|
|
|
result = writer(writer_ctx, buf, blen, &chunk_written);
|
|
if(result) {
|
|
if((result == CURLE_AGAIN) && *pwritten) {
|
|
/* blocked on subsequent write, report success */
|
|
result = CURLE_OK;
|
|
}
|
|
break;
|
|
}
|
|
if(!chunk_written) {
|
|
if(!*pwritten) {
|
|
/* treat as blocked */
|
|
result = CURLE_AGAIN;
|
|
}
|
|
break;
|
|
}
|
|
*pwritten += chunk_written;
|
|
Curl_bufq_skip(q, chunk_written);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
CURLcode Curl_bufq_write_pass(struct bufq *q,
|
|
const uint8_t *buf, size_t len,
|
|
Curl_bufq_writer *writer, void *writer_ctx,
|
|
size_t *pwritten)
|
|
{
|
|
CURLcode result = CURLE_OK;
|
|
size_t n;
|
|
|
|
*pwritten = 0;
|
|
while(len) {
|
|
if(Curl_bufq_is_full(q)) {
|
|
/* try to make room in case we are full */
|
|
result = Curl_bufq_pass(q, writer, writer_ctx, &n);
|
|
if(result) {
|
|
if(result != CURLE_AGAIN) {
|
|
/* real error, fail */
|
|
return result;
|
|
}
|
|
/* would block, bufq is full, give up */
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Add to bufq as much as there is room for */
|
|
result = Curl_bufq_write(q, buf, len, &n);
|
|
if(result) {
|
|
if(result != CURLE_AGAIN)
|
|
/* real error, fail */
|
|
return result;
|
|
/* result == CURLE_AGAIN */
|
|
if(*pwritten)
|
|
/* we did write successfully before */
|
|
result = CURLE_OK;
|
|
return result;
|
|
}
|
|
else if(n == 0)
|
|
/* edge case of writer returning 0 (and len is >0)
|
|
* break or we might enter an infinite loop here */
|
|
break;
|
|
|
|
/* Track what we added to bufq */
|
|
buf += n;
|
|
len -= n;
|
|
*pwritten += n;
|
|
}
|
|
|
|
return (!*pwritten && len) ? CURLE_AGAIN : CURLE_OK;
|
|
}
|
|
|
|
CURLcode Curl_bufq_sipn(struct bufq *q, size_t max_len,
|
|
Curl_bufq_reader *reader, void *reader_ctx,
|
|
size_t *pnread)
|
|
{
|
|
struct buf_chunk *tail = NULL;
|
|
|
|
*pnread = 0;
|
|
tail = get_non_full_tail(q);
|
|
if(!tail) {
|
|
if(q->chunk_count < q->max_chunks)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
/* full, blocked */
|
|
return CURLE_AGAIN;
|
|
}
|
|
|
|
return chunk_slurpn(tail, max_len, reader, reader_ctx, pnread);
|
|
}
|
|
|
|
/**
|
|
* Read up to `max_len` bytes and append it to the end of the buffer queue.
|
|
* if `max_len` is 0, no limit is imposed and the call behaves exactly
|
|
* the same as `Curl_bufq_slurp()`.
|
|
* Returns the total amount of buf read (may be 0) in `pnread` or error
|
|
* Note that even in case of an error chunks may have been read and
|
|
* the buffer queue will have different length than before.
|
|
*/
|
|
static CURLcode bufq_slurpn(struct bufq *q, size_t max_len,
|
|
Curl_bufq_reader *reader, void *reader_ctx,
|
|
size_t *pnread)
|
|
{
|
|
CURLcode result;
|
|
|
|
*pnread = 0;
|
|
while(1) {
|
|
size_t n;
|
|
result = Curl_bufq_sipn(q, max_len, reader, reader_ctx, &n);
|
|
if(result) {
|
|
if(!*pnread || result != CURLE_AGAIN) {
|
|
/* blocked on first read or real error, fail */
|
|
return result;
|
|
}
|
|
result = CURLE_OK;
|
|
break;
|
|
}
|
|
else if(n == 0) {
|
|
/* eof, result remains CURLE_OK */
|
|
break;
|
|
}
|
|
*pnread += n;
|
|
if(max_len) {
|
|
DEBUGASSERT(n <= max_len);
|
|
max_len -= n;
|
|
if(!max_len)
|
|
break;
|
|
}
|
|
/* give up slurping when we get less bytes than we asked for */
|
|
if(q->tail && !chunk_is_full(q->tail))
|
|
break;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
CURLcode Curl_bufq_slurp(struct bufq *q, Curl_bufq_reader *reader,
|
|
void *reader_ctx, size_t *pnread)
|
|
{
|
|
return bufq_slurpn(q, 0, reader, reader_ctx, pnread);
|
|
}
|