On Mon, May 25, 2020 at 7:16 PM James Read <jamesread5...@gmail.com> wrote:
>
>
>
> On Tue, May 26, 2020 at 12:02 AM Jeffrey Walton <noloa...@gmail.com> wrote:
>>
>> On Mon, May 25, 2020 at 6:27 PM James Read via curl-library
>> <curl-library@cool.haxx.se> wrote:
>> >
>> > ...
>> >
>> > Gmail seems to have taken out all the formatting. Apologies. It should 
>> > still compile though.
>>
>> I can't speak for others, but... You should probably reduce the code
>> to a minimal reproducer, and then put it on GitHub or another place
>> folks can 'git clone' and then 'make'.
>
>
> git clone https://github.com/JamesRead5737/libcurlmemoryleak.git
>
> No need to make. Just compile with gcc crawler.c -g -lssl -lcurl
> Run valgrind with valgrind -v --tool=memcheck --leak-check=full 
> --show-reachable=yes --track-origins=yes --log-file=memcheck.log ./a.out
>
> This should reproduce what I've been talking about.

The program has been running for about 10 minutes and it still has not ended:

    Parsed sites: 0, 367 parallel connections, 364 still running

I think you should reduce the program to a minimal reproducer.

This does not look correct (to me) in new_conn:

    conn->easy = curl_easy_init();

I thought you should only call curl_easy_init() once for the
application, not each thread(?).

Jeff
-------------------------------------------------------------------
Unsubscribe: https://cool.haxx.se/list/listinfo/curl-library
Etiquette:   https://curl.haxx.se/mail/etiquette.html

Reply via email to