Hi,

On Thu, 17 Jul 2014, Vincent Bernat wrote:
> I have tried to setup a local instance to contribute some patches but
> the initial import of the database needs more than 14GB of RAM. This

What command did you use for the initial import?
Was that with the fixture distro_tracker/core/fixtures/debian-repositories.xml
or with a custome set of repositories?
How did you evaluate that memory requirement?

I know it takes several Gb but none of the machines where I did the
initial import had so much memory so your claim seems strange
to me.

I often only run ./manage.py tracker_update_repositories for the initial
import that said (and not run_all_tasks).

> Trying to limit to unstable doesn't help either.

Huh. Reducing the number of repositories and packages should really
help... I wonder what you're hitting here.

> Maybe a fixture with only a thousand packages or a limit of 1000
> packages per source would help.

We should certainly try to optimize the memory consumption of the
repository update process.

I have zero experience in analysis of Python's memory usage but I believe
that there are good tools for this.

Among the packaged tools I found python-meliae and python-memprof.
python-objgraph might also be useful. 

And python 3 has tracemalloc...
https://docs.python.org/3/library/tracemalloc.html

Cheers,
-- 
Raphaël Hertzog ◈ Debian Developer

Discover the Debian Administrator's Handbook:
→ http://debian-handbook.info/get/


-- 
To UNSUBSCRIBE, email to debian-qa-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org
Archive: 
https://lists.debian.org/20140717102125.gc16...@x230-buxy.home.ouaza.com

Reply via email to