Thanks Thomas!

As a workaround I will divide my data in smaller chunks.

Cheers,

Osvaldo.



On Fri, Apr 4, 2014 at 4:54 PM, Thomas Holder <thomas.hol...@schrodinger.com
> wrote:

> Hi Osvaldo,
>
> you are right, I can reproduce this. We'll look into it and let you know
> when it has been fixed.
>
> Cheers,
>   Thomas
>
> On 02 Apr 2014, at 20:30, Osvaldo Martin <aloctavo...@gmail.com> wrote:
> > Hi,
> >
> > I need to run an script based on PyMOL  for a thousands of proteins. The
> problem is that even when I delete all object the RAM usage keeps
> increasing. I am using PyMOL 1.7 on ubuntu 14.04 (beta 2).
> >
> >
> > my script looks like this.
> >
> > import __main__
> > __main__.pymol_argv = ['pymol','-qc']
> > import pymol
> > from pymol import cmd, stored
> > import glob
> >
> > pdb_files = glob.glob('*.pdb')
> > for protein in pdb_files:
> >         cmd.load(protein)
> >         "some code here"
> >         cmd.delete('all')
> >
> >
> > I tried the following.
> >
> > 1) removing the  "some code here"
> > 2) adding cmd.reinitialize()
> > 3) adding cmd.set('suspend_undo', 1) (
> https://www.mail-archive.com/pymol-users@lists.sourceforge.net/msg11126.html
> )
> >
> > Cheers and thanks in advance.
>
> --
> Thomas Holder
> PyMOL Developer
> Schrödinger, Inc.
>
>
------------------------------------------------------------------------------
_______________________________________________
PyMOL-users mailing list (PyMOL-users@lists.sourceforge.net)
Info Page: https://lists.sourceforge.net/lists/listinfo/pymol-users
Archives: http://www.mail-archive.com/pymol-users@lists.sourceforge.net

Reply via email to