Thanks Thomas! As a workaround I will divide my data in smaller chunks.
Cheers, Osvaldo. On Fri, Apr 4, 2014 at 4:54 PM, Thomas Holder <thomas.hol...@schrodinger.com > wrote: > Hi Osvaldo, > > you are right, I can reproduce this. We'll look into it and let you know > when it has been fixed. > > Cheers, > Thomas > > On 02 Apr 2014, at 20:30, Osvaldo Martin <aloctavo...@gmail.com> wrote: > > Hi, > > > > I need to run an script based on PyMOL for a thousands of proteins. The > problem is that even when I delete all object the RAM usage keeps > increasing. I am using PyMOL 1.7 on ubuntu 14.04 (beta 2). > > > > > > my script looks like this. > > > > import __main__ > > __main__.pymol_argv = ['pymol','-qc'] > > import pymol > > from pymol import cmd, stored > > import glob > > > > pdb_files = glob.glob('*.pdb') > > for protein in pdb_files: > > cmd.load(protein) > > "some code here" > > cmd.delete('all') > > > > > > I tried the following. > > > > 1) removing the "some code here" > > 2) adding cmd.reinitialize() > > 3) adding cmd.set('suspend_undo', 1) ( > https://www.mail-archive.com/pymol-users@lists.sourceforge.net/msg11126.html > ) > > > > Cheers and thanks in advance. > > -- > Thomas Holder > PyMOL Developer > Schrödinger, Inc. > >
------------------------------------------------------------------------------
_______________________________________________ PyMOL-users mailing list (PyMOL-users@lists.sourceforge.net) Info Page: https://lists.sourceforge.net/lists/listinfo/pymol-users Archives: http://www.mail-archive.com/pymol-users@lists.sourceforge.net