"Joseph L. Casale" <jcas...@activenetwerx.com> writes: >> So presumably your data's small enough to fit into memory, right? If >> it isn't, going back to the database every time would be the best >> option. But if it is, can you simply keep three dictionaries in sync? > > Hi Chris, > Yeah the data can fit in memory and hence the desire to avoid a trip here. > >> row = (foo, bar, quux) # there could be duplicate quuxes but not foos or bars >> foo_dict = {} >> bar_dict = {} >> quux_dict = collections.defaultdict(set) >> >> foo_dict[row[0]] = row >> bar_dict[row[1]] = row >> quux_dict[row[2]].add(row) > > This is actually far simpler than I had started imagining, however the row > data > is duplicated. I am hacking away at an attempt with references to one copy of > the row. > > Its kind of hard to recreate an sql like object in Python with indexes and the > inherent programmability against a single copy of data. >
If you want an sql-like interface, you can simply create an in-memory sqlite3 database. import sqlite3 db = sqlite3.Connection(':memory:') You can create indexes as you need, and query using SQL. Later, if you find the data getting too big to fit in memory, you can switch to using an on-disk database instead without significant changes to the code. -- regards, kushal -- https://mail.python.org/mailman/listinfo/python-list