I wonder if any of the main project committers would like to weigh in on what a desired API would look like, or perhaps we should start an unscheduled Jira ticket?
On Thu, May 27, 2010 at 5:39 PM, Jake Luciani <jak...@gmail.com> wrote: > I had this: > > > string slice_dice_reduce(1:required list<binary> key, > 2:required ColumnParent > column_parent, > 3:required SlicePredicate predicate, > 4:required ConsistencyLevel > consistency_level=ONE, > 5:required string dice_js, > 6:required string reduce_js) > throws (1:InvalidRequestException ire, > 2:UnavailableException ue, 3:TimedOutException te), > > I guess it could use a union of sorts and return either. > > > > On Thu, May 27, 2010 at 8:36 PM, Jeremy Davis < > jerdavis.cassan...@gmail.com> wrote: > >> >> I agree, I had more than filter results in mind. >> Though I had envisioned the results to continue to use the >> List<ColumnOrSuperColumn> (and not JSON). You could still create new result >> columns that do not in any way exist in Cassandra, and you could still stuff >> JSON in to any of result columns. >> >> I had envisioned: >> list<ColumnOrSuperColumn> get_slice(keyspace, key, column_parent, predicate, >> consistency_level, >> javascript_blob ) >> >> -JD >> >> >> >> >> >> On Thu, May 27, 2010 at 5:01 PM, Jake Luciani <jak...@gmail.com> wrote: >> >>> I've secretly started working on this but nothing to show yet :( I'm >>> calling it SliceDiceReduce or SliceReduce. >>> >>> The plan is to use the js thrift bindings I've added for 0.3 release of >>> thrift (out very soon?) >>> >>> This will allow the supplied js to access the results like any other >>> thrift client. >>> >>> Adding a new verb handler and SEDA stage that will execute on a local >>> node and pass this nodes slice data into the supplied js "dice" function via >>> the thrift js bindings. >>> >>> The resulting js from each node would then be passed into another >>> supplied js reduce function on the starting node. >>> >>> The result of this would then return a single JSON or string result. >>> The reason I'm keeping the results in json is you can do more than filter. >>> You can do things like word count etc. >>> >>> Anyway this is little more than an idea now. But if people like this >>> approach maybe I'll get motivated! >>> >>> Jake >>> >>> >>> >>> >>> >>> On May 27, 2010, at 7:36 PM, Steve Lihn <stevel...@gmail.com> wrote: >>> >>> Mongo has it too. It could save a lot of development time if one can >>> figure out porting Mongo's query API and stored javascript to Cassandra. >>> It would be great if scala's list comprehension can be facilitated to >>> write query-like code against Cassandra schema. >>> >>> On Thu, May 27, 2010 at 11:05 AM, Vick Khera < <vi...@khera.org> >>> vi...@khera.org> wrote: >>> >>>> On Thu, May 27, 2010 at 9:50 AM, Jonathan Ellis < <jbel...@gmail.com> >>>> jbel...@gmail.com> wrote: >>>> > There definitely seems to be demand for something like this. Maybe >>>> for 0.8? >>>> > >>>> >>>> The Riak data store has something like this: you can submit queries >>>> (and map reduce jobs) written in javascript that run on the data nodes >>>> using data local to that node. It is a very compelling feature. >>>> >>> >>> >> >