I had this:

string slice_dice_reduce(1:required list<binary> key,
                                      2:required ColumnParent
column_parent,
                                      3:required SlicePredicate predicate,
                                      4:required ConsistencyLevel
consistency_level=ONE,
                                      5:required string dice_js,
                                      6:required string reduce_js)
                            throws (1:InvalidRequestException ire,
2:UnavailableException ue, 3:TimedOutException te),

I guess it could use a union of sorts and return either.



On Thu, May 27, 2010 at 8:36 PM, Jeremy Davis
<jerdavis.cassan...@gmail.com>wrote:

>
> I agree, I had more than filter results in mind.
> Though I had envisioned the results to continue to use the
> List<ColumnOrSuperColumn> (and not JSON). You could still create new result
> columns that do not in any way exist in Cassandra, and you could still stuff
> JSON in to any of result columns.
>
> I had envisioned:
> list<ColumnOrSuperColumn> get_slice(keyspace, key, column_parent, predicate, 
> consistency_level,
> javascript_blob )
>
> -JD
>
>
>
>
>
> On Thu, May 27, 2010 at 5:01 PM, Jake Luciani <jak...@gmail.com> wrote:
>
>> I've secretly started working on this but nothing to show yet :( I'm
>> calling it SliceDiceReduce or SliceReduce.
>>
>>  The plan is to use the js thrift bindings I've added for 0.3 release of
>> thrift (out very soon?)
>>
>> This will allow the supplied js to access the results like any other
>> thrift client.
>>
>> Adding a new verb handler and SEDA stage that will execute on a local node
>> and pass this nodes slice data into the supplied js "dice" function via the
>> thrift js bindings.
>>
>> The resulting js from each node would then be passed into another supplied
>> js reduce function on the starting node.
>>
>> The result of this would then return a single JSON or string result.  The
>> reason I'm keeping the results in json is you can do more than filter. You
>> can do things like word count etc.
>>
>> Anyway this is little more than an idea now. But if people like this
>> approach maybe I'll get motivated!
>>
>> Jake
>>
>>
>>
>>
>>
>> On May 27, 2010, at 7:36 PM, Steve Lihn <stevel...@gmail.com> wrote:
>>
>> Mongo has it too. It could save a lot of development time if one can
>> figure out porting Mongo's query API and stored javascript to Cassandra.
>> It would be great if scala's list comprehension can be facilitated to
>> write query-like code against Cassandra schema.
>>
>> On Thu, May 27, 2010 at 11:05 AM, Vick Khera < <vi...@khera.org>
>> vi...@khera.org> wrote:
>>
>>> On Thu, May 27, 2010 at 9:50 AM, Jonathan Ellis < <jbel...@gmail.com>
>>> jbel...@gmail.com> wrote:
>>> > There definitely seems to be demand for something like this.  Maybe for
>>> 0.8?
>>> >
>>>
>>> The Riak data store has something like this: you can submit queries
>>> (and map reduce jobs) written in javascript that run on the data nodes
>>> using data local to that node.  It is a very compelling feature.
>>>
>>
>>
>

Reply via email to