Here's a complete scala example
https://github.com/bbux-proteus/spark-accumulo-examples/blob/1dace96a115f29c44325903195c8135edf828c86/src/main/scala/org/bbux/spark/AccumuloMetadataCount.scala

Thanks
Best Regards

On Thu, Apr 23, 2015 at 12:19 PM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:

> Change your import from mapred to mapreduce. like :
>
> import org.apache.accumulo.core.client.mapreduce.AccumuloInputFormat;
>
> Thanks
> Best Regards
>
> On Wed, Apr 22, 2015 at 2:42 PM, madhvi <madhvi.gu...@orkash.com> wrote:
>
>> Hi,
>>
>> I am creating a spark RDD through accumulo writing like:
>>
>> JavaPairRDD<Key, Value> accumuloRDD =
>> sc.newAPIHadoopRDD(accumuloJob.getConfiguration(),AccumuloInputFormat.class,Key.class,
>> Value.class);
>>
>> But I am getting the following error and it is not getting compiled:
>>
>> Bound mismatch: The generic method newAPIHadoopRDD(Configuration,
>> Class<F>, Class<K>, Class<V>) of type JavaSparkContext is not applicable
>> for the arguments (Configuration, Class<AccumuloInputFormat>, Class<Key>,
>> Class<Value>). The inferred type AccumuloInputFormat is not a valid
>> substitute for the bounded parameter <F extends InputFormat<K,V>>
>>
>> I am using the following import statements:
>>
>> import org.apache.accumulo.core.client.mapred.AccumuloInputFormat;
>> import org.apache.accumulo.core.data.Key;
>> import org.apache.accumulo.core.data.Value;
>>
>> I am not getting what is the problem in this.
>>
>> Thanks
>> Madhvi
>>
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>

Reply via email to