Hi Everyone,
I am new to spark.may I know how to define and use User Define Function in
SPARK SQL.
I want to use defined UDF by using sql queries.
My Environment
Windows 8
spark 1.3.1
Warm Regards,
Vinod
com> wrote:
> Hi,
>
> sqlContext.udf.register("udfname", functionname _)
>
> example:
>
> def square(x:Int):Int = { x * x}
>
> register udf as below
>
> sqlContext.udf.register("square",square _)
>
> Thanks,
> Vishnu
>
> On Wed,
, VISHNU SUBRAMANIAN <
johnfedrickena...@gmail.com> wrote:
> HI Vinod,
>
> Yes If you want to use a scala or python function you need the block of
> code.
>
> Only Hive UDF's are available permanently.
>
> Thanks,
> Vishnu
>
> On Wed, Jul 8, 2015 at 5:17
Hi everyone
Shall we use UDF defined in hive using spark sql?
I've created a UDF in spark and registered it using
sqlContext.udf.register,but when I restarted a service the UDF was not
available.
I've heared that Hive UDF's are permanently stored in hive.(Please Correct
me if I am wrong).
Thanks
Hi Everyone,
I am new to spark.
Am using SQL in my application to handle data in my application.I have a
thought to move to spark now.
Is data processing speed of spark better than SQL server?
Thank,
Vinod
to process?
> On 9 Jul 2015 22:28, "vinod kumar" wrote:
>
>> Hi Everyone,
>>
>> I am new to spark.
>>
>> Am using SQL in my application to handle data in my application.I have a
>> thought to move to spark now.
>>
>> Is data processing speed of spark better than SQL server?
>>
>> Thank,
>> Vinod
>>
>
For records below 50,000 SQL is better right?
On Fri, Jul 10, 2015 at 12:18 AM, ayan guha wrote:
> With your load, either should be fine.
>
> I would suggest you to run couple of quick prototype.
>
> Best
> Ayan
>
> On Fri, Jul 10, 2015 at 2:06 PM, vinod kumar
>
Hi Everyone,
Is there is any document/material which compares spark with SQL server?
If so please share me the details.
Thanks,
Vinod
Hi Guys,
Can any one please share me how to use caching feature of spark via spark
sql queries?
-Vinod
Hi Everyone,
I am developing application which handles bulk of data around millions(This
may vary as per user's requirement) records.As of now I am using
MsSqlServer as back-end and it works fine but when I perform some
operation on large data I am getting overflow exceptions.I heard about
spark
15 at 3:50 AM, Akhil Das
wrote:
> This is where you can get started
> https://spark.apache.org/docs/latest/sql-programming-guide.html
>
> Thanks
> Best Regards
>
> On Mon, Jul 13, 2015 at 3:54 PM, vinod kumar
> wrote:
>
>>
>> Hi Everyone,
>>
>>
if
>> you have more memory, also if you have enough cores 4 records are
>> nothing.
>>
>> Thanks
>> Best Regards
>>
>> On Tue, Jul 14, 2015 at 3:09 PM, vinod kumar
>> wrote:
>>
>>> Hi Akhil
>>>
>>> Is my choice to switc
Hi Guys,
May I know whether cache is enabled in spark by default?
Thanks,
Vinod
s you tell it too.
>
> You have to explicitly call the cache method in your RDD:
> linesWithSpark.cache()
>
> Thanks,
> Natu
>
>
>
>
> On Fri, Oct 9, 2015 at 10:47 AM, vinod kumar
> wrote:
>
>> Hi Guys,
>>
>> May I know whether cache is enabled in spark by default?
>>
>> Thanks,
>> Vinod
>>
>
>
Hi Everyone,
I am in need to use the table from MsSQLSERVER in SPARK.Any one please
share me the optimized way for that?
Thanks in advance,
Vinod
Hi,
I am in need to create a table in spark.for that I have uploaded a csv file
in HDFS and created a table using following query
CREATE EXTERNAL table IF NOT EXISTS " + tableName + " (teams string,runs
int) " + "ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' LOCATION '" +
hdfspath + "'";
May I k
Hi,
May I know how to use the functions mentioned in
http://spark.apache.org/docs/1.4.0/api/scala/index.html#org.apache.spark.sql.functions$
in spark sql?
when I use like
"Select last(column) from tablename" I am getting error like
15/07/27 03:00:00 INFO exec.FunctionRegistry: Unable to lookup
to paste your test code here ? And which version of Spark
> are u using ?
>
> Best,
> Sun.
>
> ------
> fightf...@163.com
>
>
> *From:* vinod kumar
> *Date:* 2015-07-27 15:04
> *To:* User
> *Subject:* Functions in Spark SQ
Hi,
I am getting buffer over flow exception while using spark via thrifserver
base.May I know how to overcome this?
Code:
HqlConnection con = new HqlConnection("localhost", 10001,
HiveServer.HiveServer2);
con.Open();
HqlCommand createCommand = new HqlCommand(tablequery,
19 matches
Mail list logo