I read the table via spark SQL , and perform some  ML activity on the data
, and the resultant will be to update some specific columns with the ML
improvised result,
hence i do not have a option to do the whole operation in MySQL,


Thanks,
Sujeet

On Thu, Aug 11, 2016 at 3:29 PM, Mich Talebzadeh <mich.talebza...@gmail.com>
wrote:

> Ok it is clearer now.
>
> You are using Spark as the query tool on an RDBMS table? Read table via
> JDBC, write back updating certain records.
>
> I have not done this myself but I suspect the issue would be if Spark
> write will commit the transaction and maintains ACID compliance. (locking
> the rows etc).
>
> I know it cannot do this to a Hive transactional table.
>
> Any reason why you are not doing the whole operation in MySQL itself?
>
> HTH
>
>
> Dr Mich Talebzadeh
>
>
>
> LinkedIn * 
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>
>
>
> http://talebzadehmich.wordpress.com
>
>
> *Disclaimer:* Use it at your own risk. Any and all responsibility for any
> loss, damage or destruction of data or any other property which may arise
> from relying on this email's technical content is explicitly disclaimed.
> The author will in no case be liable for any monetary damages arising from
> such loss, damage or destruction.
>
>
>
> On 11 August 2016 at 10:46, sujeet jog <sujeet....@gmail.com> wrote:
>
>> 1 ) using mysql DB
>> 2 ) will be inserting/update/overwrite to the same table
>> 3 ) i want to update a specific column in a record, the data is read via
>> Spark SQL,
>>
>> on the below table which is read via sparkSQL, i would like to update the
>> NumOfSamples column .
>>
>> consider DF as the dataFrame which holds the records,  registered as
>> temporary table MS .
>>
>> spark.sqlContext.write.format("jdbc").option("url", url
>> ).option("dbtable", "update ms  set NumOfSamples = 20 where 'TimeSeriesID =
>> '1000'" As MS ).save
>>
>> I believe updating a record via sparkSQL is not supported,  the only
>> workaround is to open up a jdbc connection without using spark API's and do
>> a direct update ?..
>>
>> Sample Ex : -
>>
>> mysql> show columns from ms;
>> +--------------+-------------+------+-----+---------+-------+
>> | Field        | Type        | Null | Key | Default | Extra |
>> +--------------+-------------+------+-----+---------+-------+
>> | TimeSeriesID | varchar(20) | YES  |     | NULL    |       |
>> | NumOfSamples | int(11)     | YES  |     | NULL    |       |
>> +--------------+-------------+------+-----+---------+-------+
>>
>>
>> Thanks,
>> Sujeet
>>
>>
>>
>> On Tue, Aug 9, 2016 at 6:31 PM, Mich Talebzadeh <
>> mich.talebza...@gmail.com> wrote:
>>
>>> Hi,
>>>
>>>
>>>    1. what is the underlying DB, say Hive etc
>>>    2. Is table transactional or you are going to do insert/overwrite to
>>>    the same table
>>>    3. can you do all this in the database itself assuming it is an RDBMS
>>>    4. Can you provide the sql or pseudo code for such an update
>>>
>>>
>>> HTH
>>>
>>> Dr Mich Talebzadeh
>>>
>>>
>>>
>>> LinkedIn * 
>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>
>>>
>>>
>>> http://talebzadehmich.wordpress.com
>>>
>>>
>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>> any loss, damage or destruction of data or any other property which may
>>> arise from relying on this email's technical content is explicitly
>>> disclaimed. The author will in no case be liable for any monetary damages
>>> arising from such loss, damage or destruction.
>>>
>>>
>>>
>>> On 9 August 2016 at 13:39, sujeet jog <sujeet....@gmail.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> Is it possible to update certain columnr records  in DB  from spark,
>>>>
>>>> for example i have 10 rows with 3 columns  which are read from Spark
>>>> SQL,
>>>>
>>>> i want to update specific column entries  and write back to DB, but
>>>> since RDD"s are immutable i believe this would be difficult, is there a
>>>> workaround.
>>>>
>>>>
>>>> Thanks,
>>>> Sujeet
>>>>
>>>
>>>
>>
>

Reply via email to