No, No luck.

Thanks

On Sun, Feb 11, 2018 at 12:48 AM, Deepak Sharma <deepakmc...@gmail.com>
wrote:

> In hive cli:
> msck repair table 《table_name》;
>
> Thanks
> Deepak
>
> On Feb 11, 2018 11:14, "☼ R Nair (रविशंकर नायर)" <
> ravishankar.n...@gmail.com> wrote:
>
>> NO, can you pease explain the command ? Let me try now.
>>
>> Best,
>>
>> On Sun, Feb 11, 2018 at 12:40 AM, Deepak Sharma <deepakmc...@gmail.com>
>> wrote:
>>
>>> I am not sure about the exact issue bjt i see you are partioning while
>>> writing from spark.
>>> Did you tried msck repair on the table before reading it in hive ?
>>>
>>> Thanks
>>> Deepak
>>>
>>> On Feb 11, 2018 11:06, "☼ R Nair (रविशंकर नायर)" <
>>> ravishankar.n...@gmail.com> wrote:
>>>
>>>> All,
>>>>
>>>> Thanks for the inputs. Again I am not successful. I think, we need to
>>>> resolve this, as this is a very common requirement.
>>>>
>>>> Please go through my complete code:
>>>>
>>>> STEP 1:  Started Spark shell as spark-shell --master yarn
>>>>
>>>> STEP 2: Flowing code is being given as inout to shark shell
>>>>
>>>> import org.apache.spark.sql.Row
>>>> import org.apache.spark.sql.SparkSession
>>>> val warehouseLocation ="/user/hive/warehouse"
>>>>
>>>> val spark = SparkSession.builder().appName("Spark Hive
>>>> Example").config("spark.sql.warehouse.dir",
>>>> warehouseLocation).enableHiveSupport().getOrCreate()
>>>>
>>>> import org.apache.spark.sql._
>>>> var passion_df = spark.read.
>>>> format("jdbc").
>>>> option("url", "jdbc:mysql://localhost:3307/policies").
>>>> option("driver" ,"com.mysql.jdbc.Driver").
>>>> option("user", "root").
>>>> option("password", "root").
>>>> option("dbtable", "insurancedetails").
>>>> option("partitionColumn", "policyid").
>>>> option("lowerBound", "1").
>>>> option("upperBound", "100000").
>>>> option("numPartitions", "4").
>>>> load()
>>>> //Made sure that passion_df is created, as passion_df.show(5) shows me
>>>> correct data.
>>>> passion_df.write.saveAsTable("default.mine") //Default parquet
>>>>
>>>> STEP 3: Went to HIVE. Started HIVE prompt.
>>>>
>>>> hive> show tables;
>>>> OK
>>>> callcentervoicelogs
>>>> mine
>>>> Time taken: 0.035 seconds, Fetched: 2 row(s)
>>>> //As you can see HIVE is showing the table "mine" in default schema.
>>>>
>>>> STEP 4: HERE IS THE PROBLEM.
>>>>
>>>> hive> select * from mine;
>>>> OK
>>>> Time taken: 0.354 seconds
>>>> hive>
>>>> //Where is the data ???
>>>>
>>>> STEP 5:
>>>>
>>>> See the below command on HIVE
>>>>
>>>> describe formatted mine;
>>>> OK
>>>> # col_name             data_type           comment
>>>>
>>>> policyid             int
>>>> statecode           string
>>>> socialid             string
>>>> county               string
>>>> eq_site_limit       decimal(10,2)
>>>> hu_site_limit       decimal(10,2)
>>>> fl_site_limit       decimal(10,2)
>>>> fr_site_limit       decimal(10,2)
>>>> tiv_2014             decimal(10,2)
>>>> tiv_2015             decimal(10,2)
>>>> eq_site_deductible   int
>>>> hu_site_deductible   int
>>>> fl_site_deductible   int
>>>> fr_site_deductible   int
>>>> latitude             decimal(6,6)
>>>> longitude           decimal(6,6)
>>>> line                 string
>>>> construction         string
>>>> point_granularity   int
>>>>
>>>> # Detailed Table Information
>>>> Database:           default
>>>> Owner:               ravishankarnair
>>>> CreateTime:         Sun Feb 11 00:26:40 EST 2018
>>>> LastAccessTime:     UNKNOWN
>>>> Protect Mode:       None
>>>> Retention:           0
>>>> Location:           file:/Users/ravishankarnair/spark-warehouse/mine
>>>> Table Type:         MANAGED_TABLE
>>>> Table Parameters:
>>>> spark.sql.sources.provider parquet
>>>> spark.sql.sources.schema.numParts 1
>>>> spark.sql.sources.schema.part.0 {\"type\":\"struct\",\"fields\
>>>> ":[{\"name\":\"policyid\",\"type\":\"integer\",\"nullable\":
>>>> true,\"metadata\":{\"name\":\"policyid\",\"scale\":0}},{\"na
>>>> me\":\"statecode\",\"type\":\"string\",\"nullable\":true,\"m
>>>> etadata\":{\"name\":\"statecode\",\"scale\":0}},{\"name\":\"
>>>> Socialid\",\"type\":\"string\",\"nullable\":true,\"metadata\
>>>> ":{\"name\":\"Socialid\",\"scale\":0}},{\"name\":\"county\",
>>>> \"type\":\"string\",\"nullable\":true,\"metadata\":{\"name\"
>>>> :\"county\",\"scale\":0}},{\"name\":\"eq_site_limit\",\"
>>>> type\":\"decimal(10,2)\",\"nullable\":true,\"metadata\":{
>>>> \"name\":\"eq_site_limit\",\"scale\":2}},{\"name\":\"hu_site
>>>> _limit\",\"type\":\"decimal(10,2)\",\"nullable\":true,\"
>>>> metadata\":{\"name\":\"hu_site_limit\",\"scale\":2}},{\"
>>>> name\":\"fl_site_limit\",\"type\":\"decimal(10,2)\",\"nullab
>>>> le\":true,\"metadata\":{\"name\":\"fl_site_limit\",\"sca
>>>> le\":2}},{\"name\":\"fr_site_limit\",\"type\":\"decimal(10,
>>>> 2)\",\"nullable\":true,\"metadata\":{\"name\":\"fr_
>>>> site_limit\",\"scale\":2}},{\"name\":\"tiv_2014\",\"type\":\
>>>> "decimal(10,2)\",\"nullable\":true,\"metadata\":{\"name\":\"
>>>> tiv_2014\",\"scale\":2}},{\"name\":\"tiv_2015\",\"type\":\
>>>> "decimal(10,2)\",\"nullable\":true,\"metadata\":{\"name\":\"
>>>> tiv_2015\",\"scale\":2}},{\"name\":\"eq_site_deductible\",
>>>> \"type\":\"integer\",\"nullable\":true,\"metadata\":{
>>>> \"name\":\"eq_site_deductible\",\"scale\":0}},{\"name\":\"
>>>> hu_site_deductible\",\"type\":\"integer\",\"nullable\":true,
>>>> \"metadata\":{\"name\":\"hu_site_deductible\",\"scale\":0}
>>>> },{\"name\":\"fl_site_deductible\",\"type\":\"
>>>> integer\",\"nullable\":true,\"metadata\":{\"name\":\"fl_
>>>> site_deductible\",\"scale\":0}},{\"name\":\"fr_site_
>>>> deductible\",\"type\":\"integer\",\"nullable\":true,\"metada
>>>> ta\":{\"name\":\"fr_site_deductible\",\"scale\":0}},{\"
>>>> name\":\"latitude\",\"type\":\"decimal(6,6)\",\"nullable\":t
>>>> rue,\"metadata\":{\"name\":\"latitude\",\"scale\":6}},{\"nam
>>>> e\":\"longitude\",\"type\":\"decimal(6,6)\",\"nullable\":tru
>>>> e,\"metadata\":{\"name\":\"longitude\",\"scale\":6}},{\"name
>>>> \":\"line\",\"type\":\"string\",\"nullable\":true,\"metadata
>>>> \":{\"name\":\"line\",\"scale\":0}},{\"name\":\"construction
>>>> \",\"type\":\"string\",\"nullable\":true,\"metadata\":{
>>>> \"name\":\"construction\",\"scale\":0}},{\"name\":\"point_gr
>>>> anularity\",\"type\":\"integer\",\"nullable\":true,\"metadat
>>>> a\":{\"name\":\"point_granularity\",\"scale\":0}}]}
>>>> transient_lastDdlTime 1518326800
>>>>
>>>> # Storage Information
>>>> SerDe Library:       org.apache.hadoop.hive.ql.io.p
>>>> arquet.serde.ParquetHiveSerDe
>>>> InputFormat:         org.apache.hadoop.hive.ql.io.p
>>>> arquet.MapredParquetInputFormat
>>>> OutputFormat:       org.apache.hadoop.hive.ql.io.p
>>>> arquet.MapredParquetOutputFormat
>>>> Compressed:         No
>>>> Num Buckets:         -1
>>>> Bucket Columns:     []
>>>> Sort Columns:       []
>>>> Storage Desc Params:
>>>> path                 hdfs://localhost:8020/user/hive/warehouse/mine
>>>> serialization.format 1
>>>> Time taken: 0.077 seconds, Fetched: 48 row(s)
>>>>
>>>> Now, I see your advise and support. Whats the issue? Am I doing wrong,
>>>> it it a bug ? I am using Spark 2.2.1, HIVE 1.2.1, HADOOP 2.7.3. All class
>>>> path, configuration are set properly.
>>>>
>>>> Best,
>>>>
>>>> Ravion
>>>>
>>>> On Fri, Feb 9, 2018 at 1:29 PM, Nicholas Hakobian <
>>>> nicholas.hakob...@rallyhealth.com> wrote:
>>>>
>>>>> Its possible that the format of your table is not compatible with your
>>>>> version of hive, so Spark saved it in a way such that only Spark can read
>>>>> it. When this happens it prints out a very visible warning letting you 
>>>>> know
>>>>> this has happened.
>>>>>
>>>>> We've seen it most frequently when trying to save a parquet file with
>>>>> a column in date format into a Hive table. In older versions of hive, its
>>>>> parquet reader/writer did not support Date formats (among a couple 
>>>>> others).
>>>>>
>>>>> Nicholas Szandor Hakobian, Ph.D.
>>>>> Staff Data Scientist
>>>>> Rally Health
>>>>> nicholas.hakob...@rallyhealth.com
>>>>>
>>>>>
>>>>> On Fri, Feb 9, 2018 at 9:59 AM, Prakash Joshi <
>>>>> prakashcjos...@gmail.com> wrote:
>>>>>
>>>>>> Ravi,
>>>>>>
>>>>>> Can you send the result of
>>>>>> Show create table your_table_name
>>>>>>
>>>>>> Thanks
>>>>>> Prakash
>>>>>>
>>>>>> On Feb 9, 2018 8:20 PM, "☼ R Nair (रविशंकर नायर)" <
>>>>>> ravishankar.n...@gmail.com> wrote:
>>>>>>
>>>>>>> All,
>>>>>>>
>>>>>>> It has been three days continuously I am on this issue. Not getting
>>>>>>> any clue.
>>>>>>>
>>>>>>> Environment: Spark 2.2.x, all configurations are correct.
>>>>>>> hive-site.xml is in spark's conf.
>>>>>>>
>>>>>>> 1) Step 1: I created a data frame DF1 reading a csv file.
>>>>>>>
>>>>>>> 2) Did  manipulations on DF1. Resulting frame is passion_df.
>>>>>>>
>>>>>>> 3) passion_df.write.format("orc").saveAsTable("sampledb.passion")
>>>>>>>
>>>>>>> 4) The metastore shows the hive table., when I do "show tables" in
>>>>>>> HIVE, I can see table name
>>>>>>>
>>>>>>> 5) I can't select in HIVE, though I can select from SPARK as
>>>>>>> spark.sql("select * from sampledb.passion")
>>>>>>>
>>>>>>> Whats going on here? Please help. Why I am not seeing data from HIVE
>>>>>>> prompt?
>>>>>>> The "describe formatted " command on the table in HIVE shows he data
>>>>>>> is is in default warehouse location ( /user/hive/warehouse) since I set 
>>>>>>> it.
>>>>>>>
>>>>>>> I am not getting any definite answer anywhere. Many suggestions and
>>>>>>> answers given in Stackoverflow et al.Nothing really works.
>>>>>>>
>>>>>>> So asking experts here for some light on this, thanks
>>>>>>>
>>>>>>> Best,
>>>>>>> Ravion
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>>
>>>>
>>
>>
>> --
>>
>>


--

Reply via email to