NO, can you pease explain the command ? Let me try now.

Best,

On Sun, Feb 11, 2018 at 12:40 AM, Deepak Sharma <deepakmc...@gmail.com>
wrote:

> I am not sure about the exact issue bjt i see you are partioning while
> writing from spark.
> Did you tried msck repair on the table before reading it in hive ?
>
> Thanks
> Deepak
>
> On Feb 11, 2018 11:06, "☼ R Nair (रविशंकर नायर)" <
> ravishankar.n...@gmail.com> wrote:
>
>> All,
>>
>> Thanks for the inputs. Again I am not successful. I think, we need to
>> resolve this, as this is a very common requirement.
>>
>> Please go through my complete code:
>>
>> STEP 1:  Started Spark shell as spark-shell --master yarn
>>
>> STEP 2: Flowing code is being given as inout to shark shell
>>
>> import org.apache.spark.sql.Row
>> import org.apache.spark.sql.SparkSession
>> val warehouseLocation ="/user/hive/warehouse"
>>
>> val spark = SparkSession.builder().appName("Spark Hive
>> Example").config("spark.sql.warehouse.dir",
>> warehouseLocation).enableHiveSupport().getOrCreate()
>>
>> import org.apache.spark.sql._
>> var passion_df = spark.read.
>> format("jdbc").
>> option("url", "jdbc:mysql://localhost:3307/policies").
>> option("driver" ,"com.mysql.jdbc.Driver").
>> option("user", "root").
>> option("password", "root").
>> option("dbtable", "insurancedetails").
>> option("partitionColumn", "policyid").
>> option("lowerBound", "1").
>> option("upperBound", "100000").
>> option("numPartitions", "4").
>> load()
>> //Made sure that passion_df is created, as passion_df.show(5) shows me
>> correct data.
>> passion_df.write.saveAsTable("default.mine") //Default parquet
>>
>> STEP 3: Went to HIVE. Started HIVE prompt.
>>
>> hive> show tables;
>> OK
>> callcentervoicelogs
>> mine
>> Time taken: 0.035 seconds, Fetched: 2 row(s)
>> //As you can see HIVE is showing the table "mine" in default schema.
>>
>> STEP 4: HERE IS THE PROBLEM.
>>
>> hive> select * from mine;
>> OK
>> Time taken: 0.354 seconds
>> hive>
>> //Where is the data ???
>>
>> STEP 5:
>>
>> See the below command on HIVE
>>
>> describe formatted mine;
>> OK
>> # col_name             data_type           comment
>>
>> policyid             int
>> statecode           string
>> socialid             string
>> county               string
>> eq_site_limit       decimal(10,2)
>> hu_site_limit       decimal(10,2)
>> fl_site_limit       decimal(10,2)
>> fr_site_limit       decimal(10,2)
>> tiv_2014             decimal(10,2)
>> tiv_2015             decimal(10,2)
>> eq_site_deductible   int
>> hu_site_deductible   int
>> fl_site_deductible   int
>> fr_site_deductible   int
>> latitude             decimal(6,6)
>> longitude           decimal(6,6)
>> line                 string
>> construction         string
>> point_granularity   int
>>
>> # Detailed Table Information
>> Database:           default
>> Owner:               ravishankarnair
>> CreateTime:         Sun Feb 11 00:26:40 EST 2018
>> LastAccessTime:     UNKNOWN
>> Protect Mode:       None
>> Retention:           0
>> Location:           file:/Users/ravishankarnair/spark-warehouse/mine
>> Table Type:         MANAGED_TABLE
>> Table Parameters:
>> spark.sql.sources.provider parquet
>> spark.sql.sources.schema.numParts 1
>> spark.sql.sources.schema.part.0 {\"type\":\"struct\",\"fields\
>> ":[{\"name\":\"policyid\",\"type\":\"integer\",\"nullable\":
>> true,\"metadata\":{\"name\":\"policyid\",\"scale\":0}},{\"na
>> me\":\"statecode\",\"type\":\"string\",\"nullable\":true,\"
>> metadata\":{\"name\":\"statecode\",\"scale\":0}},{\"name\":\
>> "Socialid\",\"type\":\"string\",\"nullable\":true,\"metadata
>> \":{\"name\":\"Socialid\",\"scale\":0}},{\"name\":\"
>> county\",\"type\":\"string\",\"nullable\":true,\"metadata\":
>> {\"name\":\"county\",\"scale\":0}},{\"name\":\"eq_site_
>> limit\",\"type\":\"decimal(10,2)\",\"nullable\":true,\"
>> metadata\":{\"name\":\"eq_site_limit\",\"scale\":2}},{\"
>> name\":\"hu_site_limit\",\"type\":\"decimal(10,2)\",\"nullab
>> le\":true,\"metadata\":{\"name\":\"hu_site_limit\",\"sca
>> le\":2}},{\"name\":\"fl_site_limit\",\"type\":\"decimal(10,
>> 2)\",\"nullable\":true,\"metadata\":{\"name\":\"fl_
>> site_limit\",\"scale\":2}},{\"name\":\"fr_site_limit\",\"typ
>> e\":\"decimal(10,2)\",\"nullable\":true,\"metadata\":{\"
>> name\":\"fr_site_limit\",\"scale\":2}},{\"name\":\"tiv_2014\
>> ",\"type\":\"decimal(10,2)\",\"nullable\":true,\"metadata\":
>> {\"name\":\"tiv_2014\",\"scale\":2}},{\"name\":\"tiv_
>> 2015\",\"type\":\"decimal(10,2)\",\"nullable\":true,\"
>> metadata\":{\"name\":\"tiv_2015\",\"scale\":2}},{\"name\"
>> :\"eq_site_deductible\",\"type\":\"integer\",\"nullable\
>> ":true,\"metadata\":{\"name\":\"eq_site_deductible\",\"
>> scale\":0}},{\"name\":\"hu_site_deductible\",\"type\":\"
>> integer\",\"nullable\":true,\"metadata\":{\"name\":\"hu_site
>> _deductible\",\"scale\":0}},{\"name\":\"fl_site_deductible\"
>> ,\"type\":\"integer\",\"nullable\":true,\"metadata\":{
>> \"name\":\"fl_site_deductible\",\"scale\":0}},{\"name\":\"
>> fr_site_deductible\",\"type\":\"integer\",\"nullable\":true,
>> \"metadata\":{\"name\":\"fr_site_deductible\",\"scale\":0}},
>> {\"name\":\"latitude\",\"type\":\"decimal(6,6)\",\"nullable\
>> ":true,\"metadata\":{\"name\":\"latitude\",\"scale\":6}},{\"
>> name\":\"longitude\",\"type\":\"decimal(6,6)\",\"nullable\":
>> true,\"metadata\":{\"name\":\"longitude\",\"scale\":6}},{\"
>> name\":\"line\",\"type\":\"string\",\"nullable\":true,\"
>> metadata\":{\"name\":\"line\",\"scale\":0}},{\"name\":\"
>> construction\",\"type\":\"string\",\"nullable\":true,\"
>> metadata\":{\"name\":\"construction\",\"scale\":0}},{
>> \"name\":\"point_granularity\",\"type\":\"integer\",\"
>> nullable\":true,\"metadata\":{\"name\":\"point_granularity\"
>> ,\"scale\":0}}]}
>> transient_lastDdlTime 1518326800
>>
>> # Storage Information
>> SerDe Library:       org.apache.hadoop.hive.ql.io.p
>> arquet.serde.ParquetHiveSerDe
>> InputFormat:         org.apache.hadoop.hive.ql.io.p
>> arquet.MapredParquetInputFormat
>> OutputFormat:       org.apache.hadoop.hive.ql.io.p
>> arquet.MapredParquetOutputFormat
>> Compressed:         No
>> Num Buckets:         -1
>> Bucket Columns:     []
>> Sort Columns:       []
>> Storage Desc Params:
>> path                 hdfs://localhost:8020/user/hive/warehouse/mine
>> serialization.format 1
>> Time taken: 0.077 seconds, Fetched: 48 row(s)
>>
>> Now, I see your advise and support. Whats the issue? Am I doing wrong, it
>> it a bug ? I am using Spark 2.2.1, HIVE 1.2.1, HADOOP 2.7.3. All class
>> path, configuration are set properly.
>>
>> Best,
>>
>> Ravion
>>
>> On Fri, Feb 9, 2018 at 1:29 PM, Nicholas Hakobian <
>> nicholas.hakob...@rallyhealth.com> wrote:
>>
>>> Its possible that the format of your table is not compatible with your
>>> version of hive, so Spark saved it in a way such that only Spark can read
>>> it. When this happens it prints out a very visible warning letting you know
>>> this has happened.
>>>
>>> We've seen it most frequently when trying to save a parquet file with a
>>> column in date format into a Hive table. In older versions of hive, its
>>> parquet reader/writer did not support Date formats (among a couple others).
>>>
>>> Nicholas Szandor Hakobian, Ph.D.
>>> Staff Data Scientist
>>> Rally Health
>>> nicholas.hakob...@rallyhealth.com
>>>
>>>
>>> On Fri, Feb 9, 2018 at 9:59 AM, Prakash Joshi <prakashcjos...@gmail.com>
>>> wrote:
>>>
>>>> Ravi,
>>>>
>>>> Can you send the result of
>>>> Show create table your_table_name
>>>>
>>>> Thanks
>>>> Prakash
>>>>
>>>> On Feb 9, 2018 8:20 PM, "☼ R Nair (रविशंकर नायर)" <
>>>> ravishankar.n...@gmail.com> wrote:
>>>>
>>>>> All,
>>>>>
>>>>> It has been three days continuously I am on this issue. Not getting
>>>>> any clue.
>>>>>
>>>>> Environment: Spark 2.2.x, all configurations are correct.
>>>>> hive-site.xml is in spark's conf.
>>>>>
>>>>> 1) Step 1: I created a data frame DF1 reading a csv file.
>>>>>
>>>>> 2) Did  manipulations on DF1. Resulting frame is passion_df.
>>>>>
>>>>> 3) passion_df.write.format("orc").saveAsTable("sampledb.passion")
>>>>>
>>>>> 4) The metastore shows the hive table., when I do "show tables" in
>>>>> HIVE, I can see table name
>>>>>
>>>>> 5) I can't select in HIVE, though I can select from SPARK as
>>>>> spark.sql("select * from sampledb.passion")
>>>>>
>>>>> Whats going on here? Please help. Why I am not seeing data from HIVE
>>>>> prompt?
>>>>> The "describe formatted " command on the table in HIVE shows he data
>>>>> is is in default warehouse location ( /user/hive/warehouse) since I set 
>>>>> it.
>>>>>
>>>>> I am not getting any definite answer anywhere. Many suggestions and
>>>>> answers given in Stackoverflow et al.Nothing really works.
>>>>>
>>>>> So asking experts here for some light on this, thanks
>>>>>
>>>>> Best,
>>>>> Ravion
>>>>>
>>>>>
>>>>>
>>>
>>
>>
>> --
>>
>>


--

Reply via email to