Nirmalkumar created HIVE-14844:
----------------------------------

             Summary: Not able to create the Hive table with more than 700 
columns
                 Key: HIVE-14844
                 URL: https://issues.apache.org/jira/browse/HIVE-14844
             Project: Hive
          Issue Type: Bug
          Components: Metastore
    Affects Versions: 1.1.0
         Environment: MySQL
            Reporter: Nirmalkumar


We tried creating the Hive table with 700+ columns which leads to the below 
error:

FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. 
MetaException(message:javax.jdo.JDODataStoreException: Put request failed : 
INSERT INTO `SERDE_PARAMS` (`PARAM_VALUE`,`SERDE_ID`,`PARAM_KEY`) VALUES (?,?,?)

we use hive 1.1.0 and below is the SERDE properties we have in our MySQL.

mysql> desc SERDE_PARAMS;
+-------------+---------------+------+-----+---------+-------+
| Field       | Type          | Null | Key | Default | Extra |
+-------------+---------------+------+-----+---------+-------+
| SERDE_ID    | bigint(20)    | NO   | PRI | NULL    |       |
| PARAM_KEY   | varchar(256)  | NO   | PRI | NULL    |       |
| PARAM_VALUE | varchar(4000) | YES  |     | NULL    |       |
+-------------+---------------+------+-----+---------+-------+

As per Cloudera,
Actually it's a known limitation and there are no patches for this. 
To avoid this issue typically is increased the size of the field in the 
metastore, BUT this is not a supported Cloudera solution that can result in a 
break in a future Metastore schema upgrade (in case if we are evolving our CDH 
version to an upper one that will implement an hive version greater than the 
actual one).

So could you please provide a fix for this?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to