fwiw. i ended up re-creating the ec2 cluster with that same host name just so i could drop those tables from the metastore.
note to self. be careful - be real careful - with "sharing" hive metastores between different compute paradigms. Regards, Stephen. On Tue, May 16, 2017 at 6:38 AM, Stephen Sprague <sprag...@gmail.com> wrote: > hey guys, > here's something bizarre. i created about 200 external tables with a > location something like this 'hdfs://<ec2-host>/path'. this was three > months ago and now i'm revisiting and want to drop these tables. > > ha! no can do! > > that <ec2-host> is long gone. > > Upon issuing the drop table command i get this: > > Error while processing statement: FAILED: Execution Error, return code 1 > from org.apache.hadoop.hive.ql.exec.DDLTask. > MetaException(message:java.lang.IllegalArgumentException: > java.net.UnknownHostException: <ec2-host>) > > where <ec2-host> is that old host name. > > so i ask is there a work around for this? given they are external tables > i'm surprised it "checks" that that location exists (or not.) > > thanks, > Stephen >