I just tried enabling kerberos on 0.6.1 and its now able to talk to my hive
metastore, I see all my databases and tables. However the moment I takeout
kerberos and run zeppelin 0.6.1 in local mode by resetting the "master" &
removing spark variables, i see just one default database.

I can live with this, as enabling kerberos integration is able to make
zeppelin talk to my hive metastore.

Thanks,
Pradeep

On Wed, Aug 31, 2016 at 8:36 PM, Pradeep Reddy <pradeepreddy.a...@gmail.com>
wrote:

> No luck, even after copying the hive-site.xml in
> interpreter/spark/dep....I also tried, downloading 0.6.1 src  vs 0.5.6. for
> the same steps I followed, 0.5.6 is able to talk to my hive metastore,
> where as the other recent builds are not, they just show one "default"
> database.
>
> Thanks,
> Pradeep
>
> On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jongy...@gmail.com> wrote:
>
>> I think it's related to https://issues.apache.org/j
>> ira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin
>> launches interpreter. Could you please check your hive-site.xml is included
>> in your interpreter process? It looks like a configuration issue because
>> you can see the default database. If it doesn't exists, you should copy
>> your xml into interpreter/spark/dep/
>>
>> Regards,
>> JL
>>
>> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy <
>> pradeepreddy.a...@gmail.com> wrote:
>>
>>> Hi Jongyoul- I followed the exact same steps for compiling and setting
>>> up the new build from source as 0.5.6 (only difference is, I acquired the
>>> source for latest build using "git clone")
>>>
>>> hive-site.xml was copied to conf directory. But, the spark interpreter
>>> is not talking to the hive metastore. Both the 0.5.6 & the latest builds
>>> are running in the same machine. In 0.5.6 when i run the below command, I
>>> see 116 databases listed, as per my expectations and I'm able to run my
>>> notebooks built on those databases.
>>>
>>> [image: Inline image 1]
>>>
>>> Thanks,
>>> Pradeep
>>>
>>>
>>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jongy...@gmail.com>
>>> wrote:
>>>
>>>> Hello,
>>>>
>>>> Do you copy your hive-site.xml in a proper position?
>>>>
>>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy <
>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>
>>>>> nothing obvious. I will stick to 0.5.6 build, until the latest builds
>>>>> stabilize.
>>>>>
>>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>>
>>>>>> Then I guess maybe you are connecting to different database. Why not
>>>>>> using  'z.show(sql("databases"))' to display the databases ? Then
>>>>>> you will get a hint what's going on.
>>>>>>
>>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>
>>>>>>> Yes...I didn't wish to show the names of the databases that we have
>>>>>>> in our data lake on that screen shot. so thats why I chose to display 
>>>>>>> the
>>>>>>> count. The latest zeppelin build just shows 1 count which is "default"
>>>>>>> database.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Pradeep
>>>>>>>
>>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zjf...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> 116 is the databases count number. Do you expect a list of database
>>>>>>>> ? then you need to use 'z.show(sql("databases"))'
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Here it is Jeff
>>>>>>>>>
>>>>>>>>> [image: Inline image 1]
>>>>>>>>>
>>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Pradeep,
>>>>>>>>>>
>>>>>>>>>> I don't see the databases on your screenshot (second one for
>>>>>>>>>> 0.5.6). I think the output is correct.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin
>>>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk 
>>>>>>>>>>> to the
>>>>>>>>>>> shared hive meta store are defects in the current build.
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi Jeff-
>>>>>>>>>>>>
>>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases 
>>>>>>>>>>>> &
>>>>>>>>>>>> tables, just shows 1 database "default".
>>>>>>>>>>>>
>>>>>>>>>>>> Zeppelin Latest Build
>>>>>>>>>>>>
>>>>>>>>>>>> [image: Inline image 3]
>>>>>>>>>>>>
>>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show my
>>>>>>>>>>>> databases and tables.
>>>>>>>>>>>>
>>>>>>>>>>>> [image: Inline image 4]
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> > the spark interpreter is not showing my tables & databases,
>>>>>>>>>>>>> may be its running in an isolated mode... I'm just getting empty 
>>>>>>>>>>>>> list, so I
>>>>>>>>>>>>> attempted to do kerberos authentication to workaround that issue, 
>>>>>>>>>>>>> and
>>>>>>>>>>>>> bumped into this road block.
>>>>>>>>>>>>>
>>>>>>>>>>>>> kerberos would not help here, actually I think it would make
>>>>>>>>>>>>> the problem more complicated.  You need to first check the log 
>>>>>>>>>>>>> why you get
>>>>>>>>>>>>> empty list.
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after
>>>>>>>>>>>>>> running kdestroy, with the below command and was able to get to 
>>>>>>>>>>>>>> my hive
>>>>>>>>>>>>>> tables.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode
>>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but
>>>>>>>>>>>>>>> when I migrated from 0.5.6 to this version, the spark 
>>>>>>>>>>>>>>> interpreter is not
>>>>>>>>>>>>>>> showing my tables & databases, may be its running in an 
>>>>>>>>>>>>>>> isolated mode...
>>>>>>>>>>>>>>> I'm just getting empty list, so I attempted to do kerberos 
>>>>>>>>>>>>>>> authentication
>>>>>>>>>>>>>>> to workaround that issue, and bumped into this road block.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file and
>>>>>>>>>>>>>>> its working fine.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *Kerberos test:*
>>>>>>>>>>>>>>> $ kdestroy
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19
>>>>>>>>>>>>>>>  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master
>>>>>>>>>>>>>>> yarn --num-executors 2 --executor-memory 2g --queue 
>>>>>>>>>>>>>>> <QUEUE_NAME>"
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>>>>>>       "name": "spark",
>>>>>>>>>>>>>>>       "group": "spark",
>>>>>>>>>>>>>>>       "properties": {
>>>>>>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>>>>>>         "args": "",
>>>>>>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>>>>>>> "spark-packages,http://dl.bintray.com/spark-packages/maven,f
>>>>>>>>>>>>>>> alse;"
>>>>>>>>>>>>>>>       },
>>>>>>>>>>>>>>>       "status": "READY",
>>>>>>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "spark",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>> rkInterpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "sql",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "dep",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Dep
>>>>>>>>>>>>>>> Interpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>>>>>>> parkInterpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>>       ],
>>>>>>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>>>>>>       "option": {
>>>>>>>>>>>>>>>         "remote": true,
>>>>>>>>>>>>>>>         "port": -1,
>>>>>>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>>>>>>         "users": []
>>>>>>>>>>>>>>>       }
>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <
>>>>>>>>>>>>>>> zjf...@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you
>>>>>>>>>>>>>>>> mind to share your configuration ? Or you can first try to run 
>>>>>>>>>>>>>>>> spark-shell
>>>>>>>>>>>>>>>> using spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added 
>>>>>>>>>>>>>>>>> "spark.yarn.keytab" &
>>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  
>>>>>>>>>>>>>>>>> But I keep
>>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the 
>>>>>>>>>>>>>>>>> notebook
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>>> kInterpreter335845091
>>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>> setRules(KerberosName.java:386)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati
>>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.initialize(U
>>>>>>>>>>>>>>>>> serGroupInformation.java:227)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia
>>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica
>>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn
>>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro
>>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre
>>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ
>>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>> Job.run(Job.java:176)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>
>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Best Regards
>>>>>>>>>>
>>>>>>>>>> Jeff Zhang
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best Regards
>>>>>>>>
>>>>>>>> Jeff Zhang
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best Regards
>>>>>>
>>>>>> Jeff Zhang
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> 이종열, Jongyoul Lee, 李宗烈
>>>> http://madeng.net
>>>>
>>>
>>>
>>
>>
>> --
>> 이종열, Jongyoul Lee, 李宗烈
>> http://madeng.net
>>
>
>

Reply via email to