Re: What's column family name for columns of table created by phoniex create table statement without a specific cf name?

2015-06-23 Thread James Taylor
To add to what Gabriel said, you can also specify your own default column family name with the DEFAULT_COLUMN_FAMILY property when you create your table: https://phoenix.apache.org/language/index.html#create_table On Tue, Jun 23, 2015 at 11:07 AM, Gabriel Reid wrote: > The default column family n

Re: error: no valid quorum servers found

2015-06-23 Thread James Taylor
You can specify the zookeeper quorum in the connection string as described here: https://phoenix.apache.org/#SQL_Support. All of the hosts are expected to use the same port (which may be specified as well). On Tue, Jun 23, 2015 at 1:05 PM, Alex Kamil wrote: > it's running in standalone mode with

Re: count distinct

2015-06-23 Thread James Taylor
Michael, You're correct, count distinct doesn't support multiple arguments currently (I filed PHOENIX-2062 for this). Another workaround is to combine a.col1 and b.col2 into an expression, for example concatenating them. If order matters, you could do this: select count(distinct col1 || col2) ...

Re: count distinct

2015-06-23 Thread Michael McAllister
> Your second query should work with Phoenix 4.3 or later. > Thanks, unfortunately at the moment I’m stuck with Phoenix 4.2. > I will investigate the problem with the first one and get back to you. > Appreciate this. Michael McAllister Staff Data Warehouse Engineer | Decision Systems mmcallis..

SocketTimeoutException on Update Statistics

2015-06-23 Thread Riesland, Zack
Thanks Gabriel, Can someone please give me detailed instructions for increasing the timeout? I tried running Update Statistics and it failed with the exception below. I am running the query from a region server node by CD’ing into /user/hdp/2.2.0.0-2041/phoenix/bin and calling ./sqlline.py I cr

Re: count distinct

2015-06-23 Thread Maryann Xue
Sorry, I missed the first line. Your second query should work with Phoenix 4.3 or later. I will investigate the problem with the first one and get back to you. Thanks, Maryann On Tuesday, June 23, 2015, Michael McAllister wrote: > Hi > > (This questions relates to Phoenix 4.2 on HDP 2.2) >

Re: count distinct

2015-06-23 Thread Maryann Xue
Which version of Phoenix are you using? On Tuesday, June 23, 2015, Michael McAllister wrote: > Hi > > (This questions relates to Phoenix 4.2 on HDP 2.2) > > I have a situation where I want to count the distinct combination of a > couple of columns. > > When I try the following:- > > select

Re: CDH 5.4 and Phoenix

2015-06-23 Thread yanlin wang
I think cloudera has a phoenix parcel available to download and use. You can google “cloudera phoenix support” and you should find a instruction about how to connect to a cloudera phoenix repo and install parcel etc. Yanlin > On Jun 23, 2015, at 3:36 PM, Kevin Verhoeven > wrote: > > I f

Re: Can't understand why phoenix saves but not selects

2015-06-23 Thread James Taylor
Oops, forgot the commit: Connection conn = getJdbcFacade().createConnection(); int result = conn.prepareStatement("upsert into unique_site_visitor (visitorId, siteId, visitTs) values ('xxxyyyzzz', 1, 2)").executeUpdate(); conn.commit(); LOG.debug("executeUpdate result: {}", result); //

Re: Can't understand why phoenix saves but not selects

2015-06-23 Thread James Taylor
Make sure you run the commit on the same connection from which you do the upsert. Looks like you're opening a new connection with each statement. Instead, open it once in the beginning and include the commit like Samarth mentioned: Connection conn = getJdbcFacade().createConnection(); int result =

RE: CDH 5.4 and Phoenix

2015-06-23 Thread Kevin Verhoeven
I found the Cloudera Labs Phoenix GitHub page here: https://github.com/cloudera-labs/phoenix, if it helps. And I believe that a parcel is simply a gzipped tar file, you should be able to crack it open. You might find jars within the parcel (http://archive.cloudera.com/cloudera-labs/phoenix/parc

Re: How To Count Rows In Large Phoenix Table?

2015-06-23 Thread anil gupta
70 minutes sounds too high. You must be having very less number of MapReduce slots or very less number of regions in your table. It should not take 70 minutes for that job. I run that job on a 5TB table in around 2-5 min(table has around 1200 regions). ~Anil On Tue, Jun 23, 2015 at 1:54 PM, Vladi

Re: How To Count Rows In Large Phoenix Table?

2015-06-23 Thread Vladimir Rodionov
A couple recommendations: 1. Increase # of regions. Your servers can handle in parallel much more than 4. You will increase aggregation performance as well. 2. Setting phoenix.query.timeoutMs is not sufficient. You have to increase HBase RPC timeout as well - hbase.rpc.timeout. 3. Upgrading to HBa

Re: Can't understand why phoenix saves but not selects

2015-06-23 Thread Serega Sheypak
Hi, thanks for the reply. I've tried to invoke commit explicitly, no luck. executeUpdate still returns 1 (I assume one record has been saved). Select returns empty resultset. 2015-06-23 18:11 GMT+02:00 Samarth Jain : > Hi Serega, > > Do you know if auto-commit is on for the connection returned by

Re: CDH 5.4 and Phoenix

2015-06-23 Thread Serega Sheypak
I read that labs article. I would like to get phoenix and cdh compatible jars, since I want to run test in my maven project. Are these cdh-phoenix jars officially published somewhere? 2015-06-23 18:15 GMT+02:00 Leon Prouger : > I did this installation > You should install Phoenix parcel from the

Re: error: no valid quorum servers found

2015-06-23 Thread Alex Kamil
it's running in standalone mode with hbase managing zk, I can connect to hbase , I can also connect with phoenix jdbc client (single tenant connection). When I try using multitenant connection I'm able to connect with the first tenant and write to hbase via phoenix jdbc connection, the second attem

Re: error: no valid quorum servers found

2015-06-23 Thread Nick Dimiduk
If HBase is running in standalone mode, it's running an embedded ZK and there's no need for zoo.cfg. You can connect to hbase? If so, the master status UI or the hbase shell can tell you the ZK connection details. On Tue, Jun 23, 2015 at 12:01 PM, Alex Kamil wrote: > > getting below error when u

error: no valid quorum servers found

2015-06-23 Thread Alex Kamil
getting below error when using multitenant phoenix connection is there a way to programmatically specify zoo.cfg and hbase-site.xml properties (like *server.0=myhostname:2888:3888) *when initializing phoenix connection? adding hbase-site.xml and zoo.cfg to classpath doesnt help in this case, i

count distinct

2015-06-23 Thread Michael McAllister
Hi (This questions relates to Phoenix 4.2 on HDP 2.2) I have a situation where I want to count the distinct combination of a couple of columns. When I try the following:- select count(distinct a.col1, b.col2) from table tab1 a inner join tab2 b on b.joincol = a.joincol where a.col3 = ‘some con

Re: How to upsert data into dynamic columns in phoniex.

2015-06-23 Thread Thomas D'Silva
Since lastGCTime is a dynamic column you need to specify the dynamic column explicitly along with the table name. Select lastGCTime FROM EventLog(lastGCTime TIME) Select * will return only the regular columns (and not dynamic columns). On Tue, Jun 23, 2015 at 12:09 AM, guxiaobo1982 wrote: > Hi

Re: What's column family name for columns of table created by phoniex create table statement without a specific cf name?

2015-06-23 Thread Gabriel Reid
The default column family name is "0". This is the string containing the character representation of zero (or in other words, a single byte with value 48). And yes, it's possible to read Phoenix tables using the HBase API (although it's of course a lot easier if you go via Phoenix). - Gabriel O

Re: CsvBulkLoad output questions

2015-06-23 Thread Gabriel Reid
Ok, that log info would be great in diagnosing what's going on here. You can modify your original table by requesting that HBase split the regions (this is possible via the hbase shell, or via the web interface). I don't think that there's a way to do this from within Phoenix itself. Another opti

RE: CsvBulkLoad output questions

2015-06-23 Thread Riesland, Zack
Thanks Gabriel, That’s all very helpful. I’m not at all sure that the timeouts are related to compactions. This is just my best guess at reading the logs. When our cluster is free for an extended period of time, I’ll try to do a large ingest and capture some details that will hopefully help cl

Re: CsvBulkLoad output questions

2015-06-23 Thread Gabriel Reid
Thanks for all the additional details. The short answer (to both of your questions from your most-recent mail) is that there shouldn't be any data loss, and that the failed reducer will automatically be re-run by MapReduce. The full job is only successful (as it was in this case) when all mappers

Crash with no logs

2015-06-23 Thread Jean-Marc Spaggiari
Hi, I try to install Phoenix 4.3.0 on HBase 1.0 (CDH 5.4.2). I think installation is going well, but when starting the shell for the first time, it does many operations on HBase side and then just exit. By exit I mean it exits the RegionServer. Process ends without anything on the logs (See below)

Re: Data Model Suggestion

2015-06-23 Thread Vladimir Rodionov
> > > > ADID | COUNTRY | HOUR ID | CF.IMP | > > > > > - > > > > 1| US | 201506230

Re: Data Model Suggestion

2015-06-23 Thread James Taylor
-- > > 1| US | 2015062301 | 3000| > > 1| US | 2015062302 | 3421| > > 1| UK | 2015062302 | 1212| > - > > Is it a good schema design or shall I create alt

Re: CDH 5.4 and Phoenix

2015-06-23 Thread Leon Prouger
I did this installation You should install Phoenix parcel from the Cloudera Manager like any other parcel. And it's for Phoenix 4.3.1, 1.0 it's probably the version of the cloudera's parcel. , and it's for Phoenix 4.3.1. On Tue, Jun 23, 2015 at 2:48 PM Serega Sheypak wrote: > Hi, no one:) ? > >

Re: Can't understand why phoenix saves but not selects

2015-06-23 Thread Samarth Jain
Hi Serega, Do you know if auto-commit is on for the connection returned by getJdbcFacade().createConnection(). If not, you need to call connection.commit() after executeUpdate() -Samarth On Tuesday, June 23, 2015, Serega Sheypak wrote: > Hi, I'm testing dummy code: > > int result = getJdbcFaca

RE: Data Model Suggestion

2015-06-23 Thread Puneet Kumar Ojha
ary Key - ADID | COUNTRY | DAY ID ADID | COUNTRY | DAY ID | CF.IMP01 | CF.IMP02 | ---- 1| US | 20150623 | 300

CsvBulkLoad output questions

2015-06-23 Thread Riesland, Zack
This question is mostly a followup based on my earlier mail (below). I’m re-consuming this data, one (5GB) csv file at a time. I see that in consuming this file, there was one failed reduce task. In the output, I see a stack trace that I’m guessing is related. So, 2 questions: 1 – does this me

Re: webcast on Phoenix this Thu @ 10am

2015-06-23 Thread James Taylor
Thursday, June 25th @ 10am PST On Tuesday, June 23, 2015, Ns G wrote: > Hi James, > > Can you specify the time zone please? > > Thanks > Satya > On 23-Jun-2015 9:32 am, "James Taylor" > wrote: > >> If you're interested in learning more about Phoenix, tune in this >> Thursday @ 10am where I'll b

Re: Data Model Suggestion

2015-06-23 Thread Pariksheet Barapatre
UK | 2015062302 | 1212| > > - > > Is it a good schema design or shall I create alternate schema as below > Primary Key - *ADID | COUNTRY | DAY ID* > > ---

RE: Data Model Suggestion

2015-06-23 Thread Puneet Kumar Ojha
RY | DAY ID ADID | COUNTRY | DAY ID | CF.IMP01 | CF.IMP02 | 1| US | 20150623 | 3000|

Re: Data Model Suggestion

2015-06-23 Thread Pariksheet Barapatre
-- >> >> Is it a good schema design or shall I create alternate schema as below >> Primary Key - *ADID | COUNTRY | DAY ID* >> >> >> *ADID | COUNTRY | DAY ID*

Re: Data Model Suggestion

2015-06-23 Thread Skanda
-- > *ADID | COUNTRY | DAY ID* | CF.IMP*01* | CF.IMP*02* > | > > ---- > 1| US

RE: How To Count Rows In Large Phoenix Table?

2015-06-23 Thread Riesland, Zack
Anil: Thanks for the tip about mapreduce.RowCounter. That takes about 70 minutes, but it works! Unfortunately, I only got about 60% of the rows I’m expecting. Gabriel (and anyone interested): Thanks for your response! A few details to give context to my question: Our cluster has 6 region serv

Data Model Suggestion

2015-06-23 Thread Pariksheet Barapatre
* *ADID | COUNTRY | DAY ID* | CF.IMP*01* | CF.IMP*02* | 1| US | 20150623 | 3000| 3421 | 1| UK

Can't understand why phoenix saves but not selects

2015-06-23 Thread Serega Sheypak
Hi, I'm testing dummy code: int result = getJdbcFacade().createConnection().prepareStatement("upsert into unique_site_visitor (visitorId, siteId, visitTs) values ('xxxyyyzzz', 1, 2)").executeUpdate(); LOG.debug("executeUpdate result: {}", result); //executeUpdate result: 1

Re: CDH 5.4 and Phoenix

2015-06-23 Thread Serega Sheypak
Hi, no one:) ? 2015-06-21 22:41 GMT+02:00 Serega Sheypak : > Hi!, did anyone try to integrate Phoenix 1.0 with CDH 5.4.x? > I see weird installation path here: > > http://www.cloudera.com/content/cloudera/en/developers/home/cloudera-labs/apache-phoenix/install-apache-phoenix-cloudera-labs.pdf > I

Re: webcast on Phoenix this Thu @ 10am

2015-06-23 Thread Ns G
Hi James, Can you specify the time zone please? Thanks Satya On 23-Jun-2015 9:32 am, "James Taylor" wrote: > If you're interested in learning more about Phoenix, tune in this > Thursday @ 10am where I'll be talking about Phoenix in a free Webcast > hosted by O'Reilly: http://www.oreilly.com/pub

What's column family name for columns of table created by phoniex create table statement without a specific cf name?

2015-06-23 Thread dataware
Hi, Can we use the HBase APIs to access the raw HBase tables? Thanks.

Re: StackOverflowError

2015-06-23 Thread Bahubali Jain
Hi Maryann, EVENTS Table has id,article, and more columns. Id is the primay key MAPPING Table has id,article,category columns. Id is the primay key There is index on article column of both the tables. Below is the query. select count(MAPPING.article) as cnt,MAPPING.category from EVENTS join MAPP

Re: How to upsert data into dynamic columns in phoniex.

2015-06-23 Thread guxiaobo1982
Hi Thomas, I tried your suggestion, the result is the upsert statement with new column defined works, but subsequent queries can't find the new column, which is select * from eventlog will not show the new defined column lastGCTime and select lastGCTime from eventlog fails with undefined

Re: Phoenix case-insensitivity interaction with Excel and ODBC

2015-06-23 Thread anil gupta
Hi Alex, I dont have the authority to speak on behalf of Phoenix Committers. However, in worst case, if your contribution is not accepted by Phoenix, then also you can locally patch Phoenix and use it the way you want. Many people use open source softwares like that. HTH, Anil On Fri, Jun 19, 20