To add to what Gabriel said, you can also specify your own default
column family name with the DEFAULT_COLUMN_FAMILY property when you
create your table:
https://phoenix.apache.org/language/index.html#create_table
On Tue, Jun 23, 2015 at 11:07 AM, Gabriel Reid wrote:
> The default column family n
You can specify the zookeeper quorum in the connection string as
described here: https://phoenix.apache.org/#SQL_Support. All of the
hosts are expected to use the same port (which may be specified as
well).
On Tue, Jun 23, 2015 at 1:05 PM, Alex Kamil wrote:
> it's running in standalone mode with
Michael,
You're correct, count distinct doesn't support multiple arguments currently
(I filed PHOENIX-2062 for this). Another workaround is to combine a.col1
and b.col2 into an expression, for example concatenating them. If order
matters, you could do this:
select count(distinct col1 || col2) ...
>
Your second query should work with Phoenix 4.3 or later.
>
Thanks, unfortunately at the moment I’m stuck with Phoenix 4.2.
>
I will investigate the problem with the first one and get back to you.
>
Appreciate this.
Michael McAllister
Staff Data Warehouse Engineer | Decision Systems
mmcallis..
Thanks Gabriel,
Can someone please give me detailed instructions for increasing the timeout?
I tried running Update Statistics and it failed with the exception below.
I am running the query from a region server node by CD’ing into
/user/hdp/2.2.0.0-2041/phoenix/bin and calling ./sqlline.py
I cr
Sorry, I missed the first line. Your second query should work with Phoenix
4.3 or later.
I will investigate the problem with the first one and get back to you.
Thanks,
Maryann
On Tuesday, June 23, 2015, Michael McAllister
wrote:
> Hi
>
> (This questions relates to Phoenix 4.2 on HDP 2.2)
>
Which version of Phoenix are you using?
On Tuesday, June 23, 2015, Michael McAllister
wrote:
> Hi
>
> (This questions relates to Phoenix 4.2 on HDP 2.2)
>
> I have a situation where I want to count the distinct combination of a
> couple of columns.
>
> When I try the following:-
>
> select
I think cloudera has a phoenix parcel available to download and use. You can
google “cloudera phoenix support” and you should find a instruction about how
to connect to a cloudera phoenix repo and install parcel etc.
Yanlin
> On Jun 23, 2015, at 3:36 PM, Kevin Verhoeven
> wrote:
>
> I f
Oops, forgot the commit:
Connection conn = getJdbcFacade().createConnection();
int result = conn.prepareStatement("upsert into unique_site_visitor
(visitorId, siteId, visitTs) values ('xxxyyyzzz', 1,
2)").executeUpdate();
conn.commit();
LOG.debug("executeUpdate result: {}", result);
//
Make sure you run the commit on the same connection from which you do
the upsert. Looks like you're opening a new connection with each
statement. Instead, open it once in the beginning and include the
commit like Samarth mentioned:
Connection conn = getJdbcFacade().createConnection();
int result =
I found the Cloudera Labs Phoenix GitHub page here:
https://github.com/cloudera-labs/phoenix, if it helps. And I believe that a
parcel is simply a gzipped tar file, you should be able to crack it open. You
might find jars within the parcel
(http://archive.cloudera.com/cloudera-labs/phoenix/parc
70 minutes sounds too high. You must be having very less number of
MapReduce slots or very less number of regions in your table. It should not
take 70 minutes for that job. I run that job on a 5TB table in around 2-5
min(table has around 1200 regions).
~Anil
On Tue, Jun 23, 2015 at 1:54 PM, Vladi
A couple recommendations:
1. Increase # of regions. Your servers can handle in parallel much more
than 4. You will increase aggregation performance as well.
2. Setting phoenix.query.timeoutMs is not sufficient. You have to increase
HBase RPC timeout as well - hbase.rpc.timeout.
3. Upgrading to HBa
Hi, thanks for the reply.
I've tried to invoke commit explicitly, no luck. executeUpdate still
returns 1 (I assume one record has been saved). Select returns empty
resultset.
2015-06-23 18:11 GMT+02:00 Samarth Jain :
> Hi Serega,
>
> Do you know if auto-commit is on for the connection returned by
I read that labs article. I would like to get phoenix and cdh compatible
jars, since I want to run test in my maven project.
Are these cdh-phoenix jars officially published somewhere?
2015-06-23 18:15 GMT+02:00 Leon Prouger :
> I did this installation
> You should install Phoenix parcel from the
it's running in standalone mode with hbase managing zk, I can connect to
hbase , I can also connect with phoenix jdbc client (single tenant
connection).
When I try using multitenant connection I'm able to connect with the first
tenant and write to hbase via phoenix jdbc connection,
the second attem
If HBase is running in standalone mode, it's running an embedded ZK and
there's no need for zoo.cfg. You can connect to hbase? If so, the master
status UI or the hbase shell can tell you the ZK connection details.
On Tue, Jun 23, 2015 at 12:01 PM, Alex Kamil wrote:
>
> getting below error when u
getting below error when using multitenant phoenix connection
is there a way to programmatically specify zoo.cfg and hbase-site.xml
properties (like *server.0=myhostname:2888:3888) *when initializing phoenix
connection?
adding hbase-site.xml and zoo.cfg to classpath doesnt help in this case, i
Hi
(This questions relates to Phoenix 4.2 on HDP 2.2)
I have a situation where I want to count the distinct combination of a couple
of columns.
When I try the following:-
select count(distinct a.col1, b.col2)
from table tab1 a
inner join tab2 b on b.joincol = a.joincol
where a.col3 = ‘some con
Since lastGCTime is a dynamic column you need to specify the dynamic
column explicitly along with the table name.
Select lastGCTime FROM EventLog(lastGCTime TIME)
Select * will return only the regular columns (and not dynamic columns).
On Tue, Jun 23, 2015 at 12:09 AM, guxiaobo1982 wrote:
> Hi
The default column family name is "0". This is the string containing the
character representation of zero (or in other words, a single byte with
value 48).
And yes, it's possible to read Phoenix tables using the HBase API (although
it's of course a lot easier if you go via Phoenix).
- Gabriel
O
Ok, that log info would be great in diagnosing what's going on here.
You can modify your original table by requesting that HBase split the
regions (this is possible via the hbase shell, or via the web interface). I
don't think that there's a way to do this from within Phoenix itself.
Another opti
Thanks Gabriel,
That’s all very helpful.
I’m not at all sure that the timeouts are related to compactions. This is just
my best guess at reading the logs.
When our cluster is free for an extended period of time, I’ll try to do a large
ingest and capture some details that will hopefully help cl
Thanks for all the additional details.
The short answer (to both of your questions from your most-recent mail) is
that there shouldn't be any data loss, and that the failed reducer will
automatically be re-run by MapReduce. The full job is only successful (as
it was in this case) when all mappers
Hi,
I try to install Phoenix 4.3.0 on HBase 1.0 (CDH 5.4.2). I think
installation is going well, but when starting the shell for the first time,
it does many operations on HBase side and then just exit. By exit I mean it
exits the RegionServer. Process ends without anything on the logs (See
below)
> >
> > ADID | COUNTRY | HOUR ID | CF.IMP |
> >
> >
> -
> >
> > 1| US | 201506230
--
>
> 1| US | 2015062301 | 3000|
>
> 1| US | 2015062302 | 3421|
>
> 1| UK | 2015062302 | 1212|
> -
>
> Is it a good schema design or shall I create alt
I did this installation
You should install Phoenix parcel from the Cloudera Manager like any other
parcel. And it's for Phoenix 4.3.1, 1.0 it's probably the version of the
cloudera's parcel.
, and it's for Phoenix 4.3.1.
On Tue, Jun 23, 2015 at 2:48 PM Serega Sheypak
wrote:
> Hi, no one:) ?
>
>
Hi Serega,
Do you know if auto-commit is on for the connection returned by
getJdbcFacade().createConnection().
If not, you need to call connection.commit() after executeUpdate()
-Samarth
On Tuesday, June 23, 2015, Serega Sheypak wrote:
> Hi, I'm testing dummy code:
>
> int result = getJdbcFaca
ary Key - ADID | COUNTRY | DAY ID
ADID | COUNTRY | DAY ID | CF.IMP01 | CF.IMP02 |
----
1| US | 20150623 | 300
This question is mostly a followup based on my earlier mail (below).
I’m re-consuming this data, one (5GB) csv file at a time.
I see that in consuming this file, there was one failed reduce task. In the
output, I see a stack trace that I’m guessing is related.
So, 2 questions:
1 – does this me
Thursday, June 25th @ 10am PST
On Tuesday, June 23, 2015, Ns G wrote:
> Hi James,
>
> Can you specify the time zone please?
>
> Thanks
> Satya
> On 23-Jun-2015 9:32 am, "James Taylor" > wrote:
>
>> If you're interested in learning more about Phoenix, tune in this
>> Thursday @ 10am where I'll b
UK | 2015062302 | 1212|
>
> -
>
> Is it a good schema design or shall I create alternate schema as below
> Primary Key - *ADID | COUNTRY | DAY ID*
>
> ---
RY | DAY ID
ADID | COUNTRY | DAY ID | CF.IMP01 | CF.IMP02 |
1| US | 20150623 | 3000|
--
>>
>> Is it a good schema design or shall I create alternate schema as below
>> Primary Key - *ADID | COUNTRY | DAY ID*
>>
>>
>> *ADID | COUNTRY | DAY ID*
--
> *ADID | COUNTRY | DAY ID* | CF.IMP*01* | CF.IMP*02*
> |
>
> ----
> 1| US
Anil: Thanks for the tip about mapreduce.RowCounter. That takes about 70
minutes, but it works!
Unfortunately, I only got about 60% of the rows I’m expecting.
Gabriel (and anyone interested):
Thanks for your response!
A few details to give context to my question:
Our cluster has 6 region serv
*
*ADID | COUNTRY | DAY ID* | CF.IMP*01* | CF.IMP*02* |
1| US | 20150623 | 3000|
3421 |
1| UK
Hi, I'm testing dummy code:
int result = getJdbcFacade().createConnection().prepareStatement("upsert
into unique_site_visitor (visitorId, siteId, visitTs) values ('xxxyyyzzz',
1, 2)").executeUpdate();
LOG.debug("executeUpdate result: {}", result); //executeUpdate
result: 1
Hi, no one:) ?
2015-06-21 22:41 GMT+02:00 Serega Sheypak :
> Hi!, did anyone try to integrate Phoenix 1.0 with CDH 5.4.x?
> I see weird installation path here:
>
> http://www.cloudera.com/content/cloudera/en/developers/home/cloudera-labs/apache-phoenix/install-apache-phoenix-cloudera-labs.pdf
> I
Hi James,
Can you specify the time zone please?
Thanks
Satya
On 23-Jun-2015 9:32 am, "James Taylor" wrote:
> If you're interested in learning more about Phoenix, tune in this
> Thursday @ 10am where I'll be talking about Phoenix in a free Webcast
> hosted by O'Reilly: http://www.oreilly.com/pub
Hi,
Can we use the HBase APIs to access the raw HBase tables?
Thanks.
Hi Maryann,
EVENTS Table has id,article, and more columns. Id is the primay key
MAPPING Table has id,article,category columns. Id is the primay key
There is index on article column of both the tables.
Below is the query.
select count(MAPPING.article) as cnt,MAPPING.category from EVENTS
join
MAPP
Hi Thomas,
I tried your suggestion, the result is
the upsert statement with new column defined works,
but subsequent queries can't find the new column, which is
select * from eventlog will not show the new defined column lastGCTime
and
select lastGCTime from eventlog fails with undefined
Hi Alex,
I dont have the authority to speak on behalf of Phoenix Committers.
However, in worst case, if your contribution is not accepted by Phoenix,
then also you can locally patch Phoenix and use it the way you want. Many
people use open source softwares like that.
HTH,
Anil
On Fri, Jun 19, 20
45 matches
Mail list logo