>
> Read our latest technical blog posts here
> <https://www.instaclustr.com/blog/>.
>
> This email has been sent on behalf of Instaclustr Pty. Limited (Australia)
> and Instaclustr Inc (USA).
>
> This email and any attachments may contain confidential and legally
> privileged information. If you are not the intended recipient, do not copy
> or disclose its content, but please reply to this email immediately and
> highlight the error to the sender and then immediately delete the message.
>
--
Siddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full table
scan)
t;>> keysace.table WHERE token(id) > token(test:ODP0144-0883E-022R-002/047-052)
>>>> LIMIT 100 (see tombstone_warn_threshold)
>>>>
>>>> Read 100 live rows and 1435 tombstone cells for query SELECT * FROM
>>>> keysace.table WHERE token(id) > token(test:2565-AMK-2) LIMIT 100 (see
>>>> tombstone_warn_threshold)
>>>>
>>>> Read 96 live rows and 1385 tombstone cells for query SELECT * FROM
>>>> keysace.table WHERE token(id) > token(test:-2220-UV033/04) LIMIT 100 (see
>>>> tombstone_warn_threshold).
>>>>
>>>>
>>>>
>>>>
>>>> Can you please help me to get the total count of the table.
>>>>
>>>> --
>>>> Selvam Raman
>>>> "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"
>>>>
>>>>
>>>
>>>
>>> --
>>> Selvam Raman
>>> "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"
>>>
>>>
>>>
>
--
Siddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full table
scan)
for the input! Are you using the DataStax connector as well? Does
> it support querying against any column well (not just clustering columns)?
> I’m wondering how it could build the index around them “on-the-fly”.
>
>
>
> Regards,
>
> John
>
>
>
> *From:* sidd
closure, copying,
> distribution or use of the information contained in or attached to this
> message is strictly prohibited. If you have received this email
> transmission in error, please notify the sender by replying to this email
> and then delete it from your system.
>
>
--
Siddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full table
scan)
ble for and avoiding the replicated data.
>
> Many thanks for any help and guidance,
>
> Frank Hughes
>
--
Siddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full table
scan)
ite.
>
>
>
>
>
> *Thanks & Regards,*
> *Abhishek Kumar Maheshwari*
> *+91- 805591 (Mobile)*
>
> Times Internet Ltd. | A Times of India Group Company
>
> FC - 6, Sector 16A, Film City, Noida, U.P. 201301 | INDIA
>
> *P** Please do not print this email unless it is absolutely necessary.
> Spread environmental awareness.*
>
>
>
> Education gets Exciting with IIM Kozhikode Executive Post Graduate
> Programme in Management - 2 years (AMBA accredited with full benefits of
> IIMK Alumni status). Brought to you by IIMK in association with TSW, an
> Executive Education initiative from The Times of India Group. Learn more:
> www.timestsw.com
>
>
>
>
>
>
>
>
>
> --
>
>
>
> Benjamin Roth
>
> Prokurist
>
>
>
> Jaumo GmbH · www.jaumo.com
>
> Wehrstraße 46 · 73035 Göppingen · Germany
>
> Phone +49 7161 304880-6 · Fax +49 7161 304880-1
>
> AG Ulm · HRB 731058 · Managing Director: Jens Kammerer
>
>
>
>
>
>
>
--
Siddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full table
scan)
e through cqlsh, but I don't see a reasonable way to differentiate all
> the mutations anyway so I think this is a bust.
>
> I haven't spent a lot of time in this code, happy to be corrected if I'm
> wrong.
>
> Jon
>
>
> On Fri, Nov 11, 2016 at 10:14 AM sid
Hi Sathish,
You could look into, Change Data Capture (CDC) (
https://issues.apache.org/jira/browse/CASSANDRA-8844 .
It might help you for some of your requirements.
Regards
Siddharth Verma
On Fri, Nov 11, 2016 at 11:34 PM, Jonathan Haddad wrote:
> cqlsh uses the Python driver, I don't
Hi Sathish,
You could look into, Change Data Capture (CDC) ( https://issues.apache.org/
jira/browse/CASSANDRA-8844 .
It might help you for some of your requirements.
Regards
Siddharth Verma
On Fri, Nov 11, 2016 at 11:34 PM, Jonathan Haddad wrote:
> cqlsh uses the Python driver, I don't
Hi Sathish,
You could look into, Change Data Capture (CDC) (
https://issues.apache.org/jira/browse/CASSANDRA-8844 .
It might help you for some of your requirements.
Regards
Siddharth Verma
On Fri, Nov 11, 2016 at 11:34 PM, Jonathan Haddad wrote:
> cqlsh uses the Python driver, I don't
atures usually take 4-6 versions to solidify.
>
>
>
> On Mon, Oct 17, 2016 at 3:19 AM, siddharth verma <
> sidd.verma29.l...@gmail.com> wrote:
>
>> Hi,
>> We have a base table with ~300 million entries.
>> And in a recent sanity activity, I saw approx ~33k entire
Sorry Krishna, didn't get were you were trying to say.
Regards
SIddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full
table scan)
On Sat, Oct 15, 2016 at 11:50 PM, Krishna Chandra Prajapati <
prajapat...@gmail.com> wrote:
> Hi which side is th
icate inconsistency between base
table and materialized view. Any help would be appreciated.
C* 3.6
Regards
SIddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full table
scan)
omatic paging in either of the scenarios.
Thanks
Siddharth Verma
(Visit https://github.com/siddv29/cfs for a high speed cassandra full table
scan)
hecking the next.length(), you can decide whether it is an
insert/update(length > 0) OR delete(length == 0)
I would urge you to try the snippet once on you own, to see what kind of
data it produces in *next*. You could dump the output of next in a column
for audit table, to see that outp
Here, if the assumption stated above holds true, you could avoid a
cassandra select for that key.
Thanks
Siddharth Verma
On Wed, Oct 5, 2016 at 12:20 AM, Kant Kodali wrote:
> Thanks a lot, This helps me to make a decision on not to write one for the
> performance reasons you pointed out
Hi Jon,
It wan't allowed.
Moreover, if someone who isn't familiar with spark, and might be new to map
filter reduce etc. operations, could also use the utility for some simple
operations assuming a sequential scan of the cassandra table.
Regards
Siddharth Verma
On Tue, Oct 4, 2016
paging on result set.
Regards
Siddharth Verma
*Software Engineer I - CaMS*
*M*: +91 9013689856, *T*: 011 22791596 *EXT*: 14697
CA2125, 2nd Floor, ASF Centre-A, Jwala Mill Road,
Udyog Vihar Phase - IV, Gurgaon-122016, INDIA
Download Our App
[image: A]
<https://play.google.com/store/ap
comments. :-)
Regards,
Siddharth Verma
s and Regards,
Siddharth Verma
(previous email id on this mailing list : verma.siddha...@snapdeal.com)
erations where it can know in advance.
>
> With RF == N like your setup, every piece of data is owned by every node,
> so as long as your LBP is distributive, and outside of performance testing,
> I can't see why you'd be needing to target specific nodes for anything.
>
&
consistency LOCAL_ONE, but i guess, i doesn't guarantee that
data will be fetched from it.
Thanks
Siddharth Verma
Debugged the issue a little.
AbstractFuture.get() throws java.util..concurrent.ExecutionException
in, Uninterruptables.getUninterruptibly interrupted gets set to true, which
does Thread.interrupt()
thus in DefaultResultSetFuture
(ResultSet)Uninterruptibles.getUninterruptibly(this) throws exception.
Correction : java driver version : 3.1.0
Update :
I page on the ResultSet explicitly in the program(
*boundStatement.setPagingState(PagingState.fromString(currentPageInfo));*)
When i put a *limit 10* after select statement, there is no error, with the
token range set i provided it
However, without that, there is this error, on second toke
bably help people provide and answer if you let everyone some
> more details like:
> - cassandra version and driver version you are using
> - query that is being executed when the error occurs
> - schema of the table that is being queried
>
> Cheers
> Ben
>
> On Thu, 1 Sep
Hi,
Could someone help me out with the following exception in cassandra java
driver.
Why did it occur?
MyClass program is paging on the result set.
com.datastax.driver.core.exceptions.ServerError: An unexpected error
occurred server side on /10.0.230.25:9042: java.lang.AssertionError:
[DecoratedKe
Hi ,
Can we be sure that, token ranges in nodetool describering will be non
overlapping?
Thanks
Siddharth Verma
d be using this as a backup.
Siddharth Verma
e assumption ABSOLUTELY correct?
(Kindly suggest changes/errors, if any)
Any help would be great.
Thanks and Regards,
Siddharth Verma
,
Siddharth Verma
06888',
'4113277683763393618', '4133779736724357706', '436041502200834273',
'4409977876499706529', '4501676578785778929', '4542194500781241332',
'457304378229441003', '4625447288380261317', '4635627349070688401',
'465311781634413', '4684783897201143715', '4718159652990403843',
'4774133625331605257', '4830998188269931348', '4910055390341907857',
'4998500558397491684', '5023943523343001536', '505744292693224967',
'5162320571738032500', '5277396368371480274', '5346830336223723812',
'5360976982928138315', '5415193046016833300', '5420351854649967783',
'551728115923760971', '5611031614377902472', '564955142261319773',
'5741673630415265730', '5765467955391572803', '5777120745629762657',
'5911733864168126799', '5963370543879046708', '601095846312433394',
'6025232037324913724', '612464281122209', '6130180990267145790',
'614044759087911080', '615625082046151174', '6204536250349648494',
'6254832987275160383', '6299420497233402212', '6330219643037606275',
'6332166655931126455', '649843146990783793', '6546360696742228459',
'6667173768996479693', '6681703316740595756', '6700409845144118031',
'6768174919292984477', '6993240335038495953', '699592662848728421',
'7067614760768641586', '7464549774220819665', '7470203399706056435',
'7517126645125886173', '75179143059671689', '7532146377366822000',
'7668359789169421370', '7693365598138379140', '7693612576353655751',
'771781371808055068', '773878061649801356', '7949652196530716229',
'8101045330184519696', '8131531126072838629', '8140361357245000979',
'8302292937826438542', '8422369105285792829', '8449836158927214743',
'8499955449276612349', '850192426321138914', '8605431687486711156',
'8678800982435339567', '8687105151647096442', '8776316423500407040',
'8953364009441747611', '8979834615817990709', '8984590311736053933',
'9069513190093258312', '9080235839310378242', '9098260908744107307',
'9102600462020381711', '938833944614881458', '987385371860069729',
'990370821724370771'}
Thanks,
Siddharth Verma
trigger
example.
Regards,
Siddharth Verma
the values of partition key, clustering column, on which the
update,insert,delete is called,
this can be extracted by using the UnfilteredRowIterator of the Partition
object you receive in the trigger function.
Thanks and Regards,
Siddharth Verma
If anyone, who has done a POC for the same, could share his/her views.
Any help would be appreciated.
Thanks
Siddharth Verma
(overlapping/ non overlapping group of columns) affect
read performance?
For better performance, should we split the table into 4 sub tables, where
each table is used to serve respective read api?
Thanks and Regards,
Siddharth Verma
id is partition key,
f_name is clustering key
We weren't querying on lucene indexes.
lucene index is on id, and f_d_name (another column).
We were facing this issue on production in one column family, due to which
we had to downgrade to 3.0.3
No, all rows were not the same.
Querying only on the partition key gives 20 rows.
In the erroneous result, while querying on partition key and clustering
key, we got 16 of those 20 rows.
And for "*tombstone_threshold"* there isn't any entry at column family
level.
Thanks,
Siddharth Verma
Running nodetool compact fixed the issue.
Could someone help out as why it occurred.
d for that partition
key.
3. While upgrading, we dropped key_cache folder on some, not all.
What could be the causes and how to fix this issue?
We speculate that it might be due to cache.
Any help would be appreciated.
Thanks
Siddharth Verma
Hi Sam,
Sorry, I couldn't understand.
I am already using
UnfilteredRowIterator unfilteredRowIterator =partition.unfilteredIterator();
while(unfilteredRowIterator.hasNext()){
next.append(unfilteredRowIterator.next().toString()+"\001");
}
Is there another way to access it?
would be appreciated.
Thanks
Siddharth Verma
f22d2
date| 2016/05/25
keyspace_name | test
table_name | test
uuid| df274fc0-4362-42b1-a3bf-0030f8d2062f
primary_key| pk1:pk2
values| [[v1=v1 ts=1464184100769315], [v2=v2
ts=1464184100769315]]
How to audit ck1 and ck2 also?
Thanks,
Siddharth Verma
Hi Eduardo,
Thanks for your reply. If it is fixed in 3.0.5.1, we will shift to it.
One more question,
If instead of truncating table, we remove some rows, then
are the lucene documents and indexes for those rows deleted?
Hi, I missed out on some info
node 1,2,3 are in DC1
node 4,5,6 are in DC2
and RF is 3
so all data is on all nodes
@Carlos : There was only one query. And yes all nodes have same data for
col5 only
node 6 has
P1,100,A,val1,w1
P1,100,B,val2,w2
P1,200,C,val3,w_x
P1,200,D,val4,w4
node 1,2,3,4,5 have
or only once and then execute on all?
Thanks
Siddharth Verma
for writes is Local_Quorum.
Thanks
Siddharth Verma
@Joseph,
An incident we saw in production, and have a speculation as to how it might
have occured.
*A detailed description of use case*
*Incident*
We have a 2 DCs each with three nodes.
And our keyspace has RF 3 per DC. read_repair_chance is 0.0 for all the
tables.
After a while(we run periodic f
nnect to either of the nodes in the same DC.
Any help would be appreciated
Thanks
Siddharth Verma
nHost,9042
.build();
mainSession = mainCluster.connect();
Regards,
SIddharth Verma
On Thu, May 5, 2016 at 8:59 PM, Jeff Jirsa
wrote:
> This doesn’t actually guarantee the behavior you think it does. There’s no
> actual way to guarantee this behavior in Cassandra, as far
Hi,
We have a 3 node cluster in DC1, where replication factor of keyspace is 3.
How can i read data only from one particular node in java driver?
Thanks,
Siddharth Verma
Anyways, thanks for your reply.
On Thu, Apr 28, 2016 at 1:59 PM, Hannu Kröger wrote:
> Ok, then I don’t understand the problem.
>
> Hannu
>
> On 28 Apr 2016, at 11:19, Siddharth Verma
> wrote:
>
> Hi Hannu,
>
> Had the issue been caused due to read, the insert
Hi Hannu,
Had the issue been caused due to read, the insert, and delete statement
would have been erroneous.
"I saw the stdout from web-ui of spark, and the query along with true was
printed for both the queries.".
The statements were correct as seen on the UI.
Thanks,
Siddharth Verma
https://issues.apache.org/jira/browse/CASSANDRA-10010, but that only
> affected 2.x. Can you open a Jira ticket with your table schema, the
> problematic query, and the details you posted here?
>
> On Tue, Apr 19, 2016 at 10:25 AM, Siddharth Verma <
> verma.siddha...@snapdeal.com&g
Edit:
1. dc2 node has been removed.
nodetool status shows only active nodes.
2. Repair done on all nodes.
3. Cassandra restarted
Still it doesn't solve the problem.
On Thu, Apr 28, 2016 at 9:00 AM, Siddharth Verma <
verma.siddha...@snapdeal.com> wrote:
> Hi, If the info coul
could provide insight to the problem, it would be helpful.
Thanks
On Wed, Apr 27, 2016 at 11:11 PM, Siddharth Verma <
verma.siddha...@snapdeal.com> wrote:
> Hi,
> I dont know, if someone has faced this problem or not.
> I am running a job where some data is loaded from cassandr
sometimes the result is displayed and
sometime the changes don't take place.
I saw the stdout from web-ui of spark, and the query along with true was
printed for both the queries.
I can't understand, what could be the issue.
Any help would be appreciated.
Thanks,
Siddharth Verma
Hi,
We are using cassandra(dsc3.0.3) on production.
For some purpose, we were doing a full table scan (setPagingState and
getPagingState used on ResultSet in java program), and there has been some
discrepancy when we ran the same job multiple times.
Each time some new data was added to the output
In cassandra I have a table with the following schema.
CREATE TABLE my_keyspace.my_table1 (
col_1 text,
col_2 text,
col_3 text,
col_4 text,,
col_5 text,
col_6 text,
col_7 text,
PRIMARY KEY (col_1, col_2, col_3)
) WITH CLUSTERING ORDER BY (col_2 ASC, col_3 ASC);
For
tion(s) to filter the row set on the
server.
Thanks and Regards
Siddharth Verma
*Software Engineer*
CA2125, 2nd Floor, ASF Centre-A, Jwala Mill Road,
Udyog Vihar Phase - IV, Gurgaon-122016, INDIA
Download Our App
[image: A]
<https://play.google.com/store/apps/details?id=com.snapdeal.main&a
60 matches
Mail list logo