On Thu, Feb 14, 2019 at 6:56 PM Kenneth Brotman
<kenbrot...@yahoo.com.invalid> wrote:

> Those aren’t the same error messages so I think progress has been made.
>
>
>
> What version of C* are you running?
>
3.0.17 We will upgrade to 3.0.18 soon.

> How did you clear out the space?
>
I had a few topology changes to cleanup. `nodetool cleanup` did miracles.

Regards,

Leo

>
> *From:* Léo FERLIN SUTTON [mailto:lfer...@mailjet.com.INVALID]
> *Sent:* Thursday, February 14, 2019 7:54 AM
> *To:* user@cassandra.apache.org
> *Subject:* Re: Bootstrap keeps failing
>
>
>
> Hello again !
>
>
>
> I have managed to free a lot of disk space and now most nodes hover
> between 50% and 80%.
>
> I am still getting bootstrapping failures :(
>
>
>
> Here I have some logs :
>
> 2019-02-14T15:23:05+00:00 cass02-0001.c.company.internal user err
> cassandra  [org.apache.cassandra.streaming.StreamSession] [onError] -
> [Stream #ea8ae230-2f8f-11e9-8418-6d4f57de615d] Streaming error occurred
>
> 2019-02-14T15:23:05+00:00 cass02-0001.c.company..internal user info
> cassandra  [org.apache.cassandra.streaming.StreamResultFuture]
> [handleSessionComplete] - [Stream #ea8ae230-2f8f-11e9-8418-6d4f57de615d]
> Session with /10.10.23.1
>
> 55 is complete
>
> 2019-02-14T15:23:05+00:00 cass02-0001.c.company..internal user warning
> cassandra  [org.apache.cassandra.streaming.StreamResultFuture]
> [maybeComplete] - [Stream #ea8ae230-2f8f-11e9-8418-6d4f57de615d] Stream
> failed
>
> 2019-02-14T15:23:05+00:00 cass02-0001.c.company..internal user warning
> cassandra  [org.apache.cassandra.service.StorageService] [onFailure] -
> Error during bootstrap.
>
> 2019-02-14T15:23:05+00:00 cass02-0001.c.company..internal user err
> cassandra  [org.apache.cassandra.service.StorageService] [bootstrap] -
> Error while waiting on bootstrap to complete. Bootstrap will have to be
> restarted.
>
> 2019-02-14T15:23:05+00:00 cass02-0001.c.company..internal user warning
> cassandra  [org.apache.cassandra.service.StorageService] [joinTokenRing] -
> Some data streaming failed. Use nodetool to check bootstrap state and
> resume. For more, see `nodetool help bootstrap`. IN_PROGRESS
>
>
>
>
>
> I can see a `Streaming error occured` for all of my nodes it is trying to
> stream from. Is there a way to have more logs to know why the failures
> occurred ?
>
> I have set `<logger name="org.apache.cassandra.streaming.StreamSession"
> level="DEBUG" />` but it doesn't seem to give me more details, is there
> another class I should set to DEBUG ?
>
>
>
> Finally I have also noticed a lot of :
>
> [org.apache.cassandra.db.compaction.LeveledManifest]
> [getCompactionCandidates] - Bootstrapping - doing STCS in L0
>
> In my logs files, It might be important.
>
>
>
> Regards,
>
>
>
> Leo
>
>
>
> On Fri, Feb 8, 2019 at 3:59 PM Léo FERLIN SUTTON <lfer...@mailjet.com>
> wrote:
>
> On Fri, Feb 8, 2019 at 3:37 PM Kenneth Brotman
> <kenbrot...@yahoo.com.invalid> wrote:
>
> Thanks for the details that helps us understand the situation.  I’m pretty
> sure you’ve exceed the working capacity of some of those nodes.  Going over
> 50% - 75% depending on compaction strategy is ill-advised.
>
> 50% free disk space is a steep price to pay for disk space not used. We
> have about 90 terabytes of data on SSD and we are paying about 100$ per
> terabytes of ssd storage (on google cloud).
>
> Maybe we can get closer to 75%.
>
>
>
> Our compaction strategy is `LeveledCompactionStrategy` on our two biggest
> tables (90% of the data).
>
>
>
> You need to clear out as much room as possible to add more nodes.
>
> Are the tombstones clearing out.
>
> I think we don't have a lot of tombstones :
>
> We have 0 deletes on our two biggest tables.
>
> One of them get updated with new data (messages.messages), but the updates
> are filling columns previously empty, I am unsure but I think this doesn't
> cause any tombstones.
>
> I have joined the info from `nodetool tablestats` for our two largest
> tables.
>
>
>
> We are using cassandra-reaper that manages our repairs. A full repair
> takes about 13 days. So if we have tombstones they should not be older than
> 13 days.
>
>
>
> Are there old snap shots that you can delete.  And so on.
>
> Unfortunately no. We take a daily snapshot that we backup then drop.
>
>
>
> You have to make more room on the existing nodes.
>
>
>
> I am trying to run `nodetool cleanup` on our most "critical" nodes to see
> if it helps. If that doesn't do the trick we will only have two solutions :
>
>    - Add more disk space on each node
>    - Adding new nodes
>
> We have looked at some other companies case studies and it looks like we
> have a few very big nodes instead of a lot of smaller ones.
>
> We are currently trying to add nodes, and are hoping to eventually
> transition to a "lot of small nodes" model and be able to add nodes a lot
> faster.
>
>
>
> Thank you again for your interest,
>
>
>
> Regards,
>
>
>
> Leo
>
>
>
> *From:* Léo FERLIN SUTTON [mailto:lfer...@mailjet.com.INVALID]
> *Sent:* Friday, February 08, 2019 6:16 AM
> *To:* user@cassandra.apache.org
> *Subject:* Re: Bootstrap keeps failing
>
>
>
> On Thu, Feb 7, 2019 at 10:11 PM Kenneth Brotman
> <kenbrot...@yahoo.com.invalid> wrote:
>
> Lots of things come to mind. We need more information from you to help us
> understand:
>
> How long have you had your cluster running?
>
> A bit more than a year old. But it has been constantly growing (3 nodes to
> 6 nodes to 12 nodes, etc).
>
> We have a replication_factor of 3 on all keyspaces and 3 racks with an
> equal amount of nodes.
>
>
>
> Is it generally working ok?
>
> Works fine. Good performance, repairs managed by cassandra-reaper.
>
>
>
> Is it just one node that is misbehaving at a time?
>
> We only bootstrap nodes one at a time. Sometimes it works flawlessly,
> sometimes it fails. When it fails it tends to fail a lot in a row before we
> manage to get it bootstrapped.
>
>
>
> How many nodes do you need to replace?
>
> I am adding nodes, not replacing any. Our nodes are starting to get very
> full and we wish to add at least 6 more nodes (short-term).
>
> Adding a new node is quite slow (48 to 72 hours) and that's when the
> boostrap process works at the first try.
>
>
>
> Are you doing rolling restarts instead of simultaneously?
>
> Yes.
>
>
>
> Do you have enough capacity on your machines?  Did you say some of the
> nodes are at 90% capacity?
>
> The free disk space left fluctuates but is generally between 80% and 90%,
> this is why we are planning to add a lot more nodes.
>
>
>
> When did this problem begin?
>
>  Not sure about this one. Probably since our nodes have more than 2to
> data, I don't remember it being an issue when our nodes were smaller.
>
>
>
> Could something be causing a racing condition?
>
> We have schema changes every day.
>
> We have temporary data stored in cassandra, only used for 6 days then
> destroyed.
>
>
>
> In order to avoid tombstones we have a table rotation, every day we create
> a new table to contain the data for the next day, and we drop the oldest
> temporary table.
>
>
>
> This means that when the node starts to bootstrap it will ask other nodes
> for data that will almost certainly be dropped before the boostrap process
> is finished.
>
>
>
> Did you recheck the commands you used to make sure they are correct?
>
> What procedure do you use?
>
>
>
> Our procedure is :
>
>    1. We install cassandra on a brand new instance (debian).
>    2. We install cassandra.
>    3. We stop the default cassandra (launched by the debian package).
>    4. We empty these directories :
>    /var/lib/cassandra/commitlog
>    /var/lib/cassandra/data
>    /var/lib/cassandra/saved_caches
>    5. We put our configuration in place of the default one.
>    6. We start the cassandra.
>
> If after 3 days we see that the node hasn't joined the cluster we check
> the `nodetool netstats` command to see if the node is still streaming data.
> If it is not we launch `nodetool bootstrap resume` on the instance.
>
>
>
> Thank you for you interest in our issue !
>
>
>
> Regards,
>
>
>
> Leo
>
>
>
>
>
> *From:* Léo FERLIN SUTTON [mailto:lfer...@mailjet.com.INVALID]
> *Sent:* Thursday, February 07, 2019 9:16 AM
> *To:* user@cassandra.apache.org
> *Subject:* Re: [EXTERNAL] Re: Bootstrap keeps failing
>
>
>
> Thank you for the recommendation.
>
>
>
> We are already using datastax's recommended settings for tcp_keepalive
>
>
>
> Regards,
>
>
>
> Leo
>
>
>
> On Thu, Feb 7, 2019 at 5:49 PM Durity, Sean R <sean_r_dur...@homedepot.com>
> wrote:
>
> I have seen unreliable streaming (streaming that doesn’t finish) because
> of TCP timeouts from firewalls or switches. The default tcp_keepalive
> kernel parameters are usually not tuned for that. See
> https://docs.datastax.com/en/dse-trblshoot/doc/troubleshooting/idleFirewallLinux.html
> for more details. These “remote” timeouts are difficult to detect or prove
> if you don’t have access to the intermediate network equipment.
>
>
>
> Sean Durity
>
> *From:* Léo FERLIN SUTTON <lferlin@mailjetcomINVALID>
> *Sent:* Thursday, February 07, 2019 10:26 AM
> *To:* user@cassandra.apache.org; dinesh.jo...@yahoo.com
> *Subject:* [EXTERNAL] Re: Bootstrap keeps failing
>
>
>
> Hello !
>
> Thank you for your answers.
>
>
>
> So I have tried, multiple times, to start bootstrapping from scratch. I
> often have the same problem (on other nodes as well) but sometimes it works
> and I can move on to another node.
>
>
>
> I have joined a jstack dump and some logs.
>
>
>
> Our node was shut down at around 97% disk space used
>
> I turned it back on and it starting the bootstrap process again.
>
>
>
> The log file is the log from this attempt, same for the thread dump.
>
>
>
> Small warning, I have somewhat anonymised the log files so there may be
> some inconsistencies
>
>
>
> Regards,
>
>
>
> Leo
>
>
>
> On Thu, Feb 7, 2019 at 8:13 AM dinesh.jo...@yahoo.com.INVALID <
> dinesh.jo...@yahoo.com.invalid <dinesh.joshi@yahoocom.invalid>> wrote:
>
> Would it be possible for you to take a thread dump & logs and share them?
>
>
>
> Dinesh
>
>
>
>
>
> On Wednesday, February 6, 2019, 10:09:11 AM PST, Léo FERLIN SUTTON <
> lfer...@mailjet.com.INVALID> wrote:
>
>
>
>
>
> Hello !
>
>
>
> I am having a recurrent problem when trying to bootstrap a few new nodes.
>
>
>
> Some general info :
>
>    - I am running cassandra 3.0.17
>    - We have about 30 nodes in our cluster
>    - All healthy nodes have between 60% to 90% used disk space on
>    /var/lib/cassandra
>
> So I create a new node and let auto_bootstrap do it's job. After a few
> days the bootstrapping node stops streaming new data but is still not a
> member of the cluster.
>
>
>
> `nodetool status` says the node is still joining,
>
>
>
> When this happens I run `nodetool bootstrap resume`. This usually ends up
> in two different ways :
>
>    1. The node fills up to 100% disk space and crashes.
>    2. The bootstrap resume finishes with errors
>
> When I look at `nodetool netstats -H` is  looks like `bootstrap resume`
> does not resume but restarts a full transfer of every data from every node.
>
>
>
> This is the output I get from `nodetool resume` :
>
> [2019-02-06 01:39:14,369] received file
> /var/lib/cassandra/raw/raw_17930-d7cc0590230d11e9bc0af381b0ee7ac6/mc-225-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:39:16,821] received file
> /var/lib/cassandra/data/system_distributed/repair_history-759fffad624b318180eefa9a52d1f627/mc-88-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:39:17,003] received file
> /var/lib/cassandra/data/system_distributed/repair_history-759fffad624b318180eefa9a52d1f627/mc-89-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:39:17,032] session with /10.16.XX.YYY complete (progress:
> 2113%)
>
> [2019-02-06 01:41:15,160] received file
> /var/lib/cassandra/raw/raw_17930-d7cc0590230d11e9bc0af381b0ee7ac6/mc-220-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:42:02,864] received file
> /var/lib/cassandra/raw/raw_17930-d7cc0590230d11e9bc0af381b0ee7ac6/mc-226-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:42:09,284] received file
> /var/lib/cassandra/raw/raw_17930-d7cc0590230d11e9bc0af381b0ee7ac6/mc-227-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:42:10,522] received file
> /var/lib/cassandra/raw/raw_17930-d7cc0590230d11e9bc0af381b0ee7ac6/mc-228-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:42:10,622] received file
> /var/lib/cassandra/raw/raw_17930-d7cc0590230d11e9bc0af381b0ee7ac6/mc-229-big-Data.db
> (progress: 2113%)
>
> [2019-02-06 01:42:11,925] received file
> /var/lib/cassandra/data/system_distributed/repair_history-759fffad624b318180eefa9a52d1f627/mc-90-big-Data.db
> (progress: 2114%)
>
> [2019-02-06 01:42:14,887] received file
> /var/lib/cassandra/data/system_distributed/repair_history-759fffad624b318180eefa9a52d1f627/mc-91-big-Data.db
> (progress: 2114%)
>
> [2019-02-06 01:42:14,980] session with /10.16.XX.ZZZ complete (progress:
> 2114%)
>
> [2019-02-06 01:42:14,980] Stream failed
>
> [2019-02-06 01:42:14,982] Error during bootstrap: Stream failed
>
> [2019-02-06 01:42:14,982] Resume bootstrap complete
>
>
>
> The bootstrap `progress` goes way over 100% and eventually fails.
>
>
>
>
>
> Right now I have a node with this output from `nodetool status` :
>
> `UJ  10.16.XX.YYY  2.93 TB    256          ?
>  5788f061-a3c0-46af-b712-ebeecd397bf7  c`
>
>
>
> It is almost filled with data, yet if I look at `nodetool netstats` :
>
>         Receiving 480 files, 325.39 GB total. Already received 5 files,
> 68.32 MB total
>         Receiving 499 files, 328.96 GB total. Already received 1 files,
> 1.32 GB total
>         Receiving 506 files, 345.33 GB total. Already received 6 files,
> 24.19 MB total
>         Receiving 362 files, 206.73 GB total. Already received 7 files, 34
> MB total
>         Receiving 424 files, 281.25 GB total. Already received 1 files,
> 1.3 GB total
>         Receiving 581 files, 349.26 GB total. Already received 8 files,
> 45.96 MB total
>         Receiving 443 files, 337.26 GB total. Already received 6 files,
> 96.15 MB total
>         Receiving 424 files, 275.23 GB total. Already received 5 files,
> 42.67 MB total
>
>
>
> It is trying to pull all the data again.
>
>
>
> Am I missing something about the way `nodetool bootstrap resume` is
> supposed to be used ?
>
>
>
> Regards,
>
>
>
> Leo
>
>
>
>
> ------------------------------
>
>
> The information in this Internet Email is confidential and may be legally
> privileged. It is intended solely for the addressee. Access to this Email
> by anyone else is unauthorized. If you are not the intended recipient, any
> disclosure, copying, distribution or any action taken or omitted to be
> taken in reliance on it, is prohibited and may be unlawful. When addressed
> to our clients any opinions or advice contained in this Email are subject
> to the terms and conditions expressed in any applicable governing The Home
> Depot terms of business or client engagement letter. The Home Depot
> disclaims all responsibility and liability for the accuracy and content of
> this attachment and for any damages or losses arising from any
> inaccuracies, errors, viruses, e.g., worms, trojan horses, etc., or other
> items of a destructive nature, which may be contained in this attachment
> and shall not be liable for direct, indirect, consequential or special
> damages in connection with this e-mail message or its attachment.
>
>

Reply via email to