Hey Jason, Yeah, nodes can normally be joined without a cluster dropping its Solr Index and AAE normally rebuilds the missing KV bits.
In the case of restoring from a backup and having missing data, we can only recommend a reindex (the indexes that have the issue) with aggressive AAE settings to speed things up. It can be pretty fast. Recreating indexes are cheap in Yokozuna, but are the `data/yz` directories missing from the nodes that were force-replaced? Unless someone else wants to chime in, I’ll gather more info on what occurred from the reip vs the force-replace. Zeeshan Lakhani programmer | software engineer at @basho | org. member/founder of @papers_we_love | paperswelove.org twitter => @zeeshanlakhani > On Apr 24, 2015, at 7:02 PM, Jason Campbell <xia...@xiaclo.net> wrote: > > Is there a way to do a restore without rebuilding these indexes though? > Obviously this could take a long time depending on the amount of indexed data > in the cluster. It's a fairly big gotcha to say that Yokozuna fixes a lot of > the data access issues that Riak has, but if you restore from a backup, it > could be useless for days or weeks. > > As far as disk consistency, the nodes were stopped during the snapshot, so > I'm assuming on-disk it would be consistent within a single node. And > cluster wide, I would expect the overall data to fall somewhere between the > first and last node snapshot. AAE should still repair the bits left over, > but it shouldn't have to rebuild the entire Solr index. > > So the heart of the question can I join a node to a cluster without dropping > it's Solr index? force-replace obviously doesn't work, what is the harm in > running reip on every node instead of just the first? > > Thanks for the help, > Jason > >> On 25 Apr 2015, at 00:36, Zeeshan Lakhani <zlakh...@basho.com> wrote: >> >> Hey Jason, >> >> Here’s a little more discussion on Yokozuna backup strategies: >> http://lists.basho.com/pipermail/riak-users_lists.basho.com/2014-January/014514.html. >> >> Nonetheless, I wouldn’t say the behavior’s expected, but we’re going to be >> adding more to the docs on how to rebuild indexes. >> >> To do so, you could just remove the yz_anti_entropy directory, and make AAE >> more aggressive, via >> >> ``` >> rpc:multicall([node() | nodes()], application, set_env, [yokozuna, >> anti_entropy_build_limit, {100, 1000}]). >> rpc:multicall([node() | nodes()], application, set_env, [yokozuna, >> anti_entropy_concurrency, 4]) >> ``` >> >> and the indexes will rebuild. You can try to initialize the building of >> trees with `yz_entropy_mgr:init([])` via `riak attach`, but a restart would >> also kick AAE into gear. There’s a bit more related info on this thread: >> http://lists.basho.com/pipermail/riak-users_lists.basho.com/2015-March/016929.html. >> >> Thanks. >> >> Zeeshan Lakhani >> programmer | >> software engineer at @basho | >> org. member/founder of @papers_we_love | paperswelove.org >> twitter => @zeeshanlakhani >> >>> On Apr 24, 2015, at 1:34 AM, Jason Campbell <xia...@xiaclo.net> wrote: >>> >>> I think I figured it out. >>> >>> I followed this guide: >>> http://docs.basho.com/riak/latest/ops/running/nodes/renaming/#Clusters-from-Backups >>> >>> The first Riak node (changed with riak-admin reip) kept it's Solr index. >>> However, the other nodes when joined via riak-admin cluster force-replace, >>> dropped their Solr indexes. >>> >>> Is this expected? If so, it should really be in the docs, and there should >>> be another way to restore a cluster keeping Solr intact. >>> >>> Also, is there a way to rebuild a Solr index? >>> >>> Thanks, >>> Jason >>> >>>> On 24 Apr 2015, at 15:16, Jason Campbell <xia...@xiaclo.net> wrote: >>>> >>>> I've just done a backup and restore of our production Riak cluster, and >>>> Yokozuna has dropped from around 125 million records to 25million. >>>> Obviously the IPs have changed, and although the Riak cluster is stable, >>>> I'm not sure Solr handled the transition as nicely. >>>> >>>> Is there a way to force Solr to rebuild the indexes, or at least get back >>>> to the state it was in before the backup? >>>> >>>> Also, is this expected behaviour? >>>> >>>> Thanks, >>>> Jason >>>> _______________________________________________ >>>> riak-users mailing list >>>> riak-users@lists.basho.com >>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com >>> >>> >>> _______________________________________________ >>> riak-users mailing list >>> riak-users@lists.basho.com >>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com >> >> _______________________________________________ >> riak-users mailing list >> riak-users@lists.basho.com >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com >
_______________________________________________ riak-users mailing list riak-users@lists.basho.com http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com