Awesome. Thanks.

Zeeshan Lakhani
programmer | 
software engineer at @basho | 
org. member/founder of @papers_we_love |
twitter => @zeeshanlakhani

> On Apr 24, 2015, at 8:56 PM, Jason Campbell <xia...@xiaclo.net> wrote:
> 
> This may be a case of force-replace vs replace vs reip.  I'm happy to see if 
> I can get new cluster from backup to keep the Solr indexes.
> 
> The disk backup was all of /var/lib/riak, so definitely included the YZ 
> indexes before the force-replace, and they were kept on the first node that 
> was changed with reip.  I stopped each node before the snapshot to ensure 
> consistency.  So I would expect the final restored cluster to be somewhere 
> between the first and last node snapshot in terms of data, and AAE to repair 
> things to a consistent state for that few minute gap.
> 
> I'll experiment with different methods of rebuilding the cluster on Monday 
> and see if I can get it to keep the Solr indexes.  Maybe moving the YZ 
> indexes out of the way during the force-replace, then stopping the node and 
> putting them back could help as well.  I'll let you know the results of the 
> experiments either way.
> 
> Thanks,
> Jason
> 
>> On 25 Apr 2015, at 09:25, Zeeshan Lakhani <zlakh...@basho.com> wrote:
>> 
>> Hey Jason,
>> 
>> Yeah, nodes can normally be joined without a cluster dropping its Solr Index 
>> and AAE normally rebuilds the missing KV bits.
>> 
>> In the case of restoring from a backup and having missing data, we can only 
>> recommend a reindex (the indexes that have the issue) with aggressive AAE 
>> settings to speed things up. It can be pretty fast. Recreating indexes are 
>> cheap in Yokozuna, but are the `data/yz` directories missing from the nodes 
>> that were force-replaced? Unless someone else wants to chime in, I’ll gather 
>> more info on what occurred from the reip vs the force-replace. 
>> 
>> Zeeshan Lakhani
>> programmer | 
>> software engineer at @basho | 
>> org. member/founder of @papers_we_love | paperswelove.org
>> twitter => @zeeshanlakhani
>> 
>>> On Apr 24, 2015, at 7:02 PM, Jason Campbell <xia...@xiaclo.net> wrote:
>>> 
>>> Is there a way to do a restore without rebuilding these indexes though?  
>>> Obviously this could take a long time depending on the amount of indexed 
>>> data in the cluster.  It's a fairly big gotcha to say that Yokozuna fixes a 
>>> lot of the data access issues that Riak has, but if you restore from a 
>>> backup, it could be useless for days or weeks.
>>> 
>>> As far as disk consistency, the nodes were stopped during the snapshot, so 
>>> I'm assuming on-disk it would be consistent within a single node.  And 
>>> cluster wide, I would expect the overall data to fall somewhere between the 
>>> first and last node snapshot.  AAE should still repair the bits left over, 
>>> but it shouldn't have to rebuild the entire Solr index.
>>> 
>>> So the heart of the question can I join a node to a cluster without 
>>> dropping it's Solr index?  force-replace obviously doesn't work, what is 
>>> the harm in running reip on every node instead of just the first?
>>> 
>>> Thanks for the help,
>>> Jason
>>> 
>>>> On 25 Apr 2015, at 00:36, Zeeshan Lakhani <zlakh...@basho.com> wrote:
>>>> 
>>>> Hey Jason,
>>>> 
>>>> Here’s a little more discussion on Yokozuna backup strategies: 
>>>> http://lists.basho.com/pipermail/riak-users_lists.basho.com/2014-January/014514.html.
>>>> 
>>>> Nonetheless, I wouldn’t say the behavior’s expected, but we’re going to be 
>>>> adding more to the docs on how to rebuild indexes.
>>>> 
>>>> To do so, you could just remove the yz_anti_entropy directory, and make 
>>>> AAE more aggressive, via
>>>> 
>>>> ```
>>>> rpc:multicall([node() | nodes()], application, set_env, [yokozuna, 
>>>> anti_entropy_build_limit, {100, 1000}]).
>>>> rpc:multicall([node() | nodes()], application, set_env, [yokozuna, 
>>>> anti_entropy_concurrency, 4])
>>>> ```
>>>> 
>>>> and the indexes will rebuild. You can try to initialize the building of 
>>>> trees with `yz_entropy_mgr:init([])` via `riak attach`, but a restart 
>>>> would also kick AAE into gear. There’s a bit more related info on this 
>>>> thread: 
>>>> http://lists.basho.com/pipermail/riak-users_lists.basho.com/2015-March/016929.html.
>>>> 
>>>> Thanks.
>>>> 
>>>> Zeeshan Lakhani
>>>> programmer | 
>>>> software engineer at @basho | 
>>>> org. member/founder of @papers_we_love | paperswelove.org
>>>> twitter => @zeeshanlakhani
>>>> 
>>>>> On Apr 24, 2015, at 1:34 AM, Jason Campbell <xia...@xiaclo.net> wrote:
>>>>> 
>>>>> I think I figured it out.
>>>>> 
>>>>> I followed this guide: 
>>>>> http://docs.basho.com/riak/latest/ops/running/nodes/renaming/#Clusters-from-Backups
>>>>> 
>>>>> The first Riak node (changed with riak-admin reip) kept it's Solr index.  
>>>>> However, the other nodes when joined via riak-admin cluster 
>>>>> force-replace, dropped their Solr indexes.
>>>>> 
>>>>> Is this expected?  If so, it should really be in the docs, and there 
>>>>> should be another way to restore a cluster keeping Solr intact.
>>>>> 
>>>>> Also, is there a way to rebuild a Solr index?
>>>>> 
>>>>> Thanks,
>>>>> Jason
>>>>> 
>>>>>> On 24 Apr 2015, at 15:16, Jason Campbell <xia...@xiaclo.net> wrote:
>>>>>> 
>>>>>> I've just done a backup and restore of our production Riak cluster, and 
>>>>>> Yokozuna has dropped from around 125 million records to 25million.  
>>>>>> Obviously the IPs have changed, and although the Riak cluster is stable, 
>>>>>> I'm not sure Solr handled the transition as nicely.
>>>>>> 
>>>>>> Is there a way to force Solr to rebuild the indexes, or at least get 
>>>>>> back to the state it was in before the backup?
>>>>>> 
>>>>>> Also, is this expected behaviour?
>>>>>> 
>>>>>> Thanks,
>>>>>> Jason
>>>>>> _______________________________________________
>>>>>> riak-users mailing list
>>>>>> riak-users@lists.basho.com
>>>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> riak-users mailing list
>>>>> riak-users@lists.basho.com
>>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>> 
>>>> _______________________________________________
>>>> riak-users mailing list
>>>> riak-users@lists.basho.com
>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>> 
>> _______________________________________________
>> riak-users mailing list
>> riak-users@lists.basho.com
>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> 

_______________________________________________
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to