One interesting moment happened when I tried removing the index:

- this index was associated with a bucket type, called fs_chunks
- so I first called RpbSetBucketTypeReq to set search_index: _dont_index_
- i then tried to remove the index with RpbYokozunaIndexDeleteReq which
failed with "index is in use" and list of all buckets of the fs_chunks type
- for some reason all these buckets had their own search_index property set
to that same index

How can this happen if I definitely never set the search_index property per
bucket?

On 24 March 2016 at 22:41, Oleksiy Krivoshey <oleks...@gmail.com> wrote:

> OK!
>
> On 24 March 2016 at 21:11, Magnus Kessler <mkess...@basho.com> wrote:
>
>> Hi Oleksiy,
>>
>> On 24 March 2016 at 14:55, Oleksiy Krivoshey <oleks...@gmail.com> wrote:
>>
>>> Hi Magnus,
>>>
>>> Thanks! I guess I will go with index deletion because I've already tried
>>> expiring the trees before.
>>>
>>> Do I need to delete AAE data somehow or removing the index is enough?
>>>
>>
>> If you expire the AAE trees with the commands I posted earlier, there
>> should be no need to remove the AAE data directories manually.
>>
>> I hope this works for you. Please monitor the tree rebuild and exchanges
>> with `riak-admin search aae-status` for the next few days. In particular
>> the exchanges should be ongoing on a continuous basis once all trees have
>> been rebuilt. If they don't, please let me know. At that point you should
>> also gather `riak-debug` output from all nodes before it gets rotated out
>> after 5 days by default.
>>
>> Kind Regards,
>>
>> Magnus
>>
>>
>>>
>>> On 24 March 2016 at 13:28, Magnus Kessler <mkess...@basho.com> wrote:
>>>
>>>> Hi Oleksiy,
>>>>
>>>> As a first step, I suggest to simply expire the Yokozuna AAE trees
>>>> again if the output of `riak-admin search aae-status` still suggests that
>>>> no recent exchanges have taken place. To do this, run `riak attach` on one
>>>> node and then
>>>>
>>>> riak_core_util:rpc_every_member_ann(yz_entropy_mgr, expire_trees, [], 
>>>> 5000).
>>>>
>>>>
>>>> Exit from the riak console with `Ctrl+G q`.
>>>>
>>>> Depending on your settings and amount of data the full index should be
>>>> rebuilt within the next 2.5 days (for a cluster with ring size 128 and
>>>> default settings). You can monitor the progress with `riak-admin search
>>>> aae-status` and also in the logs, which should have messages along the
>>>> lines of
>>>>
>>>> 2016-03-24 10:28:25.372 [info]
>>>> <0.4647.6477>@yz_exchange_fsm:key_exchange:179 Repaired 83055 keys during
>>>> active anti-entropy exchange of partition
>>>> 1210306043414653979137426502093171875652569137152 for preflist
>>>> {1164634117248063262943561351070788031288321245184,3}
>>>>
>>>>
>>>> Re-indexing can put additional strain on the cluster and may cause
>>>> elevated latency on a cluster already under heavy load. Please monitor the
>>>> response times while the cluster is re-indexing data.
>>>>
>>>> If the cluster load allows it, you can force more rapid re-indexing by
>>>> changing a few parameters. Again at the `riak attach` console, run
>>>>
>>>> riak_core_util:rpc_every_member_ann(application, set_env, [yokozuna, 
>>>> anti_entropy_build_limit, {4, 60000}], 5000).
>>>> riak_core_util:rpc_every_member_ann(application, set_env, [yokozuna, 
>>>> anti_entropy_concurrency, 5], 5000).
>>>>
>>>> This will allow up to 4 trees per node to be built/exchanged per hour,
>>>> with up to 5 concurrent exchanges throughout the cluster. To return back to
>>>> the default settings, use
>>>>
>>>> riak_core_util:rpc_every_member_ann(application, set_env, [yokozuna, 
>>>> anti_entropy_build_limit, {1, 360000}], 5000).
>>>> riak_core_util:rpc_every_member_ann(application, set_env, [yokozuna, 
>>>> anti_entropy_concurrency, 2], 5000).
>>>>
>>>>
>>>> If the cluster still doesn't make any progress with automatically
>>>> re-indexing data, the next steps are pretty much what you already
>>>> suggested, to drop the existing index and re-index from scratch. I'm
>>>> assuming that losing the indexes temporarily is acceptable to you at this
>>>> point.
>>>>
>>>> Using any client API that supports RpbYokozunaIndexDeleteReq, you can
>>>> drop the index from all Solr instances, losing any data stored there
>>>> immediately. Next, you'll have to re-create the index. I have tried this
>>>> with the python API, where I deleted the index and re-created it with the
>>>> same already uploaded schema:
>>>>
>>>> from riak import RiakClient
>>>>
>>>> c = RiakClient()
>>>> c.delete_search_index('my_index')
>>>> c.create_search_index('my_index', 'my_schema')
>>>>
>>>> Note that simply deleting the index does not remove it's existing
>>>> association with any bucket or bucket type. Any PUT operations on these
>>>> buckets will lead to indexing failures being logged until the index has
>>>> been recreated. However, this also means that no separate operation in
>>>> `riak-admin` is required to associate the newly recreated index with the
>>>> buckets again.
>>>>
>>>> After recreating the index expire the trees as explained previously.
>>>>
>>>> Let us know if this solves your issue.
>>>>
>>>> Kind Regards,
>>>>
>>>> Magnus
>>>>
>>>>
>>>> On 24 March 2016 at 08:44, Oleksiy Krivoshey <oleks...@gmail.com>
>>>> wrote:
>>>>
>>>>> This is how things are looking after two weeks:
>>>>>
>>>>> - there are no solr indexing issues for a long period (2 weeks)
>>>>> - there are no yokozuna errors at all for 2 weeks
>>>>> - there is an index with all empty schema, just _yz_* fields, objects
>>>>> stored in a bucket(s) are binary and so are not analysed by yokozuna
>>>>> - same yokozuna query repeated gives different number for num_found,
>>>>> typically the difference between real number of keys in a bucket and
>>>>> num_found is about 25%
>>>>> - number of keys repaired by AAE (according to logs) is about 1-2 per
>>>>> few hours (number of keys "missing" in index is close to 1,000,000)
>>>>>
>>>>> Should I now try to delete the index and yokozuna AAE data and wait
>>>>> another 2 weeks? If yes - how should I delete the index and AAE data?
>>>>> Will RpbYokozunaIndexDeleteReq be enough?
>>>>>
>>>>>
>>>>>
>>>> --
>>>> Magnus Kessler
>>>> Client Services Engineer
>>>> Basho Technologies Limited
>>>>
>>>> Registered Office - 8 Lincoln’s Inn Fields London WC2A 3BP Reg 07970431
>>>>
>>>
>>>
>>
>>
>> --
>> Magnus Kessler
>> Client Services Engineer
>> Basho Technologies Limited
>>
>> Registered Office - 8 Lincoln’s Inn Fields London WC2A 3BP Reg 07970431
>>
>
>
_______________________________________________
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to