Hello.

Good news. After i remove folder, and start yz_entropy_mgr:init([]) and
then re-save objects it's start to return correct numFound (test on 2
buckets, about 6000 requests). This is tested on dev instance.

So, as i understand, i need to make same operation on production cluster on
each node?

And, is this bug already patched? Or in future i can have same problem
after add/remove nodes?

Thanks,
Roman Lakotko.


2015-03-12 18:56 GMT+03:00 Roma Lakotko <ro...@lakotko.ru>:

> No, its simple riak search http request.
> 12 марта 2015 г. 18:54 пользователь "Zeeshan Lakhani" <zlakh...@basho.com>
> написал:
>
> Are you running mapreduce with Solr queries?
>>
>>
>> On Mar 12, 2015, at 11:50 AM, Roma Lakotko <ro...@lakotko.ru> wrote:
>>
>> I don't see any solr errors. But each 10-20 minutes on prod and once a
>> day on dev i see strange errors:
>>
>> 2015-03-11 09:18:10.668 [error] <0.234.0> Supervisor
>> riak_pipe_fitting_sup had child undefined started with
>> riak_pipe_fitting:start_link() at <0.12060.2> exit with reason noproc in
>> context shutdown_error
>> 2015-03-12 13:12:05.200 [error] <0.379.0> Supervisor riak_kv_mrc_sink_sup
>> had child undefined started with riak_kv_mrc_sink:start_link() at
>> <0.6601.1> exit with reason noproc in context shutdown_error
>>
>> For both prod and dev instance values are:
>>
>> anti_entropy_build_limit  -> {ok,{1,3600000}}
>> anti_entropy_concurrency -> {ok,2}
>> anti_entropy_tick - > undefined
>>
>> I delete data folder and run init method, i'll  results after it rebuild
>> trees.
>>
>> 2015-03-12 18:22 GMT+03:00 Zeeshan Lakhani <zlakh...@basho.com>:
>>
>>> Are you noticing any Solr errors in the logs?
>>>
>>> For your container instance, you can attempt to clear the AAE trees and
>>> force a rebuild by removing the entropy directories in 
>>> `./data/yz_anti_entropy`
>>> and running `yz_entropy_mgr:init([])` via `riak attach`.  Or, you can
>>> let AAE occur naturally (after removing the entropy data) and up the
>>> concurrency/build_limit/tick (using set_env). You can see what you’re
>>> current settings are by calling...
>>>
>>> ```
>>> riak_core_util:rpc_every_member_ann(application, get_env, [riak_kv,
>>> anti_entropy_build_limit],infinity).
>>> riak_core_util:rpc_every_member_ann(application, get_env, [riak_kv,
>>> anti_entropy_concurrency],infinity).
>>> riak_core_util:rpc_every_member_ann(application, get_env, [yokozuna,
>>> anti_entropy_tick],infinity).
>>> ```
>>>
>>> … on any of the nodes.  Query coverage is R=1, but the values should be
>>> replicated across.
>>>
>>> Thanks.
>>>
>>>
>>> On Mar 12, 2015, at 9:51 AM, Roma Lakotko <ro...@lakotko.ru> wrote:
>>>
>>> Hello Zeeshan.
>>>
>>> While i run queries no delete object is occurs.
>>>
>>> Stats on production and developer nodes output something like this:
>>> https://gist.github.com/romulka/d0254aa193a9dbb52b67
>>>
>>> On dev container:
>>>
>>> /etc/riak# grep anti_entropy *
>>> riak.conf:anti_entropy = active
>>> riak.conf.dpkg-dist:anti_entropy = active
>>>
>>> ll -h /var/lib/riak/yz_anti_entropy/
>>> total 264K
>>> drwxrwxr-x 66 riak riak 4.0K Sep 25 12:08 ./
>>> drwxr-xr-x 12 riak riak 4.0K Dec  9 12:19 ../
>>> drwxr-xr-x  9 riak riak 4.0K Mar 12 12:01 0/
>>> drwxr-xr-x  9 riak riak 4.0K Mar 12 12:01
>>> 1004782375664995756265033322492444576013453623296/
>>> drwxr-xr-x  9 riak riak 4.0K Mar 12 12:01
>>> 1027618338748291114361965898003636498195577569280/
>>> ....
>>>
>>> On prod:
>>>
>>> grep anti_entropy * /etc/riak/ -> empty
>>>
>>> root@riak-21:/var/lib/riak/yz_anti_entropy# ll -h
>>> total 64K
>>> drwxrwxr-x 16 riak riak 4.0K Dec  4 03:44 ./
>>> drwxr-xr-x 14 riak riak 4.0K Dec  9 12:10 ../
>>> drwxr-xr-x  9 riak riak 4.0K Dec  4 03:44 0/
>>> drwxr-xr-x  9 riak riak 4.0K Mar 12 12:57
>>> 1027618338748291114361965898003636498195577569280/
>>> ....
>>>
>>> I'm already try re-save all keys, it doesn't helps.
>>>
>>> Production cluster have 7 node, start from 3. So yes, nodes was
>>> added/delete sometimes.
>>>
>>> On dev, i have 1 instance in docker container, never added to cluster.
>>> But data in that riak is imported from production cluster a while ago.
>>>
>>> I can give you a copy of container, if you need to.
>>>
>>> Thanks,
>>> Ronan Lakotko
>>>
>>>
>>>
>>> 2015-03-12 16:36 GMT+03:00 Zeeshan Lakhani <zlakh...@basho.com>:
>>>
>>>> Hello Roma,
>>>>
>>>> Have you deleted this object at some point in your runs? Please make
>>>> sure AAE is running by checking search’s AAE status, `riak-admin search
>>>> aae-status`, and that data exists in the correct directory,
>>>> `./data/yz_anti_entropy` (
>>>> http://docs.basho.com/riak/latest/ops/advanced/configs/search/). You
>>>> may just need to perform a read-repair by performing a fetch of the object
>>>> itself first, before performing search queries again.
>>>>
>>>> Also, have you left or added nodes? I’m guessing that  even your 1 node
>>>> instance is still running a cluster on that one node, right?
>>>>
>>>> Thanks.
>>>>
>>>> Zeeshan Lakhani
>>>> programmer |
>>>> software engineer at @basho |
>>>> org. member/founder of @papers_we_love | paperswelove.org
>>>> twitter => @zeeshanlakhani
>>>>
>>>> On Mar 12, 2015, at 5:59 AM, Roma Lakotko <ro...@lakotko.ru> wrote:
>>>>
>>>> Each request to riak search return different results. It's return
>>>> different numFound.
>>>>
>>>> I use request like this:
>>>>
>>>>
>>>> http://localhost:8098/search/query/assets?wt=json&q=type:*&sort=_yz_rk%20asc
>>>>
>>>> If add start offset it can return:
>>>>
>>>>
>>>> http://localhost:8098/search/query/assets?wt=json&q=type:*&sort=_yz_rk%20asc&start=1247
>>>>
>>>> "response": {
>>>>         "numFound": 1248,
>>>>         "start": 1247,
>>>>         "docs": [
>>>>             {
>>>>                 "_yz_id": 
>>>> "1*default*assets*fff63ecf-a0c4-4ecf-b24d-c493ca3a302f*44",
>>>>                 "_yz_rk": "fff63ecf-a0c4-4ecf-b24d-c493ca3a302f",
>>>>                 "_yz_rt": "default",
>>>>                 "_yz_rb": "assets"
>>>>             }
>>>>         ]
>>>>     }
>>>>
>>>>
>>>> On next request it return something like this
>>>>
>>>>
>>>> "numFound": 1224,
>>>>         "start": 1247,
>>>>         "docs": []
>>>>
>>>>
>>>> I have 1 node installation, and no process write to Riak.
>>>>
>>>> I have same problem this production cluster with 7 nodes.
>>>>
>>>>
>>>> Scheme for document
>>>>
>>>>
>>>> <?xml version="1.0" encoding="UTF-8" ?>
>>>> <schema name="schedule" version="1.5">
>>>>  <fields>
>>>>    <field name="objectId"     type="string_ci"   indexed="true" 
>>>> stored="false" />
>>>>    <field name="type"     type="string_ci"   indexed="true" stored="false" 
>>>> />
>>>>    <field name="objectType"     type="string_ci"   indexed="true" 
>>>> stored="false" />
>>>>
>>>>    <field name="contentType"     type="string_ci"   indexed="true" 
>>>> stored="false" />
>>>>    <field name="properties"    type="string_ci"   indexed="true" 
>>>> stored="false" multiValued="true" />
>>>>    <field name="tag"     type="string_ci"   indexed="true" stored="false" 
>>>> />
>>>>    <field name="isUploaded"    type="boolean"     indexed="true" 
>>>> stored="false" />
>>>>    <field name="published"    type="boolean"     indexed="true" 
>>>> stored="false" />
>>>>    <field name="drm"    type="boolean"     indexed="true" stored="false" />
>>>>    <field name="dateCreated" type="date" indexed="true" stored="false" />
>>>>
>>>>    <!-- All of these fields are required by Riak Search -->
>>>>    <field name="_yz_id"   type="_yz_str" indexed="true" stored="true"  
>>>> multiValued="false" required="true"/>
>>>>    <field name="_yz_ed"   type="_yz_str" indexed="true" stored="false" 
>>>> multiValued="false"/>
>>>>    <field name="_yz_pn"   type="_yz_str" indexed="true" stored="false" 
>>>> multiValued="false"/>
>>>>    <field name="_yz_fpn"  type="_yz_str" indexed="true" stored="false" 
>>>> multiValued="false"/>
>>>>    <field name="_yz_vtag" type="_yz_str" indexed="true" stored="false" 
>>>> multiValued="false"/>
>>>>    <field name="_yz_rk"   type="_yz_str" indexed="true" stored="true"  
>>>> multiValued="false"/>
>>>>    <field name="_yz_rt"   type="_yz_str" indexed="true" stored="true"  
>>>> multiValued="false"/>
>>>>    <field name="_yz_rb"   type="_yz_str" indexed="true" stored="true"  
>>>> multiValued="false"/>
>>>>    <field name="_yz_err"  type="_yz_str" indexed="true" stored="false" 
>>>> multiValued="false"/>
>>>>
>>>>    <dynamicField name="*" type="ignored"/>
>>>>  </fields>
>>>>
>>>>  <uniqueKey>_yz_id</uniqueKey>
>>>>
>>>>  <types>
>>>>   <!-- YZ String: Used for non-analyzed fields text_ru -->
>>>>   <fieldType name="date" class="solr.TrieDateField" sortMissingLast="true" 
>>>> omitNorms="true"/>
>>>>   <fieldType name="double" class="solr.TrieDoubleField" 
>>>> sortMissingLast="true" omitNorms="true"/>
>>>>   <fieldType name="int" class="solr.TrieIntField" sortMissingLast="true" 
>>>> omitNorms="true"/>
>>>>
>>>>   <fieldType name="boolean" class="solr.BoolField" sortMissingLast="true" 
>>>> omitNorms="true"/>
>>>>   <fieldType name="_yz_str" class="solr.StrField" sortMissingLast="true" />
>>>>   <fieldtype name="ignored" stored="false" indexed="false" 
>>>> multiValued="true" class="solr.StrField" />
>>>>   <fieldType name="string_ci" class="solr.TextField" 
>>>> sortMissingLast="true" omitNorms="true">
>>>>         <analyzer>
>>>>             <tokenizer class="solr.StandardTokenizerFactory"/>
>>>>             <filter class="solr.LowerCaseFilterFactory" />
>>>>             <filter class='solr.PatternReplaceFilterFactory' pattern='ё' 
>>>> replacement='е' replace='all'/>
>>>>         </analyzer>
>>>>     </fieldType>
>>>>   </types>
>>>>
>>>> </schema>
>>>>
>>>>
>>>> Best regards,
>>>>
>>>> Roman
>>>>
>>>> _______________________________________________
>>>> riak-users mailing list
>>>> riak-users@lists.basho.com
>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>>
>>>>
>>>>
>>>
>>>
>>
>>
_______________________________________________
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to