I repeat same on production cluster, it's work. Thanks. 2015-03-13 10:23 GMT+03:00 Roma Lakotko <ro...@lakotko.ru>:
> Hello. > > Good news. After i remove folder, and start yz_entropy_mgr:init([]) and > then re-save objects it's start to return correct numFound (test on 2 > buckets, about 6000 requests). This is tested on dev instance. > > So, as i understand, i need to make same operation on production cluster > on each node? > > And, is this bug already patched? Or in future i can have same problem > after add/remove nodes? > > Thanks, > Roman Lakotko. > > > 2015-03-12 18:56 GMT+03:00 Roma Lakotko <ro...@lakotko.ru>: > >> No, its simple riak search http request. >> 12 марта 2015 г. 18:54 пользователь "Zeeshan Lakhani" <zlakh...@basho.com> >> написал: >> >> Are you running mapreduce with Solr queries? >>> >>> >>> On Mar 12, 2015, at 11:50 AM, Roma Lakotko <ro...@lakotko.ru> wrote: >>> >>> I don't see any solr errors. But each 10-20 minutes on prod and once a >>> day on dev i see strange errors: >>> >>> 2015-03-11 09:18:10.668 [error] <0.234.0> Supervisor >>> riak_pipe_fitting_sup had child undefined started with >>> riak_pipe_fitting:start_link() at <0.12060.2> exit with reason noproc in >>> context shutdown_error >>> 2015-03-12 13:12:05.200 [error] <0.379.0> Supervisor >>> riak_kv_mrc_sink_sup had child undefined started with >>> riak_kv_mrc_sink:start_link() at <0.6601.1> exit with reason noproc in >>> context shutdown_error >>> >>> For both prod and dev instance values are: >>> >>> anti_entropy_build_limit -> {ok,{1,3600000}} >>> anti_entropy_concurrency -> {ok,2} >>> anti_entropy_tick - > undefined >>> >>> I delete data folder and run init method, i'll results after it rebuild >>> trees. >>> >>> 2015-03-12 18:22 GMT+03:00 Zeeshan Lakhani <zlakh...@basho.com>: >>> >>>> Are you noticing any Solr errors in the logs? >>>> >>>> For your container instance, you can attempt to clear the AAE trees and >>>> force a rebuild by removing the entropy directories in >>>> `./data/yz_anti_entropy` >>>> and running `yz_entropy_mgr:init([])` via `riak attach`. Or, you can >>>> let AAE occur naturally (after removing the entropy data) and up the >>>> concurrency/build_limit/tick (using set_env). You can see what you’re >>>> current settings are by calling... >>>> >>>> ``` >>>> riak_core_util:rpc_every_member_ann(application, get_env, [riak_kv, >>>> anti_entropy_build_limit],infinity). >>>> riak_core_util:rpc_every_member_ann(application, get_env, [riak_kv, >>>> anti_entropy_concurrency],infinity). >>>> riak_core_util:rpc_every_member_ann(application, get_env, [yokozuna, >>>> anti_entropy_tick],infinity). >>>> ``` >>>> >>>> … on any of the nodes. Query coverage is R=1, but the values should be >>>> replicated across. >>>> >>>> Thanks. >>>> >>>> >>>> On Mar 12, 2015, at 9:51 AM, Roma Lakotko <ro...@lakotko.ru> wrote: >>>> >>>> Hello Zeeshan. >>>> >>>> While i run queries no delete object is occurs. >>>> >>>> Stats on production and developer nodes output something like this: >>>> https://gist.github.com/romulka/d0254aa193a9dbb52b67 >>>> >>>> On dev container: >>>> >>>> /etc/riak# grep anti_entropy * >>>> riak.conf:anti_entropy = active >>>> riak.conf.dpkg-dist:anti_entropy = active >>>> >>>> ll -h /var/lib/riak/yz_anti_entropy/ >>>> total 264K >>>> drwxrwxr-x 66 riak riak 4.0K Sep 25 12:08 ./ >>>> drwxr-xr-x 12 riak riak 4.0K Dec 9 12:19 ../ >>>> drwxr-xr-x 9 riak riak 4.0K Mar 12 12:01 0/ >>>> drwxr-xr-x 9 riak riak 4.0K Mar 12 12:01 >>>> 1004782375664995756265033322492444576013453623296/ >>>> drwxr-xr-x 9 riak riak 4.0K Mar 12 12:01 >>>> 1027618338748291114361965898003636498195577569280/ >>>> .... >>>> >>>> On prod: >>>> >>>> grep anti_entropy * /etc/riak/ -> empty >>>> >>>> root@riak-21:/var/lib/riak/yz_anti_entropy# ll -h >>>> total 64K >>>> drwxrwxr-x 16 riak riak 4.0K Dec 4 03:44 ./ >>>> drwxr-xr-x 14 riak riak 4.0K Dec 9 12:10 ../ >>>> drwxr-xr-x 9 riak riak 4.0K Dec 4 03:44 0/ >>>> drwxr-xr-x 9 riak riak 4.0K Mar 12 12:57 >>>> 1027618338748291114361965898003636498195577569280/ >>>> .... >>>> >>>> I'm already try re-save all keys, it doesn't helps. >>>> >>>> Production cluster have 7 node, start from 3. So yes, nodes was >>>> added/delete sometimes. >>>> >>>> On dev, i have 1 instance in docker container, never added to cluster. >>>> But data in that riak is imported from production cluster a while ago. >>>> >>>> I can give you a copy of container, if you need to. >>>> >>>> Thanks, >>>> Ronan Lakotko >>>> >>>> >>>> >>>> 2015-03-12 16:36 GMT+03:00 Zeeshan Lakhani <zlakh...@basho.com>: >>>> >>>>> Hello Roma, >>>>> >>>>> Have you deleted this object at some point in your runs? Please make >>>>> sure AAE is running by checking search’s AAE status, `riak-admin search >>>>> aae-status`, and that data exists in the correct directory, >>>>> `./data/yz_anti_entropy` ( >>>>> http://docs.basho.com/riak/latest/ops/advanced/configs/search/). You >>>>> may just need to perform a read-repair by performing a fetch of the object >>>>> itself first, before performing search queries again. >>>>> >>>>> Also, have you left or added nodes? I’m guessing that even your 1 >>>>> node instance is still running a cluster on that one node, right? >>>>> >>>>> Thanks. >>>>> >>>>> Zeeshan Lakhani >>>>> programmer | >>>>> software engineer at @basho | >>>>> org. member/founder of @papers_we_love | paperswelove.org >>>>> twitter => @zeeshanlakhani >>>>> >>>>> On Mar 12, 2015, at 5:59 AM, Roma Lakotko <ro...@lakotko.ru> wrote: >>>>> >>>>> Each request to riak search return different results. It's return >>>>> different numFound. >>>>> >>>>> I use request like this: >>>>> >>>>> >>>>> http://localhost:8098/search/query/assets?wt=json&q=type:*&sort=_yz_rk%20asc >>>>> >>>>> If add start offset it can return: >>>>> >>>>> >>>>> http://localhost:8098/search/query/assets?wt=json&q=type:*&sort=_yz_rk%20asc&start=1247 >>>>> >>>>> "response": { >>>>> "numFound": 1248, >>>>> "start": 1247, >>>>> "docs": [ >>>>> { >>>>> "_yz_id": >>>>> "1*default*assets*fff63ecf-a0c4-4ecf-b24d-c493ca3a302f*44", >>>>> "_yz_rk": "fff63ecf-a0c4-4ecf-b24d-c493ca3a302f", >>>>> "_yz_rt": "default", >>>>> "_yz_rb": "assets" >>>>> } >>>>> ] >>>>> } >>>>> >>>>> >>>>> On next request it return something like this >>>>> >>>>> >>>>> "numFound": 1224, >>>>> "start": 1247, >>>>> "docs": [] >>>>> >>>>> >>>>> I have 1 node installation, and no process write to Riak. >>>>> >>>>> I have same problem this production cluster with 7 nodes. >>>>> >>>>> >>>>> Scheme for document >>>>> >>>>> >>>>> <?xml version="1.0" encoding="UTF-8" ?> >>>>> <schema name="schedule" version="1.5"> >>>>> <fields> >>>>> <field name="objectId" type="string_ci" indexed="true" >>>>> stored="false" /> >>>>> <field name="type" type="string_ci" indexed="true" >>>>> stored="false" /> >>>>> <field name="objectType" type="string_ci" indexed="true" >>>>> stored="false" /> >>>>> >>>>> <field name="contentType" type="string_ci" indexed="true" >>>>> stored="false" /> >>>>> <field name="properties" type="string_ci" indexed="true" >>>>> stored="false" multiValued="true" /> >>>>> <field name="tag" type="string_ci" indexed="true" stored="false" >>>>> /> >>>>> <field name="isUploaded" type="boolean" indexed="true" >>>>> stored="false" /> >>>>> <field name="published" type="boolean" indexed="true" >>>>> stored="false" /> >>>>> <field name="drm" type="boolean" indexed="true" stored="false" >>>>> /> >>>>> <field name="dateCreated" type="date" indexed="true" stored="false" /> >>>>> >>>>> <!-- All of these fields are required by Riak Search --> >>>>> <field name="_yz_id" type="_yz_str" indexed="true" stored="true" >>>>> multiValued="false" required="true"/> >>>>> <field name="_yz_ed" type="_yz_str" indexed="true" stored="false" >>>>> multiValued="false"/> >>>>> <field name="_yz_pn" type="_yz_str" indexed="true" stored="false" >>>>> multiValued="false"/> >>>>> <field name="_yz_fpn" type="_yz_str" indexed="true" stored="false" >>>>> multiValued="false"/> >>>>> <field name="_yz_vtag" type="_yz_str" indexed="true" stored="false" >>>>> multiValued="false"/> >>>>> <field name="_yz_rk" type="_yz_str" indexed="true" stored="true" >>>>> multiValued="false"/> >>>>> <field name="_yz_rt" type="_yz_str" indexed="true" stored="true" >>>>> multiValued="false"/> >>>>> <field name="_yz_rb" type="_yz_str" indexed="true" stored="true" >>>>> multiValued="false"/> >>>>> <field name="_yz_err" type="_yz_str" indexed="true" stored="false" >>>>> multiValued="false"/> >>>>> >>>>> <dynamicField name="*" type="ignored"/> >>>>> </fields> >>>>> >>>>> <uniqueKey>_yz_id</uniqueKey> >>>>> >>>>> <types> >>>>> <!-- YZ String: Used for non-analyzed fields text_ru --> >>>>> <fieldType name="date" class="solr.TrieDateField" >>>>> sortMissingLast="true" omitNorms="true"/> >>>>> <fieldType name="double" class="solr.TrieDoubleField" >>>>> sortMissingLast="true" omitNorms="true"/> >>>>> <fieldType name="int" class="solr.TrieIntField" sortMissingLast="true" >>>>> omitNorms="true"/> >>>>> >>>>> <fieldType name="boolean" class="solr.BoolField" sortMissingLast="true" >>>>> omitNorms="true"/> >>>>> <fieldType name="_yz_str" class="solr.StrField" sortMissingLast="true" >>>>> /> >>>>> <fieldtype name="ignored" stored="false" indexed="false" >>>>> multiValued="true" class="solr.StrField" /> >>>>> <fieldType name="string_ci" class="solr.TextField" >>>>> sortMissingLast="true" omitNorms="true"> >>>>> <analyzer> >>>>> <tokenizer class="solr.StandardTokenizerFactory"/> >>>>> <filter class="solr.LowerCaseFilterFactory" /> >>>>> <filter class='solr.PatternReplaceFilterFactory' pattern='ё' >>>>> replacement='е' replace='all'/> >>>>> </analyzer> >>>>> </fieldType> >>>>> </types> >>>>> >>>>> </schema> >>>>> >>>>> >>>>> Best regards, >>>>> >>>>> Roman >>>>> >>>>> _______________________________________________ >>>>> riak-users mailing list >>>>> riak-users@lists.basho.com >>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com >>>>> >>>>> >>>>> >>>> >>>> >>> >>> >
_______________________________________________ riak-users mailing list riak-users@lists.basho.com http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com