Riak.filterNotFound and Riak.reduceSort in reduce phase dramaticaly decrease performance

2014-04-17 Thread Alexander Popov
query1.sh:
=
curl -w %{time_connect}:%{time_starttransfer}:%{time_total}   -X POST \
-H "content-type: application/json" \
-d @- \
http://localhost:8098/mapred \
< 0.005:1.468:1.468


query2.sh:

curl -w %{time_connect}:%{time_starttransfer}:%{time_total} -X POST \
-H "content-type: application/json" \
-d @- \
http://localhost:8098/mapred \
<0.005:1.439:1.439



query3.sh:

curl -w %{time_connect}:%{time_starttransfer}:%{time_total} -X POST \
-H "content-type: application/json" \
-d @- \
http://localhost:8098/mapred \
< 0.005:0.218:0.218

As for me, this is not so fast too. because currently it is very simple and
 if I add some logic or  users data is grow, all things will be much
slower.



Total number of keys in bucket: 3703
Keys matched with index: 299

Environment: cluster of 3 ec2 c3.xlarge instances(debian).
storage_backend : riak_kv_eleveldb_backend
erlydtl_version : <<"0.7.0">>
riak_control_version : <<"1.4.4-0-g9a74e57">>
cluster_info_version : <<"1.2.4">>
riak_search_version : <<"1.4.8-0-gbe6e4ed">>
merge_index_version : <<"1.3.2-0-gcb38ee7">>
riak_kv_version : <<"1.4.8-0-g7545390">>
sidejob_version : <<"0.2.0">>
riak_api_version : <<"1.4.4-0-g395e6fd">>
riak_pipe_version : <<"1.4.4-0-g7f390f3">>
riak_core_version : <<"1.4.4">>
bitcask_version : <<"1.6.6-0-g230b6d6">>
basho_stats_version : <<"1.0.3">>
webmachine_version : <<"1.10.4-0-gfcff795">>
mochiweb_version : <<"1.5.1p6">>
inets_version : <<"5.9">>
erlang_js_version : <<"1.2.2">>
runtime_tools_version : <<"1.8.8">>
os_mon_version : <<"2.2.9">>
riak_sysmon_version : <<"1.1.3">>
ssl_version : <<"5.0.1">>
public_key_version : <<"0.15">>
crypto_version : <<"2.1">>
sasl_version : <<"2.2.1">>
lager_version : <<"2.0.1">>
goldrush_version : <<"0.1.5">>
compiler_version : <<"4.8.1">>
syntax_tools_version : <<"1.6.8">>
stdlib_version : <<"1.18.1">>
kernel_version : <<"2.15.1">>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Riak 2.0 search changes

2014-04-23 Thread Alexander Popov
1. Does it support wildcards in a middle or start?  *abc, a*bc,
2. Does presort supports any field instead of  key or score?

I'm not found this in 2.0 docs.
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


[no subject]

2014-04-26 Thread Alexander Popov
from documentation:

>>Currently, the wildcard must come at the end of the term in both cases,
and must be preceded by a minimum of two characters.
Does it any way to able to search with one char?

I use RiakSearch  to search contacts, and should show something for users
when they type one char only
Maybe some combination with proximity search or custom tokenizers?
Also what about riak 2.0? Does it work there?

I have just one  idea, create custom index with field first_letter  and put
first char there and when only single char comes to input  -  I will search
exact  q=first_letter:a
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


riak search presort

2014-04-29 Thread Alexander Popov
does riak support DESC order with  presort=key ?


simple presort=key desc gives me error


{error,
{error,function_clause,
[{riak_search_client,search_doc,
 [{riak_search_schema,<<"items">>,"1.1",3,<<"value">>,
  <<"id">>,
  [{riak_search_field,<<"^.*_num$">>,[],integer,10,48,false,
   false,true,
   {erlang,text_analyzers,integer_analyzer_factory},
   ["10"],
   false},
   {riak_search_field,<<"^.*_int$">>,[],integer,10,48,false,
   false,true,
   {erlang,text_analyzers,integer_analyzer_factory},
   ["10"],
   false},
   {riak_search_field,<<"^.*_dt$">>,[],date,10,48,false,
   false,true,
   {erlang,text_analyzers,noop_analyzer_factory},
   undefined,false},
   {riak_search_field,<<"^.*_date$">>,[],date,10,48,false,
   false,true,
   {erlang,text_analyzers,noop_analyzer_factory},
   undefined,false},
   {riak_search_field,<<"^.*_txt$">>,[],string,0,32,false,
   false,true,
   {erlang,text_analyzers,standard_analyzer_factory},
   undefined,false},
   {riak_search_field,<<"^.*_text$">>,[],string,0,32,false,
   false,true,
   {erlang,text_analyzers,standard_analyzer_factory},
   undefined,false},
   {riak_search_field,<<"^.*$">>,[],string,0,32,false,false,
   true,
   {erlang,text_analyzers,whitespace_analyzer_factory},
   undefined,false}],
  'or',
  {erlang,text_analyzers,whitespace_analyzer_factory}},
  {scope,#Ref<0.0.5.19421>,"items","value",
  {scope,#Ref<0.0.5.19420>,undefined,"__version",
  {term,#Ref<0.0.5.19424>,<<"1">>,
  [{'riak@127.0.0.1',162},
   {'riak@127.0.0.1',195},
   {'riak@127.0.0.1',162}],
  173.0,1.0}}},
  [],0,10,'key desc',6,
  {riak_search_client,{riak_client,'riak@127.0.0.1',undefined}}],
 [{file,"src/riak_search_client.erl"},{line,141}]},
 {riak_search_utils,run_query,7,
 [{file,"src/riak_search_utils.erl"},{line,283}]},
 {riak_solr_searcher_wm,to_xml,2,
 [{file,"src/riak_solr_searcher_wm.erl"},{line,131}]},
 {webmachine_resource,resource_call,3,
 [{file,"src/webmachine_resource.erl"},{line,186}]},
 {webmachine_resource,do,3,
 [{file,"src/webmachine_resource.erl"},{line,142}]},
 {webmachine_decision_core,resource_call,1,
 [{file,"src/webmachine_decision_core.erl"},{line,48}]},
 {webmachine_decision_core,decision,1,
 [{file,"src/webmachine_decision_core.erl"},{line,558}]},
 {webmachine_decision_core,handle_request,2,
 [{file,"src/webmachine_decision_core.erl"},{line,33}]}]}}

--
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Riak 1.4.8, Riak Search numFound inconsistency

2014-05-13 Thread Alexander Popov
solr/some_index/select?q=some_query&wt=json&rows=50

{

   - responseHeader:
   {
  - status: 0,
  - QTime: 47,
  - params:
  {
 - q: "some_query",
 - q.op: "or",
 - filter: "",
 - wt: "json"
 }
  },
   - response:
   {
  - numFound: 34,
  - start: 0,
  - maxScore: "0.353553",
  - docs: [ ]
  }

}
I was manage my index only using the Solr HTTP Interface

1. Looks i got some broken records. How can i found and cleanup it?
2. How to prevent this one in future

I was try:
1.  Reboot entire cluster( to eliminate in memory cache )
2. search-cmd delete my_index

no success
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Riak 1.4.8, Riak Search numFound inconsistency

2014-05-15 Thread Alexander Popov
I can see records ids if I add &fl=id in query:
solr/some_index/select?q=some_query&wt=json&rows=50&fl=id

{

   -
   - response:
   {
  - numFound: 33,
  - start: 0,
  - maxScore: "0.0",
  - docs:
  [
 -
 {
- id:

"1396297239219-7a5886c8d03549798f23b8c2e46739e6-660761ea9e3d472487be4e01f077f706"
},
 -
 {
- id:

"8600017235068-7a5886c8d03549798f23b8c2e46739e6-e3d0b4904161437ca47368afed842737"
},
 -
 {
- id:

"8600019503366-7a5886c8d03549798f23b8c2e46739e6-9e3990c226b44b1fb747baf6dfd868a2"
},


But when I try remove this IDs by POST to  http://db:8098/solr/some_index
/update



1396297239219-7a5886c8d03549798f23b8c2e46739e6-660761ea9e3d472487be4e01f077f706


Nothing happens. entry still returnig





On Tue, May 13, 2014 at 5:11 PM, Alexander Popov wrote:

> solr/some_index/select?q=some_query&wt=json&rows=50
>
> {
>
>- responseHeader:
>{
>   - status: 0,
>   - QTime: 47,
>   - params:
>   {
>  - q: "some_query",
>  - q.op: "or",
>  - filter: "",
>  - wt: "json"
>  }
>   },
>- response:
>{
>   - numFound: 34,
>   - start: 0,
>   - maxScore: "0.353553",
>   - docs: [ ]
>   }
>
> }
> I was manage my index only using the Solr HTTP Interface
>
> 1. Looks i got some broken records. How can i found and cleanup it?
> 2. How to prevent this one in future
>
> I was try:
> 1.  Reboot entire cluster( to eliminate in memory cache )
> 2. search-cmd delete my_index
>
> no success
>
>
>
>
>
>
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Indexing using the Solr Interface, add after delete will not applied on cluster

2014-05-21 Thread Alexander Popov
I've delete index using
http://docs.basho.com/riak/latest/dev/references/search-indexing/#Deleting-using-the-Solr-Interface
and than immediately add it by
http://docs.basho.com/riak/latest/dev/references/search-indexing/#Indexing-using-the-Solr-Interface

Item will not appears in search results in cluster setup.
On single node not seen this issue.
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Yokozuna search

2014-07-23 Thread Alexander Popov
Will queries support masks at beging and 1 char mask like  *lala and a*
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Forbidden error after upgrade on riak 2.0

2014-09-26 Thread Alexander Popov
When query Yokozuna, sometimes fails with error:

403
ForbiddenForbiddenForbiddenmochiweb+webmachine
web server

url logged with error:
/search/query/full-proof-info_index?q=owner%3A6d87f18a3dca4a60b0fc385b1f46c165%20AND%20archived%3Afalse%20AND%20NOT%20draft%3Atrue&start=0&rows=20&wt=json&sort=updated_t
imestamp%20desc


if repeat query to this url  in browser - all works fine.
No errors in logs. No clue  whats wrong
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


should solr facet.pivot works with yokozuna?

2014-10-01 Thread Alexander Popov
I provide it to query, see in response in params section:


id,owner,type


but facet output section have no pivot section






71
14
1
1







does anybody can confirm this functionality works?
probably it is disabled in SOLR?
Or need some schema setup  ?
according to
http://docs.basho.com/riak/latest/dev/advanced/search-schema/#Field-Properties-By-Use-Case
I just need indexed=true for all fields
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Proper way do delete datatypes ( map )

2014-10-02 Thread Alexander Popov
I have map bucket with some data( with neseted maps )
aprox structure( don't know does it matter or not ):
{
"update": {
"some_register": "value",
"some_flag": "disable",
"nested_map": {
"update": {
"nested1_map": {
"update": {
"some_register": "value",
"some_flag": "disable",
}
},
"nested1_map": {
"update": {
"some_register": "value",
"some_flag": "disable",
}
}
}
},
"some_counter": 13,
}
}


Updates works fine. Even simultaneous
But Sometimes  I need recreate entire value, so I Delete it using
curl -XDELETE http://host:8098/types/maps/buckets/mybucket/keys/some

After that sometimes  siblings appears.


curl -H "Accept: multipart/mixed"
http://host:8098/types/maps/buckets/mybucket/keys/some

show that confict with delete

--XZ98hy0TJbr4sVETS44XBEJf7Yt
Last-Modified: Thu, 02 Oct 2014 11:29:15 GMT

E
��A�6- some binary
--XZ98hy0TJbr4sVETS44XBEJf7Yt
Content-Type: application/octet-stream
Link: ; rel="up"
Etag: 1MqocFt6qWeQxIw8bE1B8e
Last-Modified: Thu, 02 Oct 2014 11:29:03 GMT
X-Riak-Deleted: true


--XZ98hy0TJbr4sVETS44XBEJf7Yt--


Further updates to datatype using
http://host:8098/types/maps/buckets/mybucket/datatypes/some NOT create new
sibling. It is replace previous one.

Problems:

1. I should remove datatype with different method? Or how  to resolve such
conflicts? probably
Data I receive is binary, and query like
curl -H "Accept: multipart/mixed"
http://host:8098/types/maps/buckets/mybucket/datatypes/some
to get json data not works.Should I post it as binary back?


2. I have also search index on this bucket.
  each further  updates to this datatype before resolution creates new
records in solr because
_yz_id includes sibling id:
1*maps*mybucket*some*34*46hGXxyhuW3yn3L8bRHIml

good news that when I delete record again all entries in solr deleted too.
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Proper way do delete datatypes ( map )

2014-10-02 Thread Alexander Popov
Since I use http interface I decide for now to not delete records at all.
I just empty it by
{ remove :[allmykeys], update:{timestamp}}
command.
Probably later I will implement kind of task to kill lost empty items, but
now this way work  fine for me. I will never get siblings headache and
yokozuna index is fine all time

P.s I'm realy appreciate new RIAK. It is incredible improvement.  Great
work!
02.10.2014 20:10 пользователь "Russell Brown" 
написал:

> Hi Alexander,
>
> I think you are deleting data-types the proper way.
>
> What is your `delete_mode` setting, please?
>
> I would guess that sibling you are seeing is a tombstone, which suggests
> you have some concurrent update with the delete.
>
> You will only ever have a single CRDT sibling, and 1 (or possibly more)
> tombstone siblings.
>
> If you write back just the binary CRDT sibling as the value in the normal
> K/V interface, any concurrent data type updates will merge ok to a single
> value.
>
> I don’t know about your Yokozuna question, but I think it is just that
> Yokozuna indexes each sibling as an independent document.
>
> Let me know if that works out OK for you, please?
>
> Cheers
>
> Russell
>
> On 2 Oct 2014, at 12:54, Alexander Popov  wrote:
>
> I have map bucket with some data( with neseted maps )
> aprox structure( don't know does it matter or not ):
> {
> "update": {
> "some_register": "value",
> "some_flag": "disable",
> "nested_map": {
> "update": {
> "nested1_map": {
> "update": {
> "some_register": "value",
> "some_flag": "disable",
> }
> },
> "nested1_map": {
> "update": {
> "some_register": "value",
> "some_flag": "disable",
> }
> }
> }
> },
> "some_counter": 13,
> }
> }
>
>
> Updates works fine. Even simultaneous
> But Sometimes  I need recreate entire value, so I Delete it using
> curl -XDELETE http://host:8098/types/maps/buckets/mybucket/keys/some
>
> After that sometimes  siblings appears.
>
>
> curl -H "Accept: multipart/mixed"
> http://host:8098/types/maps/buckets/mybucket/keys/some
>
> show that confict with delete
>
> --XZ98hy0TJbr4sVETS44XBEJf7Yt
> Last-Modified: Thu, 02 Oct 2014 11:29:15 GMT
>
> E
> ��A�6- some binary
> --XZ98hy0TJbr4sVETS44XBEJf7Yt
> Content-Type: application/octet-stream
> Link: ; rel="up"
> Etag: 1MqocFt6qWeQxIw8bE1B8e
> Last-Modified: Thu, 02 Oct 2014 11:29:03 GMT
> X-Riak-Deleted: true
>
>
> --XZ98hy0TJbr4sVETS44XBEJf7Yt--
>
>
> Further updates to datatype using
> http://host:8098/types/maps/buckets/mybucket/datatypes/some NOT create
> new sibling. It is replace previous one.
>
> Problems:
>
> 1. I should remove datatype with different method? Or how  to resolve such
> conflicts? probably
> Data I receive is binary, and query like
> curl -H "Accept: multipart/mixed"
> http://host:8098/types/maps/buckets/mybucket/datatypes/some
> to get json data not works.Should I post it as binary back?
>
>
> 2. I have also search index on this bucket.
>   each further  updates to this datatype before resolution creates new
> records in solr because
> _yz_id includes sibling id:
> 1*maps*mybucket*some*34*46hGXxyhuW3yn3L8bRHIml
>
> good news that when I delete record again all entries in solr deleted too.
>
>
>
>
>
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
>
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


transfer partition restarts and failed many times.

2014-10-09 Thread Alexander Popov
and seems he cannot finish transfer
part of log:

2014-10-09 07:07:27.360 [error]
<0.988.2>@riak_core_handoff_sender:start_fold:272 ownership_transfer
transfer of riak_kv_vnode from 'riak@10.0.0.218'
1370157784997721485815954530671515330927436759040 to 'ri
ak@10.0.0.152' 1370157784997721485815954530671515330927436759040 failed
because of enotconn
2014-10-09 07:07:27.360 [error]
<0.177.0>@riak_core_handoff_manager:handle_info:285 An outbound handoff of
partition riak_kv_vnode 1370157784997721485815954530671515330927436759040
was terminated for reason
: {shutdown,{error,enotconn}}
2014-10-09 07:08:30.686 [error]
<0.3648.2>@riak_core_handoff_sender:start_fold:269 ownership_transfer
transfer of riak_kv_vnode from 'riak@10.0.0.218'
913438523331814323877303020447676887284957839360 to 'ri
ak@10.0.0.152' 913438523331814323877303020447676887284957839360 failed
because of TCP recv timeout
2014-10-09 07:08:30.686 [error]
<0.177.0>@riak_core_handoff_manager:handle_info:285 An outbound handoff of
partition riak_kv_vnode 913438523331814323877303020447676887284957839360
was terminated for reason:
 {shutdown,timeout}
2014-10-09 07:08:37.386 [error]
<0.4068.2>@riak_core_handoff_sender:start_fold:272 ownership_transfer
transfer of riak_kv_vnode from 'riak@10.0.0.218'
1370157784997721485815954530671515330927436759040 to 'r
iak@10.0.0.152' 1370157784997721485815954530671515330927436759040 failed
because of enotconn
2014-10-09 07:08:37.386 [error]
<0.177.0>@riak_core_handoff_manager:handle_info:285 An outbound handoff of
partition riak_kv_vnode 1370157784997721485815954530671515330927436759040
was terminated for reason
: {shutdown,{error,enotconn}}
2014-10-09 07:09:40.646 [error]
<0.6721.2>@riak_core_handoff_sender:start_fold:269 ownership_transfer
transfer of riak_kv_vnode from 'riak@10.0.0.218'
913438523331814323877303020447676887284957839360 to 'ri
ak@10.0.0.152' 913438523331814323877303020447676887284957839360 failed
because of TCP recv timeout
2014-10-09 07:09:40.646 [error]
<0.177.0>@riak_core_handoff_manager:handle_info:285 An outbound handoff of
partition riak_kv_vnode 913438523331814323877303020447676887284957839360
was terminated for reason:
 {shutdown,timeout}
2014-10-09 07:09:47.345 [error]
<0.7132.2>@riak_core_handoff_sender:start_fold:272 ownership_transfer
transfer of riak_kv_vnode from 'riak@10.0.0.218'
1370157784997721485815954530671515330927436759040 to 'r
iak@10.0.0.152' 1370157784997721485815954530671515330927436759040 failed
because of enotconn
2014-10-09 07:09:47.345 [error]
<0.177.0>@riak_core_handoff_manager:handle_info:285 An outbound handoff of
partition riak_kv_vnode 1370157784997721485815954530671515330927436759040
was terminated for reason
: {shutdown,{error,enotconn}}
2014-10-09 07:10:50.590 [error]
<0.9901.2>@riak_core_handoff_sender:start_fold:269 ownership_transfer
transfer of riak_kv_vnode from 'riak@10.0.0.218'
913438523331814323877303020447676887284957839360 to 'ri
ak@10.0.0.152' 913438523331814323877303020447676887284957839360 failed
because of TCP recv timeout
2014-10-09 07:10:50.590 [error]
<0.177.0>@riak_core_handoff_manager:handle_info:285 An outbound handoff of
partition riak_kv_vnode 913438523331814323877303020447676887284957839360
was terminated for reason:
 {shutdown,timeout}
2014-10-09 07:10:57.412 [error]
<0.10318.2>@riak_core_handoff_sender:start_fold:272 ownership_transfer
transfer of riak_kv_vnode from 'riak@10.0.0.218'
1370157784997721485815954530671515330927436759040 to '
riak@10.0.0.152' 1370157784997721485815954530671515330927436759040 failed
because of enotconn
2014-10-09 07:10:57.413 [error]
<0.177.0>@riak_core_handoff_manager:handle_info:285 An outbound handoff of
partition riak_kv_vnode 1370157784997721485815954530671515330927436759040
was terminated for reason
: {shutdown,{error,enotconn}}
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


[SOLR] different number of result on same query

2014-10-22 Thread Alexander Popov
RIAK 2.0.1, 5 nodes on different hosts

query: 
comments_index?q=owner:6d87f18a3dca4a60b0fc385b1f46c165%20AND%20target:35975db44af44b2494751abddfcfe466&fl=id&wt=json&rows=15

RESULT1:
{
responseHeader: {
status: 0,
QTime: 3,
params: {
10.0.0.150:8093: "_yz_pn:56 OR _yz_pn:41 OR _yz_pn:26 OR _yz_pn:11",
shards: "
10.0.0.150:8093/internal_solr/comments_index,10.0.0.151:8093/internal_solr/comments_index,10.0.0.152:8093/internal_solr/comments_index,10.0.0.153:8093/internal_solr/comments_index,10.0.0.218:8093/internal_solr/comments_index
",
fl: "id",
10.0.0.152:8093: "_yz_pn:63 OR _yz_pn:53 OR _yz_pn:38 OR _yz_pn:23 OR
_yz_pn:8",
q: "owner:6d87f18a3dca4a60b0fc385b1f46c165 AND
target:35975db44af44b2494751abddfcfe466",
10.0.0.218:8093: "(_yz_pn:60 AND (_yz_fpn:60)) OR _yz_pn:50 OR _yz_pn:35 OR
_yz_pn:20 OR _yz_pn:5",
wt: "json",
10.0.0.153:8093: "_yz_pn:59 OR _yz_pn:44 OR _yz_pn:29 OR _yz_pn:14",
10.0.0.151:8093: "_yz_pn:47 OR _yz_pn:32 OR _yz_pn:17 OR _yz_pn:2",
rows: "15"
}
},
response: {
numFound: 12,
start: 0,
maxScore: 6.72534,
docs: [
.


RESULT2:
{
responseHeader: {
status: 0,
QTime: 3,
params: {
10.0.0.150:8093: "_yz_pn:61 OR _yz_pn:46 OR _yz_pn:31 OR _yz_pn:16 OR
_yz_pn:1",
shards: "
10.0.0.150:8093/internal_solr/comments_index,10.0.0.151:8093/internal_solr/comments_index,10.0.0.152:8093/internal_solr/comments_index,10.0.0.153:8093/internal_solr/comments_index,10.0.0.218:8093/internal_solr/comments_index
",
fl: "id",
10.0.0.152:8093: "_yz_pn:58 OR _yz_pn:43 OR _yz_pn:28 OR _yz_pn:13",
q: "owner:6d87f18a3dca4a60b0fc385b1f46c165 AND
target:35975db44af44b2494751abddfcfe466",
10.0.0.218:8093: "_yz_pn:55 OR _yz_pn:40 OR _yz_pn:25 OR _yz_pn:10",
wt: "json",
10.0.0.153:8093: "_yz_pn:49 OR _yz_pn:34 OR _yz_pn:19 OR _yz_pn:4",
10.0.0.151:8093: "(_yz_pn:62 AND (_yz_fpn:62)) OR _yz_pn:52 OR _yz_pn:37 OR
_yz_pn:22 OR _yz_pn:7",
rows: "15"
}
},
response: {
numFound: 11,
start: 0,
maxScore: 6.72534,
docs: [



Does it not guaranteed that all records will be found each time?
also, seems not random record missed, each time specific one
if I run query for this specific record with q=id:
50473c1239934ef29f24b87f5a6d1ca2
it is random return or not return this record
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: [SOLR] different number of result on same query

2014-10-22 Thread Alexander Popov
Also if run query with specific id, on internal_solr directly on each host,
for this problematic id it founds only on 2 hosts, but other id's found on
3 of them.

On Wed, Oct 22, 2014 at 10:18 PM, Alexander Popov 
wrote:

> RIAK 2.0.1, 5 nodes on different hosts
>
>
> query: 
> comments_index?q=owner:6d87f18a3dca4a60b0fc385b1f46c165%20AND%20target:35975db44af44b2494751abddfcfe466&fl=id&wt=json&rows=15
>
> RESULT1:
> {
> responseHeader: {
> status: 0,
> QTime: 3,
> params: {
> 10.0.0.150:8093: "_yz_pn:56 OR _yz_pn:41 OR _yz_pn:26 OR _yz_pn:11",
> shards: "
> 10.0.0.150:8093/internal_solr/comments_index,10.0.0.151:8093/internal_solr/comments_index,10.0.0.152:8093/internal_solr/comments_index,10.0.0.153:8093/internal_solr/comments_index,10.0.0.218:8093/internal_solr/comments_index
> ",
> fl: "id",
> 10.0.0.152:8093: "_yz_pn:63 OR _yz_pn:53 OR _yz_pn:38 OR _yz_pn:23 OR
> _yz_pn:8",
> q: "owner:6d87f18a3dca4a60b0fc385b1f46c165 AND
> target:35975db44af44b2494751abddfcfe466",
> 10.0.0.218:8093: "(_yz_pn:60 AND (_yz_fpn:60)) OR _yz_pn:50 OR _yz_pn:35
> OR _yz_pn:20 OR _yz_pn:5",
> wt: "json",
> 10.0.0.153:8093: "_yz_pn:59 OR _yz_pn:44 OR _yz_pn:29 OR _yz_pn:14",
> 10.0.0.151:8093: "_yz_pn:47 OR _yz_pn:32 OR _yz_pn:17 OR _yz_pn:2",
> rows: "15"
> }
> },
> response: {
> numFound: 12,
> start: 0,
> maxScore: 6.72534,
> docs: [
> .
>
>
> RESULT2:
> {
> responseHeader: {
> status: 0,
> QTime: 3,
> params: {
> 10.0.0.150:8093: "_yz_pn:61 OR _yz_pn:46 OR _yz_pn:31 OR _yz_pn:16 OR
> _yz_pn:1",
> shards: "
> 10.0.0.150:8093/internal_solr/comments_index,10.0.0.151:8093/internal_solr/comments_index,10.0.0.152:8093/internal_solr/comments_index,10.0.0.153:8093/internal_solr/comments_index,10.0.0.218:8093/internal_solr/comments_index
> ",
> fl: "id",
> 10.0.0.152:8093: "_yz_pn:58 OR _yz_pn:43 OR _yz_pn:28 OR _yz_pn:13",
> q: "owner:6d87f18a3dca4a60b0fc385b1f46c165 AND
> target:35975db44af44b2494751abddfcfe466",
> 10.0.0.218:8093: "_yz_pn:55 OR _yz_pn:40 OR _yz_pn:25 OR _yz_pn:10",
> wt: "json",
> 10.0.0.153:8093: "_yz_pn:49 OR _yz_pn:34 OR _yz_pn:19 OR _yz_pn:4",
> 10.0.0.151:8093: "(_yz_pn:62 AND (_yz_fpn:62)) OR _yz_pn:52 OR _yz_pn:37
> OR _yz_pn:22 OR _yz_pn:7",
> rows: "15"
> }
> },
> response: {
> numFound: 11,
> start: 0,
> maxScore: 6.72534,
> docs: [
> 
>
>
> Does it not guaranteed that all records will be found each time?
> also, seems not random record missed, each time specific one
> if I run query for this specific record with q=id:
> 50473c1239934ef29f24b87f5a6d1ca2
> it is random return or not return this record
>
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: [SOLR] different number of result on same query

2014-10-22 Thread Alexander Popov
Initially it was 1 node, upgraded from 1.4.8.
Than I add second node, and than added  another 3 nodes before transfer to
second is finished.

There was several hangs while transferring between nodes,  and I was
restart they  several times.

Should re-save  problematic records helps with this issue?

On Wed, Oct 22, 2014 at 10:42 PM, Eric Redmond  wrote:

> Have you added/removed nodes from the cluster, or did you start populating
> data before the cluster was done being built? You may have ran into a known
> handoff bug that we're currently working on.
>
> Eric
>
>
> On Oct 22, 2014, at 11:32 AM, Alexander Popov  wrote:
>
> Also if run query with specific id, on internal_solr directly on each
> host, for this problematic id it founds only on 2 hosts, but other id's
> found on 3 of them.
>
> On Wed, Oct 22, 2014 at 10:18 PM, Alexander Popov 
> wrote:
>
>> RIAK 2.0.1, 5 nodes on different hosts
>>
>>
>> query: 
>> comments_index?q=owner:6d87f18a3dca4a60b0fc385b1f46c165%20AND%20target:35975db44af44b2494751abddfcfe466&fl=id&wt=json&rows=15
>>
>> RESULT1:
>> {
>> responseHeader: {
>> status: 0,
>> QTime: 3,
>> params: {
>> 10.0.0.150:8093: "_yz_pn:56 OR _yz_pn:41 OR _yz_pn:26 OR _yz_pn:11",
>> shards: "
>> 10.0.0.150:8093/internal_solr/comments_index,10.0.0.151:8093/internal_solr/comments_index,10.0.0.152:8093/internal_solr/comments_index,10.0.0.153:8093/internal_solr/comments_index,10.0.0.218:8093/internal_solr/comments_index
>> ",
>> fl: "id",
>> 10.0.0.152:8093: "_yz_pn:63 OR _yz_pn:53 OR _yz_pn:38 OR _yz_pn:23 OR
>> _yz_pn:8",
>> q: "owner:6d87f18a3dca4a60b0fc385b1f46c165 AND
>> target:35975db44af44b2494751abddfcfe466",
>> 10.0.0.218:8093: "(_yz_pn:60 AND (_yz_fpn:60)) OR _yz_pn:50 OR _yz_pn:35
>> OR _yz_pn:20 OR _yz_pn:5",
>> wt: "json",
>> 10.0.0.153:8093: "_yz_pn:59 OR _yz_pn:44 OR _yz_pn:29 OR _yz_pn:14",
>> 10.0.0.151:8093: "_yz_pn:47 OR _yz_pn:32 OR _yz_pn:17 OR _yz_pn:2",
>> rows: "15"
>> }
>> },
>> response: {
>> numFound: 12,
>> start: 0,
>> maxScore: 6.72534,
>> docs: [
>> .
>>
>>
>> RESULT2:
>> {
>> responseHeader: {
>> status: 0,
>> QTime: 3,
>> params: {
>> 10.0.0.150:8093: "_yz_pn:61 OR _yz_pn:46 OR _yz_pn:31 OR _yz_pn:16 OR
>> _yz_pn:1",
>> shards: "
>> 10.0.0.150:8093/internal_solr/comments_index,10.0.0.151:8093/internal_solr/comments_index,10.0.0.152:8093/internal_solr/comments_index,10.0.0.153:8093/internal_solr/comments_index,10.0.0.218:8093/internal_solr/comments_index
>> ",
>> fl: "id",
>> 10.0.0.152:8093: "_yz_pn:58 OR _yz_pn:43 OR _yz_pn:28 OR _yz_pn:13",
>> q: "owner:6d87f18a3dca4a60b0fc385b1f46c165 AND
>> target:35975db44af44b2494751abddfcfe466",
>> 10.0.0.218:8093: "_yz_pn:55 OR _yz_pn:40 OR _yz_pn:25 OR _yz_pn:10",
>> wt: "json",
>> 10.0.0.153:8093: "_yz_pn:49 OR _yz_pn:34 OR _yz_pn:19 OR _yz_pn:4",
>> 10.0.0.151:8093: "(_yz_pn:62 AND (_yz_fpn:62)) OR _yz_pn:52 OR _yz_pn:37
>> OR _yz_pn:22 OR _yz_pn:7",
>> rows: "15"
>> }
>> },
>> response: {
>> numFound: 11,
>> start: 0,
>> maxScore: 6.72534,
>> docs: [
>> 
>>
>>
>> Does it not guaranteed that all records will be found each time?
>> also, seems not random record missed, each time specific one
>> if I run query for this specific record with q=id:
>> 50473c1239934ef29f24b87f5a6d1ca2
>> it is random return or not return this record
>>
>>
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
>
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


update schema without reindex all data

2014-11-02 Thread Alexander Popov
does it possible?

now I use flow:

curl -XPUT "http://localhost:8098/buckets/some_bucket/props";
-H'content-type: application/json'
-d'{"props":{"search_index":"_dont_index_", "search":false}}'
curl -XDELETE "http://localhost:8098/search/index/some_index";

curl -XPUT "http://localhost:8098/search/schema/some_schema";
-H'content-type:application/xml'  --data-binary @ some_schema.xml

curl -XPUT "http://localhost:8098/search/index/some_index";
-H'content-type:application/json' -d"{\"schema\":\"some_schema\"}"

curl -XPUT "http://localhosy:8098/buckets/some_bucket/props";
-H'content-type: application/json'
-d"{\"props\":{\"search_index\":\"some_index\"}}"



after that All indexed data removed, and I need resave all data.

It works when there is mid size bucket. but how I should update schema
in production when it have millions records.
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


[SOLR]Does indexing is async operatoin?

2014-11-12 Thread Alexander Popov
I put data to some bucket with indexing schema,
Immediately after that I'm  trying to query this data using search interface
very often just inserted data is missed.

Does it any way to know when it will be done? or any other ways how to
achieve this?
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


How you are dealing with spikes?

2014-12-09 Thread Alexander Popov
I have Riak 2.0.1 cluster with 5 nodes ( ec2 m3-large ) with elnm in front
sometimes I got spikes  up to 10 seconds

I can't say that I have  huge load at this time,  max 200 requests per
second for all 5 nodes.

Most expensive queries is
* list by secondary index ( usually returns from 0 to 100 records  )
* and solr queries( max 10 records )

save operations  is slowdown sometimes but not so much ( up to 1 sec )

It's slowdown not for specific requests, same one work pretty fast later.

Does it any possibilities to profile|log somehow to determine reason
why this happen?

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: How you are dealing with spikes?

2014-12-09 Thread Alexander Popov
Stats when  recent spike  happens for 15 minutes around it
 get  (826)
 save  (341)
 listByIndex  (1161)
 mapReduce  (621)  //Input is IDs list
 SOLR  (4294)

6 Solr requests was longer than 9sec ( all returns 0 rows )
4 Solr requests was longer within 4-5s ( both returns 0 rows )
11 listByIndex requests was longer than within 4-5s ( both returns 0 rows )
all another requests was less than 300ms


Sometimes more load do not make such spikes
Some graphs from  maintanance tasks:
1. http://i.imgur.com/xAE6B06.png
3 simple tasks, first 2  of them reads  all keys, decide to do
nothing and continue so just read happens, third task resave all data
in bucket.
since  rate is pretty good, some peaks happens

2. More complex task
http://i.imgur.com/7nwHb3Q.png,  it have  more serious computing, and
updating typed bucked( map ), but no peaks to 9s



sysctl -a | fgrep vm.dirty_:

vm.dirty_background_bytes = 0
vm.dirty_background_ratio = 10
vm.dirty_bytes = 0
vm.dirty_expire_centisecs = 3000
vm.dirty_ratio = 20
vm.dirty_writeback_centisecs = 500

On Tue, Dec 9, 2014 at 5:46 PM, Luke Bakken  wrote:
> Hi Alexander,
>
> Can you comment on the read vs. write load of this cluster/
>
> Could you please run the following command and reply with the output?
>
> sysctl -a | fgrep vm.dirty_
>
> We've seen cases where dirty pages get written in a synchronous manner
> all at once, causing latency spikes due to I/O blocking.
> --
> Luke Bakken
> Engineer / CSE
> lbak...@basho.com
>
>
> On Tue, Dec 9, 2014 at 4:58 AM, Alexander Popov  wrote:
>> I have Riak 2.0.1 cluster with 5 nodes ( ec2 m3-large ) with elnm in front
>> sometimes I got spikes  up to 10 seconds
>>
>> I can't say that I have  huge load at this time,  max 200 requests per
>> second for all 5 nodes.
>>
>> Most expensive queries is
>> * list by secondary index ( usually returns from 0 to 100 records  )
>> * and solr queries( max 10 records )
>>
>> save operations  is slowdown sometimes but not so much ( up to 1 sec )
>>
>> It's slowdown not for specific requests, same one work pretty fast later.
>>
>> Does it any possibilities to profile|log somehow to determine reason
>> why this happen?
>>
>> ___
>> riak-users mailing list
>> riak-users@lists.basho.com
>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: How you are dealing with spikes?

2014-12-21 Thread Alexander Popov
Left graph show counts, Right graph show times, graphs are synchronized by
time
What about SOLR requests instead 2i? should it be faster?
Or what you recommend to use for populating lists of users data? for
example now we have files,
 that have 2i like owner, so when user request his files,   we populating
buckets/files/owner_bin/user_id, If we change this query to SOLR analog ->
can we gave some boost?

Also, does key length matter for  2i performance? Does number of 2i indexes
per object matter for 2i?

On Tue, Dec 9, 2014 at 7:54 PM, Alexander Popov  wrote:

> Stats when  recent spike  happens for 15 minutes around it
>  get  (826)
>  save  (341)
>  listByIndex  (1161)
>  mapReduce  (621)  //Input is IDs list
>  SOLR  (4294)
>
> 6 Solr requests was longer than 9sec ( all returns 0 rows )
> 4 Solr requests was longer within 4-5s ( both returns 0 rows )
> 11 listByIndex requests was longer than within 4-5s ( both returns 0 rows )
> all another requests was less than 300ms
>
>
> Sometimes more load do not make such spikes
> Some graphs from  maintanance tasks:
> 1. http://i.imgur.com/xAE6B06.png
> 3 simple tasks, first 2  of them reads  all keys, decide to do
> nothing and continue so just read happens, third task resave all data
> in bucket.
> since  rate is pretty good, some peaks happens
>
> 2. More complex task
> http://i.imgur.com/7nwHb3Q.png,  it have  more serious computing, and
> updating typed bucked( map ), but no peaks to 9s
>
>
>
> sysctl -a | fgrep vm.dirty_:
>
> vm.dirty_background_bytes = 0
> vm.dirty_background_ratio = 10
> vm.dirty_bytes = 0
> vm.dirty_expire_centisecs = 3000
> vm.dirty_ratio = 20
> vm.dirty_writeback_centisecs = 500
>
> On Tue, Dec 9, 2014 at 5:46 PM, Luke Bakken  wrote:
> > Hi Alexander,
> >
> > Can you comment on the read vs. write load of this cluster/
> >
> > Could you please run the following command and reply with the output?
> >
> > sysctl -a | fgrep vm.dirty_
> >
> > We've seen cases where dirty pages get written in a synchronous manner
> > all at once, causing latency spikes due to I/O blocking.
> > --
> > Luke Bakken
> > Engineer / CSE
> > lbak...@basho.com
> >
> >
> > On Tue, Dec 9, 2014 at 4:58 AM, Alexander Popov 
> wrote:
> >> I have Riak 2.0.1 cluster with 5 nodes ( ec2 m3-large ) with elnm in
> front
> >> sometimes I got spikes  up to 10 seconds
> >>
> >> I can't say that I have  huge load at this time,  max 200 requests per
> >> second for all 5 nodes.
> >>
> >> Most expensive queries is
> >> * list by secondary index ( usually returns from 0 to 100 records  )
> >> * and solr queries( max 10 records )
> >>
> >> save operations  is slowdown sometimes but not so much ( up to 1 sec )
> >>
> >> It's slowdown not for specific requests, same one work pretty fast
> later.
> >>
> >> Does it any possibilities to profile|log somehow to determine reason
> >> why this happen?
> >>
> >> ___
> >> riak-users mailing list
> >> riak-users@lists.basho.com
> >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: How you are dealing with spikes?

2014-12-31 Thread Alexander Popov
Was replaced most 2i and MapReduce calls to SOLR requests,
seems not helps to much. Now Solr request has peaks sometimes


looking on /stats :
what is difference between search_index_latency search_query_latency


   - search_query_throughput_count: 364711,
   - search_query_throughput_one: 0,
   - search_query_fail_count: 6,
   - search_query_fail_one: 0,
   - search_query_latency_min: 0,
   - search_query_latency_max: 0,
   - search_query_latency_median: 0,
   - search_query_latency_95: 0,
   - search_query_latency_99: 0,
   - search_query_latency_999: 0,
   - search_index_throughput_count: 300612,
   - search_index_throughtput_one: 2585,
   - search_index_fail_count: 367,
   - search_index_fail_one: 12,
   - search_index_latency_min: 765,
   - search_index_latency_max: 49859,
   - search_index_latency_median: 1097,
   - search_index_latency_95: 2801,
   - search_index_latency_99: 18763,
   - search_index_latency_999: 37138,



On Mon, Dec 22, 2014 at 1:39 AM, Alexander Popov 
wrote:

> Left graph show counts, Right graph show times, graphs are synchronized by
> time
> What about SOLR requests instead 2i? should it be faster?
> Or what you recommend to use for populating lists of users data? for
> example now we have files,
>  that have 2i like owner, so when user request his files,   we populating
> buckets/files/owner_bin/user_id, If we change this query to SOLR analog ->
> can we gave some boost?
>
> Also, does key length matter for  2i performance? Does number of 2i
> indexes per object matter for 2i?
>
> On Tue, Dec 9, 2014 at 7:54 PM, Alexander Popov 
> wrote:
>
>> Stats when  recent spike  happens for 15 minutes around it
>>  get  (826)
>>  save  (341)
>>  listByIndex  (1161)
>>  mapReduce  (621)  //Input is IDs list
>>  SOLR  (4294)
>>
>> 6 Solr requests was longer than 9sec ( all returns 0 rows )
>> 4 Solr requests was longer within 4-5s ( both returns 0 rows )
>> 11 listByIndex requests was longer than within 4-5s ( both returns 0 rows
>> )
>> all another requests was less than 300ms
>>
>>
>> Sometimes more load do not make such spikes
>> Some graphs from  maintanance tasks:
>> 1. http://i.imgur.com/xAE6B06.png
>> 3 simple tasks, first 2  of them reads  all keys, decide to do
>> nothing and continue so just read happens, third task resave all data
>> in bucket.
>> since  rate is pretty good, some peaks happens
>>
>> 2. More complex task
>> http://i.imgur.com/7nwHb3Q.png,  it have  more serious computing, and
>> updating typed bucked( map ), but no peaks to 9s
>>
>>
>>
>> sysctl -a | fgrep vm.dirty_:
>>
>> vm.dirty_background_bytes = 0
>> vm.dirty_background_ratio = 10
>> vm.dirty_bytes = 0
>> vm.dirty_expire_centisecs = 3000
>> vm.dirty_ratio = 20
>> vm.dirty_writeback_centisecs = 500
>>
>> On Tue, Dec 9, 2014 at 5:46 PM, Luke Bakken  wrote:
>> > Hi Alexander,
>> >
>> > Can you comment on the read vs. write load of this cluster/
>> >
>> > Could you please run the following command and reply with the output?
>> >
>> > sysctl -a | fgrep vm.dirty_
>> >
>> > We've seen cases where dirty pages get written in a synchronous manner
>> > all at once, causing latency spikes due to I/O blocking.
>> > --
>> > Luke Bakken
>> > Engineer / CSE
>> > lbak...@basho.com
>> >
>> >
>> > On Tue, Dec 9, 2014 at 4:58 AM, Alexander Popov 
>> wrote:
>> >> I have Riak 2.0.1 cluster with 5 nodes ( ec2 m3-large ) with elnm in
>> front
>> >> sometimes I got spikes  up to 10 seconds
>> >>
>> >> I can't say that I have  huge load at this time,  max 200 requests per
>> >> second for all 5 nodes.
>> >>
>> >> Most expensive queries is
>> >> * list by secondary index ( usually returns from 0 to 100 records  )
>> >> * and solr queries( max 10 records )
>> >>
>> >> save operations  is slowdown sometimes but not so much ( up to 1 sec )
>> >>
>> >> It's slowdown not for specific requests, same one work pretty fast
>> later.
>> >>
>> >> Does it any possibilities to profile|log somehow to determine reason
>> >> why this happen?
>> >>
>> >> ___
>> >> riak-users mailing list
>> >> riak-users@lists.basho.com
>> >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>
>
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


adding nodes to cluster

2015-01-20 Thread Alexander Popov
Upgraded riak 2.0.2-> 2.0.4
trying to add nodes to cluster( was single node )
all new nodes located on another host, but multi-nodes installation from
source
each new node have config settings tuned like

nodename = riak1@10.0.0.28
platform_data_dir = /var/lib/riak1
listener.http.internal = 0.0.0.0:18098
listener.protobuf.internal = 0.0.0.0:18087
handoff.port = 18099
search.solr.port = 18093
search.solr.jmx_port = 18985
search=on

after adding this nodes to cluster for long time was  nothing was
transferred for new nodes.
I  was  down this nodes and force-remove it from cluster.

After this strange things happens.
On new nodes:
Even after restart cleaning entire data dir and start again, in folder yz
indexes somehow appears again.

logs contains only
2015-01-20 13:15:04.815 [info] <0.352.0>@riak_core:wait_for_service:483
Wait complete for service riak_kv (10 seconds)
2015-01-20 13:15:06.660 [info] <0.547.0>@yz_index:local_create:189 Created
index contacts_index with schema contacts
2015-01-20 13:15:08.683 [info] <0.547.0>@yz_index:local_create:189 Created
index users_index with schema users


Main Node now have bunch of errors  on start and  periodically after:

015-01-20 13:29:00.382 [error] <0.871.0>@riak_core_vnode:vnode_command:348
riak_kv_vnode command failed
{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_map,'-filter_unique/4-fun-1-',4,[{file,"src/ria
k_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{l
ine,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"orddict.erl"},{line,170}]}]}
2015-01-20 13:29:00.382 [error] <0.650.0>@riak_core_vnode:vnode_command:348
riak_kv_vnode command failed
{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_map,'-filter_unique/4-fun-1-',4,[{file,"src/ria
k_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{l
ine,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"orddict.erl"},{line,170}]}]}
2015-01-20 13:29:00.383 [error] <0.4520.0> gen_fsm <0.4520.0> in state
waiting_local_vnode terminated with reason: no case clause matching
{vnode_error,{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_
map,'-filter_unique/4-fun-1-',4,[{file,"src/riak_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_
dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{line,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"or..."},...]}]}}
in riak_kv_put_fsm:waiting_local_vnode/2 line 5
67
2015-01-20 13:29:00.384 [error] <0.4520.0> CRASH REPORT Process <0.4520.0>
with 0 neighbours exited with reason: no case clause matching
{vnode_error,{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_ma
p,'-filter_unique/4-fun-1-',4,[{file,"src/riak_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_dt
_map,merge,2,[{file,"src/riak_dt_map.erl"},{line,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"or..."},...]}]}}
in riak_kv_put_fsm:waiting_local_vnode/2 line 567
 in gen_fsm:terminate/7 line 622
2015-01-20 13:29:00.384 [error] <0.4521.0> gen_fsm <0.4521.0> in state
waiting_local_vnode terminated with reason: no case clause matching
{vnode_error,{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_
map,'-filter_unique/4-fun-1-',4,[{file,"src/riak_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_
dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{line,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"or..."},...]}]}}
in riak_kv_put_fsm:waiting_local_vnode/2 line 5
67
2015-01-20 13:29:00.384 [error] <0.4521.0> CRASH REPORT Process <0.4521.0>
with 0 neighbours exited with reason: no case clause matching
{vnode_error,{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_ma
p,'-filter_unique/4-fun-1-',4,[{file,"src/riak_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,3

Re: adding nodes to cluster

2015-01-20 Thread Alexander Popov
Thanks.

On Tue, Jan 20, 2015 at 5:07 PM, Russell Brown  wrote:

> FYI https://github.com/basho/riak/issues/667
>
> I’ll get to work on it at once, we might be able to hit the 2.0.5 build
> date at the end of the week.
>
> On 20 Jan 2015, at 14:04, Russell Brown  wrote:
>
> > I’ll open a ticket for it. We changed the internal structure of both
> Maps and Sets, and tested that they were backwards compatible, but clearly
> missed something.
> >
> > On 20 Jan 2015, at 13:54, Alexander Popov  wrote:
> >
> >> Upgraded riak 2.0.2-> 2.0.4
> >> trying to add nodes to cluster( was single node )
> >> all new nodes located on another host, but multi-nodes installation
> from source
> >> each new node have config settings tuned like
> >>
> >> nodename = riak1@10.0.0.28
> >> platform_data_dir = /var/lib/riak1
> >> listener.http.internal = 0.0.0.0:18098
> >> listener.protobuf.internal = 0.0.0.0:18087
> >> handoff.port = 18099
> >> search.solr.port = 18093
> >> search.solr.jmx_port = 18985
> >> search=on
> >>
> >> after adding this nodes to cluster for long time was  nothing was
> transferred for new nodes.
> >> I  was  down this nodes and force-remove it from cluster.
> >>
> >> After this strange things happens.
> >> On new nodes:
> >> Even after restart cleaning entire data dir and start again, in folder
> yz indexes somehow appears again.
> >>
> >> logs contains only
> >> 2015-01-20 13:15:04.815 [info] <0.352.0>@riak_core:wait_for_service:483
> Wait complete for service riak_kv (10 seconds)
> >> 2015-01-20 13:15:06.660 [info] <0.547.0>@yz_index:local_create:189
> Created index contacts_index with schema contacts
> >> 2015-01-20 13:15:08.683 [info] <0.547.0>@yz_index:local_create:189
> Created index users_index with schema users
> >>
> >>
> >> Main Node now have bunch of errors  on start and  periodically after:
> >>
> >> 015-01-20 13:29:00.382 [error]
> <0.871.0>@riak_core_vnode:vnode_command:348 riak_kv_vnode command failed
> {{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_map,'-filter_unique/4-fun-1-',4,[{file,"src/ria
> >>
> k_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{l
> >>
> ine,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"orddict.erl"},{line,170}]}]}
> >> 2015-01-20 13:29:00.382 [error]
> <0.650.0>@riak_core_vnode:vnode_command:348 riak_kv_vnode command failed
> {{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_map,'-filter_unique/4-fun-1-',4,[{file,"src/ria
> >>
> k_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{l
> >>
> ine,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"orddict.erl"},{line,170}]}]}
> >> 2015-01-20 13:29:00.383 [error] <0.4520.0> gen_fsm <0.4520.0> in state
> waiting_local_vnode terminated with reason: no case clause matching
> {vnode_error,{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_
> >>
> map,'-filter_unique/4-fun-1-',4,[{file,"src/riak_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_
> >>
> dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{line,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"or..."},...]}]}}
> in riak_kv_put_fsm:waiting_local_vnode/2 line 5
> >> 67
> >> 2015-01-20 13:29:00.384 [error] <0.4520.0> CRASH REPORT Process
> <0.4520.0> with 0 neighbours exited with reason: no case clause matching
> {vnode_error,{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_ma
> >>
> p,'-filter_unique/4-fun-1-',4,[{file,&quo

Re: adding nodes to cluster

2015-01-21 Thread Alexander Popov
Actually, problematic node is production. but when shit happens i was
migrate all data ( export/import) to clean setup.
so i can experiment  with old node.
It have not source installation, from package.
I will try compile it tomorrow.

On Wed, Jan 21, 2015 at 9:01 PM, Russell Brown  wrote:

> There’s a branch of riak_dt here https://github.com/basho/riak_dt/pull/111
>
> Since you’re running against a source install (I assume in development, or
> test?) would you be so kind as to try this branch and see if it solves your
> issue?
>
> Many thanks
>
> Russell
>
> On 20 Jan 2015, at 15:12, Alexander Popov  wrote:
>
> > Thanks.
> >
> > On Tue, Jan 20, 2015 at 5:07 PM, Russell Brown 
> wrote:
> > FYI https://github.com/basho/riak/issues/667
> >
> > I’ll get to work on it at once, we might be able to hit the 2.0.5 build
> date at the end of the week.
> >
> > On 20 Jan 2015, at 14:04, Russell Brown  wrote:
> >
> > > I’ll open a ticket for it. We changed the internal structure of both
> Maps and Sets, and tested that they were backwards compatible, but clearly
> missed something.
> > >
> > > On 20 Jan 2015, at 13:54, Alexander Popov  wrote:
> > >
> > >> Upgraded riak 2.0.2-> 2.0.4
> > >> trying to add nodes to cluster( was single node )
> > >> all new nodes located on another host, but multi-nodes installation
> from source
> > >> each new node have config settings tuned like
> > >>
> > >> nodename = riak1@10.0.0.28
> > >> platform_data_dir = /var/lib/riak1
> > >> listener.http.internal = 0.0.0.0:18098
> > >> listener.protobuf.internal = 0.0.0.0:18087
> > >> handoff.port = 18099
> > >> search.solr.port = 18093
> > >> search.solr.jmx_port = 18985
> > >> search=on
> > >>
> > >> after adding this nodes to cluster for long time was  nothing was
> transferred for new nodes.
> > >> I  was  down this nodes and force-remove it from cluster.
> > >>
> > >> After this strange things happens.
> > >> On new nodes:
> > >> Even after restart cleaning entire data dir and start again, in
> folder yz indexes somehow appears again.
> > >>
> > >> logs contains only
> > >> 2015-01-20 13:15:04.815 [info]
> <0.352.0>@riak_core:wait_for_service:483 Wait complete for service riak_kv
> (10 seconds)
> > >> 2015-01-20 13:15:06.660 [info] <0.547.0>@yz_index:local_create:189
> Created index contacts_index with schema contacts
> > >> 2015-01-20 13:15:08.683 [info] <0.547.0>@yz_index:local_create:189
> Created index users_index with schema users
> > >>
> > >>
> > >> Main Node now have bunch of errors  on start and  periodically after:
> > >>
> > >> 015-01-20 13:29:00.382 [error]
> <0.871.0>@riak_core_vnode:vnode_command:348 riak_kv_vnode command failed
> {{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_map,'-filter_unique/4-fun-1-',4,[{file,"src/ria
> > >>
> k_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{l
> > >>
> ine,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"orddict.erl"},{line,170}]}]}
> > >> 2015-01-20 13:29:00.382 [error]
> <0.650.0>@riak_core_vnode:vnode_command:348 riak_kv_vnode command failed
> {{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_map,'-filter_unique/4-fun-1-',4,[{file,"src/ria
> > >>
> k_dt_map.erl"},{line,466}]},{sets,fold_bucket,3,[{file,"sets.erl"},{line,313}]},{sets,fold_seg,4,[{file,"sets.erl"},{line,309}]},{sets,fold_segs,4,[{file,"sets.erl"},{line,305}]},{riak_dt_map,merge,2,[{file,"src/riak_dt_map.erl"},{l
> > >>
> ine,454}]},{riak_kv_crdt,'-merge_value/2-fun-0-',5,[{file,"src/riak_kv_crdt.erl"},{line,204}]},{orddict,update,4,[{file,"orddict.erl"},{line,170}]}]}
> > >> 2015-01-20 13:29:00.383 [error] <0.4520.0> gen_fsm <0.4520.0> in
> state waiting_local_vnode terminated with reason: no case clause matching
> {vnode_error,{{badrecord,dict},[{dict,filter_dict,2,[{file,"dict.erl"},{line,464}]},{riak_dt_
> > >>
> map,'-filt

2.0.5 compiled from source SOLR failed to start

2015-02-27 Thread Alexander Popov
>
> 2015-02-27 10:38:27.960 [info] <0.563.0>@yz_solr_proc:ensure_data_dir:219
> No solr config found, creating a new one
> 2015-02-27 10:38:27.961 [info] <0.563.0>@yz_solr_proc:init:96 Starting
> solr: "/usr/bin/java"
> ["-Djava.awt.headless=true","-Djetty.home=/var/riak-dist/riak1/bin/../lib/yokozuna-2.0.0-34-g122659d/priv/solr","
>
> -Djetty.port=18093","-Dsolr.solr.home=/var/lib/riak1/yz","-DhostContext=/internal_solr","-cp","/var/riak-dist/riak1/bin/../lib/yokozuna-2.0.0-34-g122659d/priv/solr/start.jar","-Dlog4j.configuration=file:///
>
> var/riak-dist/riak1/etc/solr-log4j.properties","-Dyz.lib.dir=/var/riak-dist/riak1/bin/../lib/yokozuna-2.0.0-34-g122659d/priv/java_lib","-d64","-Xms1g","-Xmx3g","-XX:+UseStringCache","-XX:+UseCompressedOops"
>
> ,"-Dcom.sun.management.jmxremote.port=18985","-Dcom.sun.management.jmxremote.authenticate=false","-Dcom.sun.management.jmxremote.ssl=false","org.eclipse.jetty.start.Main"]
> 2015-02-27 10:38:28.004 [info] <0.7.0> Application yokozuna started on
> node 'riak1@10.0.0.133'
> 2015-02-27 10:38:28.007 [info] <0.7.0> Application cluster_info started on
> node 'riak1@10.0.0.133'
> 2015-02-27 10:38:28.033 [info]
> <0.198.0>@riak_core_capability:process_capability_changes:555 New
> capability: {riak_control,member_info_version} = v1
> 2015-02-27 10:38:28.035 [info] <0.7.0> Application riak_control started on
> node 'riak1@10.0.0.133'
> 2015-02-27 10:38:28.035 [info] <0.7.0> Application erlydtl started on node
> 'riak1@10.0.0.133'
> 2015-02-27 10:38:28.043 [info] <0.7.0> Application riak_auth_mods started
> on node 'riak1@10.0.0.133'
> 2015-02-27 10:38:28.307 [info] <0.563.0>@yz_solr_proc:handle_info:135 solr
> stdout/err: java.io.FileNotFoundException: No XML configuration files
> specified in start.config or command line.
>
> 2015-02-27 10:38:28.308 [info] <0.563.0>@yz_solr_proc:handle_info:135 solr
> stdout/err:  at org.eclipse.jetty.start.Main.start(Main.java:502)
> at org.eclipse.jetty.start.Main.main(Main.java:96)
>
> Usage: java -jar start.jar [options] [properties] [configs]
>java -jar start.jar --help  # for more information
> 2015-02-27 10:38:28.625 [error] <0.563.0> gen_server yz_solr_proc
> terminated with reason: {"solr OS process exited",251}
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Still getting peaks on solr queries

2015-03-06 Thread Alexander Popov
I not have stable reproduce but peaks happens several times in day and
takes 10-50 sec.
some times it happens on very low traffic, so I think it not related on
traffic.
Looks like query is locked somehow.
What can I do for troubleshoot peaks? Enable addition logs or Monitoring ?
There is no errors in logs at time of the peaks
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Solr indexes becomes broken after node replacement

2015-06-08 Thread Alexander Popov
I was perform all nodes upgrade  in x6 nodes cluster
and at same time  4 of them was prooceed with node name changed by steps:
http://docs.basho.com/riak/latest/ops/running/recovery/failed-node/#Node-Name-Changed

after that search indexes becomes completely broken - many data missed,
 returns different results on every requests.

Does it have any options now except re-save  entire data in bucket to fix
it?
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Solr indexes becomes broken after node replacement

2015-06-08 Thread Alexander Popov
Does  it require remove yz_anti_entropy directory on all nodes at same time?
Does  it require to restart nodes?

On Mon, Jun 8, 2015 at 4:01 PM, Zeeshan Lakhani  wrote:

> Hello Alexander,
>
> Can you send us the logs? I’d like to see the errors that occurred after
> the time of the node name change.
>
> In terms of getting your data back into the indexes, I’d recommend letting
> AAE handle the situation by removing the `yz_anti_entropy` directory and
> letting the keys re-index. You can even remove the search index
> directories, e.g.
> https://github.com/basho/riak_test/blob/master/tests/yz_core_properties_create_unload.erl#L94,
>   and
> let the them get recreated, which will still re-index via AAE.
>
> Nonetheless, we’d like to see the logs and try and reproduce the issue in
> a test when the node name changes.
>
> Thanks.
>
> Zeeshan Lakhani
> programmer |
> software engineer at @basho
>
> On Jun 8, 2015, at 8:30 AM, Alexander Popov  wrote:
>
> I was perform all nodes upgrade  in x6 nodes cluster
> and at same time  4 of them was prooceed with node name changed by steps:
>
> http://docs.basho.com/riak/latest/ops/running/recovery/failed-node/#Node-Name-Changed
>
> after that search indexes becomes completely broken - many data missed,
>  returns different results on every requests.
>
> Does it have any options now except re-save  entire data in bucket to fix
> it?
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
>
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


How you are manage your clusters on Amazon aws

2015-06-09 Thread Alexander Popov
does it any reasonable alternatives to chef/ops works?
Use cases:
easy to scale up/scale down
Easy to manage - run commands, apply configuration to all nodes,  restart
nodes, buck up.
Monitor cluster, audit all actions, collect logs
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


riak restart does not restart yokozuka well

2015-09-14 Thread Alexander Popov
i'm doing  riak restart
and got 'ok' in answer, but node appears in shutdown state.

in proccess list i found that solr is still running, and in logs found :

2015-09-14 09:21:03.939 [info] <0.579.0>@yz_solr_proc:init:96 Starting
solr: "/usr/bin/java"
["-Djava.awt.headless=true","-Djetty.home=/usr/lib/riak/lib/yokozuna-2.1.0-0-gcb41c27/priv/solr","-Djetty.port=80
2015-09-14 09:21:04.006 [warning] <0.6.0> lager_error_logger_h dropped 6
messages in the last second that exceeded the limit of 100 messages/sec
2015-09-14 09:21:04.006 [info] <0.7.0> Application yokozuna started on node
'riak@ip-10-0-0-147.ec2.internal'
2015-09-14 09:21:04.010 [info] <0.7.0> Application cluster_info started on
node 'riak@ip-10-0-0-147.ec2.internal'
2015-09-14 09:21:04.047 [info]
<0.209.0>@riak_core_capability:process_capability_changes:555 New
capability: {riak_control,member_info_version} = v1
2015-09-14 09:21:04.051 [info] <0.7.0> Application riak_control started on
node 'riak@ip-10-0-0-147.ec2.internal'
2015-09-14 09:21:04.051 [info] <0.7.0> Application erlydtl started on node
'riak@ip-10-0-0-147.ec2.internal'
2015-09-14 09:21:04.062 [info] <0.7.0> Application riak_auth_mods started
on node 'riak@ip-10-0-0-147.ec2.internal'
2015-09-14 09:21:04.156 [info] <0.579.0>@yz_solr_proc:handle_info:135 solr
stdout/err: Error: Exception thrown by the agent.
2015-09-14 09:21:04.156 [info] <0.579.0>@yz_solr_proc:handle_info:135 solr
stdout/err: : java.rmi.server.ExportException: Port already in use: 8985;
nested exception is:.
<-->java.net.BindException: Address already in use

2015-09-14 09:21:04.476 [error] <0.579.0> gen_server yz_solr_proc
terminated with reason: {"solr OS process exited",1}
2015-09-14 09:21:04.476 [error] <0.579.0> CRASH REPORT Process yz_solr_proc
with 0 neighbours exited with reason: {"solr OS process exited",1} in
gen_server:terminate/6 line 744
2015-09-14 09:21:04.476 [error] <0.578.0> Supervisor yz_solr_sup had child
yz_solr_proc started with yz_solr_proc:start_link("/var/lib/riak/yz", 8093,
8985) at <0.579.0> exit with reason {"solr OS process e
2015-09-14 09:21:04.477 [info] <0.622.0>@yz_solr_proc:init:96 Starting
solr: "/usr/bin/java"
["-Djava.awt.headless=true","-Djetty.home=/usr/lib/riak/lib/yokozuna-2.1.0-0-gcb41c27/priv/solr","-Djetty.port=80
2015-09-14 09:21:04.710 [info] <0.622.0>@yz_solr_proc:handle_info:135 solr
stdout/err: Error: Exception thrown by the agent.
2015-09-14 09:21:04.710 [info] <0.622.0>@yz_solr_proc:handle_info:135 solr
stdout/err: : java.rmi.server.ExportException: Port already in use: 8985;
nested exception is:.
<-->java.net.BindException: Address already in use

2015-09-14 09:21:05.029 [error] <0.622.0> gen_server yz_solr_proc
terminated with reason: {"solr OS process exited",1}
2015-09-14 09:21:05.029 [error] <0.622.0> CRASH REPORT Process yz_solr_proc
with 0 neighbours exited with reason: {"solr OS process exited",1} in
gen_server:terminate/6 line 744
2015-09-14 09:21:05.030 [error] <0.578.0> Supervisor yz_solr_sup had child
yz_solr_proc started with yz_solr_proc:start_link("/var/lib/riak/yz", 8093,
8985) at <0.622.0> exit with reason {"solr OS process e
2015-09-14 09:21:05.030 [error] <0.578.0> Supervisor yz_solr_sup had child
yz_solr_proc started with yz_solr_proc:start_link("/var/lib/riak/yz", 8093,
8985) at <0.622.0> exit with reason reached_max_restart
2015-09-14 09:21:05.030 [error] <0.577.0> Supervisor yz_sup had child
yz_solr_sup started with yz_solr_sup:start_link() at <0.578.0> exit with
reason shutdown in context child_terminated
2015-09-14 09:21:05.030 [error] <0.577.0> Supervisor yz_sup had child
yz_solr_sup started with yz_solr_sup:start_link() at <0.578.0> exit with
reason reached_max_restart_intensity in context shutdown
2015-09-14 09:21:05.031 [info] <0.7.0> Application yokozuna exited with
reason: shutdown
2015-09-14 09:21:05.033 [info] <0.340.0>@riak_kv_app:prep_stop:232 Stopping
application riak_kv - marked service down.
2015-09-14 09:21:05.034 [info] <0.340.0>@riak_kv_app:prep_stop:236
Unregistered pb services
2015-09-14 09:21:05.034 [info] <0.340.0>@riak_kv_app:prep_stop:241
unregistered webmachine routes
2015-09-14 09:21:05.037 [info] <0.340.0>@riak_kv_app:prep_stop:243 all
active put FSMs completed
2015-09-14 09:21:05.038 [info] <0.415.0>@riak_kv_js_vm:terminate:237
Spidermonkey VM (pool: riak_kv_js_hook) host stopping (<0.415.0>)
2015-09-14 09:21:05.038 [info] <0.416.0>@riak_kv_js_vm:terminate:237
Spidermonkey VM (pool: riak_kv_js_hook) host stopping (<0.416.0>)
2015-09-14 09:21:05.038 [info] <0.408.0>@riak_kv_js_vm:terminate:237
Spidermonkey VM (pool: riak_kv_js_reduce) host stopping (<0.408.0>)
2015-09-14 09:21:05.038 [info] <0.410.0>@riak_kv_js_vm:terminate:237
Spidermonkey VM (pool: riak_kv_js_reduce) host stopping (<0.410.0>)
2015-09-14 09:21:05.038 [info] <0.409.0>@riak_kv_js_vm:terminate:237
Spidermonkey VM (pool: riak_kv_js_reduce) host stopping (<0.409.0>)
2015-09-14 09:21:05.038 [info] <0.412.0>@riak_kv_js_vm:terminate:237
Spidermonkey VM (p

Re: riak restart does not restart yokozuka well

2015-09-14 Thread Alexander Popov
Yes there is plenty of errors there  like
Committed before 500 {msg=GC overhead limit
exceeded,trace=java.lang.OutOfMemoryError: GC overhead limit exceeded
 null:org.eclipse.jetty.io.EofException

and so on,  this is reason why I try to restart  node

My concerns is:
* search on this node come to un-working state and not repaired itself
* halted node, requires manual  actions
* false positive report or* riak restart *


On Mon, Sep 14, 2015 at 6:22 PM, Fred Dushin  wrote:

> Check the solr logs to see why it failed to shut down.  If necessary, find
> the pid bound to port 8985 and kill it.
>
> -Fred
>
> On Sep 14, 2015, at 5:28 AM, Alexander Popov  wrote:
>
> i'm doing  riak restart
> and got 'ok' in answer, but node appears in shutdown state.
>
> in proccess list i found that solr is still running, and in logs found :
>
> 2015-09-14 09:21:03.939 [info] <0.579.0>@yz_solr_proc:init:96 Starting
> solr: "/usr/bin/java"
> ["-Djava.awt.headless=true","-Djetty.home=/usr/lib/riak/lib/yokozuna-2.1.0-0-gcb41c27/priv/solr","-Djetty.port=80
> 2015-09-14 09:21:04.006 [warning] <0.6.0> lager_error_logger_h dropped 6
> messages in the last second that exceeded the limit of 100 messages/sec
> 2015-09-14 09:21:04.006 [info] <0.7.0> Application yokozuna started on
> node 'riak@ip-10-0-0-147.ec2.internal'
> 2015-09-14 09:21:04.010 [info] <0.7.0> Application cluster_info started on
> node 'riak@ip-10-0-0-147.ec2.internal'
> 2015-09-14 09:21:04.047 [info]
> <0.209.0>@riak_core_capability:process_capability_changes:555 New
> capability: {riak_control,member_info_version} = v1
> 2015-09-14 09:21:04.051 [info] <0.7.0> Application riak_control started on
> node 'riak@ip-10-0-0-147.ec2.internal'
> 2015-09-14 09:21:04.051 [info] <0.7.0> Application erlydtl started on node
> 'riak@ip-10-0-0-147.ec2.internal'
> 2015-09-14 09:21:04.062 [info] <0.7.0> Application riak_auth_mods started
> on node 'riak@ip-10-0-0-147.ec2.internal'
> 2015-09-14 09:21:04.156 [info] <0.579.0>@yz_solr_proc:handle_info:135 solr
> stdout/err: Error: Exception thrown by the agent.
> 2015-09-14 09:21:04.156 [info] <0.579.0>@yz_solr_proc:handle_info:135 solr
> stdout/err: : java.rmi.server.ExportException: Port already in use: 8985;
> nested exception is:.
> <-->java.net.BindException: Address already in use
>
> 2015-09-14 09:21:04.476 [error] <0.579.0> gen_server yz_solr_proc
> terminated with reason: {"solr OS process exited",1}
> 2015-09-14 09:21:04.476 [error] <0.579.0> CRASH REPORT Process
> yz_solr_proc with 0 neighbours exited with reason: {"solr OS process
> exited",1} in gen_server:terminate/6 line 744
> 2015-09-14 09:21:04.476 [error] <0.578.0> Supervisor yz_solr_sup had child
> yz_solr_proc started with yz_solr_proc:start_link("/var/lib/riak/yz", 8093,
> 8985) at <0.579.0> exit with reason {"solr OS process e
> 2015-09-14 09:21:04.477 [info] <0.622.0>@yz_solr_proc:init:96 Starting
> solr: "/usr/bin/java"
> ["-Djava.awt.headless=true","-Djetty.home=/usr/lib/riak/lib/yokozuna-2.1.0-0-gcb41c27/priv/solr","-Djetty.port=80
> 2015-09-14 09:21:04.710 [info] <0.622.0>@yz_solr_proc:handle_info:135 solr
> stdout/err: Error: Exception thrown by the agent.
> 2015-09-14 09:21:04.710 [info] <0.622.0>@yz_solr_proc:handle_info:135 solr
> stdout/err: : java.rmi.server.ExportException: Port already in use: 8985;
> nested exception is:.
> <-->java.net.BindException: Address already in use
>
> 2015-09-14 09:21:05.029 [error] <0.622.0> gen_server yz_solr_proc
> terminated with reason: {"solr OS process exited",1}
> 2015-09-14 09:21:05.029 [error] <0.622.0> CRASH REPORT Process
> yz_solr_proc with 0 neighbours exited with reason: {"solr OS process
> exited",1} in gen_server:terminate/6 line 744
> 2015-09-14 09:21:05.030 [error] <0.578.0> Supervisor yz_solr_sup had child
> yz_solr_proc started with yz_solr_proc:start_link("/var/lib/riak/yz", 8093,
> 8985) at <0.622.0> exit with reason {"solr OS process e
> 2015-09-14 09:21:05.030 [error] <0.578.0> Supervisor yz_solr_sup had child
> yz_solr_proc started with yz_solr_proc:start_link("/var/lib/riak/yz", 8093,
> 8985) at <0.622.0> exit with reason reached_max_restart
> 2015-09-14 09:21:05.030 [error] <0.577.0> Supervisor yz_sup had child
> yz_solr_sup started with yz_solr_sup:start_link() at <0.578.0> exit with
> reason shutdown in context child_terminated
> 2015-09-14 09:21:05.030 [error] <0.577.0> Supervisor yz_sup had ch

limit of bucket/solr indexes

2015-11-18 Thread Alexander Popov
1. does Riak have restrictions on bucket numbers?
2. same for SOLR indexes?

Interesting not theoretical, but practical limits when it start works  bad.

Use case: we want split users bucket to bucket-per-team  model,
this bucket have also solr index on top. Can it work for  1000 teams? 10k?
100k?
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


AWS instance types

2015-12-08 Thread Alexander Popov
What to choose C4 or M4 ?

M4 benefits:
* twice more memory

C4 benefits:
* can be launched in one placement group, that potential can improve nodes
communications performance
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


2i indexes and keys request inconsistencies

2016-03-08 Thread Alexander Popov
Noticied that sometimes 2i query and all keys requesrs returns extra
records ~2% of all records.

When call this items by get request after,  it returns 404 and after that
key stops to returns in 2i and keys requests.

Does it normally or my database is corrupted?
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: 2i indexes and keys request inconsistencies

2016-03-09 Thread Alexander Popov
@Matthew

No, db activity was very low at this time, and keys returned by this
queries  was deleted long time ago ( some of them at Dec 2015 )

I got this  issue when proceed maintenance task  which touch all keys in
DB, list all keys by /keys?keys=true query, read, upgrade, save.

We have some logic depends on 2i indexes results -  ( count number of
related   keys ). But if it returns phantom keys,
 I cannot trust this data, need to double check by getting each object, but
this is more expensive operation




On Tue, Mar 8, 2016 at 10:21 PM, Matthew Von-Maszewski 
wrote:

> Is the database being actively modified during your queries?
>
> Queries can lock down a "snapshot" within leveldb.  The query operation
> can return keys that existed at the time of the snapshot, but have been
> subsequently deleted by normal operations.
>
> In such a case, the query is correct in giving you the key and the 404
> afterward is also correct.  They represent two different versions of the
> database over time.
>
> Not sure if this is a valid scenario for you or not.
>
> Matthew
>
>
> On Mar 8, 2016, at 1:22 PM, Alexander Popov  wrote:
>
> Noticied that sometimes 2i query and all keys requesrs returns extra
> records ~2% of all records.
>
> When call this items by get request after,  it returns 404 and after that
> key stops to returns in 2i and keys requests.
>
> Does it normally or my database is corrupted?
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
>
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Solr http endpoint and POST

2016-04-26 Thread Alexander Popov
Does it possible?
if yes - which encoding? form-data or multipart?

if no - does it possible to enlarge GET query length? (looks like  maximum
is around 6500)
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Update solr schema

2017-03-06 Thread Alexander Popov
After release riak 2.2  does it any new possibilities to update search
 schema without  re-index all data?
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


2.2 CPU usage

2017-03-06 Thread Alexander Popov
Did Upgrade  production cluster to  2.2.0
looks that it eat more CPU,
in idle time 10-30%  always( on 2.1.4 was  up to 10%  with same load )
in Peaks  it takes up to 100% ( on 2.1.4 was  peaks was up to 75%   )

2.2:

[image: Inline image 1]
[image: Inline image 2]

2.1.4:

[image: Inline image 4]

[image: Inline image 3]
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com