In Solr 9.2.1, is there a way to tweak solrconfig.xml so that
hl.method=original is (once again) the default?
Wouldn’t using an invariant definition work?
https://solr.apache.org/guide/solr/latest/configuration-guide/requesthandlers-searchcomponents.html#invariants
> On Nov 11, 2023, at 12:18 PM, Oakley, Craig (NIH/NLM/NCBI) [C]
> wrote:
>
> In Solr 9.2.1, is there a way to tweak solrconfig.xml so
Hi,
I have a CVS file with 200 fields and 100 million rows of historical and
latest data.
The current processing is taking 20+ hours.
The schema is liked:
...
...
In terms of hardware, I have 3 identical servers. One of them is used to
load this CSV to create a core.
What is the fastest
Are you using copy from?
On Sat, Nov 11, 2023, 2:33 PM Vince McMahon
wrote:
> Hi,
>
> I have a CVS file with 200 fields and 100 million rows of historical and
> latest data.
>
> The current processing is taking 20+ hours.
>
> The schema is liked:
>
> ...
>
>
>
> ...
>
>
>
> In terms of har
I'm not querying with catch_all at the moment, but, other developers may.
I am new. Mind sharing how it matters, esp. How it makes loading n idx
fast?
On Sat, Nov 11, 2023, 3:05 PM Benedict Holland
wrote:
> Are you using copy from?
>
> On Sat, Nov 11, 2023, 2:33 PM Vince McMahon >
> wrote:
It really depends on how you are loafing data. If you are going line by
line then it's going to be very slow. Upi shpuld load datasets like this
with a copy from. If you have any issues with your csv file though, it's
going to be a problem. Things like commas without quotes tend to make up
the most
Oh, also, this matters because copy from is a batch job that will stream
your data into the table. It's extremely fast. Your indexes are not the
problem. They are extremely efficient. The problem is likely how you are
loading the data.
So actually, how are you loading the data?
Thanks,
Ben
On Sa
On 11/11/2023 12:32, Vince McMahon wrote:
What is the fastest way to load and index this large and wide CSV file? It
is taking too long, 20+ hours, now.
I am assuming here that you are sending the CSV data directly to Solr
and letting Solr parse it into documents. If that is incorrect, pleas
>From the Solr UI, how can I tell the number of threads are set for indexing
?
On Sat, Nov 11, 2023, 5:31 PM Shawn Heisey
wrote:
> On 11/11/2023 12:32, Vince McMahon wrote:
> > What is the fastest way to load and index this large and wide CSV file?
> It
> > is taking too long, 20+ hours, now.
>
Benedict,
Thanks for your replies.
I am trying to load to Solr Core and index there, not postgres.
Would you happen to know the fastest way to load and index Solr Core?
Thanks.
On Sat, Nov 11, 2023, 4:41 PM Benedict Holland
wrote:
> Oh, also, this matters because copy from is a batch job th
Shawn,
Thanks for helping me out. Solr documentation has a lot of bells and
whistles and I am overwhelmed.
The total number of documents is 200 millions. Each line of the csv will
be a document. There are 200 million lines.
I have the 2 options on load-n-index
The current way of getting dat
Hi,
I would like to find out the unit of the QUERY./dataImport.totalTime. Is
it in mill-seconds or seconds?
[image: image.png]
12 matches
Mail list logo