Hello, I came across some new information regarding the original architecture. We have a file on a website that basically contains all the links of all the articles that are searchable. This file is meant to be a crawler starting point. The articles already have metadata that can be used for indexing. The data retrieval from the database is handled by something else which I currently do not have access to (so I'm not exaclty sure how it's done).
Would a crawler have to be written from scratch or would something like Nutch be useful in this case? Basically I want to build an index from the metadata of all the articles that are available. Thanks for all your help/suggestions Cheers P.S. Wasn't sure if I need a new topic for a new question, so I just used this one -- View this message in context: http://www.nabble.com/Help-with-installing-Lucene-tp21332541p21353560.html Sent from the Lucene - Java Users mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org For additional commands, e-mail: java-user-h...@lucene.apache.org