My site's pages have exceeded the limit of pages for Google Custom Search so many of the results are not found in our site search.
I've been reading about Lucene, Nutch, Solr, etc and I'm wondering if I'd have the requirements for running those on a single server that also runs the site (on nginx) and our mysql server. We hae 2 gigs of RAM.
I'd appreciate any suggestions for migrating to a new site search.
How many pages do you have to have exceeded the limit of a Google Custom Search (just wondering)?
I recommend using Sphinx, Lucene was fine up to a few thousand items being indexed but beyond that it was unusable, searches were unbearably slow and re-building the index would take hours.
We have Sphinx running on a Rackspace Cloud Server with 1gb of RAM alongside the rest of the services required to run the site (Apache, PHP, MySQL, Memcached etc.) and it performs great.
The website we have Sphinx running on at present has >70,000 articles, searches complete very quickly and it can rebuild its entire index in ~11 seconds. I chose Sphinx based on recommendations from other developers and the knowledge that a few big sites rely on it for their search engines (Neowin being one of them).