Jason Pruim
li...@pruimphotography.com


On Oct 27, 2011, at 4:08 PM, Nathan Nobbe wrote:

> On Mon, Oct 24, 2011 at 6:50 PM, Jason Pruim <li...@pruimphotography.com> 
> wrote:
> Now that I've managed to list 3 separate programming languages and somewhat 
> tie it back into php here's the question...
> 
> I have about 89 million records in mysql... the initial load of the page 
> takes 2 to 3 minutes, I am using pagination, so I have LIMIT's on the SQL 
> query's... But they just aren't going fast enough...
> 
> What I would like to do, is pull the data out of MySQL and store it in the 
> HTML files, and then update the HTML files once a day/week/month... I can 
> figure most of it out... BUT... How do I automatically link to the individual 
> pages?
> 
> I have the site working when you pull it from MySQL... Just the load time 
> sucks... Any suggestions on where I can pull some more info from? :)
> 
> Thanks in advance!
> 
> dial in the db schema (think keys) and queries; then investigate a reverse 
> proxy like varnish to cache the generated html.
> 
> you'll be able to handle a couple thousand requests per second against the 
> proxy in no time.
> 
> might be worth pre-generating some of the pages if they are still really slow 
> after db optimization.

Would it make sense that COUNT()'ing 89 million records would kill the 
performance? Once that was gone... It worked perfectly! 


Reply via email to