Hello, On my site I paginate query results by limiting rows output to a value, say LIMIT, and then the 2nd, 3rd pages run the same query with $skip=LIMIT, $skip=(LIMIT*2) value posted back. I use the following code to "skip" these result rows, which is just fetching the next row to an unused array.
//if there are rows to skip if ($result_rows > $rows_to_skip) { while ( $rows_to_skip ) { // eat a row mysql_fetch_array($result); $rows_to_skip--; $total_results_shown++; } } Can I make this more efficient? Is there a way to eliminate this data before it leaves the mySQL server (and would it be faster)? If it makes any difference the average result row is probably around 40-50 bytes but some of these queries can return up to 850 rows.. Steve -- [EMAIL PROTECTED] ** http://mrclay.org -- PHP General Mailing List (http://www.php.net/) To unsubscribe, visit: http://www.php.net/unsub.php