250k columns?    As in, you have a CREATE TABLE statement that would have
over 250K separate, typed fields?

On Fri, Aug 14, 2015 at 11:07 AM Ahmed Ferdous <ahmed.ferd...@ze.com> wrote:

> Hi Guys,
>
>
>
> We have designed a table to have rows with large number of columns (more
> than 250k). One of my colleagues, mistakenly ran a select on the  and that
> caused the nodes to go out of memory. I was just wondering if there are
> ways to configure Cassandra 1. To limit number of columns that can be read
> 2. To gracefully reject a read request if it appears to be consuming a lot
> of memory. Otherwise, we are leaving too much open to human mistakes.
>
>
>
> Cheers,
>
>
>
> Ahmed
>
>
>
> ------------------------------
>
> *Ahmed Ferdous*
> Systems Architect
> Corporate: 604-244-1469     Email: ahmed.ferd...@ze.com
> ------------------------------
> [image: 2015 ZEMA User Forum]
> <http://www.ze.com/events/zemauserforum2015/>
>
> *Z**E PowerGroup Inc.*
> 130 - 5920 No. Two Road, Richmond, BC, Canada V7C 4R9     Web: www.ze.com
> North America: 1-866-944-1469      Europe: 0-800-520-0193       Singapore:
> 800-130-1609
> ------------------------------
>
> ZE PowerGroup Inc. Confidentiality Notice: This e-mail and any attachments
> are for the exclusive and confidential use of the intended recipient and
> contain confidential and proprietary information. If you are not the
> intended recipient, be aware that any reading, distribution, disclosure,
> copying, printing or taking action in reliance upon this e-mail is
> prohibited. If you have received this in error, please notify us
> immediately by replying to this e-mail and promptly delete this e-mail and
> its attachments from your computer system.
>
>

Reply via email to