> On Nov 4, 2020, at 9:46 AM, Medvedkin, Vladimir
> <vladimir.medved...@intel.com> wrote:
>
> Hi Thakkar,
>
> On 03/11/2020 22:23, Dharmik Thakkar wrote:
>> Avoid code duplication by combining single and multi threaded tests
>> Also, enable support for more than 2 writers
>> Signed-off-by: Dharmik Thakkar <dharmik.thak...@arm.com>
>> Reviewed-by: Ruifeng Wang <ruifeng.w...@arm.com>
>> ---
>> app/test/test_lpm_perf.c | 359 +++++++++------------------------------
>> 1 file changed, 84 insertions(+), 275 deletions(-)
>> diff --git a/app/test/test_lpm_perf.c b/app/test/test_lpm_perf.c
>> index c8e70ec89ff5..a1485e74e77f 100644
>> --- a/app/test/test_lpm_perf.c
>> +++ b/app/test/test_lpm_perf.c
>> @@ -23,6 +23,7 @@ static struct rte_rcu_qsbr *rv;
>> static volatile uint8_t writer_done;
>> static volatile uint32_t thr_id;
>> static uint64_t gwrite_cycles;
>> +static uint32_t single_insert;
>> /* LPM APIs are not thread safe, use mutex to provide thread safety */
>> static pthread_mutex_t lpm_mutex = PTHREAD_MUTEX_INITIALIZER;
>> @@ -430,24 +431,21 @@ test_lpm_rcu_qsbr_writer(void *arg)
>> {
>> unsigned int i, j, si, ei;
>> uint64_t begin, total_cycles;
>> - uint8_t core_id = (uint8_t)((uintptr_t)arg);
>> uint32_t next_hop_add = 0xAA;
>> + bool single_writer = (single_insert == NUM_LDEPTH_ROUTE_ENTRIES) ?
>> + true : false;
>> + uint8_t pos_core = (uint8_t)((uintptr_t)arg);
>> - /* 2 writer threads are used */
>> - if (core_id % 2 == 0) {
>> - si = 0;
>> - ei = NUM_LDEPTH_ROUTE_ENTRIES / 2;
>> - } else {
>> - si = NUM_LDEPTH_ROUTE_ENTRIES / 2;
>> - ei = NUM_LDEPTH_ROUTE_ENTRIES;
>> - }
>> + si = pos_core * single_insert;
>> + ei = si + single_insert;
>
> In this case, given that you are doing
> "single_insert = NUM_LDEPTH_ROUTE_ENTRIES / j;"
> below, the number of ldepth_routes must be a multiple of the number of
> writers, so some number of routes can be skipped in the opposite case.
> Consider something like:
>
> number_of_writers = j;
> ...
> si = (pos_core * NUM_LDEPTH_ROUTE_ENTRIES)/number_of_writers;
> ei = ((pos_core + 1 ) * NUM_LDEPTH_ROUTE_ENTRIES)/number_of_writers;
>
Yes, agreed some routes can be skipped. I will update the patch with the above
changes. Thanks!
>
>> /* Measure add/delete. */
>> begin = rte_rdtsc_precise();
>> for (i = 0; i < RCU_ITERATIONS; i++) {
>> /* Add all the entries */
>> for (j = si; j < ei; j++) {
>> - pthread_mutex_lock(&lpm_mutex);
>> + if (!single_writer)
>> + pthread_mutex_lock(&lpm_mutex);
>> if (rte_lpm_add(lpm, large_ldepth_route_table[j].ip,
>> large_ldepth_route_table[j].depth,
>> next_hop_add) != 0) {
>> @@ -455,19 +453,22 @@ test_lpm_rcu_qsbr_writer(void *arg)
>> i, j);
>> goto error;
>> }
>> - pthread_mutex_unlock(&lpm_mutex);
>> + if (!single_writer)
>> + pthread_mutex_unlock(&lpm_mutex);
>> }
>> /* Delete all the entries */
>> for (j = si; j < ei; j++) {
>> - pthread_mutex_lock(&lpm_mutex);
>> + if (!single_writer)
>> + pthread_mutex_lock(&lpm_mutex);
>> if (rte_lpm_delete(lpm, large_ldepth_route_table[j].ip,
>> large_ldepth_route_table[j].depth) != 0) {
>> printf("Failed to delete iteration %d, route#
>> %d\n",
>> i, j);
>> goto error;
>> }
>> - pthread_mutex_unlock(&lpm_mutex);
>> + if (!single_writer)
>> + pthread_mutex_unlock(&lpm_mutex);
>> }
>> }
>> @@ -478,22 +479,24 @@ test_lpm_rcu_qsbr_writer(void *arg)
>> return 0;
>> error:
>> - pthread_mutex_unlock(&lpm_mutex);
>> + if (!single_writer)
>> + pthread_mutex_unlock(&lpm_mutex);
>> return -1;
>> }
>> /*
>> * Functional test:
>> - * 2 writers, rest are readers
>> + * 1/2 writers, rest are readers
>> */
>> static int
>> -test_lpm_rcu_perf_multi_writer(void)
>> +test_lpm_rcu_perf_multi_writer(uint8_t use_rcu)
>> {
>> struct rte_lpm_config config;
>> size_t sz;
>> - unsigned int i;
>> + unsigned int i, j;
>> uint16_t core_id;
>> struct rte_lpm_rcu_config rcu_cfg = {0};
>> + int (*reader_f)(void *arg) = NULL;
>> if (rte_lcore_count() < 3) {
>> printf("Not enough cores for lpm_rcu_perf_autotest, expecting
>> at least 3\n");
>> @@ -506,273 +509,79 @@ test_lpm_rcu_perf_multi_writer(void)
>> num_cores++;
>> }
>> - printf("\nPerf test: 2 writers, %d readers, RCU integration enabled\n",
>> - num_cores - 2);
>> -
>> - /* Create LPM table */
>> - config.max_rules = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.number_tbl8s = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.flags = 0;
>> - lpm = rte_lpm_create(__func__, SOCKET_ID_ANY, &config);
>> - TEST_LPM_ASSERT(lpm != NULL);
>> -
>> - /* Init RCU variable */
>> - sz = rte_rcu_qsbr_get_memsize(num_cores);
>> - rv = (struct rte_rcu_qsbr *)rte_zmalloc("rcu0", sz,
>> - RTE_CACHE_LINE_SIZE);
>> - rte_rcu_qsbr_init(rv, num_cores);
>> -
>> - rcu_cfg.v = rv;
>> - /* Assign the RCU variable to LPM */
>> - if (rte_lpm_rcu_qsbr_add(lpm, &rcu_cfg) != 0) {
>> - printf("RCU variable assignment failed\n");
>> - goto error;
>> - }
>> -
>> - writer_done = 0;
>> - __atomic_store_n(&gwrite_cycles, 0, __ATOMIC_RELAXED);
>> -
>> - __atomic_store_n(&thr_id, 0, __ATOMIC_SEQ_CST);
>> -
>> - /* Launch reader threads */
>> - for (i = 2; i < num_cores; i++)
>> - rte_eal_remote_launch(test_lpm_rcu_qsbr_reader, NULL,
>> - enabled_core_ids[i]);
>> -
>> - /* Launch writer threads */
>> - for (i = 0; i < 2; i++)
>> - rte_eal_remote_launch(test_lpm_rcu_qsbr_writer,
>> - (void *)(uintptr_t)i,
>> - enabled_core_ids[i]);
>> -
>> - /* Wait for writer threads */
>> - for (i = 0; i < 2; i++)
>> - if (rte_eal_wait_lcore(enabled_core_ids[i]) < 0)
>> - goto error;
>> -
>> - printf("Total LPM Adds: %d\n", TOTAL_WRITES);
>> - printf("Total LPM Deletes: %d\n", TOTAL_WRITES);
>> - printf("Average LPM Add/Del: %"PRIu64" cycles\n",
>> - __atomic_load_n(&gwrite_cycles, __ATOMIC_RELAXED)
>> - / TOTAL_WRITES);
>> -
>> - writer_done = 1;
>> - /* Wait until all readers have exited */
>> - for (i = 2; i < num_cores; i++)
>> - rte_eal_wait_lcore(enabled_core_ids[i]);
>> -
>> - rte_lpm_free(lpm);
>> - rte_free(rv);
>> - lpm = NULL;
>> - rv = NULL;
>> -
>> - /* Test without RCU integration */
>> - printf("\nPerf test: 2 writers, %d readers, RCU integration disabled\n",
>> - num_cores - 2);
>> -
>> - /* Create LPM table */
>> - config.max_rules = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.number_tbl8s = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.flags = 0;
>> - lpm = rte_lpm_create(__func__, SOCKET_ID_ANY, &config);
>> - TEST_LPM_ASSERT(lpm != NULL);
>> -
>> - writer_done = 0;
>> - __atomic_store_n(&gwrite_cycles, 0, __ATOMIC_RELAXED);
>> - __atomic_store_n(&thr_id, 0, __ATOMIC_SEQ_CST);
>> -
>> - /* Launch reader threads */
>> - for (i = 2; i < num_cores; i++)
>> - rte_eal_remote_launch(test_lpm_reader, NULL,
>> - enabled_core_ids[i]);
>> -
>> - /* Launch writer threads */
>> - for (i = 0; i < 2; i++)
>> - rte_eal_remote_launch(test_lpm_rcu_qsbr_writer,
>> - (void *)(uintptr_t)i,
>> - enabled_core_ids[i]);
>> -
>> - /* Wait for writer threads */
>> - for (i = 0; i < 2; i++)
>> - if (rte_eal_wait_lcore(enabled_core_ids[i]) < 0)
>> - goto error;
>> -
>> - printf("Total LPM Adds: %d\n", TOTAL_WRITES);
>> - printf("Total LPM Deletes: %d\n", TOTAL_WRITES);
>> - printf("Average LPM Add/Del: %"PRIu64" cycles\n",
>> - __atomic_load_n(&gwrite_cycles, __ATOMIC_RELAXED)
>> - / TOTAL_WRITES);
>> -
>> - writer_done = 1;
>> - /* Wait until all readers have exited */
>> - for (i = 2; i < num_cores; i++)
>> - rte_eal_wait_lcore(enabled_core_ids[i]);
>> -
>> - rte_lpm_free(lpm);
>> -
>> - return 0;
>> -
>> -error:
>> - writer_done = 1;
>> - /* Wait until all readers have exited */
>> - rte_eal_mp_wait_lcore();
>> -
>> - rte_lpm_free(lpm);
>> - rte_free(rv);
>> -
>> - return -1;
>> -}
>> -
>> -/*
>> - * Functional test:
>> - * Single writer, rest are readers
>> - */
>> -static int
>> -test_lpm_rcu_perf(void)
>> -{
>> - struct rte_lpm_config config;
>> - uint64_t begin, total_cycles;
>> - size_t sz;
>> - unsigned int i, j;
>> - uint16_t core_id;
>> - uint32_t next_hop_add = 0xAA;
>> - struct rte_lpm_rcu_config rcu_cfg = {0};
>> -
>> - if (rte_lcore_count() < 2) {
>> - printf("Not enough cores for lpm_rcu_perf_autotest, expecting
>> at least 2\n");
>> - return TEST_SKIPPED;
>> - }
>> -
>> - num_cores = 0;
>> - RTE_LCORE_FOREACH_WORKER(core_id) {
>> - enabled_core_ids[num_cores] = core_id;
>> - num_cores++;
>> - }
>> -
>> - printf("\nPerf test: 1 writer, %d readers, RCU integration enabled\n",
>> - num_cores);
>> -
>> - /* Create LPM table */
>> - config.max_rules = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.number_tbl8s = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.flags = 0;
>> - lpm = rte_lpm_create(__func__, SOCKET_ID_ANY, &config);
>> - TEST_LPM_ASSERT(lpm != NULL);
>> -
>> - /* Init RCU variable */
>> - sz = rte_rcu_qsbr_get_memsize(num_cores);
>> - rv = (struct rte_rcu_qsbr *)rte_zmalloc("rcu0", sz,
>> - RTE_CACHE_LINE_SIZE);
>> - rte_rcu_qsbr_init(rv, num_cores);
>> -
>> - rcu_cfg.v = rv;
>> - /* Assign the RCU variable to LPM */
>> - if (rte_lpm_rcu_qsbr_add(lpm, &rcu_cfg) != 0) {
>> - printf("RCU variable assignment failed\n");
>> - goto error;
>> - }
>> -
>> - writer_done = 0;
>> - __atomic_store_n(&thr_id, 0, __ATOMIC_SEQ_CST);
>> -
>> - /* Launch reader threads */
>> - for (i = 0; i < num_cores; i++)
>> - rte_eal_remote_launch(test_lpm_rcu_qsbr_reader, NULL,
>> - enabled_core_ids[i]);
>> -
>> - /* Measure add/delete. */
>> - begin = rte_rdtsc_precise();
>> - for (i = 0; i < RCU_ITERATIONS; i++) {
>> - /* Add all the entries */
>> - for (j = 0; j < NUM_LDEPTH_ROUTE_ENTRIES; j++)
>> - if (rte_lpm_add(lpm, large_ldepth_route_table[j].ip,
>> - large_ldepth_route_table[j].depth,
>> - next_hop_add) != 0) {
>> - printf("Failed to add iteration %d, route#
>> %d\n",
>> - i, j);
>> - goto error;
>> - }
>> -
>> - /* Delete all the entries */
>> - for (j = 0; j < NUM_LDEPTH_ROUTE_ENTRIES; j++)
>> - if (rte_lpm_delete(lpm, large_ldepth_route_table[j].ip,
>> - large_ldepth_route_table[j].depth) != 0) {
>> - printf("Failed to delete iteration %d, route#
>> %d\n",
>> - i, j);
>> + for (j = 1; j < 3; j++) {
>> + if (use_rcu)
>> + printf("\nPerf test: %d writer(s), %d reader(s),"
>> + " RCU integration enabled\n", j, num_cores - j);
>> + else
>> + printf("\nPerf test: %d writer(s), %d reader(s),"
>> + " RCU integration disabled\n", j, num_cores - j);
>> +
>> + /* Calculate writes by each writer */
>> + single_insert = NUM_LDEPTH_ROUTE_ENTRIES / j;
>> +
>> + /* Create LPM table */
>> + config.max_rules = NUM_LDEPTH_ROUTE_ENTRIES;
>> + config.number_tbl8s = NUM_LDEPTH_ROUTE_ENTRIES;
>> + config.flags = 0;
>> + lpm = rte_lpm_create(__func__, SOCKET_ID_ANY, &config);
>> + TEST_LPM_ASSERT(lpm != NULL);
>> +
>> + /* Init RCU variable */
>> + if (use_rcu) {
>> + sz = rte_rcu_qsbr_get_memsize(num_cores);
>> + rv = (struct rte_rcu_qsbr *)rte_zmalloc("rcu0", sz,
>> + RTE_CACHE_LINE_SIZE);
>> + rte_rcu_qsbr_init(rv, num_cores);
>> +
>> + rcu_cfg.v = rv;
>> + /* Assign the RCU variable to LPM */
>> + if (rte_lpm_rcu_qsbr_add(lpm, &rcu_cfg) != 0) {
>> + printf("RCU variable assignment failed\n");
>> goto error;
>> }
>> - }
>> - total_cycles = rte_rdtsc_precise() - begin;
>> - printf("Total LPM Adds: %d\n", TOTAL_WRITES);
>> - printf("Total LPM Deletes: %d\n", TOTAL_WRITES);
>> - printf("Average LPM Add/Del: %g cycles\n",
>> - (double)total_cycles / TOTAL_WRITES);
>> + reader_f = test_lpm_rcu_qsbr_reader;
>> + } else
>> + reader_f = test_lpm_reader;
>> - writer_done = 1;
>> - /* Wait until all readers have exited */
>> - for (i = 0; i < num_cores; i++)
>> - rte_eal_wait_lcore(enabled_core_ids[i]);
>> -
>> - rte_lpm_free(lpm);
>> - rte_free(rv);
>> - lpm = NULL;
>> - rv = NULL;
>> + writer_done = 0;
>> + __atomic_store_n(&gwrite_cycles, 0, __ATOMIC_RELAXED);
>> - /* Test without RCU integration */
>> - printf("\nPerf test: 1 writer, %d readers, RCU integration disabled\n",
>> - num_cores);
>> -
>> - /* Create LPM table */
>> - config.max_rules = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.number_tbl8s = NUM_LDEPTH_ROUTE_ENTRIES;
>> - config.flags = 0;
>> - lpm = rte_lpm_create(__func__, SOCKET_ID_ANY, &config);
>> - TEST_LPM_ASSERT(lpm != NULL);
>> + __atomic_store_n(&thr_id, 0, __ATOMIC_SEQ_CST);
>> - writer_done = 0;
>> - __atomic_store_n(&thr_id, 0, __ATOMIC_SEQ_CST);
>> + /* Launch reader threads */
>> + for (i = j; i < num_cores; i++)
>> + rte_eal_remote_launch(reader_f, NULL,
>> + enabled_core_ids[i]);
>> - /* Launch reader threads */
>> - for (i = 0; i < num_cores; i++)
>> - rte_eal_remote_launch(test_lpm_reader, NULL,
>> - enabled_core_ids[i]);
>> + /* Launch writer threads */
>> + for (i = 0; i < j; i++)
>> + rte_eal_remote_launch(test_lpm_rcu_qsbr_writer,
>> + (void *)(uintptr_t)i,
>> + enabled_core_ids[i]);
>> - /* Measure add/delete. */
>> - begin = rte_rdtsc_precise();
>> - for (i = 0; i < RCU_ITERATIONS; i++) {
>> - /* Add all the entries */
>> - for (j = 0; j < NUM_LDEPTH_ROUTE_ENTRIES; j++)
>> - if (rte_lpm_add(lpm, large_ldepth_route_table[j].ip,
>> - large_ldepth_route_table[j].depth,
>> - next_hop_add) != 0) {
>> - printf("Failed to add iteration %d, route#
>> %d\n",
>> - i, j);
>> + /* Wait for writer threads */
>> + for (i = 0; i < j; i++)
>> + if (rte_eal_wait_lcore(enabled_core_ids[i]) < 0)
>> goto error;
>> - }
>> - /* Delete all the entries */
>> - for (j = 0; j < NUM_LDEPTH_ROUTE_ENTRIES; j++)
>> - if (rte_lpm_delete(lpm, large_ldepth_route_table[j].ip,
>> - large_ldepth_route_table[j].depth) != 0) {
>> - printf("Failed to delete iteration %d, route#
>> %d\n",
>> - i, j);
>> - goto error;
>> - }
>> + printf("Total LPM Adds: %d\n", TOTAL_WRITES);
>> + printf("Total LPM Deletes: %d\n", TOTAL_WRITES);
>> + printf("Average LPM Add/Del: %"PRIu64" cycles\n",
>> + __atomic_load_n(&gwrite_cycles, __ATOMIC_RELAXED)
>> + / TOTAL_WRITES);
>> +
>> + writer_done = 1;
>> + /* Wait until all readers have exited */
>> + for (i = j; i < num_cores; i++)
>> + rte_eal_wait_lcore(enabled_core_ids[i]);
>> +
>> + rte_lpm_free(lpm);
>> + rte_free(rv);
>> + lpm = NULL;
>> + rv = NULL;
>> }
>> - total_cycles = rte_rdtsc_precise() - begin;
>> -
>> - printf("Total LPM Adds: %d\n", TOTAL_WRITES);
>> - printf("Total LPM Deletes: %d\n", TOTAL_WRITES);
>> - printf("Average LPM Add/Del: %g cycles\n",
>> - (double)total_cycles / TOTAL_WRITES);
>> -
>> - writer_done = 1;
>> - /* Wait until all readers have exited */
>> - for (i = 0; i < num_cores; i++)
>> - rte_eal_wait_lcore(enabled_core_ids[i]);
>> -
>> - rte_lpm_free(lpm);
>> return 0;
>> @@ -948,10 +757,10 @@ test_lpm_perf(void)
>> rte_lpm_delete_all(lpm);
>> rte_lpm_free(lpm);
>> - if (test_lpm_rcu_perf() < 0)
>> + if (test_lpm_rcu_perf_multi_writer(0) < 0)
>> return -1;
>> - if (test_lpm_rcu_perf_multi_writer() < 0)
>> + if (test_lpm_rcu_perf_multi_writer(1) < 0)
>> return -1;
>> return 0;
>
> --
> Regards,
> Vladimir