@@ -3,4 +3,5 @@
odp_init
odp_queue
odp_crypto
+odp_schedule
odp_shm
@@ -6,13 +6,15 @@ AM_LDFLAGS += -static
if ODP_CUNIT_ENABLED
TESTS = ${bin_PROGRAMS}
check_PROGRAMS = ${bin_PROGRAMS}
-bin_PROGRAMS = odp_init odp_queue odp_crypto odp_shm
+bin_PROGRAMS = odp_init odp_queue odp_crypto odp_shm odp_schedule
odp_init_LDFLAGS = $(AM_LDFLAGS)
odp_queue_LDFLAGS = $(AM_LDFLAGS)
odp_crypto_CFLAGS = $(AM_CFLAGS) -I$(srcdir)/crypto
odp_crypto_LDFLAGS = $(AM_LDFLAGS)
odp_shm_CFLAGS = $(AM_CFLAGS)
odp_shm_LDFLAGS = $(AM_LDFLAGS)
+odp_schedule_CFLAGS = $(AM_CFLAGS)
+odp_schedule_LDFLAGS = $(AM_LDFLAGS)
endif
dist_odp_init_SOURCES = odp_init.c
@@ -22,3 +24,4 @@ dist_odp_crypto_SOURCES = crypto/odp_crypto_test_async_inp.c \
crypto/odp_crypto_test_rng.c \
odp_crypto.c common/odp_cunit_common.c
dist_odp_shm_SOURCES = odp_shm.c common/odp_cunit_common.c
+dist_odp_schedule_SOURCES = odp_schedule.c common/odp_cunit_common.c
new file mode 100644
@@ -0,0 +1,607 @@
+/* Copyright (c) 2014, Linaro Limited
+ * All rights reserved.
+ *
+ * SPDX-License-Identifier: BSD-3-Clause
+ */
+
+#include <odp.h>
+#include "odp_cunit_common.h"
+
+#define MAX_WORKERS_THREADS 32
+#define MSG_POOL_SIZE (4*1024*1024)
+#define QUEUES_PER_PRIO 16
+#define BUF_SIZE 64
+#define TEST_NUM_BUFS 100
+#define BURST_BUF_SIZE 4
+#define TEST_NUM_BUFS_EXCL 10000
+
+#define GLOBALS_SHM_NAME "test_globals"
+#define MSG_POOL_NAME "msg_pool"
+#define SHM_MSG_POOL_NAME "shm_msg_pool"
+#define SHM_THR_ARGS_NAME "shm_thr_args"
+
+#define ONE_Q 1
+#define MANY_QS QUEUES_PER_PRIO
+
+#define ONE_PRIO 1
+
+#define SCHD_ONE 0
+#define SCHD_MULTI 1
+
+#define DISABLE_EXCL_ATOMIC 0
+#define ENABLE_EXCL_ATOMIC 1
+
+
+/* Test global variables */
+typedef struct {
+ int core_count;
+ odp_barrier_t barrier;
+ odp_schedule_prio_t current_prio;
+ int prio_buf_count;
+ odp_ticketlock_t count_lock;
+ odp_spinlock_t atomic_lock;
+} test_globals_t;
+
+typedef struct ODP_PACKED {
+ pthrd_arg thrdarg;
+ odp_schedule_sync_t sync;
+ int num_queues;
+ int num_prio;
+ int num_bufs;
+ int num_cores;
+ int enable_schd_multi;
+ int enable_excl_atomic;
+} thread_args_t;
+
+odp_buffer_pool_t pool;
+
+static void test_schedule_wait_time(void)
+{
+ uint64_t wait_time;
+
+ wait_time = odp_schedule_wait_time(0);
+
+ wait_time = odp_schedule_wait_time(1);
+ CU_ASSERT(wait_time > 0);
+
+ wait_time = odp_schedule_wait_time((uint64_t)-1LL);
+ CU_ASSERT(wait_time > 0);
+}
+
+static void test_schedule_num_prio(void)
+{
+ int prio;
+
+ prio = odp_schedule_num_prio();
+
+ CU_ASSERT(prio > 0);
+ CU_ASSERT(prio == odp_schedule_num_prio());
+}
+
+static void *schedule_common_(void *arg)
+{
+ thread_args_t *args = (thread_args_t *)arg;
+ odp_schedule_sync_t sync;
+ int num_queues, num_prio, num_bufs, num_cores;
+ odp_shm_t shm;
+ test_globals_t *globals;
+
+ sync = args->sync;
+ num_queues = args->num_queues;
+ num_prio = args->num_prio;
+ num_bufs = args->num_bufs;
+ num_cores = args->num_cores;
+
+ shm = odp_shm_lookup(GLOBALS_SHM_NAME);
+ CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
+ globals = odp_shm_addr(shm);
+ CU_ASSERT_FATAL(globals != NULL);
+
+
+ if (num_cores == globals->core_count)
+ odp_barrier_wait(&globals->barrier);
+
+ while (1) {
+ odp_buffer_t buf;
+ odp_queue_t from;
+ int num = 0;
+ int locked;
+
+ odp_ticketlock_lock(&globals->count_lock);
+ if (globals->prio_buf_count ==
+ num_bufs * num_queues * num_prio) {
+ odp_ticketlock_unlock(&globals->count_lock);
+ break;
+ }
+ odp_ticketlock_unlock(&globals->count_lock);
+
+ if (args->enable_schd_multi) {
+ odp_buffer_t bufs[BURST_BUF_SIZE];
+ int j;
+ num = odp_schedule_multi(&from, ODP_SCHED_NO_WAIT, bufs,
+ BURST_BUF_SIZE);
+ CU_ASSERT(num >= 0);
+ CU_ASSERT(num <= BURST_BUF_SIZE);
+ if (num == 0)
+ continue;
+ for (j = 0; j < num; j++)
+ odp_buffer_free(bufs[j]);
+ } else {
+ buf = odp_schedule(&from, ODP_SCHED_NO_WAIT);
+ if (buf == ODP_BUFFER_INVALID)
+ continue;
+ num = 1;
+ odp_buffer_free(buf);
+ }
+
+ if (args->enable_excl_atomic) {
+ locked = odp_spinlock_trylock(&globals->atomic_lock);
+ CU_ASSERT(locked == 1);
+ CU_ASSERT(from != ODP_QUEUE_INVALID);
+ if (locked) {
+ int cnt;
+ uint64_t cycles = 0;
+ /* Do some work here to keep the thread busy */
+ for (cnt = 0; cnt < 1000; cnt++)
+ cycles += odp_time_cycles();
+
+ odp_spinlock_unlock(&globals->atomic_lock);
+ }
+ }
+
+ odp_ticketlock_lock(&globals->count_lock);
+ globals->prio_buf_count += num;
+
+ if (sync == ODP_SCHED_SYNC_ATOMIC)
+ odp_schedule_release_atomic();
+
+ odp_ticketlock_unlock(&globals->count_lock);
+ }
+
+ return NULL;
+}
+
+static void fill_queues(thread_args_t *args)
+{
+ odp_schedule_sync_t sync;
+ int num_queues, num_prio;
+ odp_buffer_pool_t pool;
+ int i, j, k;
+ char name[32];
+
+ sync = args->sync;
+ num_queues = args->num_queues;
+ num_prio = args->num_prio;
+
+ pool = odp_buffer_pool_lookup(MSG_POOL_NAME);
+ CU_ASSERT_FATAL(pool != ODP_BUFFER_POOL_INVALID);
+
+ for (i = 0; i < num_prio; i++) {
+ for (j = 0; j < num_queues; j++) {
+ odp_queue_t queue;
+
+ switch (sync) {
+ case ODP_SCHED_SYNC_NONE:
+ snprintf(name, sizeof(name),
+ "sched_%d_%d_n", i, j);
+ break;
+ case ODP_SCHED_SYNC_ATOMIC:
+ snprintf(name, sizeof(name),
+ "sched_%d_%d_a", i, j);
+ break;
+ case ODP_SCHED_SYNC_ORDERED:
+ snprintf(name, sizeof(name),
+ "sched_%d_%d_o", i, j);
+ break;
+ default:
+ CU_ASSERT(0);
+ break;
+ }
+
+ queue = odp_queue_lookup(name);
+ CU_ASSERT_FATAL(queue != ODP_QUEUE_INVALID);
+
+ for (k = 0; k < args->num_bufs; k++) {
+ odp_buffer_t buf;
+ buf = odp_buffer_alloc(pool);
+ CU_ASSERT(buf != ODP_BUFFER_INVALID);
+ CU_ASSERT(odp_queue_enq(queue, buf) == 0);
+ }
+ }
+ }
+}
+
+static void schedule_common(odp_schedule_sync_t sync, int num_queues,
+ int num_prio, int enable_schd_multi)
+{
+ thread_args_t args;
+ odp_shm_t shm;
+ test_globals_t *globals;
+
+ shm = odp_shm_lookup(GLOBALS_SHM_NAME);
+ CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
+ globals = odp_shm_addr(shm);
+ CU_ASSERT_FATAL(globals != NULL);
+
+ globals->current_prio = ODP_SCHED_PRIO_HIGHEST;
+ globals->prio_buf_count = 0;
+
+ args.sync = sync;
+ args.num_queues = num_queues;
+ args.num_prio = num_prio;
+ args.num_bufs = TEST_NUM_BUFS;
+ args.num_cores = 1;
+ args.enable_schd_multi = enable_schd_multi;
+ args.enable_excl_atomic = 0; /* Not needed with a single core */
+
+ fill_queues(&args);
+
+ schedule_common_(&args);
+}
+
+static void parallel_execute(odp_schedule_sync_t sync, int num_queues,
+ int num_prio, int enable_schd_multi,
+ int enable_excl_atomic)
+{
+ odp_shm_t shm;
+ test_globals_t *globals;
+ thread_args_t *thr_args;
+
+ shm = odp_shm_lookup(GLOBALS_SHM_NAME);
+ CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
+ globals = odp_shm_addr(shm);
+ CU_ASSERT_FATAL(globals != NULL);
+
+ shm = odp_shm_lookup(SHM_THR_ARGS_NAME);
+ CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
+ thr_args = odp_shm_addr(shm);
+ CU_ASSERT_FATAL(thr_args != NULL);
+
+ thr_args->sync = sync;
+ thr_args->num_queues = num_queues;
+ thr_args->num_prio = num_prio;
+ if (enable_excl_atomic)
+ thr_args->num_bufs = TEST_NUM_BUFS_EXCL;
+ else
+ thr_args->num_bufs = TEST_NUM_BUFS;
+ thr_args->num_cores = globals->core_count;
+ thr_args->enable_schd_multi = enable_schd_multi;
+ thr_args->enable_excl_atomic = enable_excl_atomic;
+
+ fill_queues(thr_args);
+
+ /* Reset buffer counters from the main thread */
+ globals->current_prio = ODP_SCHED_PRIO_HIGHEST;
+ globals->prio_buf_count = 0;
+
+ /* Create and launch worker threads */
+ thr_args->thrdarg.numthrds = globals->core_count;
+ odp_cunit_thread_create(schedule_common_, &thr_args->thrdarg);
+
+ /* Wait for worker threads to terminate */
+ odp_cunit_thread_exit(&thr_args->thrdarg);
+}
+
+/* 1 queue 1 thread ODP_SCHED_SYNC_NONE */
+static void test_schedule_1q_1t_n(void)
+{
+ schedule_common(ODP_SCHED_SYNC_NONE, ONE_Q, ONE_PRIO, SCHD_ONE);
+}
+
+/* 1 queue 1 thread ODP_SCHED_SYNC_ATOMIC */
+static void test_schedule_1q_1t_a(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_ONE);
+}
+
+/* 1 queue 1 thread ODP_SCHED_SYNC_ORDERED */
+static void test_schedule_1q_1t_o(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ORDERED, ONE_Q, ONE_PRIO, SCHD_ONE);
+}
+
+/* Many queues 1 thread ODP_SCHED_SYNC_NONE */
+static void test_schedule_mq_1t_n(void)
+{
+ /* Only one priority involved in these tests, but use
+ the same number of queues the more general case uses */
+ schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, ONE_PRIO, SCHD_ONE);
+}
+
+/* Many queues 1 thread ODP_SCHED_SYNC_ATOMIC */
+static void test_schedule_mq_1t_a(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, ONE_PRIO, SCHD_ONE);
+}
+
+/* Many queues 1 thread ODP_SCHED_SYNC_ORDERED */
+static void test_schedule_mq_1t_o(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, ONE_PRIO, SCHD_ONE);
+}
+
+/* Many queues 1 thread check priority ODP_SCHED_SYNC_NONE */
+static void test_schedule_mq_1t_prio_n(void)
+{
+ int prio = odp_schedule_num_prio();
+ schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_ONE);
+}
+
+/* Many queues 1 thread check priority ODP_SCHED_SYNC_ATOMIC */
+static void test_schedule_mq_1t_prio_a(void)
+{
+ int prio = odp_schedule_num_prio();
+ schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_ONE);
+}
+
+/* Many queues 1 thread check priority ODP_SCHED_SYNC_ORDERED */
+static void test_schedule_mq_1t_prio_o(void)
+{
+ int prio = odp_schedule_num_prio();
+ schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_ONE);
+}
+
+/* Many queues many threads check priority ODP_SCHED_SYNC_NONE */
+static void test_schedule_mq_mt_prio_n(void)
+{
+ int prio = odp_schedule_num_prio();
+ parallel_execute(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_ONE,
+ DISABLE_EXCL_ATOMIC);
+}
+
+/* Many queues many threads check priority ODP_SCHED_SYNC_ATOMIC */
+static void test_schedule_mq_mt_prio_a(void)
+{
+ int prio = odp_schedule_num_prio();
+ parallel_execute(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_ONE,
+ DISABLE_EXCL_ATOMIC);
+}
+
+/* Many queues many threads check priority ODP_SCHED_SYNC_ORDERED */
+static void test_schedule_mq_mt_prio_o(void)
+{
+ int prio = odp_schedule_num_prio();
+ parallel_execute(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_ONE,
+ DISABLE_EXCL_ATOMIC);
+}
+
+/* 1 queue many threads check exclusive access on ATOMIC queues */
+static void test_schedule_1q_mt_a_excl(void)
+{
+ parallel_execute(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_ONE,
+ ENABLE_EXCL_ATOMIC);
+}
+
+/* 1 queue 1 thread ODP_SCHED_SYNC_NONE multi */
+static void test_schedule_multi_1q_1t_n(void)
+{
+ schedule_common(ODP_SCHED_SYNC_NONE, ONE_Q, ONE_PRIO, SCHD_MULTI);
+}
+
+/* 1 queue 1 thread ODP_SCHED_SYNC_ATOMIC multi */
+static void test_schedule_multi_1q_1t_a(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_MULTI);
+}
+
+/* 1 queue 1 thread ODP_SCHED_SYNC_ORDERED multi */
+static void test_schedule_multi_1q_1t_o(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ORDERED, ONE_Q, ONE_PRIO, SCHD_MULTI);
+}
+
+/* Many queues 1 thread ODP_SCHED_SYNC_NONE multi */
+static void test_schedule_multi_mq_1t_n(void)
+{
+ /* Only one priority involved in these tests, but use
+ the same number of queues the more general case uses */
+ schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, ONE_PRIO, SCHD_MULTI);
+}
+
+/* Many queues 1 thread ODP_SCHED_SYNC_ATOMIC multi */
+static void test_schedule_multi_mq_1t_a(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, ONE_PRIO, SCHD_MULTI);
+}
+
+/* Many queues 1 thread ODP_SCHED_SYNC_ORDERED multi */
+static void test_schedule_multi_mq_1t_o(void)
+{
+ schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, ONE_PRIO, SCHD_MULTI);
+}
+
+/* Many queues 1 thread check priority ODP_SCHED_SYNC_NONE multi */
+static void test_schedule_multi_mq_1t_prio_n(void)
+{
+ int prio = odp_schedule_num_prio();
+ schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_MULTI);
+}
+
+/* Many queues 1 thread check priority ODP_SCHED_SYNC_ATOMIC multi */
+static void test_schedule_multi_mq_1t_prio_a(void)
+{
+ int prio = odp_schedule_num_prio();
+ schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_MULTI);
+}
+
+/* Many queues 1 thread check priority ODP_SCHED_SYNC_ORDERED multi */
+static void test_schedule_multi_mq_1t_prio_o(void)
+{
+ int prio = odp_schedule_num_prio();
+ schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_MULTI);
+}
+
+/* Many queues many threads check priority ODP_SCHED_SYNC_NONE multi */
+static void test_schedule_multi_mq_mt_prio_n(void)
+{
+ int prio = odp_schedule_num_prio();
+ parallel_execute(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_MULTI, 0);
+}
+
+/* Many queues many threads check priority ODP_SCHED_SYNC_ATOMIC multi */
+static void test_schedule_multi_mq_mt_prio_a(void)
+{
+ int prio = odp_schedule_num_prio();
+ parallel_execute(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_MULTI, 0);
+}
+
+/* Many queues many threads check priority ODP_SCHED_SYNC_ORDERED multi */
+static void test_schedule_multi_mq_mt_prio_o(void)
+{
+ int prio = odp_schedule_num_prio();
+ parallel_execute(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_MULTI, 0);
+}
+
+/* 1 queue many threads check exclusive access on ATOMIC queues multi */
+static void test_schedule_multi_1q_mt_a_excl(void)
+{
+ parallel_execute(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_MULTI,
+ ENABLE_EXCL_ATOMIC);
+}
+
+static int create_queues(void)
+{
+ int i, j, prios;
+
+ prios = odp_schedule_num_prio();
+
+ for (i = 0; i < prios; i++) {
+ odp_queue_param_t p;
+ p.sched.prio = i;
+ p.sched.group = ODP_SCHED_GROUP_DEFAULT;
+
+ for (j = 0; j < QUEUES_PER_PRIO; j++) {
+ /* Per sched sync type */
+ char name[32];
+ odp_queue_t q;
+
+ snprintf(name, sizeof(name), "sched_%d_%d_n", i, j);
+ p.sched.sync = ODP_SCHED_SYNC_NONE;
+ q = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED, &p);
+
+ if (q == ODP_QUEUE_INVALID) {
+ printf("Schedule queue create failed.\n");
+ return -1;
+ }
+
+ snprintf(name, sizeof(name), "sched_%d_%d_a", i, j);
+ p.sched.sync = ODP_SCHED_SYNC_ATOMIC;
+ q = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED, &p);
+
+ if (q == ODP_QUEUE_INVALID) {
+ printf("Schedule queue create failed.\n");
+ return -1;
+ }
+
+ snprintf(name, sizeof(name), "sched_%d_%d_o", i, j);
+ p.sched.sync = ODP_SCHED_SYNC_ORDERED;
+ q = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED, &p);
+
+ if (q == ODP_QUEUE_INVALID) {
+ printf("Schedule queue create failed.\n");
+ return -1;
+ }
+ }
+ }
+
+ return 0;
+}
+
+static int schd_suite_init(void)
+{
+ odp_shm_t shm;
+ void *pool_base;
+ odp_buffer_pool_t pool;
+ test_globals_t *globals;
+ thread_args_t *thr_args;
+
+ shm = odp_shm_reserve(SHM_MSG_POOL_NAME, MSG_POOL_SIZE,
+ ODP_CACHE_LINE_SIZE, 0);
+ pool_base = odp_shm_addr(shm);
+ if (pool_base == NULL) {
+ printf("Shared memory reserve failed.\n");
+ return -1;
+ }
+
+ pool = odp_buffer_pool_create(MSG_POOL_NAME, pool_base, MSG_POOL_SIZE,
+ BUF_SIZE, ODP_CACHE_LINE_SIZE,
+ ODP_BUFFER_TYPE_RAW);
+ if (pool == ODP_BUFFER_POOL_INVALID) {
+ printf("Pool creation failed (msg).\n");
+ return -1;
+ }
+
+ shm = odp_shm_reserve(GLOBALS_SHM_NAME,
+ sizeof(test_globals_t), ODP_CACHE_LINE_SIZE, 0);
+
+ globals = odp_shm_addr(shm);
+
+ if (globals == NULL) {
+ printf("Shared memory reserve failed (globals).\n");
+ return -1;
+ }
+
+ memset(globals, 0, sizeof(test_globals_t));
+
+ globals->core_count = odp_sys_core_count();
+ if (globals->core_count > MAX_WORKERS)
+ globals->core_count = MAX_WORKERS;
+
+ shm = odp_shm_reserve(SHM_THR_ARGS_NAME, sizeof(thread_args_t),
+ ODP_CACHE_LINE_SIZE, 0);
+ thr_args = odp_shm_addr(shm);
+
+ if (thr_args == NULL) {
+ printf("Shared memory reserve failed (thr_args).\n");
+ return -1;
+ }
+
+ memset(thr_args, 0, sizeof(thread_args_t));
+
+ /* Barrier to sync test case execution */
+ odp_barrier_init(&globals->barrier, globals->core_count);
+ odp_ticketlock_init(&globals->count_lock);
+ odp_spinlock_init(&globals->atomic_lock);
+
+ if (create_queues() != 0)
+ return -1;
+
+ return 0;
+}
+
+struct CU_TestInfo test_odp_schedule[] = {
+ {"schedule_wait_time", test_schedule_wait_time},
+ {"schedule_num_prio", test_schedule_num_prio},
+ {"schedule_1q_1t_n", test_schedule_1q_1t_n},
+ {"schedule_1q_1t_a", test_schedule_1q_1t_a},
+ {"schedule_1q_1t_o", test_schedule_1q_1t_o},
+ {"schedule_mq_1t_n", test_schedule_mq_1t_n},
+ {"schedule_mq_1t_a", test_schedule_mq_1t_a},
+ {"schedule_mq_1t_o", test_schedule_mq_1t_o},
+ {"schedule_mq_1t_prio_n", test_schedule_mq_1t_prio_n},
+ {"schedule_mq_1t_prio_a", test_schedule_mq_1t_prio_a},
+ {"schedule_mq_1t_prio_o", test_schedule_mq_1t_prio_o},
+ {"schedule_mq_mt_prio_n", test_schedule_mq_mt_prio_n},
+ {"schedule_mq_mt_prio_a", test_schedule_mq_mt_prio_a},
+ {"schedule_mq_mt_prio_o", test_schedule_mq_mt_prio_o},
+ {"schedule_1q_mt_a_excl", test_schedule_1q_mt_a_excl},
+ {"schedule_multi_1q_1t_n", test_schedule_multi_1q_1t_n},
+ {"schedule_multi_1q_1t_a", test_schedule_multi_1q_1t_a},
+ {"schedule_multi_1q_1t_o", test_schedule_multi_1q_1t_o},
+ {"schedule_multi_mq_1t_n", test_schedule_multi_mq_1t_n},
+ {"schedule_multi_mq_1t_a", test_schedule_multi_mq_1t_a},
+ {"schedule_multi_mq_1t_o", test_schedule_multi_mq_1t_o},
+ {"schedule_multi_mq_1t_prio_n", test_schedule_multi_mq_1t_prio_n},
+ {"schedule_multi_mq_1t_prio_a", test_schedule_multi_mq_1t_prio_a},
+ {"schedule_multi_mq_1t_prio_o", test_schedule_multi_mq_1t_prio_o},
+ {"schedule_multi_mq_mt_prio_n", test_schedule_multi_mq_mt_prio_n},
+ {"schedule_multi_mq_mt_prio_a", test_schedule_multi_mq_mt_prio_a},
+ {"schedule_multi_mq_mt_prio_o", test_schedule_multi_mq_mt_prio_o},
+ {"schedule_multi_1q_mt_a_excl", test_schedule_multi_1q_mt_a_excl},
+ CU_TEST_INFO_NULL,
+};
+
+CU_SuiteInfo odp_testsuites[] = {
+ {"Scheduler", schd_suite_init, NULL, NULL, NULL, test_odp_schedule},
+ CU_SUITE_INFO_NULL,
+};
Signed-off-by: Ciprian Barbu <ciprian.barbu@linaro.org> --- v4: - fixes after Jerin's comments - removed tests_global_init and made it suite init function v3: - changes after Mike's review - removed duplicate check of end of test in schedule_common_ v2: - rebased against ODP tip - fixed some bugs - added some defines to clearly see the testcase parameters test/validation/.gitignore | 1 + test/validation/Makefile.am | 5 +- test/validation/odp_schedule.c | 607 +++++++++++++++++++++++++++++++++++++++++ 3 files changed, 612 insertions(+), 1 deletion(-) create mode 100644 test/validation/odp_schedule.c