pthread_mutex_init(&sync_mutex, NULL);
pthread_mutex_init(&sync_mutex_tx, NULL);
pthread_cond_init(&signal_rx_condition, NULL);
- pthread_cond_init(&signal_tx_condition, NULL);
}
static void pthread_destroy_mutex(void)
pthread_mutex_destroy(&sync_mutex);
pthread_mutex_destroy(&sync_mutex_tx);
pthread_cond_destroy(&signal_rx_condition);
- pthread_cond_destroy(&signal_tx_condition);
}
static void *memset32_htonl(void *dest, u32 val, u32 size)
}
}
-static void thread_common_ops(struct ifobject *ifobject, void *bufs, pthread_mutex_t *mutexptr,
- atomic_int *spinningptr)
+static void thread_common_ops(struct ifobject *ifobject, void *bufs, pthread_mutex_t *mutexptr)
{
int ctr = 0;
int ret;
*/
pthread_mutex_lock(mutexptr);
while (ret && ctr < SOCK_RECONF_CTR) {
- atomic_store(spinningptr, 1);
+ if (ifobject->fv.vector == rx)
+ atomic_store(&spinning_rx, 1);
xsk_configure_umem(ifobject, bufs, num_frames * XSK_UMEM__DEFAULT_FRAME_SIZE);
ret = xsk_configure_socket(ifobject);
usleep(USLEEP_MAX);
ctr++;
}
- atomic_store(spinningptr, 0);
+ if (ifobject->fv.vector == rx)
+ atomic_store(&spinning_rx, 0);
pthread_mutex_unlock(mutexptr);
if (ctr >= SOCK_RECONF_CTR)
void *bufs = NULL;
if (!bidi_pass)
- thread_common_ops(ifobject, bufs, &sync_mutex_tx, &spinning_tx);
+ thread_common_ops(ifobject, bufs, &sync_mutex_tx);
while (atomic_load(&spinning_rx) && spinningrxctr < SOCK_RECONF_CTR) {
spinningrxctr++;
void *bufs = NULL;
if (!bidi_pass)
- thread_common_ops(ifobject, bufs, &sync_mutex_tx, &spinning_rx);
+ thread_common_ops(ifobject, bufs, &sync_mutex_tx);
if (stat_test_type != STAT_TEST_RX_FILL_EMPTY)
xsk_populate_fill_ring(ifobject->umem);