[PATCH 1/2] event/cnxk: fix burst timer arm routine
Jerin Jacob
jerinjacobk at gmail.com
Tue Feb 7 10:27:17 CET 2023
On Thu, Feb 2, 2023 at 1:40 PM <pbhagavatula at marvell.com> wrote:
>
> From: Pavan Nikhilesh <pbhagavatula at marvell.com>
>
> Fix timer burst arm routine writing improper updates to
> the bucket.
>
> Fixes: 5f644e1bd14c ("event/cnxk: add timer arm timeout burst")
>
> Signed-off-by: Pavan Nikhilesh <pbhagavatula at marvell.com>
> ---
Series applied to dpdk-next-net-eventdev/for-main. Thanks
> drivers/event/cnxk/cnxk_tim_worker.h | 81 ++++++++++++++--------------
> 1 file changed, 42 insertions(+), 39 deletions(-)
>
> diff --git a/drivers/event/cnxk/cnxk_tim_worker.h b/drivers/event/cnxk/cnxk_tim_worker.h
> index eda84c6f31..6be31f6f9d 100644
> --- a/drivers/event/cnxk/cnxk_tim_worker.h
> +++ b/drivers/event/cnxk/cnxk_tim_worker.h
> @@ -270,7 +270,8 @@ cnxk_tim_add_entry_sp(struct cnxk_tim_ring *const tim_ring,
> } while (hbt_state & BIT_ULL(33));
> #endif
>
> - if (!(hbt_state & BIT_ULL(34))) {
> + if (!(hbt_state & BIT_ULL(34)) ||
> + !(hbt_state & GENMASK(31, 0))) {
> cnxk_tim_bkt_dec_lock(bkt);
> goto __retry;
> }
> @@ -352,7 +353,8 @@ cnxk_tim_add_entry_mp(struct cnxk_tim_ring *const tim_ring,
> } while (hbt_state & BIT_ULL(33));
> #endif
>
> - if (!(hbt_state & BIT_ULL(34))) {
> + if (!(hbt_state & BIT_ULL(34)) ||
> + !(hbt_state & GENMASK(31, 0))) {
> cnxk_tim_bkt_dec_lock(bkt);
> goto __retry;
> }
> @@ -449,10 +451,10 @@ cnxk_tim_add_entry_brst(struct cnxk_tim_ring *const tim_ring,
> struct cnxk_tim_ent *chunk = NULL;
> struct cnxk_tim_bkt *mirr_bkt;
> struct cnxk_tim_bkt *bkt;
> - uint16_t chunk_remainder;
> + int16_t chunk_remainder;
> uint16_t index = 0;
> uint64_t lock_sema;
> - int16_t rem, crem;
> + int16_t rem;
> uint8_t lock_cnt;
>
> __retry:
> @@ -460,31 +462,6 @@ cnxk_tim_add_entry_brst(struct cnxk_tim_ring *const tim_ring,
>
> /* Only one thread beyond this. */
> lock_sema = cnxk_tim_bkt_inc_lock(bkt);
> - lock_cnt = (uint8_t)((lock_sema >> TIM_BUCKET_W1_S_LOCK) &
> - TIM_BUCKET_W1_M_LOCK);
> -
> - if (lock_cnt) {
> - cnxk_tim_bkt_dec_lock(bkt);
> -#ifdef RTE_ARCH_ARM64
> - asm volatile(PLT_CPU_FEATURE_PREAMBLE
> - " ldxrb %w[lock_cnt], [%[lock]] \n"
> - " tst %w[lock_cnt], 255 \n"
> - " beq dne%= \n"
> - " sevl \n"
> - "rty%=: wfe \n"
> - " ldxrb %w[lock_cnt], [%[lock]] \n"
> - " tst %w[lock_cnt], 255 \n"
> - " bne rty%= \n"
> - "dne%=: \n"
> - : [lock_cnt] "=&r"(lock_cnt)
> - : [lock] "r"(&bkt->lock)
> - : "memory");
> -#else
> - while (__atomic_load_n(&bkt->lock, __ATOMIC_RELAXED))
> - ;
> -#endif
> - goto __retry;
> - }
>
> /* Bucket related checks. */
> if (unlikely(cnxk_tim_bkt_get_hbt(lock_sema))) {
> @@ -509,21 +486,46 @@ cnxk_tim_add_entry_brst(struct cnxk_tim_ring *const tim_ring,
> } while (hbt_state & BIT_ULL(33));
> #endif
>
> - if (!(hbt_state & BIT_ULL(34))) {
> + if (!(hbt_state & BIT_ULL(34)) ||
> + !(hbt_state & GENMASK(31, 0))) {
> cnxk_tim_bkt_dec_lock(bkt);
> goto __retry;
> }
> }
> }
>
> + lock_cnt = (uint8_t)((lock_sema >> TIM_BUCKET_W1_S_LOCK) &
> + TIM_BUCKET_W1_M_LOCK);
> + if (lock_cnt) {
> + cnxk_tim_bkt_dec_lock(bkt);
> +#ifdef RTE_ARCH_ARM64
> + asm volatile(PLT_CPU_FEATURE_PREAMBLE
> + " ldxrb %w[lock_cnt], [%[lock]] \n"
> + " tst %w[lock_cnt], 255 \n"
> + " beq dne%= \n"
> + " sevl \n"
> + "rty%=: wfe \n"
> + " ldxrb %w[lock_cnt], [%[lock]] \n"
> + " tst %w[lock_cnt], 255 \n"
> + " bne rty%= \n"
> + "dne%=: \n"
> + : [lock_cnt] "=&r"(lock_cnt)
> + : [lock] "r"(&bkt->lock)
> + : "memory");
> +#else
> + while (__atomic_load_n(&bkt->lock, __ATOMIC_RELAXED))
> + ;
> +#endif
> + goto __retry;
> + }
> +
> chunk_remainder = cnxk_tim_bkt_fetch_rem(lock_sema);
> rem = chunk_remainder - nb_timers;
> if (rem < 0) {
> - crem = tim_ring->nb_chunk_slots - chunk_remainder;
> - if (chunk_remainder && crem) {
> + if (chunk_remainder > 0) {
> chunk = ((struct cnxk_tim_ent *)
> mirr_bkt->current_chunk) +
> - crem;
> + tim_ring->nb_chunk_slots - chunk_remainder;
>
> index = cnxk_tim_cpy_wrk(index, chunk_remainder, chunk,
> tim, ents, bkt);
> @@ -537,18 +539,19 @@ cnxk_tim_add_entry_brst(struct cnxk_tim_ring *const tim_ring,
> chunk = cnxk_tim_insert_chunk(bkt, mirr_bkt, tim_ring);
>
> if (unlikely(chunk == NULL)) {
> - cnxk_tim_bkt_dec_lock(bkt);
> + cnxk_tim_bkt_dec_lock_relaxed(bkt);
> rte_errno = ENOMEM;
> tim[index]->state = RTE_EVENT_TIMER_ERROR;
> - return crem;
> + return index;
> }
> *(uint64_t *)(chunk + tim_ring->nb_chunk_slots) = 0;
> mirr_bkt->current_chunk = (uintptr_t)chunk;
> - cnxk_tim_cpy_wrk(index, nb_timers, chunk, tim, ents, bkt);
> + index = cnxk_tim_cpy_wrk(index, nb_timers, chunk, tim, ents,
> + bkt) -
> + index;
>
> - rem = nb_timers - chunk_remainder;
> - cnxk_tim_bkt_set_rem(bkt, tim_ring->nb_chunk_slots - rem);
> - cnxk_tim_bkt_add_nent(bkt, rem);
> + cnxk_tim_bkt_set_rem(bkt, tim_ring->nb_chunk_slots - index);
> + cnxk_tim_bkt_add_nent(bkt, index);
> } else {
> chunk = (struct cnxk_tim_ent *)mirr_bkt->current_chunk;
> chunk += (tim_ring->nb_chunk_slots - chunk_remainder);
> --
> 2.25.1
>
More information about the dev
mailing list