58 lines
1.5 KiB
C
58 lines
1.5 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only
|
|
* Copyright (C) 2020 Marvell.
|
|
*/
|
|
|
|
#ifndef __SOC_OTX2_ASM_H
|
|
#define __SOC_OTX2_ASM_H
|
|
|
|
#include <linux/types.h>
|
|
#if defined(CONFIG_ARM64)
|
|
/*
|
|
* otx2_lmt_flush is used for LMT store operation.
|
|
* On octeontx2 platform CPT instruction enqueue and
|
|
* NIX packet send are only possible via LMTST
|
|
* operations and it uses LDEOR instruction targeting
|
|
* the coprocessor address.
|
|
*/
|
|
#define otx2_lmt_flush(ioaddr) \
|
|
({ \
|
|
u64 result = 0; \
|
|
__asm__ volatile(".cpu generic+lse\n" \
|
|
"ldeor xzr, %x[rf], [%[rs]]" \
|
|
: [rf]"=r" (result) \
|
|
: [rs]"r" (ioaddr)); \
|
|
(result); \
|
|
})
|
|
/*
|
|
* STEORL store to memory with release semantics.
|
|
* This will avoid using DMB barrier after each LMTST
|
|
* operation.
|
|
*/
|
|
#define cn10k_lmt_flush(val, addr) \
|
|
({ \
|
|
__asm__ volatile(".cpu generic+lse\n" \
|
|
"steorl %x[rf],[%[rs]]" \
|
|
: [rf] "+r"(val) \
|
|
: [rs] "r"(addr)); \
|
|
})
|
|
|
|
static inline u64 otx2_atomic64_fetch_add(u64 incr, u64 *ptr)
|
|
{
|
|
u64 result;
|
|
|
|
asm volatile (".cpu generic+lse\n"
|
|
"ldadda %x[i], %x[r], [%[b]]"
|
|
: [r] "=r" (result), "+m" (*ptr)
|
|
: [i] "r" (incr), [b] "r" (ptr)
|
|
: "memory");
|
|
return result;
|
|
}
|
|
|
|
#else
|
|
#define otx2_lmt_flush(ioaddr) ({ 0; })
|
|
#define cn10k_lmt_flush(val, addr) ({ addr = val; })
|
|
#define otx2_atomic64_fetch_add(incr, ptr) ({ incr; })
|
|
#endif
|
|
|
|
#endif /* __SOC_OTX2_ASM_H */
|