linux/include/linux/soc/marvell/octeontx2/asm.h

/* SPDX-License-Identifier: GPL-2.0-only
 * Copyright (C) 2020 Marvell.
 */

#ifndef __SOC_OTX2_ASM_H
#define __SOC_OTX2_ASM_H

#include <linux/types.h>
#if defined(CONFIG_ARM64)
/*
 * otx2_lmt_flush is used for LMT store operation.
 * On octeontx2 platform CPT instruction enqueue and
 * NIX packet send are only possible via LMTST
 * operations and it uses LDEOR instruction targeting
 * the coprocessor address.
 */
#define otx2_lmt_flush(ioaddr)                          \
({                                                      \
	u64 result = 0;                                 \
	__asm__ volatile(".cpu  generic+lse\n"          \
			 "ldeor xzr, %x[rf], [%[rs]]"   \
			 : [rf]"=r" (result)            \
			 : [rs]"r" (ioaddr));           \
	(result);                                       \
})
/*
 * STEORL store to memory with release semantics.
 * This will avoid using DMB barrier after each LMTST
 * operation.
 */
#define cn10k_lmt_flush(val, addr)			\
({							\
	__asm__ volatile(".cpu  generic+lse\n"		\
			 "steorl %x[rf],[%[rs]]"		\
			 : [rf] "+r"(val)		\
			 : [rs] "r"(addr));		\
})

static inline u64 otx2_atomic64_fetch_add(u64 incr, u64 *ptr)
{
	u64 result;

	asm volatile (".cpu  generic+lse\n"
		      "ldadda %x[i], %x[r], [%[b]]"
		      : [r] "=r" (result), "+m" (*ptr)
		      : [i] "r" (incr), [b] "r" (ptr)
		      : "memory");
	return result;
}

#else
#define otx2_lmt_flush(ioaddr)          ({ 0; })
#define cn10k_lmt_flush(val, addr)	({ addr = val; })
#define otx2_atomic64_fetch_add(incr, ptr)	({ incr; })
#endif

#endif /* __SOC_OTX2_ASM_H */