/* This test program is part of GDB, the GNU debugger.
Copyright 2022-2023 Free Software Foundation, Inc.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see . */
/* Exercise AArch64's Memory Tagging Extension corefile support. We allocate
multiple memory mappings with PROT_MTE and assign tag values for all the
existing MTE granules. */
/* This test was based on the documentation for the AArch64 Memory Tagging
Extension from the Linux Kernel, found in the sources in
Documentation/arm64/memory-tagging-extension.rst. */
#include
#include
#include
#include
#include
#include
#include
/* From arch/arm64/include/uapi/asm/hwcap.h */
#ifndef HWCAP2_MTE
#define HWCAP2_MTE (1 << 18)
#endif
/* From arch/arm64/include/uapi/asm/mman.h */
#ifndef PROT_MTE
#define PROT_MTE 0x20
#endif
#ifndef PR_SET_TAGGED_ADDR_CTRL
#define PR_SET_TAGGED_ADDR_CTRL 55
#define PR_TAGGED_ADDR_ENABLE (1UL << 0)
#endif
/* From include/uapi/linux/prctl.h */
#ifndef PR_MTE_TCF_SHIFT
#define PR_MTE_TCF_SHIFT 1
#define PR_MTE_TCF_SYNC (1UL << PR_MTE_TCF_SHIFT)
#define PR_MTE_TCF_ASYNC (2UL << PR_MTE_TCF_SHIFT)
#define PR_MTE_TAG_SHIFT 3
#define PR_MTE_TAG_MASK (0xffffUL << PR_MTE_TAG_SHIFT)
#endif
#ifdef ASYNC
#define TCF_MODE PR_MTE_TCF_ASYNC
#else
#define TCF_MODE PR_MTE_TCF_SYNC
#endif
#define NMAPS 5
/* We store the pointers and sizes of the memory maps we requested. Each
of them has a different size. */
unsigned char *mmap_pointers[NMAPS];
/* Set the allocation tag on the destination address. */
#define set_tag(tagged_addr) do { \
asm volatile("stg %0, [%0]" : : "r" (tagged_addr) : "memory"); \
} while (0)
uintptr_t
set_logical_tag (uintptr_t ptr, unsigned char tag)
{
ptr &= ~0xFF00000000000000ULL;
ptr |= ((uintptr_t) tag << 56);
return ptr;
}
void
fill_map_with_tags (unsigned char *ptr, size_t size, unsigned char *tag)
{
for (size_t start = 0; start < size; start += 16)
{
set_tag (set_logical_tag (((uintptr_t)ptr + start) & ~(0xFULL), *tag));
*tag = (*tag + 1) % 16;
}
}
int
main (int argc, char **argv)
{
unsigned char *tagged_ptr;
unsigned long page_sz = sysconf (_SC_PAGESIZE);
unsigned long hwcap2 = getauxval (AT_HWCAP2);
/* Bail out if MTE is not supported. */
if (!(hwcap2 & HWCAP2_MTE))
return 1;
/* Enable the tagged address ABI, synchronous MTE tag check faults and
allow all non-zero tags in the randomly generated set. */
if (prctl (PR_SET_TAGGED_ADDR_CTRL,
PR_TAGGED_ADDR_ENABLE | TCF_MODE
| (0xfffe << PR_MTE_TAG_SHIFT),
0, 0, 0))
{
perror ("prctl () failed");
return 1;
}
/* Map a big area of NMAPS * 2 pages. */
unsigned char *big_map = mmap (0, NMAPS * 2 * page_sz, PROT_NONE,
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
if (big_map == MAP_FAILED)
{
perror ("mmap () failed");
return 1;
}
/* Start with a tag of 0x1 so we can crash later. */
unsigned char tag = 1;
/* From that big area of NMAPS * 2 pages, go through each page and protect
alternating pages. This should prevent the kernel from merging different
mmap's and force the creation of multiple individual MTE-protected entries
in /proc//smaps. */
for (int i = 0; i < NMAPS; i++)
{
mmap_pointers[i] = big_map + (i * 2 * page_sz);
/* Enable MTE on alternating pages. */
if (mprotect (mmap_pointers[i], page_sz,
PROT_READ | PROT_WRITE | PROT_MTE))
{
perror ("mprotect () failed");
return 1;
}
fill_map_with_tags (mmap_pointers[i], page_sz, &tag);
}
/* The following line causes a crash on purpose. */
*mmap_pointers[0] = 0x4;
return 0;
}