mirror of
https://git.hardenedbsd.org/hardenedbsd/HardenedBSD.git
synced 2024-11-14 22:32:30 +01:00
6bb132ba1e
Add sys/errno.h, sys/malloc.h, sys/queue.h, and vm/uma.h as needed. sys/sysproto.h currently includes sys/acl.h which currently includes sys/param.h, sys/queue.h, and vm/uma.h which in turn bring in sys/errno.h sys/malloc.h. Reviewed by: kib Differential Revision: https://reviews.freebsd.org/D44465
426 lines
13 KiB
C
426 lines
13 KiB
C
/*-
|
|
* Copyright (c) 2004 Tim J. Robbins
|
|
* Copyright (c) 2002 Doug Rabson
|
|
* Copyright (c) 2000 Marcel Moolenaar
|
|
* Copyright (c) 1994-1995 Søren Schmidt
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer
|
|
* in this position and unchanged.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. The name of the author may not be used to endorse or promote products
|
|
* derived from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/fcntl.h>
|
|
#include <sys/file.h>
|
|
#include <sys/ktr.h>
|
|
#include <sys/lock.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/proc.h>
|
|
#include <sys/resourcevar.h>
|
|
#include <sys/rwlock.h>
|
|
#include <sys/syscallsubr.h>
|
|
#include <sys/sysent.h>
|
|
#include <sys/sysproto.h>
|
|
|
|
#include <vm/pmap.h>
|
|
#include <vm/vm_extern.h>
|
|
#include <vm/vm_map.h>
|
|
#include <vm/vm_object.h>
|
|
|
|
#include <compat/linux/linux_emul.h>
|
|
#include <compat/linux/linux_mmap.h>
|
|
#include <compat/linux/linux_persona.h>
|
|
#include <compat/linux/linux_util.h>
|
|
|
|
#define STACK_SIZE (2 * 1024 * 1024)
|
|
#define GUARD_SIZE (4 * PAGE_SIZE)
|
|
|
|
#if defined(__amd64__)
|
|
static void linux_fixup_prot(struct thread *td, int *prot);
|
|
#endif
|
|
|
|
static int
|
|
linux_mmap_check_fp(struct file *fp, int flags, int prot, int maxprot)
|
|
{
|
|
|
|
/* Linux mmap() just fails for O_WRONLY files */
|
|
if ((fp->f_flag & FREAD) == 0)
|
|
return (EACCES);
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
linux_mmap_common(struct thread *td, uintptr_t addr, size_t len, int prot,
|
|
int flags, int fd, off_t pos)
|
|
{
|
|
struct mmap_req mr, mr_fixed;
|
|
struct proc *p = td->td_proc;
|
|
struct vmspace *vms = td->td_proc->p_vmspace;
|
|
int bsd_flags, error;
|
|
|
|
LINUX_CTR6(mmap2, "0x%lx, %ld, %ld, 0x%08lx, %ld, 0x%lx",
|
|
addr, len, prot, flags, fd, pos);
|
|
|
|
error = 0;
|
|
bsd_flags = 0;
|
|
|
|
/*
|
|
* Linux mmap(2):
|
|
* You must specify exactly one of MAP_SHARED and MAP_PRIVATE
|
|
*/
|
|
if (!((flags & LINUX_MAP_SHARED) ^ (flags & LINUX_MAP_PRIVATE)))
|
|
return (EINVAL);
|
|
|
|
if (flags & LINUX_MAP_SHARED)
|
|
bsd_flags |= MAP_SHARED;
|
|
if (flags & LINUX_MAP_PRIVATE)
|
|
bsd_flags |= MAP_PRIVATE;
|
|
if (flags & LINUX_MAP_FIXED)
|
|
bsd_flags |= MAP_FIXED;
|
|
if (flags & LINUX_MAP_ANON) {
|
|
/* Enforce pos to be on page boundary, then ignore. */
|
|
if ((pos & PAGE_MASK) != 0)
|
|
return (EINVAL);
|
|
pos = 0;
|
|
bsd_flags |= MAP_ANON;
|
|
} else
|
|
bsd_flags |= MAP_NOSYNC;
|
|
if (flags & LINUX_MAP_GROWSDOWN)
|
|
bsd_flags |= MAP_STACK;
|
|
|
|
#if defined(__amd64__)
|
|
/*
|
|
* According to the Linux mmap(2) man page, "MAP_32BIT flag
|
|
* is ignored when MAP_FIXED is set."
|
|
*/
|
|
if ((flags & LINUX_MAP_32BIT) && (flags & LINUX_MAP_FIXED) == 0)
|
|
bsd_flags |= MAP_32BIT;
|
|
|
|
/*
|
|
* PROT_READ, PROT_WRITE, or PROT_EXEC implies PROT_READ and PROT_EXEC
|
|
* on Linux/i386 if the binary requires executable stack.
|
|
* We do this only for IA32 emulation as on native i386 this is does not
|
|
* make sense without PAE.
|
|
*
|
|
* XXX. Linux checks that the file system is not mounted with noexec.
|
|
*/
|
|
linux_fixup_prot(td, &prot);
|
|
#endif
|
|
|
|
/* Linux does not check file descriptor when MAP_ANONYMOUS is set. */
|
|
fd = (bsd_flags & MAP_ANON) ? -1 : fd;
|
|
if (flags & LINUX_MAP_GROWSDOWN) {
|
|
/*
|
|
* The Linux MAP_GROWSDOWN option does not limit auto
|
|
* growth of the region. Linux mmap with this option
|
|
* takes as addr the initial BOS, and as len, the initial
|
|
* region size. It can then grow down from addr without
|
|
* limit. However, Linux threads has an implicit internal
|
|
* limit to stack size of STACK_SIZE. Its just not
|
|
* enforced explicitly in Linux. But, here we impose
|
|
* a limit of (STACK_SIZE - GUARD_SIZE) on the stack
|
|
* region, since we can do this with our mmap.
|
|
*
|
|
* Our mmap with MAP_STACK takes addr as the maximum
|
|
* downsize limit on BOS, and as len the max size of
|
|
* the region. It then maps the top SGROWSIZ bytes,
|
|
* and auto grows the region down, up to the limit
|
|
* in addr.
|
|
*
|
|
* If we don't use the MAP_STACK option, the effect
|
|
* of this code is to allocate a stack region of a
|
|
* fixed size of (STACK_SIZE - GUARD_SIZE).
|
|
*/
|
|
|
|
if ((caddr_t)addr + len > vms->vm_maxsaddr) {
|
|
/*
|
|
* Some Linux apps will attempt to mmap
|
|
* thread stacks near the top of their
|
|
* address space. If their TOS is greater
|
|
* than vm_maxsaddr, vm_map_growstack()
|
|
* will confuse the thread stack with the
|
|
* process stack and deliver a SEGV if they
|
|
* attempt to grow the thread stack past their
|
|
* current stacksize rlimit. To avoid this,
|
|
* adjust vm_maxsaddr upwards to reflect
|
|
* the current stacksize rlimit rather
|
|
* than the maximum possible stacksize.
|
|
* It would be better to adjust the
|
|
* mmap'ed region, but some apps do not check
|
|
* mmap's return value.
|
|
*/
|
|
PROC_LOCK(p);
|
|
vms->vm_maxsaddr = (char *)round_page(vms->vm_stacktop) -
|
|
lim_cur_proc(p, RLIMIT_STACK);
|
|
PROC_UNLOCK(p);
|
|
}
|
|
|
|
/*
|
|
* This gives us our maximum stack size and a new BOS.
|
|
* If we're using VM_STACK, then mmap will just map
|
|
* the top SGROWSIZ bytes, and let the stack grow down
|
|
* to the limit at BOS. If we're not using VM_STACK
|
|
* we map the full stack, since we don't have a way
|
|
* to autogrow it.
|
|
*/
|
|
if (len <= STACK_SIZE - GUARD_SIZE) {
|
|
addr = addr - (STACK_SIZE - GUARD_SIZE - len);
|
|
len = STACK_SIZE - GUARD_SIZE;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* FreeBSD is free to ignore the address hint if MAP_FIXED wasn't
|
|
* passed. However, some Linux applications, like the ART runtime,
|
|
* depend on the hint. If the MAP_FIXED wasn't passed, but the
|
|
* address is not zero, try with MAP_FIXED and MAP_EXCL first,
|
|
* and fall back to the normal behaviour if that fails.
|
|
*/
|
|
mr = (struct mmap_req) {
|
|
.mr_hint = addr,
|
|
.mr_len = len,
|
|
.mr_prot = prot,
|
|
.mr_flags = bsd_flags,
|
|
.mr_fd = fd,
|
|
.mr_pos = pos,
|
|
.mr_check_fp_fn = linux_mmap_check_fp,
|
|
};
|
|
if (addr != 0 && (bsd_flags & MAP_FIXED) == 0 &&
|
|
(bsd_flags & MAP_EXCL) == 0) {
|
|
mr_fixed = mr;
|
|
mr_fixed.mr_flags |= MAP_FIXED | MAP_EXCL;
|
|
error = kern_mmap(td, &mr_fixed);
|
|
if (error == 0)
|
|
goto out;
|
|
}
|
|
|
|
error = kern_mmap(td, &mr);
|
|
out:
|
|
LINUX_CTR2(mmap2, "return: %d (%p)", error, td->td_retval[0]);
|
|
|
|
return (error);
|
|
}
|
|
|
|
int
|
|
linux_mprotect_common(struct thread *td, uintptr_t addr, size_t len, int prot)
|
|
{
|
|
int flags = 0;
|
|
|
|
/* XXX Ignore PROT_GROWSUP for now. */
|
|
prot &= ~LINUX_PROT_GROWSUP;
|
|
if ((prot & ~(LINUX_PROT_GROWSDOWN | PROT_READ | PROT_WRITE |
|
|
PROT_EXEC)) != 0)
|
|
return (EINVAL);
|
|
if ((prot & LINUX_PROT_GROWSDOWN) != 0) {
|
|
prot &= ~LINUX_PROT_GROWSDOWN;
|
|
flags |= VM_MAP_PROTECT_GROWSDOWN;
|
|
}
|
|
|
|
#if defined(__amd64__)
|
|
linux_fixup_prot(td, &prot);
|
|
#endif
|
|
return (kern_mprotect(td, addr, len, prot, flags));
|
|
}
|
|
|
|
/*
|
|
* Implement Linux madvise(MADV_DONTNEED), which has unusual semantics: for
|
|
* anonymous memory, pages in the range are immediately discarded.
|
|
*/
|
|
static int
|
|
linux_madvise_dontneed(struct thread *td, vm_offset_t start, vm_offset_t end)
|
|
{
|
|
vm_map_t map;
|
|
vm_map_entry_t entry;
|
|
vm_object_t backing_object, object;
|
|
vm_offset_t estart, eend;
|
|
vm_pindex_t pstart, pend;
|
|
int error;
|
|
|
|
map = &td->td_proc->p_vmspace->vm_map;
|
|
|
|
if (!vm_map_range_valid(map, start, end))
|
|
return (EINVAL);
|
|
start = trunc_page(start);
|
|
end = round_page(end);
|
|
|
|
error = 0;
|
|
vm_map_lock_read(map);
|
|
if (!vm_map_lookup_entry(map, start, &entry))
|
|
entry = vm_map_entry_succ(entry);
|
|
for (; entry->start < end; entry = vm_map_entry_succ(entry)) {
|
|
if ((entry->eflags & MAP_ENTRY_IS_SUB_MAP) != 0)
|
|
continue;
|
|
|
|
if (entry->wired_count != 0) {
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
|
|
object = entry->object.vm_object;
|
|
if (object == NULL)
|
|
continue;
|
|
if ((object->flags & (OBJ_UNMANAGED | OBJ_FICTITIOUS)) != 0)
|
|
continue;
|
|
|
|
pstart = OFF_TO_IDX(entry->offset);
|
|
if (start > entry->start) {
|
|
pstart += atop(start - entry->start);
|
|
estart = start;
|
|
} else {
|
|
estart = entry->start;
|
|
}
|
|
pend = OFF_TO_IDX(entry->offset) +
|
|
atop(entry->end - entry->start);
|
|
if (entry->end > end) {
|
|
pend -= atop(entry->end - end);
|
|
eend = end;
|
|
} else {
|
|
eend = entry->end;
|
|
}
|
|
|
|
if ((object->flags & (OBJ_ANON | OBJ_ONEMAPPING)) ==
|
|
(OBJ_ANON | OBJ_ONEMAPPING)) {
|
|
/*
|
|
* Singly-mapped anonymous memory is discarded. This
|
|
* does not match Linux's semantics when the object
|
|
* belongs to a shadow chain of length > 1, since
|
|
* subsequent faults may retrieve pages from an
|
|
* intermediate anonymous object. However, handling
|
|
* this case correctly introduces a fair bit of
|
|
* complexity.
|
|
*/
|
|
VM_OBJECT_WLOCK(object);
|
|
if ((object->flags & OBJ_ONEMAPPING) != 0) {
|
|
vm_object_collapse(object);
|
|
vm_object_page_remove(object, pstart, pend, 0);
|
|
backing_object = object->backing_object;
|
|
if (backing_object != NULL &&
|
|
(backing_object->flags & OBJ_ANON) != 0)
|
|
linux_msg(td,
|
|
"possibly incorrect MADV_DONTNEED");
|
|
VM_OBJECT_WUNLOCK(object);
|
|
continue;
|
|
}
|
|
VM_OBJECT_WUNLOCK(object);
|
|
}
|
|
|
|
/*
|
|
* Handle shared mappings. Remove them outright instead of
|
|
* calling pmap_advise(), for consistency with Linux.
|
|
*/
|
|
pmap_remove(map->pmap, estart, eend);
|
|
vm_object_madvise(object, pstart, pend, MADV_DONTNEED);
|
|
}
|
|
vm_map_unlock_read(map);
|
|
|
|
return (error);
|
|
}
|
|
|
|
int
|
|
linux_madvise_common(struct thread *td, uintptr_t addr, size_t len, int behav)
|
|
{
|
|
|
|
switch (behav) {
|
|
case LINUX_MADV_NORMAL:
|
|
return (kern_madvise(td, addr, len, MADV_NORMAL));
|
|
case LINUX_MADV_RANDOM:
|
|
return (kern_madvise(td, addr, len, MADV_RANDOM));
|
|
case LINUX_MADV_SEQUENTIAL:
|
|
return (kern_madvise(td, addr, len, MADV_SEQUENTIAL));
|
|
case LINUX_MADV_WILLNEED:
|
|
return (kern_madvise(td, addr, len, MADV_WILLNEED));
|
|
case LINUX_MADV_DONTNEED:
|
|
return (linux_madvise_dontneed(td, addr, addr + len));
|
|
case LINUX_MADV_FREE:
|
|
return (kern_madvise(td, addr, len, MADV_FREE));
|
|
case LINUX_MADV_REMOVE:
|
|
linux_msg(curthread, "unsupported madvise MADV_REMOVE");
|
|
return (EINVAL);
|
|
case LINUX_MADV_DONTFORK:
|
|
return (kern_minherit(td, addr, len, INHERIT_NONE));
|
|
case LINUX_MADV_DOFORK:
|
|
return (kern_minherit(td, addr, len, INHERIT_COPY));
|
|
case LINUX_MADV_MERGEABLE:
|
|
linux_msg(curthread, "unsupported madvise MADV_MERGEABLE");
|
|
return (EINVAL);
|
|
case LINUX_MADV_UNMERGEABLE:
|
|
/* We don't merge anyway. */
|
|
return (0);
|
|
case LINUX_MADV_HUGEPAGE:
|
|
/* Ignored; on FreeBSD huge pages are always on. */
|
|
return (0);
|
|
case LINUX_MADV_NOHUGEPAGE:
|
|
#if 0
|
|
/*
|
|
* Don't warn - Firefox uses it a lot, and in real Linux it's
|
|
* an optional feature.
|
|
*/
|
|
linux_msg(curthread, "unsupported madvise MADV_NOHUGEPAGE");
|
|
#endif
|
|
return (EINVAL);
|
|
case LINUX_MADV_DONTDUMP:
|
|
return (kern_madvise(td, addr, len, MADV_NOCORE));
|
|
case LINUX_MADV_DODUMP:
|
|
return (kern_madvise(td, addr, len, MADV_CORE));
|
|
case LINUX_MADV_WIPEONFORK:
|
|
return (kern_minherit(td, addr, len, INHERIT_ZERO));
|
|
case LINUX_MADV_KEEPONFORK:
|
|
return (kern_minherit(td, addr, len, INHERIT_COPY));
|
|
case LINUX_MADV_HWPOISON:
|
|
linux_msg(curthread, "unsupported madvise MADV_HWPOISON");
|
|
return (EINVAL);
|
|
case LINUX_MADV_SOFT_OFFLINE:
|
|
linux_msg(curthread, "unsupported madvise MADV_SOFT_OFFLINE");
|
|
return (EINVAL);
|
|
case -1:
|
|
/*
|
|
* -1 is sometimes used as a dummy value to detect simplistic
|
|
* madvise(2) stub implementations. This safeguard is used by
|
|
* BoringSSL, for example, before assuming MADV_WIPEONFORK is
|
|
* safe to use. Don't produce an "unsupported" error message
|
|
* for this special dummy value, which is unlikely to be used
|
|
* by any new advisory behavior feature.
|
|
*/
|
|
return (EINVAL);
|
|
default:
|
|
linux_msg(curthread, "unsupported madvise behav %d", behav);
|
|
return (EINVAL);
|
|
}
|
|
}
|
|
|
|
#if defined(__amd64__)
|
|
static void
|
|
linux_fixup_prot(struct thread *td, int *prot)
|
|
{
|
|
struct linux_pemuldata *pem;
|
|
|
|
if (SV_PROC_FLAG(td->td_proc, SV_ILP32) && *prot & PROT_READ) {
|
|
pem = pem_find(td->td_proc);
|
|
if (pem->persona & LINUX_READ_IMPLIES_EXEC)
|
|
*prot |= PROT_EXEC;
|
|
}
|
|
|
|
}
|
|
#endif
|