// Copyright 2010 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.

package runtime

import (
	
	
)

const (
	_EACCES = 13
	_EINVAL = 22
)

// Don't split the stack as this method may be invoked without a valid G, which
// prevents us from allocating more stack.
//
//go:nosplit
func sysAllocOS( uintptr) unsafe.Pointer {
	,  := mmap(nil, , _PROT_READ|_PROT_WRITE, _MAP_ANON|_MAP_PRIVATE, -1, 0)
	if  != 0 {
		if  == _EACCES {
			print("runtime: mmap: access denied\n")
			exit(2)
		}
		if  == _EAGAIN {
			print("runtime: mmap: too much locked memory (check 'ulimit -l').\n")
			exit(2)
		}
		return nil
	}
	return 
}

var adviseUnused = uint32(_MADV_FREE)

const madviseUnsupported = 0

func sysUnusedOS( unsafe.Pointer,  uintptr) {
	if uintptr()&(physPageSize-1) != 0 || &(physPageSize-1) != 0 {
		// madvise will round this to any physical page
		// *covered* by this range, so an unaligned madvise
		// will release more memory than intended.
		throw("unaligned sysUnused")
	}

	 := atomic.Load(&adviseUnused)
	if debug.madvdontneed != 0 &&  != madviseUnsupported {
		 = _MADV_DONTNEED
	}
	switch  {
	case _MADV_FREE:
		if madvise(, , _MADV_FREE) == 0 {
			break
		}
		atomic.Store(&adviseUnused, _MADV_DONTNEED)
		fallthrough
	case _MADV_DONTNEED:
		// MADV_FREE was added in Linux 4.5. Fall back on MADV_DONTNEED if it's
		// not supported.
		if madvise(, , _MADV_DONTNEED) == 0 {
			break
		}
		atomic.Store(&adviseUnused, madviseUnsupported)
		fallthrough
	case madviseUnsupported:
		// Since Linux 3.18, support for madvise is optional.
		// Fall back on mmap if it's not supported.
		// _MAP_ANON|_MAP_FIXED|_MAP_PRIVATE will unmap all the
		// pages in the old mapping, and remap the memory region.
		mmap(, , _PROT_READ|_PROT_WRITE, _MAP_ANON|_MAP_FIXED|_MAP_PRIVATE, -1, 0)
	}

	if debug.harddecommit > 0 {
		,  := mmap(, , _PROT_NONE, _MAP_ANON|_MAP_FIXED|_MAP_PRIVATE, -1, 0)
		if  !=  ||  != 0 {
			throw("runtime: cannot disable permissions in address space")
		}
	}
}

func sysUsedOS( unsafe.Pointer,  uintptr) {
	if debug.harddecommit > 0 {
		,  := mmap(, , _PROT_READ|_PROT_WRITE, _MAP_ANON|_MAP_FIXED|_MAP_PRIVATE, -1, 0)
		if  == _ENOMEM {
			throw("runtime: out of memory")
		}
		if  !=  ||  != 0 {
			throw("runtime: cannot remap pages in address space")
		}
		return
	}
}

func sysHugePageOS( unsafe.Pointer,  uintptr) {
	if physHugePageSize != 0 {
		// Round v up to a huge page boundary.
		 := alignUp(uintptr(), physHugePageSize)
		// Round v+n down to a huge page boundary.
		 := alignDown(uintptr()+, physHugePageSize)

		if  <  {
			madvise(unsafe.Pointer(), -, _MADV_HUGEPAGE)
		}
	}
}

func sysNoHugePageOS( unsafe.Pointer,  uintptr) {
	if uintptr()&(physPageSize-1) != 0 {
		// The Linux implementation requires that the address
		// addr be page-aligned, and allows length to be zero.
		throw("unaligned sysNoHugePageOS")
	}
	madvise(, , _MADV_NOHUGEPAGE)
}

func sysHugePageCollapseOS( unsafe.Pointer,  uintptr) {
	if uintptr()&(physPageSize-1) != 0 {
		// The Linux implementation requires that the address
		// addr be page-aligned, and allows length to be zero.
		throw("unaligned sysHugePageCollapseOS")
	}
	if physHugePageSize == 0 {
		return
	}
	// N.B. If you find yourself debugging this code, note that
	// this call can fail with EAGAIN because it's best-effort.
	// Also, when it returns an error, it's only for the last
	// huge page in the region requested.
	//
	// It can also sometimes return EINVAL if the corresponding
	// region hasn't been backed by physical memory. This is
	// difficult to guarantee in general, and it also means
	// there's no way to distinguish whether this syscall is
	// actually available. Oops.
	//
	// Anyway, that's why this call just doesn't bother checking
	// any errors.
	madvise(, , _MADV_COLLAPSE)
}

// Don't split the stack as this function may be invoked without a valid G,
// which prevents us from allocating more stack.
//
//go:nosplit
func sysFreeOS( unsafe.Pointer,  uintptr) {
	munmap(, )
}

func sysFaultOS( unsafe.Pointer,  uintptr) {
	mmap(, , _PROT_NONE, _MAP_ANON|_MAP_PRIVATE|_MAP_FIXED, -1, 0)
}

func sysReserveOS( unsafe.Pointer,  uintptr) unsafe.Pointer {
	,  := mmap(, , _PROT_NONE, _MAP_ANON|_MAP_PRIVATE, -1, 0)
	if  != 0 {
		return nil
	}
	return 
}

func sysMapOS( unsafe.Pointer,  uintptr) {
	,  := mmap(, , _PROT_READ|_PROT_WRITE, _MAP_ANON|_MAP_FIXED|_MAP_PRIVATE, -1, 0)
	if  == _ENOMEM {
		throw("runtime: out of memory")
	}
	if  !=  ||  != 0 {
		print("runtime: mmap(", , ", ", , ") returned ", , ", ", , "\n")
		throw("runtime: cannot map pages in arena address space")
	}

	// Disable huge pages if the GODEBUG for it is set.
	//
	// Note that there are a few sysHugePage calls that can override this, but
	// they're all for GC metadata.
	if debug.disablethp != 0 {
		sysNoHugePageOS(, )
	}
}