SIGN IN SIGN UP
// SPDX-License-Identifier: MIT
// Copyright (c) 2015-2021 Zig Contributors
// This file is part of [zig](https://ziglang.org/), which is MIT licensed.
// The MIT license requires this copyright notice to be included in all copies
// and substantial portions of the software.
//! This struct represents a kernel thread, and acts as a namespace for concurrency
//! primitives that operate on kernel threads. For concurrency primitives that support
//! both evented I/O and async I/O, see the respective names in the top level std namespace.
2021-06-19 17:08:56 -05:00
const std = @import("std.zig");
const os = std.os;
2021-06-20 15:39:23 -05:00
const assert = std.debug.assert;
2021-06-19 17:08:56 -05:00
const target = std.Target.current;
const Atomic = std.atomic.Atomic;
pub const AutoResetEvent = @import("Thread/AutoResetEvent.zig");
pub const Futex = @import("Thread/Futex.zig");
pub const ResetEvent = @import("Thread/ResetEvent.zig");
pub const StaticResetEvent = @import("Thread/StaticResetEvent.zig");
pub const Mutex = @import("Thread/Mutex.zig");
pub const Semaphore = @import("Thread/Semaphore.zig");
pub const Condition = @import("Thread/Condition.zig");
2021-06-19 17:08:56 -05:00
pub const spinLoopHint = @compileError("deprecated: use std.atomic.spinLoopHint");
test "std.Thread" {
2021-06-20 09:56:30 -05:00
// Doesn't use testing.refAllDecls() since that would pull in the compileError spinLoopHint.
_ = AutoResetEvent;
_ = Futex;
_ = ResetEvent;
_ = StaticResetEvent;
_ = Mutex;
_ = Semaphore;
_ = Condition;
}
2021-06-19 17:08:56 -05:00
pub const use_pthreads = target.os.tag != .windows and std.builtin.link_libc;
2021-06-19 22:01:54 -05:00
const Thread = @This();
2021-06-19 17:08:56 -05:00
const Impl = if (target.os.tag == .windows)
WindowsThreadImpl
else if (use_pthreads)
PosixThreadImpl
else if (target.os.tag == .linux)
LinuxThreadImpl
else
@compileLog("Unsupported operating system", target.os.tag);
impl: Impl,
/// Represents a unique ID per thread.
/// May be an integer or pointer depending on the platform.
pub const Id = u64;
2021-06-19 17:08:56 -05:00
/// Returns the platform ID of the callers thread.
/// Attempts to use thread locals and avoid syscalls when possible.
pub fn getCurrentId() Id {
2021-06-19 17:08:56 -05:00
return Impl.getCurrentId();
}
2021-06-19 17:08:56 -05:00
pub const CpuCountError = error{
PermissionDenied,
SystemResources,
Unexpected,
};
2021-06-19 17:08:56 -05:00
/// Returns the platforms view on the number of logical CPU cores available.
pub fn getCpuCount() CpuCountError!usize {
return Impl.getCpuCount();
}
2021-06-19 17:08:56 -05:00
/// Configuration options for hints on how to spawn threads.
pub const SpawnConfig = struct {
// TODO compile-time call graph analysis to determine stack upper bound
// https://github.com/ziglang/zig/issues/157
/// Size in bytes of the Thread's stack
stack_size: usize = 16 * 1024 * 1024,
};
pub const SpawnError = error {
/// A system-imposed limit on the number of threads was encountered.
/// There are a number of limits that may trigger this error:
/// * the RLIMIT_NPROC soft resource limit (set via setrlimit(2)),
/// which limits the number of processes and threads for a real
/// user ID, was reached;
/// * the kernel's system-wide limit on the number of processes and
/// threads, /proc/sys/kernel/threads-max, was reached (see
/// proc(5));
/// * the maximum number of PIDs, /proc/sys/kernel/pid_max, was
/// reached (see proc(5)); or
/// * the PID limit (pids.max) imposed by the cgroup "process num
/// ber" (PIDs) controller was reached.
ThreadQuotaExceeded,
/// The kernel cannot allocate sufficient memory to allocate a task structure
/// for the child, or to copy those parts of the caller's context that need to
/// be copied.
SystemResources,
/// Not enough userland memory to spawn the thread.
OutOfMemory,
/// `mlockall` is enabled, and the memory needed to spawn the thread
/// would exceed the limit.
LockedMemoryLimitExceeded,
Unexpected,
};
2021-06-19 17:08:56 -05:00
/// Spawns a new thread which executes `function` using `args` and returns a handle the spawned thread.
/// `config` can be used as hints to the platform for now to spawn and execute the `function`.
/// The caller must eventually either call `join()` to wait for the thread to finish and free its resources
/// or call `detach()` to excuse the caller from calling `join()` and have the thread clean up its resources on completion`.
2021-06-20 09:56:30 -05:00
pub fn spawn(config: SpawnConfig, comptime function: anytype, args: anytype) SpawnError!Thread {
2021-06-19 17:08:56 -05:00
if (std.builtin.single_threaded) {
@compileError("cannot spawn thread when building in single-threaded mode");
}
2021-06-20 09:56:30 -05:00
const impl = try Impl.spawn(config, function, args);
return Thread{ .impl = impl };
}
/// Represents a kernel thread handle.
/// May be an integer or a pointer depending on the platform.
pub const Handle = Impl.ThreadHandle;
2021-06-19 22:01:54 -05:00
/// Retrns the handle of this thread
pub fn getHandle(self: Thread) Handle {
return self.impl.getHandle();
}
/// Release the obligation of the caller to call `join()` and have the thread clean up its own resources on completion.
pub fn detach(self: Thread) void {
return self.impl.detach();
}
/// Waits for the thread to complete, then deallocates any resources created on `spawn()`.
pub fn join(self: Thread) void {
return self.impl.join();
}
/// State to synchronize detachment of spawner thread to spawned thread
const Completion = Atomic(enum(u8) {
2021-06-19 22:01:54 -05:00
running,
detached,
completed,
});
2021-06-19 17:08:56 -05:00
/// Used by the Thread implementations to call the spawned function with the arguments.
fn callFn(comptime f: anytype, args: anytype) switch (Impl) {
2021-06-22 11:27:58 -05:00
WindowsThreadImpl => std.os.windows.DWORD,
2021-06-19 17:08:56 -05:00
LinuxThreadImpl => u8,
PosixThreadImpl => ?*c_void,
else => unreachable,
} {
const default_value = if (Impl == PosixThreadImpl) null else 0;
const bad_fn_ret = "expected return type of startFn to be 'u8', 'noreturn', 'void', or '!void'";
switch (@typeInfo(@typeInfo(@TypeOf(f)).Fn.return_type.?)) {
.NoReturn => {
@call(.{}, f, args);
},
.Void => {
@call(.{}, f, args);
return default_value;
},
.Int => |info| {
if (info.bits != 8) {
@compileError(bad_fn_ret);
}
2021-06-19 17:08:56 -05:00
const status = @call(.{}, f, args);
if (Impl != PosixThreadImpl) {
return status;
}
2021-06-19 17:08:56 -05:00
// pthreads don't support exit status, ignore value
_ = status;
return default_value;
},
.ErrorUnion => |info| {
if (info.payload != void) {
@compileError(bad_fn_ret);
}
2021-06-19 17:08:56 -05:00
@call(.{}, f, args) catch |err| {
std.debug.warn("error: {s}\n", .{@errorName(err)});
if (@errorReturnTrace()) |trace| {
std.debug.dumpStackTrace(trace.*);
}
2021-06-19 17:08:56 -05:00
};
return default_value;
},
else => {
@compileError(bad_fn_ret);
},
}
}
const WindowsThreadImpl = struct {
const windows = os.windows;
pub const ThreadHandle = windows.HANDLE;
fn getCurrentId() u64 {
2021-06-20 17:12:28 -05:00
return windows.kernel32.GetCurrentThreadId();
2021-06-19 17:08:56 -05:00
}
fn getCpuCount() !usize {
return windows.peb().NumberOfProcessors;
}
thread: *ThreadCompletion,
const ThreadCompletion = struct {
completion: Completion,
heap_ptr: windows.PVOID,
heap_handle: windows.HANDLE,
thread_handle: windows.HANDLE = undefined,
fn free(self: ThreadCompletion) void {
const status = windows.kernel32.HeapFree(self.heap_handle, 0, self.heap_ptr);
2021-06-22 11:27:58 -05:00
assert(status != 0);
2021-06-19 17:08:56 -05:00
}
};
fn spawn(config: SpawnConfig, comptime f: anytype, args: anytype) !Impl {
const Args = @TypeOf(args);
const Instance = struct {
fn_args: Args,
thread: ThreadCompletion,
2021-06-22 11:27:58 -05:00
fn entryFn(raw_ptr: windows.PVOID) callconv(.C) windows.DWORD {
2021-06-19 17:08:56 -05:00
const self = @ptrCast(*@This(), @alignCast(@alignOf(@This()), raw_ptr));
2021-06-22 11:27:58 -05:00
defer switch (self.thread.completion.swap(.completed, .SeqCst)) {
2021-06-19 17:08:56 -05:00
.running => {},
.completed => unreachable,
.detached => self.thread.free(),
};
return callFn(f, self.fn_args);
}
};
const heap_handle = windows.kernel32.GetProcessHeap() orelse return error.OutOfMemory;
2021-06-19 17:08:56 -05:00
const alloc_bytes = @alignOf(Instance) + @sizeOf(Instance);
const alloc_ptr = windows.kernel32.HeapAlloc(heap_handle, 0, alloc_bytes) orelse return error.OutOfMemory;
errdefer assert(windows.kernel32.HeapFree(heap_handle, 0, alloc_ptr) != 0);
const instance_bytes = @ptrCast([*]u8, alloc_ptr)[0..alloc_bytes];
const instance = std.heap.FixedBufferAllocator.init(instance_bytes).allocator.create(Instance) catch unreachable;
instance.* = .{
.fn_args = args,
.thread = .{
.completion = Completion.init(.running),
.heap_ptr = alloc_ptr,
.heap_handle = heap_handle,
},
};
2021-06-20 15:39:23 -05:00
// Windows appears to only support SYSTEM_INFO.dwAllocationGranularity minimum stack size.
// Going lower makes it default to that specified in the executable (~1mb).
// Its also fine if the limit here is incorrect as stack size is only a hint.
var stack_size = std.math.cast(u32, config.stack_size) catch std.math.maxInt(u32);
stack_size = std.math.max(64 * 1024, stack_size);
2021-06-19 17:08:56 -05:00
2021-06-20 17:12:28 -05:00
instance.thread.thread_handle = windows.kernel32.CreateThread(
2021-06-20 15:39:23 -05:00
null,
stack_size,
2021-06-22 11:27:58 -05:00
Instance.entryFn,
2021-06-20 15:39:23 -05:00
@ptrCast(*c_void, instance),
0,
null,
) orelse {
2021-06-19 17:08:56 -05:00
return windows.unexpectedError(windows.kernel32.GetLastError());
};
2021-06-19 17:08:56 -05:00
2021-06-22 11:27:58 -05:00
return Impl{ .thread = &instance.thread };
}
2021-06-19 17:08:56 -05:00
fn getHandle(self: Impl) ThreadHandle {
return self.thread.thread_handle;
}
fn detach(self: Impl) void {
windows.CloseHandle(self.thread.thread_handle);
2021-06-22 11:27:58 -05:00
switch (self.thread.completion.swap(.detached, .SeqCst)) {
2021-06-19 17:08:56 -05:00
.running => {},
.completed => self.thread.free(),
.detached => unreachable,
}
2021-06-19 17:08:56 -05:00
}
fn join(self: Impl) void {
windows.WaitForSingleObjectEx(self.thread.thread_handle, windows.INFINITE, false) catch unreachable;
windows.CloseHandle(self.thread.thread_handle);
2021-06-22 11:27:58 -05:00
assert(self.thread.completion.load(.SeqCst) == .completed);
2021-06-19 17:08:56 -05:00
self.thread.free();
}
};
const PosixThreadImpl = struct {
const c = std.c;
pub const ThreadHandle = c.pthread_t;
fn getCurrentId() Id {
switch (target.os.tag) {
.linux => {
return LinuxThreadImpl.getCurrentId();
},
.macos, .ios, .watchos, .tvos => {
var thread_id: u64 = undefined;
// Pass thread=null to get the current thread ID.
assert(c.pthread_threadid_np(null, &thread_id) == 0);
return thread_id;
},
.dragonfly => {
return @bitCast(u32, c.lwp_gettid());
},
.netbsd => {
return @bitCast(u32, c._lwp_self());
},
.freebsd => {
return @bitCast(u32, c.pthread_getthreadid_np());
},
.openbsd => {
return @bitCast(u32, c.getthrid());
},
.haiku => {
return @bitCast(u32, c.find_thread(null));
},
else => {
return @ptrToInt(c.pthread_self());
},
}
2021-06-19 17:08:56 -05:00
}
fn getCpuCount() !usize {
switch (target.os.tag) {
.linux => return LinuxThreadImpl.getCpuCount(),
.openbsd => {
var count: c_int = undefined;
var count_size: usize = @sizeOf(c_int);
const mib = [_]c_int{ os.CTL_HW, os.HW_NCPUONLINE };
os.sysctl(&mib, &count, &count_size, null, 0) catch |err| switch (err) {
error.NameTooLong, error.UnknownName => unreachable,
else => |e| return e,
};
return @intCast(usize, count);
},
.haiku => {
var count: u32 = undefined;
var system_info: os.system_info = undefined;
_ = os.system.get_system_info(&system_info); // always returns B_OK
count = system_info.cpu_count;
return @intCast(usize, count);
},
else => {
var count: c_int = undefined;
var count_len: usize = @sizeOf(c_int);
const name = if (comptime target.isDarwin()) "hw.logicalcpu" else "hw.ncpu";
os.sysctlbynameZ(name, &count, &count_len, null, 0) catch |err| switch (err) {
error.NameTooLong, error.UnknownName => unreachable,
else => |e| return e,
};
return @intCast(usize, count);
},
}
2021-06-19 17:08:56 -05:00
}
handle: ThreadHandle,
fn spawn(config: SpawnConfig, comptime f: anytype, args: anytype) !Impl {
const Args = @TypeOf(args);
const allocator = std.heap.c_allocator;
const Instance = struct {
fn entryFn(raw_arg: ?*c_void) callconv(.C) ?*c_void {
const args_ptr = @ptrCast(*Args, @alignCast(@alignOf(Args), raw_arg orelse unreachable));
defer allocator.destroy(args_ptr);
return callFn(f, args_ptr.*);
}
};
const args_ptr = try allocator.create(Args);
args_ptr.* = args;
2021-06-19 17:08:56 -05:00
errdefer allocator.destroy(args_ptr);
var attr: c.pthread_attr_t = undefined;
if (c.pthread_attr_init(&attr) != 0) return error.SystemResources;
defer assert(c.pthread_attr_destroy(&attr) == 0);
// Use the same set of parameters used by the libc-less impl.
2021-06-19 17:08:56 -05:00
const stack_size = std.math.max(config.stack_size, 16 * 1024);
assert(c.pthread_attr_setstacksize(&attr, stack_size) == 0);
assert(c.pthread_attr_setguardsize(&attr, std.mem.page_size) == 0);
2021-06-19 17:08:56 -05:00
var handle: c.pthread_t = undefined;
2021-06-25 12:43:03 -05:00
switch (c.pthread_create(
2021-06-19 17:08:56 -05:00
&handle,
&attr,
2021-06-19 17:08:56 -05:00
Instance.entryFn,
@ptrCast(*c_void, args_ptr),
)) {
2021-06-25 12:43:03 -05:00
0 => return Impl{ .handle = handle },
os.EAGAIN => return error.SystemResources,
os.EPERM => unreachable,
os.EINVAL => unreachable,
2021-06-25 12:43:03 -05:00
else => |err| return os.unexpectedErrno(err),
2021-06-25 12:59:28 -05:00
}
2021-06-19 17:08:56 -05:00
}
fn getHandle(self: Impl) ThreadHandle {
return self.handle;
}
2021-06-19 17:08:56 -05:00
fn detach(self: Impl) void {
switch (c.pthread_detach(self.handle)) {
os.EINVAL => unreachable,
os.ESRCH => unreachable,
else => unreachable,
}
}
2021-06-19 17:08:56 -05:00
fn join(self: Impl) void {
switch (c.pthread_join(self.handle, null)) {
0 => {},
os.EINVAL => unreachable,
os.ESRCH => unreachable,
os.EDEADLK => unreachable,
else => unreachable,
}
2021-06-19 17:08:56 -05:00
}
};
const LinuxThreadImpl = struct {
const linux = os.linux;
pub const ThreadHandle = i32;
threadlocal var tls_thread_id: ?Id = null;
2021-06-19 17:08:56 -05:00
fn getCurrentId() Id {
2021-06-19 17:08:56 -05:00
return tls_thread_id orelse {
2021-06-20 09:56:30 -05:00
const tid = @bitCast(u32, linux.gettid());
2021-06-19 17:08:56 -05:00
tls_thread_id = tid;
return tid;
};
}
fn getCpuCount() !usize {
const cpu_set = try os.sched_getaffinity(0);
return @as(usize, os.CPU_COUNT(cpu_set)); // TODO should not need this usize cast
}
thread: *ThreadCompletion,
const ThreadCompletion = struct {
completion: Completion = Completion.init(.running),
child_tid: Atomic(i32) = Atomic(i32).init(0),
parent_tid: i32 = undefined,
mapped: []align(std.mem.page_size) u8,
};
2021-06-19 17:08:56 -05:00
fn spawn(config: SpawnConfig, comptime f: anytype, args: anytype) !Impl {
const Args = @TypeOf(args);
const Instance = struct {
fn_args: Args,
thread: ThreadCompletion,
fn entryFn(raw_arg: usize) callconv(.C) u8 {
const self = @intToPtr(*@This(), raw_arg);
defer switch (self.thread.completion.swap(.completed, .Acquire)) {
.running => {},
.completed => unreachable,
.detached => {
const memory = self.thread.mapped;
__unmap_and_exit(@ptrToInt(memory.ptr), memory.len);
},
};
return callFn(f, self.fn_args);
}
};
var guard_offset: usize = undefined;
var stack_offset: usize = undefined;
var tls_offset: usize = undefined;
var instance_offset: usize = undefined;
const map_bytes = blk: {
var bytes: usize = std.mem.page_size;
guard_offset = bytes;
bytes += std.math.max(std.mem.page_size, config.stack_size);
bytes = std.mem.alignForward(bytes, std.mem.page_size);
stack_offset = bytes;
bytes = std.mem.alignForward(bytes, linux.tls.tls_image.alloc_align);
tls_offset = bytes;
bytes += linux.tls.tls_image.alloc_size;
bytes = std.mem.alignForward(bytes, @alignOf(Instance));
instance_offset = bytes;
bytes += @sizeOf(Instance);
bytes = std.mem.alignForward(bytes, std.mem.page_size);
break :blk bytes;
};
// map all memory needed without read/write permissions
// to avoid committing the whole region right away
const mapped = os.mmap(
null,
2021-06-19 17:08:56 -05:00
map_bytes,
os.PROT_NONE,
os.MAP_PRIVATE | os.MAP_ANONYMOUS,
-1,
0,
) catch |err| switch (err) {
error.MemoryMappingNotSupported => unreachable,
error.AccessDenied => unreachable,
error.PermissionDenied => unreachable,
else => |e| return e,
};
assert(mapped.len >= map_bytes);
2021-06-19 17:08:56 -05:00
errdefer os.munmap(mapped);
2021-06-19 17:08:56 -05:00
// map everything but the guard page as read/write
os.mprotect(
2021-06-19 17:08:56 -05:00
mapped[guard_offset..],
os.PROT_READ | os.PROT_WRITE,
) catch |err| switch (err) {
error.AccessDenied => unreachable,
else => |e| return e,
};
2021-06-19 17:08:56 -05:00
// Prepare the TLS segment and prepare a user_desc struct when needed on i386
var tls_ptr = os.linux.tls.prepareTLS(mapped[tls_offset..]);
var user_desc: if (target.cpu.arch == .i386) os.linux.user_desc else void = undefined;
if (target.cpu.arch == .i386) {
defer tls_ptr = @ptrToInt(&user_desc);
user_desc = .{
.entry_number = os.linux.tls.tls_image.gdt_entry_number,
.base_addr = tks_ptr,
.limit = 0xfffff,
.seg_32bit = 1,
.contents = 0, // Data
.read_exec_only = 0,
.limit_in_pages = 1,
.seg_not_present = 0,
.useable = 1,
};
}
2021-06-19 17:08:56 -05:00
const instance = @ptrCast(*Instance, @alignCast(@alignOf(Instance), &mapped[instance_offset]));
instance.* = .{
.fn_args = args,
2021-06-20 09:56:30 -05:00
.thread = .{ .mapped = mapped },
2021-06-19 17:08:56 -05:00
};
const flags: u32 = os.CLONE_VM | os.CLONE_FS | os.CLONE_FILES |
os.CLONE_SIGHAND | os.CLONE_THREAD | os.CLONE_SYSVSEM |
os.CLONE_PARENT_SETTID | os.CLONE_CHILD_CLEARTID |
os.CLONE_DETACHED | os.CLONE_SETTLS;
2021-06-25 12:43:03 -05:00
switch (linux.getErrno(linux.clone(
2021-06-19 17:08:56 -05:00
Instance.entryFn,
@ptrToInt(&mapped[stack_offset]),
flags,
2021-06-19 17:08:56 -05:00
@ptrToInt(instance),
&instance.thread.parent_tid,
tls_ptr,
&instance.thread.child_tid.value,
))) {
2021-06-25 12:43:03 -05:00
0 => return Impl{ .thread = &instance.thread },
os.EAGAIN => return error.ThreadQuotaExceeded,
os.EINVAL => unreachable,
2021-06-25 12:43:03 -05:00
os.ENOMEM => return error.SystemResources,
os.ENOSPC => unreachable,
os.EPERM => unreachable,
os.EUSERS => unreachable,
2021-06-25 12:43:03 -05:00
else => |err| return os.unexpectedErrno(err),
2021-06-25 12:59:28 -05:00
}
}
2021-06-19 17:08:56 -05:00
fn getHandle(self: Impl) ThreadHandle {
return self.thread.parent_tid;
}
2021-06-19 17:08:56 -05:00
fn detach(self: Impl) void {
switch (self.thread.completion.swap(.detached, .AcqRel)) {
.running => {},
.completed => self.join(),
.detached => unreachable,
}
}
2021-06-19 17:08:56 -05:00
fn join(self: Impl) void {
2021-06-20 09:56:30 -05:00
defer os.munmap(self.thread.mapped);
2021-06-19 17:08:56 -05:00
var spin: u8 = 10;
while (true) {
const tid = self.thread.child_tid.load(.Acquire);
if (tid == 0) {
break;
}
if (spin > 0) {
spin -= 1;
std.atomic.spinLoopHint();
continue;
}
switch (linux.getErrno(linux.futex_wait(
&self.thread.child_tid.value,
linux.FUTEX_WAIT,
tid,
null,
))) {
0 => continue,
os.EINTR => continue,
os.EAGAIN => continue,
else => unreachable,
}
}
}
2021-06-19 17:08:56 -05:00
// Calls `munmap(ptr, len)` then `exit(1)` without touching the stack (which lives in `ptr`).
2021-06-22 11:27:58 -05:00
// Ported over from musl libc's pthread detached implementation (`__unmapself`).
2021-06-19 17:08:56 -05:00
extern fn __unmap_and_exit(ptr: usize, len: usize) callconv(.C) noreturn;
comptime {
2021-06-22 11:27:58 -05:00
if (target.os.tag == .linux) {
asm(switch (target.cpu.arch) {
.i386 => (
\\.text
\\.global __unmap_and_exit
\\.type __unmap_and_exit, @function
\\__unmap_and_exit:
\\ movl $91, %eax
\\ movl 4(%esp), %ebx
\\ movl 8(%esp), %ecx
\\ int $128
\\ xorl %ebx, %ebx
\\ movl $1, %eax
\\ int $128
),
.x86_64 => (
\\.text
\\.global __unmap_and_exit
\\.type __unmap_and_exit, @function
\\__unmap_and_exit:
\\ movl $11, %eax
\\ syscall
\\ xor %rdi, %rdi
\\ movl $60, %eax
\\ syscall
),
2021-06-25 12:43:03 -05:00
.arm, .armeb, .thumb, .thumbeb => (
2021-06-22 11:27:58 -05:00
\\.syntax unified
\\.text
\\.global __unmap_and_exit
\\.type __unmap_and_exit, %function
\\__unmap_and_exit:
\\ mov r7, #91
\\ svc 0
\\ mov r7, #1
\\ svc 0
),
.aarch64, .aarch64_be, .aarch64_32 => (
\\.global __unmap_and_exit
\\.type __unmap_and_exit, %function
\\__unmap_and_exit:
\\ mov x8, #215
\\ svc 0
\\ mov x8, #93
\\ svc 0
),
.mips, .mipsel, => (
\\.set noreorder
\\.global __unmap_and_exit
\\.type __unmap_and_exit,@function
\\__unmap_and_exit:
\\ move $sp, $25
\\ li $2, 4091
\\ syscall
\\ li $4, 0
\\ li $2, 4001
\\ syscall
),
.mips64, .mips64el => (
\\.set noreorder
\\.global __unmap_and_exit
\\.type __unmap_and_exit, @function
\\__unmap_and_exit:
\\ li $2, 4091
\\ syscall
\\ li $4, 0
\\ li $2, 4001
\\ syscall
),
.powerpc, .powerpc64, .powerpc64le => (
\\.text
\\.global __unmap_and_exit
\\.type __unmap_and_exit, %function
\\__unmap_and_exit:
\\ li 0, 91
\\ sc
\\ li 0, 1
\\ sc
\\ blr
),
.riscv64 => (
\\.global __unmap_and_exit
\\.type __unmap_and_exit, %function
\\__unmap_and_exit:
\\ li a7, 215
\\ ecall
\\ li a7, 93
\\ ecall
),
else => @compileError("Platform not supported"),
});
}
}
2021-06-19 17:08:56 -05:00
};