Linux vmi284606.contaboserver.net 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 13:30:12 UTC 2023 x86_64
Apache/2.4.57 (Ubuntu)
: 167.86.127.34 | : 216.73.217.31
Cant Read [ /etc/named.conf ]
7.2.24-0ubuntu0.18.04.17
root
Terminal
AUTO ROOT
Adminer
Backdoor Destroyer
Linux Exploit
Lock Shell
Lock File
Create User
CREATE RDP
PHP Mailer
BACKCONNECT
UNLOCK SHELL
HASH IDENTIFIER
README
+ Create Folder
+ Create File
/
usr /
local /
go /
src /
runtime /
internal /
atomic /
[ HOME SHELL ]
Name
Size
Permission
Action
asm_386.s
5.27
KB
-rw-r--r--
asm_amd64.s
3.89
KB
-rw-r--r--
asm_arm.s
5.22
KB
-rw-r--r--
asm_arm64.s
1.63
KB
-rw-r--r--
asm_mips64x.s
3.96
KB
-rw-r--r--
asm_mipsx.s
2.88
KB
-rw-r--r--
asm_ppc64x.s
4.48
KB
-rw-r--r--
asm_s390x.s
4.79
KB
-rw-r--r--
atomic_386.go
1.62
KB
-rw-r--r--
atomic_amd64.go
1.73
KB
-rw-r--r--
atomic_arm.go
4.31
KB
-rw-r--r--
atomic_arm64.go
1.38
KB
-rw-r--r--
atomic_arm64.s
3.21
KB
-rw-r--r--
atomic_mips64x.go
1.45
KB
-rw-r--r--
atomic_mips64x.s
1.06
KB
-rw-r--r--
atomic_mipsx.go
2.49
KB
-rw-r--r--
atomic_mipsx.s
482
B
-rw-r--r--
atomic_ppc64x.go
1.45
KB
-rw-r--r--
atomic_ppc64x.s
1.34
KB
-rw-r--r--
atomic_riscv64.go
1.37
KB
-rw-r--r--
atomic_riscv64.s
5.48
KB
-rw-r--r--
atomic_s390x.go
1.68
KB
-rw-r--r--
atomic_test.go
5.1
KB
-rw-r--r--
atomic_wasm.go
3.69
KB
-rw-r--r--
bench_test.go
1.78
KB
-rw-r--r--
stubs.go
795
B
-rw-r--r--
sys_linux_arm.s
3.14
KB
-rw-r--r--
sys_nonlinux_arm.s
1.34
KB
-rw-r--r--
Delete
Unzip
Zip
${this.title}
Close
Code Editor : asm_386.s
// Copyright 2015 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. #include "textflag.h" // bool Cas(int32 *val, int32 old, int32 new) // Atomically: // if(*val == old){ // *val = new; // return 1; // }else // return 0; TEXT runtime∕internal∕atomic·Cas(SB), NOSPLIT, $0-13 MOVL ptr+0(FP), BX MOVL old+4(FP), AX MOVL new+8(FP), CX LOCK CMPXCHGL CX, 0(BX) SETEQ ret+12(FP) RET TEXT runtime∕internal∕atomic·Casuintptr(SB), NOSPLIT, $0-13 JMP runtime∕internal∕atomic·Cas(SB) TEXT runtime∕internal∕atomic·CasRel(SB), NOSPLIT, $0-13 JMP runtime∕internal∕atomic·Cas(SB) TEXT runtime∕internal∕atomic·Loaduintptr(SB), NOSPLIT, $0-8 JMP runtime∕internal∕atomic·Load(SB) TEXT runtime∕internal∕atomic·Loaduint(SB), NOSPLIT, $0-8 JMP runtime∕internal∕atomic·Load(SB) TEXT runtime∕internal∕atomic·Storeuintptr(SB), NOSPLIT, $0-8 JMP runtime∕internal∕atomic·Store(SB) TEXT runtime∕internal∕atomic·Xadduintptr(SB), NOSPLIT, $0-12 JMP runtime∕internal∕atomic·Xadd(SB) TEXT runtime∕internal∕atomic·Loadint64(SB), NOSPLIT, $0-12 JMP runtime∕internal∕atomic·Load64(SB) TEXT runtime∕internal∕atomic·Xaddint64(SB), NOSPLIT, $0-20 JMP runtime∕internal∕atomic·Xadd64(SB) // bool runtime∕internal∕atomic·Cas64(uint64 *val, uint64 old, uint64 new) // Atomically: // if(*val == *old){ // *val = new; // return 1; // } else { // return 0; // } TEXT runtime∕internal∕atomic·Cas64(SB), NOSPLIT, $0-21 MOVL ptr+0(FP), BP TESTL $7, BP JZ 2(PC) MOVL 0, BP // crash with nil ptr deref MOVL old_lo+4(FP), AX MOVL old_hi+8(FP), DX MOVL new_lo+12(FP), BX MOVL new_hi+16(FP), CX LOCK CMPXCHG8B 0(BP) SETEQ ret+20(FP) RET // bool Casp1(void **p, void *old, void *new) // Atomically: // if(*p == old){ // *p = new; // return 1; // }else // return 0; TEXT runtime∕internal∕atomic·Casp1(SB), NOSPLIT, $0-13 MOVL ptr+0(FP), BX MOVL old+4(FP), AX MOVL new+8(FP), CX LOCK CMPXCHGL CX, 0(BX) SETEQ ret+12(FP) RET // uint32 Xadd(uint32 volatile *val, int32 delta) // Atomically: // *val += delta; // return *val; TEXT runtime∕internal∕atomic·Xadd(SB), NOSPLIT, $0-12 MOVL ptr+0(FP), BX MOVL delta+4(FP), AX MOVL AX, CX LOCK XADDL AX, 0(BX) ADDL CX, AX MOVL AX, ret+8(FP) RET TEXT runtime∕internal∕atomic·Xadd64(SB), NOSPLIT, $0-20 // no XADDQ so use CMPXCHG8B loop MOVL ptr+0(FP), BP TESTL $7, BP JZ 2(PC) MOVL 0, AX // crash when unaligned // DI:SI = delta MOVL delta_lo+4(FP), SI MOVL delta_hi+8(FP), DI // DX:AX = *addr MOVL 0(BP), AX MOVL 4(BP), DX addloop: // CX:BX = DX:AX (*addr) + DI:SI (delta) MOVL AX, BX MOVL DX, CX ADDL SI, BX ADCL DI, CX // if *addr == DX:AX { // *addr = CX:BX // } else { // DX:AX = *addr // } // all in one instruction LOCK CMPXCHG8B 0(BP) JNZ addloop // success // return CX:BX MOVL BX, ret_lo+12(FP) MOVL CX, ret_hi+16(FP) RET TEXT runtime∕internal∕atomic·Xchg(SB), NOSPLIT, $0-12 MOVL ptr+0(FP), BX MOVL new+4(FP), AX XCHGL AX, 0(BX) MOVL AX, ret+8(FP) RET TEXT runtime∕internal∕atomic·Xchguintptr(SB), NOSPLIT, $0-12 JMP runtime∕internal∕atomic·Xchg(SB) TEXT runtime∕internal∕atomic·Xchg64(SB),NOSPLIT,$0-20 // no XCHGQ so use CMPXCHG8B loop MOVL ptr+0(FP), BP TESTL $7, BP JZ 2(PC) MOVL 0, AX // crash when unaligned // CX:BX = new MOVL new_lo+4(FP), BX MOVL new_hi+8(FP), CX // DX:AX = *addr MOVL 0(BP), AX MOVL 4(BP), DX swaploop: // if *addr == DX:AX // *addr = CX:BX // else // DX:AX = *addr // all in one instruction LOCK CMPXCHG8B 0(BP) JNZ swaploop // success // return DX:AX MOVL AX, ret_lo+12(FP) MOVL DX, ret_hi+16(FP) RET TEXT runtime∕internal∕atomic·StorepNoWB(SB), NOSPLIT, $0-8 MOVL ptr+0(FP), BX MOVL val+4(FP), AX XCHGL AX, 0(BX) RET TEXT runtime∕internal∕atomic·Store(SB), NOSPLIT, $0-8 MOVL ptr+0(FP), BX MOVL val+4(FP), AX XCHGL AX, 0(BX) RET TEXT runtime∕internal∕atomic·StoreRel(SB), NOSPLIT, $0-8 JMP runtime∕internal∕atomic·Store(SB) // uint64 atomicload64(uint64 volatile* addr); TEXT runtime∕internal∕atomic·Load64(SB), NOSPLIT, $0-12 MOVL ptr+0(FP), AX TESTL $7, AX JZ 2(PC) MOVL 0, AX // crash with nil ptr deref MOVQ (AX), M0 MOVQ M0, ret+4(FP) EMMS RET // void runtime∕internal∕atomic·Store64(uint64 volatile* addr, uint64 v); TEXT runtime∕internal∕atomic·Store64(SB), NOSPLIT, $0-12 MOVL ptr+0(FP), AX TESTL $7, AX JZ 2(PC) MOVL 0, AX // crash with nil ptr deref // MOVQ and EMMS were introduced on the Pentium MMX. MOVQ val+4(FP), M0 MOVQ M0, (AX) EMMS // This is essentially a no-op, but it provides required memory fencing. // It can be replaced with MFENCE, but MFENCE was introduced only on the Pentium4 (SSE2). XORL AX, AX LOCK XADDL AX, (SP) RET // void runtime∕internal∕atomic·Or8(byte volatile*, byte); TEXT runtime∕internal∕atomic·Or8(SB), NOSPLIT, $0-5 MOVL ptr+0(FP), AX MOVB val+4(FP), BX LOCK ORB BX, (AX) RET // void runtime∕internal∕atomic·And8(byte volatile*, byte); TEXT runtime∕internal∕atomic·And8(SB), NOSPLIT, $0-5 MOVL ptr+0(FP), AX MOVB val+4(FP), BX LOCK ANDB BX, (AX) RET TEXT runtime∕internal∕atomic·Store8(SB), NOSPLIT, $0-5 MOVL ptr+0(FP), BX MOVB val+4(FP), AX XCHGB AX, 0(BX) RET
Close