Text file src/internal/runtime/atomic/atomic_loong64.s

     1  // Copyright 2022 The Go Authors. All rights reserved.
     2  // Use of this source code is governed by a BSD-style
     3  // license that can be found in the LICENSE file.
     4  
     5  #include "go_asm.h"
     6  #include "textflag.h"
     7  
     8  // func Cas(ptr *int32, old, new int32) bool
     9  // Atomically:
    10  //	if *ptr == old {
    11  //		*ptr = new
    12  //		return true
    13  //	} else {
    14  //		return false
    15  //	}
    16  TEXT ·Cas(SB), NOSPLIT, $0-17
    17  	MOVV	ptr+0(FP), R4
    18  	MOVW	old+8(FP), R5
    19  	MOVW	new+12(FP), R6
    20  
    21  	MOVBU	internal∕cpu·Loong64+const_offsetLOONG64HasLAMCAS(SB), R8
    22  	BEQ	R8, ll_sc
    23  	MOVV	R5, R7  // backup old value
    24  	AMCASDBW	R6, (R4), R5
    25  	BNE	R7, R5, cas_fail0
    26  	MOVV	$1, R4
    27  	MOVB	R4, ret+16(FP)
    28  	RET
    29  cas_fail0:
    30  	MOVB	R0, ret+16(FP)
    31  	RET
    32  
    33  ll_sc:
    34  	// Implemented using the ll-sc instruction pair
    35  	DBAR	$0x14	// LoadAcquire barrier
    36  cas_again:
    37  	MOVV	R6, R7
    38  	LL	(R4), R8
    39  	BNE	R5, R8, cas_fail1
    40  	SC	R7, (R4)
    41  	BEQ	R7, cas_again
    42  	MOVV	$1, R4
    43  	MOVB	R4, ret+16(FP)
    44  	DBAR	$0x12	// StoreRelease barrier
    45  	RET
    46  cas_fail1:
    47  	MOVV	$0, R4
    48  	JMP	-4(PC)
    49  
    50  // func Cas64(ptr *uint64, old, new uint64) bool
    51  // Atomically:
    52  //	if *ptr == old {
    53  //		*ptr = new
    54  //		return true
    55  //	} else {
    56  //		return false
    57  //	}
    58  TEXT ·Cas64(SB), NOSPLIT, $0-25
    59  	MOVV	ptr+0(FP), R4
    60  	MOVV	old+8(FP), R5
    61  	MOVV	new+16(FP), R6
    62  
    63  	MOVBU	internal∕cpu·Loong64+const_offsetLOONG64HasLAMCAS(SB), R8
    64  	BEQ	R8, ll_sc_64
    65  	MOVV	R5, R7  // backup old value
    66  	AMCASDBV	R6, (R4), R5
    67  	BNE	R7, R5, cas64_fail0
    68  	MOVV	$1, R4
    69  	MOVB	R4, ret+24(FP)
    70  	RET
    71  cas64_fail0:
    72  	MOVB	R0, ret+24(FP)
    73  	RET
    74  
    75  ll_sc_64:
    76  	// Implemented using the ll-sc instruction pair
    77  	DBAR	$0x14
    78  cas64_again:
    79  	MOVV	R6, R7
    80  	LLV	(R4), R8
    81  	BNE	R5, R8, cas64_fail1
    82  	SCV	R7, (R4)
    83  	BEQ	R7, cas64_again
    84  	MOVV	$1, R4
    85  	MOVB	R4, ret+24(FP)
    86  	DBAR	$0x12
    87  	RET
    88  cas64_fail1:
    89  	MOVV	$0, R4
    90  	JMP	-4(PC)
    91  
    92  TEXT ·Casint32(SB),NOSPLIT,$0-17
    93  	JMP	·Cas(SB)
    94  
    95  TEXT ·Casint64(SB),NOSPLIT,$0-25
    96  	JMP	·Cas64(SB)
    97  
    98  TEXT ·Casuintptr(SB), NOSPLIT, $0-25
    99  	JMP	·Cas64(SB)
   100  
   101  TEXT ·CasRel(SB), NOSPLIT, $0-17
   102  	JMP	·Cas(SB)
   103  
   104  TEXT ·Loaduintptr(SB),  NOSPLIT|NOFRAME, $0-16
   105  	JMP	·Load64(SB)
   106  
   107  TEXT ·Loaduint(SB), NOSPLIT|NOFRAME, $0-16
   108  	JMP	·Load64(SB)
   109  
   110  TEXT ·Storeuintptr(SB), NOSPLIT, $0-16
   111  	JMP	·Store64(SB)
   112  
   113  TEXT ·Xadduintptr(SB), NOSPLIT, $0-24
   114  	JMP	·Xadd64(SB)
   115  
   116  TEXT ·Loadint64(SB), NOSPLIT, $0-16
   117  	JMP	·Load64(SB)
   118  
   119  TEXT ·Xaddint32(SB),NOSPLIT,$0-20
   120  	JMP	·Xadd(SB)
   121  
   122  TEXT ·Xaddint64(SB), NOSPLIT, $0-24
   123  	JMP	·Xadd64(SB)
   124  
   125  // func Casp(ptr *unsafe.Pointer, old, new unsafe.Pointer) bool
   126  // Atomically:
   127  //	if *ptr == old {
   128  //		*ptr = new
   129  //		return true
   130  //	} else {
   131  //		return false
   132  //	}
   133  TEXT ·Casp1(SB), NOSPLIT, $0-25
   134  	JMP	·Cas64(SB)
   135  
   136  // uint32 Xadd(uint32 volatile *ptr, int32 delta)
   137  // Atomically:
   138  //	*val += delta;
   139  //	return *val;
   140  TEXT ·Xadd(SB), NOSPLIT, $0-20
   141  	MOVV	ptr+0(FP), R4
   142  	MOVW	delta+8(FP), R5
   143  	AMADDDBW	R5, (R4), R6
   144  	ADDV	R6, R5, R4
   145  	MOVW	R4, ret+16(FP)
   146  	RET
   147  
   148  // func Xadd64(ptr *uint64, delta int64) uint64
   149  TEXT ·Xadd64(SB), NOSPLIT, $0-24
   150  	MOVV	ptr+0(FP), R4
   151  	MOVV	delta+8(FP), R5
   152  	AMADDDBV	R5, (R4), R6
   153  	ADDV	R6, R5, R4
   154  	MOVV	R4, ret+16(FP)
   155  	RET
   156  
   157  // uint8 Xchg8(ptr *uint8, new uint8)
   158  // Atomically:
   159  //     old := *ptr;
   160  //     *ptr = new;
   161  //     return old;
   162  TEXT ·Xchg8(SB), NOSPLIT, $0-17
   163  	MOVV	ptr+0(FP), R4
   164  	MOVBU	new+8(FP), R5
   165  
   166  	// R6 = ((ptr & 3) * 8)
   167  	AND	$3, R4, R6
   168  	SLLV	$3, R6
   169  
   170  	// R7 = ((0xFF) << R6) ^ (-1)
   171  	MOVV	$0xFF, R8
   172  	SLLV	R6, R8, R7
   173  	XOR	$-1, R7
   174  
   175  	// R4 = ptr & (~3)
   176  	MOVV	$~3, R8
   177  	AND	R8, R4
   178  
   179  	// R5 = ((val) << R6)
   180  	SLLV	R6, R5
   181  
   182  	DBAR	$0x14	// LoadAcquire barrier
   183  _xchg8_again:
   184  	LL	(R4), R8
   185  	MOVV	R8, R9	// backup old val
   186  	AND	R7, R8
   187  	OR	R5, R8
   188  	SC	R8, (R4)
   189  	BEQ	R8, _xchg8_again
   190  	DBAR	$0x12	// StoreRelease barrier
   191  	SRLV	R6, R9, R9
   192  	MOVBU	R9, ret+16(FP)
   193  	RET
   194  
   195  // func Xchg(ptr *uint32, new uint32) uint32
   196  TEXT ·Xchg(SB), NOSPLIT, $0-20
   197  	MOVV	ptr+0(FP), R4
   198  	MOVW	new+8(FP), R5
   199  	AMSWAPDBW	R5, (R4), R6
   200  	MOVW	R6, ret+16(FP)
   201  	RET
   202  
   203  // func Xchg64(ptr *uint64, new uint64) uint64
   204  TEXT ·Xchg64(SB), NOSPLIT, $0-24
   205  	MOVV	ptr+0(FP), R4
   206  	MOVV	new+8(FP), R5
   207  	AMSWAPDBV	R5, (R4), R6
   208  	MOVV	R6, ret+16(FP)
   209  	RET
   210  
   211  TEXT ·Xchguintptr(SB), NOSPLIT, $0-24
   212  	JMP	·Xchg64(SB)
   213  
   214  // func Xchgint32(ptr *int32, new int32) int32
   215  TEXT ·Xchgint32(SB), NOSPLIT, $0-20
   216  	JMP	·Xchg(SB)
   217  
   218  // func Xchgint64(ptr *int64, new int64) int64
   219  TEXT ·Xchgint64(SB), NOSPLIT, $0-24
   220  	JMP	·Xchg64(SB)
   221  
   222  TEXT ·StorepNoWB(SB), NOSPLIT, $0-16
   223  	JMP	·Store64(SB)
   224  
   225  TEXT ·StoreRel(SB), NOSPLIT, $0-12
   226  	JMP	·Store(SB)
   227  
   228  TEXT ·StoreRel64(SB), NOSPLIT, $0-16
   229  	JMP	·Store64(SB)
   230  
   231  TEXT ·StoreReluintptr(SB), NOSPLIT, $0-16
   232  	JMP     ·Store64(SB)
   233  
   234  TEXT ·Store(SB), NOSPLIT, $0-12
   235  	MOVV	ptr+0(FP), R4
   236  	MOVW	val+8(FP), R5
   237  	AMSWAPDBW	R5, (R4), R0
   238  	RET
   239  
   240  TEXT ·Store8(SB), NOSPLIT, $0-9
   241  	MOVV	ptr+0(FP), R4
   242  	MOVB	val+8(FP), R5
   243  	MOVBU	internal∕cpu·Loong64+const_offsetLoong64HasLAM_BH(SB), R6
   244  	BEQ	R6, _legacy_store8_
   245  	AMSWAPDBB	R5, (R4), R0
   246  	RET
   247  _legacy_store8_:
   248  	// StoreRelease barrier
   249  	DBAR	$0x12
   250  	MOVB	R5, 0(R4)
   251  	DBAR	$0x18
   252  	RET
   253  
   254  TEXT ·Store64(SB), NOSPLIT, $0-16
   255  	MOVV	ptr+0(FP), R4
   256  	MOVV	val+8(FP), R5
   257  	AMSWAPDBV	R5, (R4), R0
   258  	RET
   259  
   260  // void	Or8(byte volatile*, byte);
   261  TEXT ·Or8(SB), NOSPLIT, $0-9
   262  	MOVV	ptr+0(FP), R4
   263  	MOVBU	val+8(FP), R5
   264  	// R6 = ptr & (~3)
   265  	MOVV	$~3, R6
   266  	AND	R4, R6
   267  	// R7 = ((ptr & 3) * 8)
   268  	AND	$3, R4, R7
   269  	SLLV	$3, R7
   270  	// R5 = val << R7
   271  	SLLV	R7, R5
   272  	AMORDBW	R5, (R6), R0
   273  	RET
   274  
   275  // void	And8(byte volatile*, byte);
   276  TEXT ·And8(SB), NOSPLIT, $0-9
   277  	MOVV	ptr+0(FP), R4
   278  	MOVBU	val+8(FP), R5
   279  	// R6 = ptr & (~3)
   280  	MOVV	$~3, R6
   281  	AND	R4, R6
   282  	// R7 = ((ptr & 3) * 8)
   283  	AND	$3, R4, R7
   284  	SLLV	$3, R7
   285  	// R5 = ((val ^ 0xFF) << R7) ^ (-1)
   286  	XOR	$255, R5
   287  	SLLV	R7,  R5
   288  	XOR	$-1, R5
   289  	AMANDDBW	R5, (R6), R0
   290  	RET
   291  
   292  // func Or(addr *uint32, v uint32)
   293  TEXT ·Or(SB), NOSPLIT, $0-12
   294  	MOVV	ptr+0(FP), R4
   295  	MOVW	val+8(FP), R5
   296  	AMORDBW	R5, (R4), R0
   297  	RET
   298  
   299  // func And(addr *uint32, v uint32)
   300  TEXT ·And(SB), NOSPLIT, $0-12
   301  	MOVV	ptr+0(FP), R4
   302  	MOVW	val+8(FP), R5
   303  	AMANDDBW	R5, (R4), R0
   304  	RET
   305  
   306  // func Or32(addr *uint32, v uint32) old uint32
   307  TEXT ·Or32(SB), NOSPLIT, $0-20
   308  	MOVV	ptr+0(FP), R4
   309  	MOVW	val+8(FP), R5
   310  	AMORDBW R5, (R4), R6
   311  	MOVW	R6, ret+16(FP)
   312  	RET
   313  
   314  // func And32(addr *uint32, v uint32) old uint32
   315  TEXT ·And32(SB), NOSPLIT, $0-20
   316  	MOVV	ptr+0(FP), R4
   317  	MOVW	val+8(FP), R5
   318  	AMANDDBW	R5, (R4), R6
   319  	MOVW	R6, ret+16(FP)
   320  	RET
   321  
   322  // func Or64(addr *uint64, v uint64) old uint64
   323  TEXT ·Or64(SB), NOSPLIT, $0-24
   324  	MOVV	ptr+0(FP), R4
   325  	MOVV	val+8(FP), R5
   326  	AMORDBV	R5, (R4), R6
   327  	MOVV	R6, ret+16(FP)
   328  	RET
   329  
   330  // func And64(addr *uint64, v uint64) old uint64
   331  TEXT ·And64(SB), NOSPLIT, $0-24
   332  	MOVV	ptr+0(FP), R4
   333  	MOVV	val+8(FP), R5
   334  	AMANDDBV	R5, (R4), R6
   335  	MOVV	R6, ret+16(FP)
   336  	RET
   337  
   338  // func Anduintptr(addr *uintptr, v uintptr) old uintptr
   339  TEXT ·Anduintptr(SB), NOSPLIT, $0-24
   340  	JMP	·And64(SB)
   341  
   342  // func Oruintptr(addr *uintptr, v uintptr) old uintptr
   343  TEXT ·Oruintptr(SB), NOSPLIT, $0-24
   344  	JMP	·Or64(SB)
   345  
   346  // uint32 internal∕runtime∕atomic·Load(uint32 volatile* ptr)
   347  TEXT ·Load(SB),NOSPLIT|NOFRAME,$0-12
   348  	MOVV	ptr+0(FP), R19
   349  	MOVWU	0(R19), R19
   350  	DBAR	$0x14	// LoadAcquire barrier
   351  	MOVW	R19, ret+8(FP)
   352  	RET
   353  
   354  // uint8 internal∕runtime∕atomic·Load8(uint8 volatile* ptr)
   355  TEXT ·Load8(SB),NOSPLIT|NOFRAME,$0-9
   356  	MOVV	ptr+0(FP), R19
   357  	MOVBU	0(R19), R19
   358  	DBAR	$0x14
   359  	MOVB	R19, ret+8(FP)
   360  	RET
   361  
   362  // uint64 internal∕runtime∕atomic·Load64(uint64 volatile* ptr)
   363  TEXT ·Load64(SB),NOSPLIT|NOFRAME,$0-16
   364  	MOVV	ptr+0(FP), R19
   365  	MOVV	0(R19), R19
   366  	DBAR	$0x14
   367  	MOVV	R19, ret+8(FP)
   368  	RET
   369  
   370  // void *internal∕runtime∕atomic·Loadp(void *volatile *ptr)
   371  TEXT ·Loadp(SB),NOSPLIT|NOFRAME,$0-16
   372  	JMP     ·Load64(SB)
   373  
   374  // uint32 internal∕runtime∕atomic·LoadAcq(uint32 volatile* ptr)
   375  TEXT ·LoadAcq(SB),NOSPLIT|NOFRAME,$0-12
   376  	JMP	·Load(SB)
   377  
   378  // uint64 ·LoadAcq64(uint64 volatile* ptr)
   379  TEXT ·LoadAcq64(SB),NOSPLIT|NOFRAME,$0-16
   380  	JMP	·Load64(SB)
   381  
   382  // uintptr ·LoadAcquintptr(uintptr volatile* ptr)
   383  TEXT ·LoadAcquintptr(SB),NOSPLIT|NOFRAME,$0-16
   384  	JMP	·Load64(SB)
   385  
   386  

View as plain text