// Copyright 2023 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package runtime import ( "runtime/internal/atomic" "unsafe" ) // A Pinner is a set of Go objects each pinned to a fixed location in memory. The // [Pinner.Pin] method pins one object, while [Pinner.Unpin] unpins all pinned // objects. See their comments for more information. type Pinner struct { *pinner } // Pin pins a Go object, preventing it from being moved or freed by the garbage // collector until the [Pinner.Unpin] method has been called. // // A pointer to a pinned object can be directly stored in C memory or can be // contained in Go memory passed to C functions. If the pinned object itself // contains pointers to Go objects, these objects must be pinned separately if they // are going to be accessed from C code. // // The argument must be a pointer of any type or an [unsafe.Pointer]. // It's safe to call Pin on non-Go pointers, in which case Pin will do nothing. func (p *Pinner) Pin(pointer any) { if p.pinner == nil { // Check the pinner cache first. mp := acquirem() if pp := mp.p.ptr(); pp != nil { p.pinner = pp.pinnerCache pp.pinnerCache = nil } releasem(mp) if p.pinner == nil { // Didn't get anything from the pinner cache. p.pinner = new(pinner) p.refs = p.refStore[:0] // We set this finalizer once and never clear it. Thus, if the // pinner gets cached, we'll reuse it, along with its finalizer. // This lets us avoid the relatively expensive SetFinalizer call // when reusing from the cache. The finalizer however has to be // resilient to an empty pinner being finalized, which is done // by checking p.refs' length. SetFinalizer(p.pinner, func(i *pinner) { if len(i.refs) != 0 { i.unpin() // only required to make the test idempotent pinnerLeakPanic() } }) } } ptr := pinnerGetPtr(&pointer) if setPinned(ptr, true) { p.refs = append(p.refs, ptr) } } // Unpin unpins all pinned objects of the [Pinner]. func (p *Pinner) Unpin() { p.pinner.unpin() mp := acquirem() if pp := mp.p.ptr(); pp != nil && pp.pinnerCache == nil { // Put the pinner back in the cache, but only if the // cache is empty. If application code is reusing Pinners // on its own, we want to leave the backing store in place // so reuse is more efficient. pp.pinnerCache = p.pinner p.pinner = nil } releasem(mp) } const ( pinnerSize = 64 pinnerRefStoreSize = (pinnerSize - unsafe.Sizeof([]unsafe.Pointer{})) / unsafe.Sizeof(unsafe.Pointer(nil)) ) type pinner struct { refs []unsafe.Pointer refStore [pinnerRefStoreSize]unsafe.Pointer } func (p *pinner) unpin() { if p == nil || p.refs == nil { return } for i := range p.refs { setPinned(p.refs[i], false) } // The following two lines make all pointers to references // in p.refs unreachable, either by deleting them or dropping // p.refs' backing store (if it was not backed by refStore). p.refStore = [pinnerRefStoreSize]unsafe.Pointer{} p.refs = p.refStore[:0] } func pinnerGetPtr(i *any) unsafe.Pointer { e := efaceOf(i) etyp := e._type if etyp == nil { panic(errorString("runtime.Pinner: argument is nil")) } if kind := etyp.Kind_ & kindMask; kind != kindPtr && kind != kindUnsafePointer { panic(errorString("runtime.Pinner: argument is not a pointer: " + toRType(etyp).string())) } if inUserArenaChunk(uintptr(e.data)) { // Arena-allocated objects are not eligible for pinning. panic(errorString("runtime.Pinner: object was allocated into an arena")) } return e.data } // isPinned checks if a Go pointer is pinned. // nosplit, because it's called from nosplit code in cgocheck. // //go:nosplit func isPinned(ptr unsafe.Pointer) bool { span := spanOfHeap(uintptr(ptr)) if span == nil { // this code is only called for Go pointer, so this must be a // linker-allocated global object. return true } pinnerBits := span.getPinnerBits() // these pinnerBits might get unlinked by a concurrently running sweep, but // that's OK because gcBits don't get cleared until the following GC cycle // (nextMarkBitArenaEpoch) if pinnerBits == nil { return false } objIndex := span.objIndex(uintptr(ptr)) pinState := pinnerBits.ofObject(objIndex) KeepAlive(ptr) // make sure ptr is alive until we are done so the span can't be freed return pinState.isPinned() } // setPinned marks or unmarks a Go pointer as pinned, when the ptr is a Go pointer. // It will be ignored while try to pin a non-Go pointer, // and it will be panic while try to unpin a non-Go pointer, // which should not happen in normal usage. func setPinned(ptr unsafe.Pointer, pin bool) bool { span := spanOfHeap(uintptr(ptr)) if span == nil { if !pin { panic(errorString("tried to unpin non-Go pointer")) } // This is a linker-allocated, zero size object or other object, // nothing to do, silently ignore it. return false } // ensure that the span is swept, b/c sweeping accesses the specials list // w/o locks. mp := acquirem() span.ensureSwept() KeepAlive(ptr) // make sure ptr is still alive after span is swept objIndex := span.objIndex(uintptr(ptr)) lock(&span.speciallock) // guard against concurrent calls of setPinned on same span pinnerBits := span.getPinnerBits() if pinnerBits == nil { pinnerBits = span.newPinnerBits() span.setPinnerBits(pinnerBits) } pinState := pinnerBits.ofObject(objIndex) if pin { if pinState.isPinned() { // multiple pins on same object, set multipin bit pinState.setMultiPinned(true) // and increase the pin counter // TODO(mknyszek): investigate if systemstack is necessary here systemstack(func() { offset := objIndex * span.elemsize span.incPinCounter(offset) }) } else { // set pin bit pinState.setPinned(true) } } else { // unpin if pinState.isPinned() { if pinState.isMultiPinned() { var exists bool // TODO(mknyszek): investigate if systemstack is necessary here systemstack(func() { offset := objIndex * span.elemsize exists = span.decPinCounter(offset) }) if !exists { // counter is 0, clear multipin bit pinState.setMultiPinned(false) } } else { // no multipins recorded. unpin object. pinState.setPinned(false) } } else { // unpinning unpinned object, bail out throw("runtime.Pinner: object already unpinned") } } unlock(&span.speciallock) releasem(mp) return true } type pinState struct { bytep *uint8 byteVal uint8 mask uint8 } // nosplit, because it's called by isPinned, which is nosplit // //go:nosplit func (v *pinState) isPinned() bool { return (v.byteVal & v.mask) != 0 } func (v *pinState) isMultiPinned() bool { return (v.byteVal & (v.mask << 1)) != 0 } func (v *pinState) setPinned(val bool) { v.set(val, false) } func (v *pinState) setMultiPinned(val bool) { v.set(val, true) } // set sets the pin bit of the pinState to val. If multipin is true, it // sets/unsets the multipin bit instead. func (v *pinState) set(val bool, multipin bool) { mask := v.mask if multipin { mask <<= 1 } if val { atomic.Or8(v.bytep, mask) } else { atomic.And8(v.bytep, ^mask) } } // pinnerBits is the same type as gcBits but has different methods. type pinnerBits gcBits // ofObject returns the pinState of the n'th object. // nosplit, because it's called by isPinned, which is nosplit // //go:nosplit func (p *pinnerBits) ofObject(n uintptr) pinState { bytep, mask := (*gcBits)(p).bitp(n * 2) byteVal := atomic.Load8(bytep) return pinState{bytep, byteVal, mask} } func (s *mspan) pinnerBitSize() uintptr { return divRoundUp(uintptr(s.nelems)*2, 8) } // newPinnerBits returns a pointer to 8 byte aligned bytes to be used for this // span's pinner bits. newPinneBits is used to mark objects that are pinned. // They are copied when the span is swept. func (s *mspan) newPinnerBits() *pinnerBits { return (*pinnerBits)(newMarkBits(uintptr(s.nelems) * 2)) } // nosplit, because it's called by isPinned, which is nosplit // //go:nosplit func (s *mspan) getPinnerBits() *pinnerBits { return (*pinnerBits)(atomic.Loadp(unsafe.Pointer(&s.pinnerBits))) } func (s *mspan) setPinnerBits(p *pinnerBits) { atomicstorep(unsafe.Pointer(&s.pinnerBits), unsafe.Pointer(p)) } // refreshPinnerBits replaces pinnerBits with a fresh copy in the arenas for the // next GC cycle. If it does not contain any pinned objects, pinnerBits of the // span is set to nil. func (s *mspan) refreshPinnerBits() { p := s.getPinnerBits() if p == nil { return } hasPins := false bytes := alignUp(s.pinnerBitSize(), 8) // Iterate over each 8-byte chunk and check for pins. Note that // newPinnerBits guarantees that pinnerBits will be 8-byte aligned, so we // don't have to worry about edge cases, irrelevant bits will simply be // zero. for _, x := range unsafe.Slice((*uint64)(unsafe.Pointer(&p.x)), bytes/8) { if x != 0 { hasPins = true break } } if hasPins { newPinnerBits := s.newPinnerBits() memmove(unsafe.Pointer(&newPinnerBits.x), unsafe.Pointer(&p.x), bytes) s.setPinnerBits(newPinnerBits) } else { s.setPinnerBits(nil) } } // incPinCounter is only called for multiple pins of the same object and records // the _additional_ pins. func (span *mspan) incPinCounter(offset uintptr) { var rec *specialPinCounter ref, exists := span.specialFindSplicePoint(offset, _KindSpecialPinCounter) if !exists { lock(&mheap_.speciallock) rec = (*specialPinCounter)(mheap_.specialPinCounterAlloc.alloc()) unlock(&mheap_.speciallock) // splice in record, fill in offset. rec.special.offset = uint16(offset) rec.special.kind = _KindSpecialPinCounter rec.special.next = *ref *ref = (*special)(unsafe.Pointer(rec)) spanHasSpecials(span) } else { rec = (*specialPinCounter)(unsafe.Pointer(*ref)) } rec.counter++ } // decPinCounter decreases the counter. If the counter reaches 0, the counter // special is deleted and false is returned. Otherwise true is returned. func (span *mspan) decPinCounter(offset uintptr) bool { ref, exists := span.specialFindSplicePoint(offset, _KindSpecialPinCounter) if !exists { throw("runtime.Pinner: decreased non-existing pin counter") } counter := (*specialPinCounter)(unsafe.Pointer(*ref)) counter.counter-- if counter.counter == 0 { *ref = counter.special.next if span.specials == nil { spanHasNoSpecials(span) } lock(&mheap_.speciallock) mheap_.specialPinCounterAlloc.free(unsafe.Pointer(counter)) unlock(&mheap_.speciallock) return false } return true } // only for tests func pinnerGetPinCounter(addr unsafe.Pointer) *uintptr { _, span, objIndex := findObject(uintptr(addr), 0, 0) offset := objIndex * span.elemsize t, exists := span.specialFindSplicePoint(offset, _KindSpecialPinCounter) if !exists { return nil } counter := (*specialPinCounter)(unsafe.Pointer(*t)) return &counter.counter } // to be able to test that the GC panics when a pinned pointer is leaking, this // panic function is a variable, that can be overwritten by a test. var pinnerLeakPanic = func() { panic(errorString("runtime.Pinner: found leaking pinned pointer; forgot to call Unpin()?")) }