1 | /* $NetBSD: mutex.h,v 1.20 2010/02/08 09:54:27 skrll Exp $ */ |
2 | |
3 | /*- |
4 | * Copyright (c) 2002, 2006, 2007, 2008, 2009 The NetBSD Foundation, Inc. |
5 | * All rights reserved. |
6 | * |
7 | * This code is derived from software contributed to The NetBSD Foundation |
8 | * by Jason R. Thorpe and Andrew Doran. |
9 | * |
10 | * Redistribution and use in source and binary forms, with or without |
11 | * modification, are permitted provided that the following conditions |
12 | * are met: |
13 | * 1. Redistributions of source code must retain the above copyright |
14 | * notice, this list of conditions and the following disclaimer. |
15 | * 2. Redistributions in binary form must reproduce the above copyright |
16 | * notice, this list of conditions and the following disclaimer in the |
17 | * documentation and/or other materials provided with the distribution. |
18 | * |
19 | * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
20 | * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
21 | * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
22 | * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS |
23 | * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
24 | * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
25 | * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
26 | * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
27 | * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
28 | * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
29 | * POSSIBILITY OF SUCH DAMAGE. |
30 | */ |
31 | |
32 | #ifndef _SYS_MUTEX_H_ |
33 | #define _SYS_MUTEX_H_ |
34 | |
35 | /* |
36 | * There are 2 types of mutexes: |
37 | * |
38 | * * Adaptive -- If the lock is already held, the thread attempting |
39 | * to acquire the lock determines if the thread that holds it is |
40 | * currently running. If so, it spins, else it sleeps. |
41 | * |
42 | * * Spin -- If the lock is already held, the thread attempting to |
43 | * acquire the lock spins. The IPL will be raised on entry. |
44 | * |
45 | * Machine dependent code must provide the following: |
46 | * |
47 | * struct mutex |
48 | * The actual mutex structure. This structure is mostly |
49 | * opaque to machine-independent code; most access are done |
50 | * through macros. However, machine-independent code must |
51 | * be able to access the following members: |
52 | * |
53 | * uintptr_t mtx_owner |
54 | * ipl_cookie_t mtx_ipl |
55 | * __cpu_simple_lock_t mtx_lock |
56 | * |
57 | * If an architecture can be considered 'simple' (no interlock required in |
58 | * the MP case, or no MP) it need only define __HAVE_SIMPLE_MUTEXES and |
59 | * provide the following: |
60 | * |
61 | * struct mutex |
62 | * |
63 | * [additionally:] |
64 | * volatile integer mtx_id |
65 | * |
66 | * MUTEX_RECEIVE(mtx) |
67 | * Post a load fence after acquiring the mutex, if necessary. |
68 | * |
69 | * MUTEX_GIVE(mtx) |
70 | * Post a load/store fence after releasing the mutex, if |
71 | * necessary. |
72 | * |
73 | * MUTEX_CAS(ptr, old, new) |
74 | * Perform an atomic "compare and swap" operation and |
75 | * evaluate to true or false according to the success |
76 | * |
77 | * Otherwise, the following must be defined: |
78 | * |
79 | * MUTEX_INITIALIZE_SPIN(mtx, dodebug, minipl) |
80 | * Initialize a spin mutex. |
81 | * |
82 | * MUTEX_INITIALIZE_ADAPTIVE(mtx, dodebug) |
83 | * Initialize an adaptive mutex. |
84 | * |
85 | * MUTEX_DESTROY(mtx) |
86 | * Tear down a mutex. |
87 | * |
88 | * MUTEX_ADAPTIVE_P(mtx) |
89 | * Evaluates to true if the mutex is an adaptive mutex. |
90 | * |
91 | * MUTEX_SPIN_P(mtx) |
92 | * Evaluates to true if the mutex is a spin mutex. |
93 | * |
94 | * MUTEX_OWNER(owner) |
95 | * Returns the owner of the adaptive mutex (LWP address). |
96 | * |
97 | * MUTEX_OWNED(owner) |
98 | * Returns non-zero if an adaptive mutex is currently |
99 | * held by an LWP. |
100 | * |
101 | * MUTEX_HAS_WAITERS(mtx) |
102 | * Returns true if the mutex has waiters. |
103 | * |
104 | * MUTEX_SET_WAITERS(mtx) |
105 | * Mark the mutex has having waiters. |
106 | * |
107 | * MUTEX_ACQUIRE(mtx, owner) |
108 | * Try to acquire an adaptive mutex such that: |
109 | * if (lock held OR waiters) |
110 | * return 0; |
111 | * else |
112 | * return 1; |
113 | * Must be MP/interrupt atomic. |
114 | * |
115 | * MUTEX_RELEASE(mtx) |
116 | * Release the lock and clear the "has waiters" indication. |
117 | * Must be interrupt atomic, need not be MP safe. |
118 | * |
119 | * MUTEX_DEBUG_P(mtx) |
120 | * Evaluates to true if the mutex is initialized with |
121 | * dodebug==true. Only used in the LOCKDEBUG case. |
122 | * |
123 | * Machine dependent code may optionally provide stubs for the following |
124 | * functions to implement the easy (unlocked / no waiters) cases. If |
125 | * these stubs are provided, __HAVE_MUTEX_STUBS should be defined. |
126 | * |
127 | * mutex_enter() |
128 | * mutex_exit() |
129 | * |
130 | * Two additional stubs may be implemented that handle only the spinlock |
131 | * case, primarily for the scheduler. __HAVE_SPIN_MUTEX_STUBS should be |
132 | * defined if these are provided: |
133 | * |
134 | * mutex_spin_enter() |
135 | * mutex_spin_exit() |
136 | */ |
137 | |
138 | #if defined(_KERNEL_OPT) |
139 | #include "opt_lockdebug.h" |
140 | #endif |
141 | |
142 | #if !defined(_KERNEL) |
143 | #include <sys/types.h> |
144 | #include <sys/inttypes.h> |
145 | #endif |
146 | |
147 | typedef enum kmutex_type_t { |
148 | MUTEX_SPIN = 0, /* To get a spin mutex at IPL_NONE */ |
149 | MUTEX_ADAPTIVE = 1, /* For porting code written for Solaris */ |
150 | MUTEX_DEFAULT = 2, /* The only native, endorsed type */ |
151 | MUTEX_DRIVER = 3, /* For porting code written for Solaris */ |
152 | MUTEX_NODEBUG = 4 /* Disables LOCKDEBUG; use with care */ |
153 | } kmutex_type_t; |
154 | |
155 | typedef struct kmutex kmutex_t; |
156 | |
157 | #if defined(__MUTEX_PRIVATE) |
158 | |
159 | #define MUTEX_THREAD ((uintptr_t)-16L) |
160 | |
161 | #define MUTEX_BIT_SPIN 0x01 |
162 | #define MUTEX_BIT_WAITERS 0x02 |
163 | |
164 | #if defined(LOCKDEBUG) |
165 | #define MUTEX_BIT_NODEBUG 0x04 /* LOCKDEBUG disabled */ |
166 | #else |
167 | #define MUTEX_BIT_NODEBUG 0x00 /* do nothing */ |
168 | #endif /* LOCKDEBUG */ |
169 | |
170 | #define MUTEX_SPIN_IPL(mtx) ((mtx)->mtx_ipl) |
171 | #define MUTEX_SPIN_OLDSPL(ci) ((ci)->ci_mtx_oldspl) |
172 | |
173 | void mutex_vector_enter(kmutex_t *); |
174 | void mutex_vector_exit(kmutex_t *); |
175 | void mutex_spin_retry(kmutex_t *); |
176 | void mutex_wakeup(kmutex_t *); |
177 | |
178 | #endif /* __MUTEX_PRIVATE */ |
179 | |
180 | #ifdef _KERNEL |
181 | #include <sys/intr.h> |
182 | #endif |
183 | |
184 | #include <machine/mutex.h> |
185 | |
186 | /* |
187 | * Return true if no spin mutexes are held by the current CPU. |
188 | */ |
189 | #ifndef MUTEX_NO_SPIN_ACTIVE_P |
190 | #define MUTEX_NO_SPIN_ACTIVE_P(ci) ((ci)->ci_mtx_count == 0) |
191 | #endif |
192 | |
193 | #ifdef _KERNEL |
194 | |
195 | void mutex_init(kmutex_t *, kmutex_type_t, int); |
196 | void mutex_destroy(kmutex_t *); |
197 | |
198 | void mutex_enter(kmutex_t *); |
199 | void mutex_exit(kmutex_t *); |
200 | |
201 | void mutex_spin_enter(kmutex_t *); |
202 | void mutex_spin_exit(kmutex_t *); |
203 | |
204 | int mutex_tryenter(kmutex_t *); |
205 | |
206 | int mutex_owned(kmutex_t *); |
207 | lwp_t *mutex_owner(kmutex_t *); |
208 | |
209 | void mutex_obj_init(void); |
210 | kmutex_t *mutex_obj_alloc(kmutex_type_t, int); |
211 | void mutex_obj_hold(kmutex_t *); |
212 | bool mutex_obj_free(kmutex_t *); |
213 | |
214 | #endif /* _KERNEL */ |
215 | |
216 | #endif /* _SYS_MUTEX_H_ */ |
217 | |