X-Git-Url: https://git.ucc.asn.au/?a=blobdiff_plain;f=Kernel%2Farch%2Fx86%2Finclude%2Farch.h;h=c4b3a2e6813523630e149038e6de2eb7e8bef7cc;hb=41769c02317835472d7678d3531ecfc23df8e17a;hp=b3f7545ebc045972964cbe4a0dcb4b6dd9758a01;hpb=55048ddf380ad9a4ac6d4ee2cabba160a187c876;p=tpg%2Facess2.git diff --git a/Kernel/arch/x86/include/arch.h b/Kernel/arch/x86/include/arch.h index b3f7545e..c4b3a2e6 100644 --- a/Kernel/arch/x86/include/arch.h +++ b/Kernel/arch/x86/include/arch.h @@ -11,7 +11,8 @@ #define BITS 32 // Allow nested spinlocks? -#define STACKED_LOCKS 1 +#define STACKED_LOCKS 2 // 0: No, 1: Per-CPU, 2: Per-Thread +#define LOCK_DISABLE_INTS 1 // - Processor/Machine Specific Features #if ARCH != i386 && ARCH != i486 && ARCH != i586 @@ -37,87 +38,19 @@ * \brief Short Spinlock structure */ struct sShortSpinlock { + #if STACKED_LOCKS == 2 + volatile void *Lock; //!< Lock value + #else volatile int Lock; //!< Lock value - int IF; //!< Interrupt state on call to SHORTLOCK - #if STACKED_LOCKS - int Depth; - #endif -}; -/** - * \brief Determine if a short spinlock is locked - * \param Lock Lock pointer - */ -static inline int IS_LOCKED(struct sShortSpinlock *Lock) { - return !!Lock->Lock; -} -/** - * \brief Acquire a Short Spinlock - * \param Lock Lock pointer - * - * This type of mutex should only be used for very short sections of code, - * or in places where a Mutex_* would be overkill, such as appending - * an element to linked list (usually two assignement lines in C) - * - * \note This type of lock halts interrupts, so ensure that no timing - * functions are called while it is held. As a matter of fact, spend as - * little time as possible with this lock held - */ -static inline void SHORTLOCK(struct sShortSpinlock *Lock) { - int v = 1; - int IF; - #if STACKED_LOCKS - extern int GetCPUNum(void); - int cpu = GetCPUNum() + 1; #endif - // Save interrupt state and clear interrupts - __ASM__ ("pushf;\n\tpop %%eax\n\tcli" : "=a"(IF)); - IF &= 0x200; // AND out all but the interrupt flag - - #if STACKED_LOCKS - if( Lock->Lock == cpu ) { - Lock->Depth ++; - return ; - } + #if LOCK_DISABLE_INTS + int IF; //!< Interrupt state on call to SHORTLOCK #endif - - // Wait for another CPU to release - while(v) { - #if STACKED_LOCKS - // CMPXCHG: - // If r/m32 == EAX, set ZF and set r/m32 = r32 - // Else, clear ZF and set EAX = r/m32 - __ASM__("lock cmpxchgl %%ecx, (%%edi)" - : "=a"(v) - : "a"(0), "c"(cpu), "D"(&Lock->Lock) - ); - #else - __ASM__("xchgl %%eax, (%%edi)":"=a"(v):"a"(1),"D"(&Lock->Lock)); - #endif - } - - Lock->IF = IF; -} -/** - * \brief Release a short lock - * \param Lock Lock pointer - */ -static inline void SHORTREL(struct sShortSpinlock *Lock) { #if STACKED_LOCKS - if( Lock->Depth ) { - Lock->Depth --; - return ; - } + int Depth; #endif - // Lock->IF can change anytime once Lock->Lock is zeroed - if(Lock->IF) { - Lock->Lock = 0; - __ASM__ ("sti"); - } - else { - Lock->Lock = 0; - } -} +}; // === MACROS === /** @@ -141,16 +74,17 @@ typedef signed short Sint16; typedef signed long Sint32; typedef signed long long Sint64; typedef Uint size_t; +typedef char BOOL; typedef Uint64 tPAddr; typedef Uint32 tVAddr; typedef struct { - Uint gs, fs, es, ds; - Uint edi, esi, ebp, kesp; + Uint gs, fs, es, ds; + Uint edi, esi, ebp, kesp; Uint ebx, edx, ecx, eax; - Uint int_num, err_code; - Uint eip, cs; + Uint int_num, err_code; + Uint eip, cs; Uint eflags, esp, ss; } tRegs; @@ -189,4 +123,10 @@ typedef struct { Uint EIP, ESP, EBP; } tTaskState; +// === FUNCTIONS === +extern int IS_LOCKED(struct sShortSpinlock *Lock); +extern int CPU_HAS_LOCK(struct sShortSpinlock *Lock); +extern void SHORTLOCK(struct sShortSpinlock *Lock); +extern void SHORTREL(struct sShortSpinlock *Lock); + #endif // !defined(_ARCH_H_)