mirror of
https://github.com/ptitSeb/Serious-Engine
synced 2024-12-27 07:54:51 +01:00
24cb244d43
This was a _ton_ of changes, made 15 years ago, so there are probably some problems to work out still. Among others: Engine/Base/Stream.* was mostly abandoned and will need to be re-ported. Still, this is a pretty good start, and probably holds a world record for lines of changes or something. :)
145 lines
3.1 KiB
C++
145 lines
3.1 KiB
C++
/* Copyright (c) 2002-2012 Croteam Ltd. All rights reserved. */
|
|
|
|
#include "SDL.h"
|
|
#include "SDL_thread.h"
|
|
|
|
#include "Engine/StdH.h"
|
|
#include <Engine/Base/Synchronization.h>
|
|
|
|
// !!! FIXME: rcg10142001 Most of CTSingleLock is platform-independent.
|
|
|
|
CTCriticalSection::CTCriticalSection(void)
|
|
{
|
|
LockCounter = 0;
|
|
cs_pvObject = (void *) SDL_CreateMutex();
|
|
ASSERT(cs_pvObject != NULL);
|
|
}
|
|
|
|
CTCriticalSection::~CTCriticalSection(void)
|
|
{
|
|
SDL_DestroyMutex((SDL_mutex *) cs_pvObject);
|
|
}
|
|
|
|
INDEX CTCriticalSection::Lock(void)
|
|
{
|
|
LockCounter++;
|
|
if (LockCounter == 1)
|
|
SDL_LockMutex((SDL_mutex *) cs_pvObject);
|
|
return(LockCounter);
|
|
}
|
|
|
|
INDEX CTCriticalSection::TryToLock(void)
|
|
{
|
|
if (LockCounter > 0) // !!! race condition. Ironic, eh?
|
|
return(0);
|
|
Lock();
|
|
return(1);
|
|
}
|
|
|
|
INDEX CTCriticalSection::Unlock(void)
|
|
{
|
|
if (LockCounter > 0)
|
|
{
|
|
LockCounter--;
|
|
if (LockCounter == 0)
|
|
SDL_UnlockMutex((SDL_mutex *) cs_pvObject);
|
|
}
|
|
|
|
return(LockCounter);
|
|
}
|
|
|
|
CTSingleLock::CTSingleLock(CTCriticalSection *pcs, BOOL bLock) : sl_cs(*pcs)
|
|
{
|
|
// initially not locked
|
|
sl_bLocked = FALSE;
|
|
sl_iLastLockedIndex = -2;
|
|
// critical section must have index assigned
|
|
//ASSERT(sl_cs.cs_iIndex>=1||sl_cs.cs_iIndex==-1);
|
|
// if should lock immediately
|
|
if (bLock) {
|
|
Lock();
|
|
}
|
|
}
|
|
CTSingleLock::~CTSingleLock(void)
|
|
{
|
|
// if locked
|
|
if (sl_bLocked) {
|
|
// unlock
|
|
Unlock();
|
|
}
|
|
}
|
|
void CTSingleLock::Lock(void)
|
|
{
|
|
// must not be locked
|
|
ASSERT(!sl_bLocked);
|
|
ASSERT(sl_iLastLockedIndex==-2);
|
|
|
|
// if not locked
|
|
if (!sl_bLocked) {
|
|
// lock
|
|
INDEX ctLocks = sl_cs.Lock();
|
|
// if this mutex was not locked already
|
|
// if (ctLocks==1) {
|
|
// // check that locking in given order
|
|
// if (sl_cs.cs_iIndex!=-1) {
|
|
// ASSERT(_iLastLockedMutex<sl_cs.cs_iIndex);
|
|
// sl_iLastLockedIndex = _iLastLockedMutex;
|
|
// _iLastLockedMutex = sl_cs.cs_iIndex;
|
|
// }
|
|
// }
|
|
}
|
|
sl_bLocked = TRUE;
|
|
}
|
|
|
|
BOOL CTSingleLock::TryToLock(void)
|
|
{
|
|
// must not be locked
|
|
ASSERT(!sl_bLocked);
|
|
// if not locked
|
|
if (!sl_bLocked) {
|
|
// if can lock
|
|
INDEX ctLocks = sl_cs.TryToLock();
|
|
if (ctLocks>=1) {
|
|
sl_bLocked = TRUE;
|
|
|
|
// if this mutex was not locked already
|
|
// if (ctLocks==1) {
|
|
// // check that locking in given order
|
|
// if (sl_cs.cs_iIndex!=-1) {
|
|
// ASSERT(_iLastLockedMutex<sl_cs.cs_iIndex);
|
|
// sl_iLastLockedIndex = _iLastLockedMutex;
|
|
// _iLastLockedMutex = sl_cs.cs_iIndex;
|
|
// }
|
|
// }
|
|
}
|
|
}
|
|
return sl_bLocked;
|
|
}
|
|
BOOL CTSingleLock::IsLocked(void)
|
|
{
|
|
return sl_bLocked;
|
|
}
|
|
|
|
void CTSingleLock::Unlock(void)
|
|
{
|
|
// must be locked
|
|
ASSERT(sl_bLocked);
|
|
// if locked
|
|
if (sl_bLocked) {
|
|
// unlock
|
|
INDEX ctLocks = sl_cs.Unlock();
|
|
// if unlocked completely
|
|
if (ctLocks==0) {
|
|
// check that unlocking in exact reverse order
|
|
// if (sl_cs.cs_iIndex!=-1) {
|
|
// ASSERT(_iLastLockedMutex==sl_cs.cs_iIndex);
|
|
// _iLastLockedMutex = sl_iLastLockedIndex;
|
|
// sl_iLastLockedIndex = -2;
|
|
// }
|
|
}
|
|
}
|
|
sl_bLocked = FALSE;
|
|
}
|
|
|
|
|