You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							240 lines
						
					
					
						
							7.1 KiB
						
					
					
				
			
		
		
		
			
			
			
				
					
				
				
					
				
			
		
		
	
	
							240 lines
						
					
					
						
							7.1 KiB
						
					
					
				| /* | |
|     Copyright 2005-2014 Intel Corporation.  All Rights Reserved. | |
|  | |
|     This file is part of Threading Building Blocks. | |
|  | |
|     Threading Building Blocks is free software; you can redistribute it | |
|     and/or modify it under the terms of the GNU General Public License | |
|     version 2 as published by the Free Software Foundation. | |
|  | |
|     Threading Building Blocks is distributed in the hope that it will be | |
|     useful, but WITHOUT ANY WARRANTY; without even the implied warranty | |
|     of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | |
|     GNU General Public License for more details. | |
|  | |
|     You should have received a copy of the GNU General Public License | |
|     along with Threading Building Blocks; if not, write to the Free Software | |
|     Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA | |
|  | |
|     As a special exception, you may use this file as part of a free software | |
|     library without restriction.  Specifically, if other files instantiate | |
|     templates or use macros or inline functions from this file, or you compile | |
|     this file and link it with other files to produce an executable, this | |
|     file does not by itself cause the resulting executable to be covered by | |
|     the GNU General Public License.  This exception does not however | |
|     invalidate any other reasons why the executable file might be covered by | |
|     the GNU General Public License. | |
| */ | |
| 
 | |
| #ifndef __TBB_recursive_mutex_H | |
| #define __TBB_recursive_mutex_H | |
|  | |
| #if _WIN32||_WIN64 | |
| #include "machine/windows_api.h" | |
| #else | |
| #include <pthread.h> | |
| #endif /* _WIN32||_WIN64 */ | |
|  | |
| #include <new> | |
| #include "aligned_space.h" | |
| #include "tbb_stddef.h" | |
| #include "tbb_profiling.h" | |
|  | |
| namespace tbb { | |
| //! Mutex that allows recursive mutex acquisition. | |
| /** Mutex that allows recursive mutex acquisition. | |
|     @ingroup synchronization */ | |
| class recursive_mutex { | |
| public: | |
|     //! Construct unacquired recursive_mutex. | |
|     recursive_mutex() { | |
| #if TBB_USE_ASSERT || TBB_USE_THREADING_TOOLS | |
|         internal_construct(); | |
| #else | |
|   #if _WIN32||_WIN64 | |
|         InitializeCriticalSectionEx(&impl, 4000, 0); | |
|   #else | |
|         pthread_mutexattr_t mtx_attr; | |
|         int error_code = pthread_mutexattr_init( &mtx_attr ); | |
|         if( error_code ) | |
|             tbb::internal::handle_perror(error_code,"recursive_mutex: pthread_mutexattr_init failed"); | |
| 
 | |
|         pthread_mutexattr_settype( &mtx_attr, PTHREAD_MUTEX_RECURSIVE ); | |
|         error_code = pthread_mutex_init( &impl, &mtx_attr ); | |
|         if( error_code ) | |
|             tbb::internal::handle_perror(error_code,"recursive_mutex: pthread_mutex_init failed"); | |
| 
 | |
|         pthread_mutexattr_destroy( &mtx_attr ); | |
|   #endif /* _WIN32||_WIN64*/ | |
| #endif /* TBB_USE_ASSERT */ | |
|     }; | |
| 
 | |
|     ~recursive_mutex() { | |
| #if TBB_USE_ASSERT | |
|         internal_destroy(); | |
| #else | |
|   #if _WIN32||_WIN64 | |
|         DeleteCriticalSection(&impl); | |
|   #else | |
|         pthread_mutex_destroy(&impl);  | |
| 
 | |
|   #endif /* _WIN32||_WIN64 */ | |
| #endif /* TBB_USE_ASSERT */ | |
|     }; | |
| 
 | |
|     class scoped_lock; | |
|     friend class scoped_lock; | |
| 
 | |
|     //! The scoped locking pattern | |
|     /** It helps to avoid the common problem of forgetting to release lock. | |
|         It also nicely provides the "node" for queuing locks. */ | |
|     class scoped_lock: internal::no_copy { | |
|     public: | |
|         //! Construct lock that has not acquired a recursive_mutex.  | |
|         scoped_lock() : my_mutex(NULL) {}; | |
| 
 | |
|         //! Acquire lock on given mutex. | |
|         scoped_lock( recursive_mutex& mutex ) { | |
| #if TBB_USE_ASSERT | |
|             my_mutex = &mutex;  | |
| #endif /* TBB_USE_ASSERT */ | |
|             acquire( mutex ); | |
|         } | |
| 
 | |
|         //! Release lock (if lock is held). | |
|         ~scoped_lock() { | |
|             if( my_mutex )  | |
|                 release(); | |
|         } | |
| 
 | |
|         //! Acquire lock on given mutex. | |
|         void acquire( recursive_mutex& mutex ) { | |
| #if TBB_USE_ASSERT | |
|             internal_acquire( mutex ); | |
| #else | |
|             my_mutex = &mutex; | |
|             mutex.lock(); | |
| #endif /* TBB_USE_ASSERT */ | |
|         } | |
| 
 | |
|         //! Try acquire lock on given recursive_mutex. | |
|         bool try_acquire( recursive_mutex& mutex ) { | |
| #if TBB_USE_ASSERT | |
|             return internal_try_acquire( mutex ); | |
| #else | |
|             bool result = mutex.try_lock(); | |
|             if( result ) | |
|                 my_mutex = &mutex; | |
|             return result; | |
| #endif /* TBB_USE_ASSERT */ | |
|         } | |
| 
 | |
|         //! Release lock | |
|         void release() { | |
| #if TBB_USE_ASSERT | |
|             internal_release(); | |
| #else | |
|             my_mutex->unlock(); | |
|             my_mutex = NULL; | |
| #endif /* TBB_USE_ASSERT */ | |
|         } | |
| 
 | |
|     private: | |
|         //! The pointer to the current recursive_mutex to work | |
|         recursive_mutex* my_mutex; | |
| 
 | |
|         //! All checks from acquire using mutex.state were moved here | |
|         void __TBB_EXPORTED_METHOD internal_acquire( recursive_mutex& m ); | |
| 
 | |
|         //! All checks from try_acquire using mutex.state were moved here | |
|         bool __TBB_EXPORTED_METHOD internal_try_acquire( recursive_mutex& m ); | |
| 
 | |
|         //! All checks from release using mutex.state were moved here | |
|         void __TBB_EXPORTED_METHOD internal_release(); | |
| 
 | |
|         friend class recursive_mutex; | |
|     }; | |
| 
 | |
|     // Mutex traits | |
|     static const bool is_rw_mutex = false; | |
|     static const bool is_recursive_mutex = true; | |
|     static const bool is_fair_mutex = false; | |
| 
 | |
|     // C++0x compatibility interface | |
|      | |
|     //! Acquire lock | |
|     void lock() { | |
| #if TBB_USE_ASSERT | |
|         aligned_space<scoped_lock,1> tmp; | |
|         new(tmp.begin()) scoped_lock(*this); | |
| #else | |
|   #if _WIN32||_WIN64 | |
|         EnterCriticalSection(&impl); | |
|   #else | |
|         pthread_mutex_lock(&impl); | |
|   #endif /* _WIN32||_WIN64 */ | |
| #endif /* TBB_USE_ASSERT */ | |
|     } | |
| 
 | |
|     //! Try acquiring lock (non-blocking) | |
|     /** Return true if lock acquired; false otherwise. */ | |
|     bool try_lock() { | |
| #if TBB_USE_ASSERT | |
|         aligned_space<scoped_lock,1> tmp; | |
|         return (new(tmp.begin()) scoped_lock)->internal_try_acquire(*this); | |
| #else         | |
|   #if _WIN32||_WIN64 | |
|         return TryEnterCriticalSection(&impl)!=0; | |
|   #else | |
|         return pthread_mutex_trylock(&impl)==0; | |
|   #endif /* _WIN32||_WIN64 */ | |
| #endif /* TBB_USE_ASSERT */ | |
|     } | |
| 
 | |
|     //! Release lock | |
|     void unlock() { | |
| #if TBB_USE_ASSERT | |
|         aligned_space<scoped_lock,1> tmp; | |
|         scoped_lock& s = *tmp.begin(); | |
|         s.my_mutex = this; | |
|         s.internal_release(); | |
| #else | |
|   #if _WIN32||_WIN64 | |
|         LeaveCriticalSection(&impl); | |
|   #else | |
|         pthread_mutex_unlock(&impl); | |
|   #endif /* _WIN32||_WIN64 */ | |
| #endif /* TBB_USE_ASSERT */ | |
|     } | |
| 
 | |
|     //! Return native_handle | |
|   #if _WIN32||_WIN64 | |
|     typedef LPCRITICAL_SECTION native_handle_type; | |
|   #else | |
|     typedef pthread_mutex_t* native_handle_type; | |
|   #endif | |
|     native_handle_type native_handle() { return (native_handle_type) &impl; } | |
| 
 | |
| private: | |
| #if _WIN32||_WIN64 | |
|     CRITICAL_SECTION impl; | |
|     enum state_t { | |
|         INITIALIZED=0x1234, | |
|         DESTROYED=0x789A, | |
|     } state; | |
| #else | |
|     pthread_mutex_t impl; | |
| #endif /* _WIN32||_WIN64 */ | |
|  | |
|     //! All checks from mutex constructor using mutex.state were moved here | |
|     void __TBB_EXPORTED_METHOD internal_construct(); | |
| 
 | |
|     //! All checks from mutex destructor using mutex.state were moved here | |
|     void __TBB_EXPORTED_METHOD internal_destroy(); | |
| }; | |
| 
 | |
| __TBB_DEFINE_PROFILING_SET_NAME(recursive_mutex) | |
| 
 | |
| } // namespace tbb  | |
|  | |
| #endif /* __TBB_recursive_mutex_H */
 |