/* * Copyright (C) 2010 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /* ThreadPool */ #include "sles_allinclusive.h" // Entry point for each worker thread static void *ThreadPool_start(void *context) { ThreadPool *tp = (ThreadPool *) context; assert(NULL != tp); for (;;) { Closure *pClosure = ThreadPool_remove(tp); // closure is NULL when thread pool is being destroyed if (NULL == pClosure) { break; } // make a copy of parameters, then free the parameters const Closure closure = *pClosure; free(pClosure); // extract parameters and call the right method depending on kind ClosureKind kind = closure.mKind; void *context1 = closure.mContext1; void *context2 = closure.mContext2; int parameter1 = closure.mParameter1; switch (kind) { case CLOSURE_KIND_PPI: { ClosureHandler_ppi handler_ppi = closure.mHandler.mHandler_ppi; assert(NULL != handler_ppi); (*handler_ppi)(context1, context2, parameter1); } break; case CLOSURE_KIND_PPII: { ClosureHandler_ppii handler_ppii = closure.mHandler.mHandler_ppii; assert(NULL != handler_ppii); int parameter2 = closure.mParameter2; (*handler_ppii)(context1, context2, parameter1, parameter2); } break; case CLOSURE_KIND_PIIPP: { ClosureHandler_piipp handler_piipp = closure.mHandler.mHandler_piipp; assert(NULL != handler_piipp); int parameter2 = closure.mParameter2; void *context3 = closure.mContext3; (*handler_piipp)(context1, parameter1, parameter2, context2, context3); } break; default: SL_LOGE("Unexpected callback kind %d", kind); assert(false); break; } } return NULL; } #define INITIALIZED_NONE 0 #define INITIALIZED_MUTEX 1 #define INITIALIZED_CONDNOTFULL 2 #define INITIALIZED_CONDNOTEMPTY 4 #define INITIALIZED_ALL 7 static void ThreadPool_deinit_internal(ThreadPool *tp, unsigned initialized, unsigned nThreads); // Initialize a ThreadPool // maxClosures defaults to CLOSURE_TYPICAL if 0 // maxThreads defaults to THREAD_TYPICAL if 0 SLresult ThreadPool_init(ThreadPool *tp, unsigned maxClosures, unsigned maxThreads) { assert(NULL != tp); memset(tp, 0, sizeof(ThreadPool)); tp->mShutdown = SL_BOOLEAN_FALSE; unsigned initialized = INITIALIZED_NONE; // which objects were successfully initialized unsigned nThreads = 0; // number of threads successfully created int err; SLresult result; // initialize mutex and condition variables err = pthread_mutex_init(&tp->mMutex, (const pthread_mutexattr_t *) NULL); result = err_to_result(err); if (SL_RESULT_SUCCESS != result) goto fail; initialized |= INITIALIZED_MUTEX; err = pthread_cond_init(&tp->mCondNotFull, (const pthread_condattr_t *) NULL); result = err_to_result(err); if (SL_RESULT_SUCCESS != result) goto fail; initialized |= INITIALIZED_CONDNOTFULL; err = pthread_cond_init(&tp->mCondNotEmpty, (const pthread_condattr_t *) NULL); result = err_to_result(err); if (SL_RESULT_SUCCESS != result) goto fail; initialized |= INITIALIZED_CONDNOTEMPTY; // use default values for parameters, if not specified explicitly tp->mWaitingNotFull = 0; tp->mWaitingNotEmpty = 0; if (0 == maxClosures) maxClosures = CLOSURE_TYPICAL; tp->mMaxClosures = maxClosures; if (0 == maxThreads) maxThreads = THREAD_TYPICAL; tp->mMaxThreads = maxThreads; // initialize circular buffer for closures if (CLOSURE_TYPICAL >= maxClosures) { tp->mClosureArray = tp->mClosureTypical; } else { tp->mClosureArray = (Closure **) malloc((maxClosures + 1) * sizeof(Closure *)); if (NULL == tp->mClosureArray) { result = SL_RESULT_RESOURCE_ERROR; goto fail; } } tp->mClosureFront = tp->mClosureArray; tp->mClosureRear = tp->mClosureArray; // initialize thread pool if (THREAD_TYPICAL >= maxThreads) { tp->mThreadArray = tp->mThreadTypical; } else { tp->mThreadArray = (pthread_t *) malloc(maxThreads * sizeof(pthread_t)); if (NULL == tp->mThreadArray) { result = SL_RESULT_RESOURCE_ERROR; goto fail; } } unsigned i; for (i = 0; i < maxThreads; ++i) { int err = pthread_create(&tp->mThreadArray[i], (const pthread_attr_t *) NULL, ThreadPool_start, tp); result = err_to_result(err); if (SL_RESULT_SUCCESS != result) goto fail; ++nThreads; } tp->mInitialized = initialized; // done return SL_RESULT_SUCCESS; // here on any kind of error fail: ThreadPool_deinit_internal(tp, initialized, nThreads); return result; } static void ThreadPool_deinit_internal(ThreadPool *tp, unsigned initialized, unsigned nThreads) { int ok; assert(NULL != tp); // Destroy all threads if (0 < nThreads) { assert(INITIALIZED_ALL == initialized); ok = pthread_mutex_lock(&tp->mMutex); assert(0 == ok); tp->mShutdown = SL_BOOLEAN_TRUE; ok = pthread_cond_broadcast(&tp->mCondNotEmpty); assert(0 == ok); ok = pthread_cond_broadcast(&tp->mCondNotFull); assert(0 == ok); ok = pthread_mutex_unlock(&tp->mMutex); assert(0 == ok); unsigned i; for (i = 0; i < nThreads; ++i) { ok = pthread_join(tp->mThreadArray[i], (void **) NULL); assert(ok == 0); } // Empty out the circular buffer of closures ok = pthread_mutex_lock(&tp->mMutex); assert(0 == ok); Closure **oldFront = tp->mClosureFront; while (oldFront != tp->mClosureRear) { Closure **newFront = oldFront; if (++newFront == &tp->mClosureArray[tp->mMaxClosures + 1]) newFront = tp->mClosureArray; Closure *pClosure = *oldFront; assert(NULL != pClosure); *oldFront = NULL; tp->mClosureFront = newFront; ok = pthread_mutex_unlock(&tp->mMutex); assert(0 == ok); free(pClosure); ok = pthread_mutex_lock(&tp->mMutex); assert(0 == ok); } ok = pthread_mutex_unlock(&tp->mMutex); assert(0 == ok); // Note that we can't be sure when mWaitingNotFull will drop to zero } // destroy the mutex and condition variables if (initialized & INITIALIZED_CONDNOTEMPTY) { ok = pthread_cond_destroy(&tp->mCondNotEmpty); assert(0 == ok); } if (initialized & INITIALIZED_CONDNOTFULL) { ok = pthread_cond_destroy(&tp->mCondNotFull); assert(0 == ok); } if (initialized & INITIALIZED_MUTEX) { ok = pthread_mutex_destroy(&tp->mMutex); assert(0 == ok); } tp->mInitialized = INITIALIZED_NONE; // release the closure circular buffer if (tp->mClosureTypical != tp->mClosureArray && NULL != tp->mClosureArray) { free(tp->mClosureArray); tp->mClosureArray = NULL; } // release the thread pool if (tp->mThreadTypical != tp->mThreadArray && NULL != tp->mThreadArray) { free(tp->mThreadArray); tp->mThreadArray = NULL; } } void ThreadPool_deinit(ThreadPool *tp) { ThreadPool_deinit_internal(tp, tp->mInitialized, tp->mMaxThreads); } // Enqueue a closure to be executed later by a worker thread. // Note that this raw interface requires an explicit "kind" and full parameter list. // There are convenience methods below that make this easier to use. SLresult ThreadPool_add(ThreadPool *tp, ClosureKind kind, ClosureHandler_generic handler, void *context1, void *context2, void *context3, int parameter1, int parameter2) { assert(NULL != tp); assert(NULL != handler); Closure *closure = (Closure *) malloc(sizeof(Closure)); if (NULL == closure) { return SL_RESULT_RESOURCE_ERROR; } closure->mKind = kind; switch(kind) { case CLOSURE_KIND_PPI: closure->mHandler.mHandler_ppi = (ClosureHandler_ppi)handler; break; case CLOSURE_KIND_PPII: closure->mHandler.mHandler_ppii = (ClosureHandler_ppii)handler; break; case CLOSURE_KIND_PIIPP: closure->mHandler.mHandler_piipp = (ClosureHandler_piipp)handler; break; default: SL_LOGE("ThreadPool_add() invalid closure kind %d", kind); assert(false); } closure->mContext1 = context1; closure->mContext2 = context2; closure->mContext3 = context3; closure->mParameter1 = parameter1; closure->mParameter2 = parameter2; int ok; ok = pthread_mutex_lock(&tp->mMutex); assert(0 == ok); // can't enqueue while thread pool shutting down if (tp->mShutdown) { ok = pthread_mutex_unlock(&tp->mMutex); assert(0 == ok); free(closure); return SL_RESULT_PRECONDITIONS_VIOLATED; } for (;;) { Closure **oldRear = tp->mClosureRear; Closure **newRear = oldRear; if (++newRear == &tp->mClosureArray[tp->mMaxClosures + 1]) newRear = tp->mClosureArray; // if closure circular buffer is full, then wait for it to become non-full if (newRear == tp->mClosureFront) { ++tp->mWaitingNotFull; ok = pthread_cond_wait(&tp->mCondNotFull, &tp->mMutex); assert(0 == ok); // can't enqueue while thread pool shutting down if (tp->mShutdown) { assert(0 < tp->mWaitingNotFull); --tp->mWaitingNotFull; ok = pthread_mutex_unlock(&tp->mMutex); assert(0 == ok); free(closure); return SL_RESULT_PRECONDITIONS_VIOLATED; } continue; } assert(NULL == *oldRear); *oldRear = closure; tp->mClosureRear = newRear; // if a worker thread was waiting to dequeue, then suggest that it try again if (0 < tp->mWaitingNotEmpty) { --tp->mWaitingNotEmpty; ok = pthread_cond_signal(&tp->mCondNotEmpty); assert(0 == ok); } break; } ok = pthread_mutex_unlock(&tp->mMutex); assert(0 == ok); return SL_RESULT_SUCCESS; } // Called by a worker thread when it is ready to accept the next closure to execute Closure *ThreadPool_remove(ThreadPool *tp) { Closure *pClosure; int ok; ok = pthread_mutex_lock(&tp->mMutex); assert(0 == ok); for (;;) { // fail if thread pool is shutting down if (tp->mShutdown) { pClosure = NULL; break; } Closure **oldFront = tp->mClosureFront; // if closure circular buffer is empty, then wait for it to become non-empty if (oldFront == tp->mClosureRear) { ++tp->mWaitingNotEmpty; ok = pthread_cond_wait(&tp->mCondNotEmpty, &tp->mMutex); assert(0 == ok); // try again continue; } // dequeue the closure at front of circular buffer Closure **newFront = oldFront; if (++newFront == &tp->mClosureArray[tp->mMaxClosures + 1]) { newFront = tp->mClosureArray; } pClosure = *oldFront; assert(NULL != pClosure); *oldFront = NULL; tp->mClosureFront = newFront; // if a client thread was waiting to enqueue, then suggest that it try again if (0 < tp->mWaitingNotFull) { --tp->mWaitingNotFull; ok = pthread_cond_signal(&tp->mCondNotFull); assert(0 == ok); } break; } ok = pthread_mutex_unlock(&tp->mMutex); assert(0 == ok); return pClosure; } // Convenience methods for applications SLresult ThreadPool_add_ppi(ThreadPool *tp, ClosureHandler_ppi handler, void *context1, void *context2, int parameter1) { // function pointers are the same size so this is a safe cast return ThreadPool_add(tp, CLOSURE_KIND_PPI, (ClosureHandler_generic) handler, context1, context2, NULL, parameter1, 0); } SLresult ThreadPool_add_ppii(ThreadPool *tp, ClosureHandler_ppii handler, void *context1, void *context2, int parameter1, int parameter2) { // function pointers are the same size so this is a safe cast return ThreadPool_add(tp, CLOSURE_KIND_PPII, (ClosureHandler_generic) handler, context1, context2, NULL, parameter1, parameter2); } SLresult ThreadPool_add_piipp(ThreadPool *tp, ClosureHandler_piipp handler, void *cntxt1, int param1, int param2, void *cntxt2, void *cntxt3) { // function pointers are the same size so this is a safe cast return ThreadPool_add(tp, CLOSURE_KIND_PIIPP, (ClosureHandler_generic) handler, cntxt1, cntxt2, cntxt3, param1, param2); }